code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class A_ (UpperCAmelCase_ ): """simple docstring""" def _A ( self :Optional[int] ) -> Optional[Any]: '''simple docstring''' snake_case_ : List[str] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_snake_case , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(_snake_case , "num_attention_heads" ) ) self.parent.assertTrue(hasattr(_snake_case , "num_encoder_blocks" ) ) class A_ : """simple docstring""" def __init__( self :List[Any] , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :Union[str, Any]=13 , lowerCAmelCase__ :Union[str, Any]=64 , lowerCAmelCase__ :int=3 , lowerCAmelCase__ :Dict=4 , lowerCAmelCase__ :Dict=[2, 2, 2, 2] , lowerCAmelCase__ :Dict=[8, 4, 2, 1] , lowerCAmelCase__ :Any=[16, 32, 64, 128] , lowerCAmelCase__ :int=[1, 4, 8, 16] , lowerCAmelCase__ :Any=[1, 2, 4, 8] , lowerCAmelCase__ :Union[str, Any]=True , lowerCAmelCase__ :List[Any]=True , lowerCAmelCase__ :Any="gelu" , lowerCAmelCase__ :str=0.1 , lowerCAmelCase__ :Any=0.1 , lowerCAmelCase__ :Optional[int]=0.0_2 , lowerCAmelCase__ :str=3 , lowerCAmelCase__ :str=None , ) -> Union[str, Any]: '''simple docstring''' snake_case_ : int = parent snake_case_ : Union[str, Any] = batch_size snake_case_ : List[Any] = image_size snake_case_ : Dict = num_channels snake_case_ : Optional[Any] = num_encoder_blocks snake_case_ : List[Any] = sr_ratios snake_case_ : Union[str, Any] = depths snake_case_ : int = hidden_sizes snake_case_ : Optional[Any] = downsampling_rates snake_case_ : List[str] = num_attention_heads snake_case_ : Any = is_training snake_case_ : Dict = use_labels snake_case_ : List[Any] = hidden_act snake_case_ : Dict = hidden_dropout_prob snake_case_ : int = attention_probs_dropout_prob snake_case_ : Optional[Any] = initializer_range snake_case_ : int = num_labels snake_case_ : Dict = scope def _A ( self :Optional[Any] ) -> str: '''simple docstring''' snake_case_ : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ : str = None if self.use_labels: snake_case_ : str = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) snake_case_ : List[Any] = self.get_config() return config, pixel_values, labels def _A ( self :List[Any] ) -> Optional[int]: '''simple docstring''' return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def _A ( self :Union[str, Any] , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Any , lowerCAmelCase__ :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : str = SegformerModel(config=_snake_case ) model.to(_snake_case ) model.eval() snake_case_ : Optional[Any] = model(_snake_case ) snake_case_ : List[Any] = self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def _A ( self :int , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Optional[int] ) -> Dict: '''simple docstring''' snake_case_ : Optional[int] = self.num_labels snake_case_ : Any = SegformerForSemanticSegmentation(_snake_case ) model.to(_snake_case ) model.eval() snake_case_ : int = model(_snake_case ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) snake_case_ : Union[str, Any] = model(_snake_case , labels=_snake_case ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss , 0.0 ) def _A ( self :int , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Any ) -> str: '''simple docstring''' snake_case_ : List[str] = 1 snake_case_ : Optional[Any] = SegformerForSemanticSegmentation(config=_snake_case ) model.to(_snake_case ) model.eval() snake_case_ : List[str] = torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size) ).to(_snake_case ) snake_case_ : Optional[int] = model(_snake_case , labels=_snake_case ) self.parent.assertGreater(result.loss , 0.0 ) def _A ( self :Optional[int] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Dict = self.prepare_config_and_inputs() snake_case_, snake_case_, snake_case_ : Tuple = config_and_inputs snake_case_ : List[Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class A_ (UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): """simple docstring""" a__ = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) a__ = ( { "feature-extraction": SegformerModel, "image-classification": SegformerForImageClassification, "image-segmentation": SegformerForSemanticSegmentation, } if is_torch_available() else {} ) a__ = True a__ = False a__ = False a__ = False def _A ( self :str ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[int] = SegformerModelTester(self ) snake_case_ : str = SegformerConfigTester(self , config_class=_snake_case ) def _A ( self :int ) -> Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() def _A ( self :Optional[Any] ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case ) def _A ( self :Dict ) -> str: '''simple docstring''' snake_case_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*_snake_case ) def _A ( self :int ) -> List[str]: '''simple docstring''' snake_case_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*_snake_case ) @unittest.skip("SegFormer does not use inputs_embeds" ) def _A ( self :Dict ) -> Tuple: '''simple docstring''' pass @unittest.skip("SegFormer does not have get_input_embeddings method and get_output_embeddings methods" ) def _A ( self :Dict ) -> List[str]: '''simple docstring''' pass def _A ( self :str ) -> Optional[int]: '''simple docstring''' snake_case_, snake_case_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ : str = model_class(_snake_case ) snake_case_ : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ : Any = [*signature.parameters.keys()] snake_case_ : Optional[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , _snake_case ) def _A ( self :Union[str, Any] ) -> Optional[Any]: '''simple docstring''' snake_case_, snake_case_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ : Optional[Any] = True for model_class in self.all_model_classes: snake_case_ : Tuple = True snake_case_ : int = False snake_case_ : List[Any] = True snake_case_ : Optional[int] = model_class(_snake_case ) model.to(_snake_case ) model.eval() with torch.no_grad(): snake_case_ : int = model(**self._prepare_for_class(_snake_case , _snake_case ) ) snake_case_ : List[Any] = outputs.attentions snake_case_ : List[str] = sum(self.model_tester.depths ) self.assertEqual(len(_snake_case ) , _snake_case ) # check that output_attentions also work using config del inputs_dict["output_attentions"] snake_case_ : Any = True snake_case_ : Dict = model_class(_snake_case ) model.to(_snake_case ) model.eval() with torch.no_grad(): snake_case_ : List[Any] = model(**self._prepare_for_class(_snake_case , _snake_case ) ) snake_case_ : List[Any] = outputs.attentions self.assertEqual(len(_snake_case ) , _snake_case ) # verify the first attentions (first block, first layer) snake_case_ : Dict = (self.model_tester.image_size // 4) ** 2 snake_case_ : List[str] = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) snake_case_ : Optional[int] = (self.model_tester.image_size // 32) ** 2 snake_case_ : Dict = (self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) snake_case_ : int = len(_snake_case ) # Check attention is always last and order is fine snake_case_ : List[Any] = True snake_case_ : Any = True snake_case_ : Dict = model_class(_snake_case ) model.to(_snake_case ) model.eval() with torch.no_grad(): snake_case_ : Optional[int] = model(**self._prepare_for_class(_snake_case , _snake_case ) ) self.assertEqual(out_len + 1 , len(_snake_case ) ) snake_case_ : Any = outputs.attentions self.assertEqual(len(_snake_case ) , _snake_case ) # verify the first attentions (first block, first layer) snake_case_ : int = (self.model_tester.image_size // 4) ** 2 snake_case_ : Optional[int] = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def _A ( self :Union[str, Any] ) -> Any: '''simple docstring''' def check_hidden_states_output(lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :str ): snake_case_ : Tuple = model_class(_snake_case ) model.to(_snake_case ) model.eval() with torch.no_grad(): snake_case_ : Optional[Any] = model(**self._prepare_for_class(_snake_case , _snake_case ) ) snake_case_ : int = outputs.hidden_states snake_case_ : Union[str, Any] = self.model_tester.num_encoder_blocks self.assertEqual(len(_snake_case ) , _snake_case ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) snake_case_, snake_case_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ : Tuple = True check_hidden_states_output(_snake_case , _snake_case , _snake_case ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ : List[Any] = True check_hidden_states_output(_snake_case , _snake_case , _snake_case ) def _A ( self :Tuple ) -> Optional[Any]: '''simple docstring''' if not self.model_tester.is_training: return snake_case_, snake_case_ : str = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ : List[str] = True for model_class in self.all_model_classes: if model_class in get_values(_snake_case ): continue snake_case_ : Optional[int] = model_class(_snake_case ) model.to(_snake_case ) model.train() snake_case_ : Union[str, Any] = self._prepare_for_class(_snake_case , _snake_case , return_labels=_snake_case ) snake_case_ : Optional[Any] = model(**_snake_case ).loss loss.backward() @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def _A ( self :Optional[Any] ) -> Union[str, Any]: '''simple docstring''' pass @slow def _A ( self :Dict ) -> Tuple: '''simple docstring''' for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ : int = SegformerModel.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) def __UpperCAmelCase ( )-> Optional[Any]: """simple docstring""" snake_case_ : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch class A_ (unittest.TestCase ): """simple docstring""" @slow def _A ( self :Dict ) -> Optional[int]: '''simple docstring''' snake_case_ : List[Any] = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_snake_case , align=_snake_case , do_random_crop=_snake_case ) snake_case_ : Tuple = SegformerForSemanticSegmentation.from_pretrained("nvidia/segformer-b0-finetuned-ade-512-512" ).to( _snake_case ) snake_case_ : str = prepare_img() snake_case_ : List[Any] = image_processor(images=_snake_case , return_tensors="pt" ) snake_case_ : str = encoded_inputs.pixel_values.to(_snake_case ) with torch.no_grad(): snake_case_ : List[Any] = model(_snake_case ) snake_case_ : Optional[int] = torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , _snake_case ) snake_case_ : Dict = torch.tensor( [ [[-4.6_3_1_0, -5.5_2_3_2, -6.2_3_5_6], [-5.1_9_2_1, -6.1_4_4_4, -6.5_9_9_6], [-5.4_4_2_4, -6.2_7_9_0, -6.7_5_7_4]], [[-1_2.1_3_9_1, -1_3.3_1_2_2, -1_3.9_5_5_4], [-1_2.8_7_3_2, -1_3.9_3_5_2, -1_4.3_5_6_3], [-1_2.9_4_3_8, -1_3.8_2_2_6, -1_4.2_5_1_3]], [[-1_2.5_1_3_4, -1_3.4_6_8_6, -1_4.4_9_1_5], [-1_2.8_6_6_9, -1_4.4_3_4_3, -1_4.7_7_5_8], [-1_3.2_5_2_3, -1_4.5_8_1_9, -1_5.0_6_9_4]], ] ).to(_snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , _snake_case , atol=1E-4 ) ) @slow def _A ( self :List[Any] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[Any] = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_snake_case , align=_snake_case , do_random_crop=_snake_case ) snake_case_ : Optional[Any] = SegformerForSemanticSegmentation.from_pretrained( "nvidia/segformer-b1-finetuned-cityscapes-1024-1024" ).to(_snake_case ) snake_case_ : Union[str, Any] = prepare_img() snake_case_ : List[str] = image_processor(images=_snake_case , return_tensors="pt" ) snake_case_ : Any = encoded_inputs.pixel_values.to(_snake_case ) with torch.no_grad(): snake_case_ : List[Any] = model(_snake_case ) snake_case_ : List[str] = torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , _snake_case ) snake_case_ : str = torch.tensor( [ [[-1_3.5_7_4_8, -1_3.9_1_1_1, -1_2.6_5_0_0], [-1_4.3_5_0_0, -1_5.3_6_8_3, -1_4.2_3_2_8], [-1_4.7_5_3_2, -1_6.0_4_2_4, -1_5.6_0_8_7]], [[-1_7.1_6_5_1, -1_5.8_7_2_5, -1_2.9_6_5_3], [-1_7.2_5_8_0, -1_7.3_7_1_8, -1_4.8_2_2_3], [-1_6.6_0_5_8, -1_6.8_7_8_3, -1_6.7_4_5_2]], [[-3.6_4_5_6, -3.0_2_0_9, -1.4_2_0_3], [-3.0_7_9_7, -3.1_9_5_9, -2.0_0_0_0], [-1.8_7_5_7, -1.9_2_1_7, -1.6_9_9_7]], ] ).to(_snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , _snake_case , atol=1E-1 ) ) @slow def _A ( self :List[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Optional[int] = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_snake_case , align=_snake_case , do_random_crop=_snake_case ) snake_case_ : int = SegformerForSemanticSegmentation.from_pretrained("nvidia/segformer-b0-finetuned-ade-512-512" ).to( _snake_case ) snake_case_ : Optional[int] = prepare_img() snake_case_ : Optional[Any] = image_processor(images=_snake_case , return_tensors="pt" ) snake_case_ : Tuple = encoded_inputs.pixel_values.to(_snake_case ) with torch.no_grad(): snake_case_ : List[Any] = model(_snake_case ) snake_case_ : Any = outputs.logits.detach().cpu() snake_case_ : Union[str, Any] = image_processor.post_process_semantic_segmentation(outputs=_snake_case , target_sizes=[(500, 300)] ) snake_case_ : int = torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , _snake_case ) snake_case_ : Dict = image_processor.post_process_semantic_segmentation(outputs=_snake_case ) snake_case_ : Optional[Any] = torch.Size((128, 128) ) self.assertEqual(segmentation[0].shape , _snake_case )
708
'''simple docstring''' from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable __lowerCamelCase : Dict = { '''configuration_gpt_neox_japanese''': ['''GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoXJapaneseConfig'''], '''tokenization_gpt_neox_japanese''': ['''GPTNeoXJapaneseTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ '''GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoXJapaneseForCausalLM''', '''GPTNeoXJapaneseLayer''', '''GPTNeoXJapaneseModel''', '''GPTNeoXJapanesePreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neox_japanese import GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXJapaneseConfig from .tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox_japanese import ( GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseLayer, GPTNeoXJapaneseModel, GPTNeoXJapanesePreTrainedModel, ) else: import sys __lowerCamelCase : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
656
0
'''simple docstring''' from __future__ import annotations import math from collections.abc import Callable def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ = 100 ,)-> Tuple: """simple docstring""" snake_case_ : Dict = x_start snake_case_ : Any = fnc(__magic_name__ ) snake_case_ : Dict = 0.0 for _ in range(__magic_name__ ): # Approximates curve as a sequence of linear lines and sums their length snake_case_ : int = (x_end - x_start) / steps + xa snake_case_ : str = fnc(__magic_name__ ) length += math.hypot(xa - xa ,fxa - fxa ) # Increment step snake_case_ : str = xa snake_case_ : Union[str, Any] = fxa return length if __name__ == "__main__": def __UpperCAmelCase ( __magic_name__ )-> Tuple: """simple docstring""" return math.sin(10 * x ) print('''f(x) = sin(10 * x)''') print('''The length of the curve from x = -10 to x = 10 is:''') __lowerCamelCase : Any = 10 while i <= 100000: print(f'''With {i} steps: {line_length(f, -10, 10, i)}''') i *= 10
709
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ )-> list[int]: """simple docstring""" if length <= 0 or not isinstance(__magic_name__ ,__magic_name__ ): raise ValueError("Length must be a positive integer." ) return [n * (2 * n - 1) for n in range(__magic_name__ )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=10))
656
0
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ = 50 )-> str: """simple docstring""" snake_case_ : str = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 ,5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(f'''{solution() = }''')
710
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def __UpperCAmelCase ( __magic_name__=None )-> List[str]: """simple docstring""" if subparsers is not None: snake_case_ : List[str] = subparsers.add_parser("test" ) else: snake_case_ : List[Any] = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" ,default=__magic_name__ ,help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) ,) if subparsers is not None: parser.set_defaults(func=__magic_name__ ) return parser def __UpperCAmelCase ( __magic_name__ )-> Tuple: """simple docstring""" snake_case_ : Optional[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: snake_case_ : str = script_name else: snake_case_ : Any = F'''--config_file={args.config_file} {script_name}''' snake_case_ : Union[str, Any] = ["accelerate-launch"] + test_args.split() snake_case_ : Optional[int] = execute_subprocess_async(__magic_name__ ,env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : Dict = test_command_parser() snake_case_ : Dict = parser.parse_args() test_command(__magic_name__ ) if __name__ == "__main__": main()
656
0
'''simple docstring''' from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if is_torch_available(): import torch __lowerCamelCase : Tuple = logging.get_logger(__name__) @add_end_docstrings( SCREAMING_SNAKE_CASE_ , r'''\n top_k (`int`, defaults to 5):\n The number of predictions to return.\n targets (`str` or `List[str]`, *optional*):\n When passed, the model will limit the scores to the passed targets instead of looking up in the whole\n vocab. If the provided targets are not in the model vocab, they will be tokenized and the first resulting\n token will be used (with a warning, and that might be slower).\n\n ''' , ) class A_ (SCREAMING_SNAKE_CASE_ ): """simple docstring""" def _A ( self :Dict , lowerCAmelCase__ :Tuple ) -> Any: '''simple docstring''' if self.framework == "tf": snake_case_ : List[str] = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy() elif self.framework == "pt": snake_case_ : Any = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=UpperCamelCase__ ) else: raise ValueError("Unsupported framework" ) return masked_index def _A ( self :str , lowerCAmelCase__ :List[Any] ) -> Any: '''simple docstring''' snake_case_ : Dict = self.get_masked_index(UpperCamelCase__ ) snake_case_ : str = np.prod(masked_index.shape ) if numel < 1: raise PipelineException( "fill-mask" , self.model.base_model_prefix , F'''No mask_token ({self.tokenizer.mask_token}) found on the input''' , ) def _A ( self :List[Any] , lowerCAmelCase__ :List[str] ) -> Optional[Any]: '''simple docstring''' if isinstance(UpperCamelCase__ , UpperCamelCase__ ): for model_input in model_inputs: self._ensure_exactly_one_mask_token(model_input["input_ids"][0] ) else: for input_ids in model_inputs["input_ids"]: self._ensure_exactly_one_mask_token(UpperCamelCase__ ) def _A ( self :List[Any] , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :int=None , **lowerCAmelCase__ :List[Any] ) -> int: '''simple docstring''' if return_tensors is None: snake_case_ : str = self.framework snake_case_ : str = self.tokenizer(UpperCamelCase__ , return_tensors=UpperCamelCase__ ) self.ensure_exactly_one_mask_token(UpperCamelCase__ ) return model_inputs def _A ( self :Optional[int] , lowerCAmelCase__ :int ) -> Optional[int]: '''simple docstring''' snake_case_ : Tuple = self.model(**UpperCamelCase__ ) snake_case_ : List[str] = model_inputs['''input_ids'''] return model_outputs def _A ( self :Dict , lowerCAmelCase__ :Dict , lowerCAmelCase__ :Tuple=5 , lowerCAmelCase__ :List[str]=None ) -> List[Any]: '''simple docstring''' if target_ids is not None and target_ids.shape[0] < top_k: snake_case_ : Union[str, Any] = target_ids.shape[0] snake_case_ : str = model_outputs['''input_ids'''][0] snake_case_ : Tuple = model_outputs['''logits'''] if self.framework == "tf": snake_case_ : Tuple = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy()[:, 0] snake_case_ : Dict = outputs.numpy() snake_case_ : Tuple = outputs[0, masked_index, :] snake_case_ : str = stable_softmax(UpperCamelCase__ , axis=-1 ) if target_ids is not None: snake_case_ : int = tf.gather_nd(tf.squeeze(UpperCamelCase__ , 0 ) , target_ids.reshape(-1 , 1 ) ) snake_case_ : str = tf.expand_dims(UpperCamelCase__ , 0 ) snake_case_ : str = tf.math.top_k(UpperCamelCase__ , k=UpperCamelCase__ ) snake_case_ : List[str] = topk.values.numpy(), topk.indices.numpy() else: snake_case_ : Optional[Any] = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=UpperCamelCase__ ).squeeze(-1 ) # Fill mask pipeline supports only one ${mask_token} per sample snake_case_ : List[Any] = outputs[0, masked_index, :] snake_case_ : Tuple = logits.softmax(dim=-1 ) if target_ids is not None: snake_case_ : str = probs[..., target_ids] snake_case_ : Tuple = probs.topk(UpperCamelCase__ ) snake_case_ : Union[str, Any] = [] snake_case_ : str = values.shape[0] == 1 for i, (_values, _predictions) in enumerate(zip(values.tolist() , predictions.tolist() ) ): snake_case_ : Optional[Any] = [] for v, p in zip(_values , _predictions ): # Copy is important since we're going to modify this array in place snake_case_ : List[str] = input_ids.numpy().copy() if target_ids is not None: snake_case_ : Any = target_ids[p].tolist() snake_case_ : List[str] = p # Filter padding out: snake_case_ : Dict = tokens[np.where(tokens != self.tokenizer.pad_token_id )] # Originally we skip special tokens to give readable output. # For multi masks though, the other [MASK] would be removed otherwise # making the output look odd, so we add them back snake_case_ : Any = self.tokenizer.decode(UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ ) snake_case_ : str = {'''score''': v, '''token''': p, '''token_str''': self.tokenizer.decode([p] ), '''sequence''': sequence} row.append(UpperCamelCase__ ) result.append(UpperCamelCase__ ) if single_mask: return result[0] return result def _A ( self :Dict , lowerCAmelCase__ :Any , lowerCAmelCase__ :Dict=None ) -> List[str]: '''simple docstring''' if isinstance(UpperCamelCase__ , UpperCamelCase__ ): snake_case_ : Dict = [targets] try: snake_case_ : Any = self.tokenizer.get_vocab() except Exception: snake_case_ : Tuple = {} snake_case_ : str = [] for target in targets: snake_case_ : int = vocab.get(UpperCamelCase__ , UpperCamelCase__ ) if id_ is None: snake_case_ : Tuple = self.tokenizer( UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , return_attention_mask=UpperCamelCase__ , return_token_type_ids=UpperCamelCase__ , max_length=1 , truncation=UpperCamelCase__ , )['''input_ids'''] if len(UpperCamelCase__ ) == 0: logger.warning( F'''The specified target token `{target}` does not exist in the model vocabulary. ''' "We cannot replace it with anything meaningful, ignoring it" ) continue snake_case_ : int = input_ids[0] # XXX: If users encounter this pass # it becomes pretty slow, so let's make sure # The warning enables them to fix the input to # get faster performance. logger.warning( F'''The specified target token `{target}` does not exist in the model vocabulary. ''' F'''Replacing with `{self.tokenizer.convert_ids_to_tokens(id_ )}`.''' ) target_ids.append(id_ ) snake_case_ : str = list(set(UpperCamelCase__ ) ) if len(UpperCamelCase__ ) == 0: raise ValueError("At least one target must be provided when passed." ) snake_case_ : Optional[int] = np.array(UpperCamelCase__ ) return target_ids def _A ( self :Tuple , lowerCAmelCase__ :str=None , lowerCAmelCase__ :int=None ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = {} if targets is not None: snake_case_ : Any = self.get_target_ids(UpperCamelCase__ , UpperCamelCase__ ) snake_case_ : List[Any] = target_ids if top_k is not None: snake_case_ : Union[str, Any] = top_k if self.tokenizer.mask_token_id is None: raise PipelineException( "fill-mask" , self.model.base_model_prefix , "The tokenizer does not define a `mask_token`." ) return {}, {}, postprocess_params def __call__( self :Union[str, Any] , lowerCAmelCase__ :Optional[Any] , *lowerCAmelCase__ :Union[str, Any] , **lowerCAmelCase__ :List[str] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Any = super().__call__(UpperCamelCase__ , **UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) and len(UpperCamelCase__ ) == 1: return outputs[0] return outputs
711
'''simple docstring''' from scipy.stats import spearmanr import datasets __lowerCamelCase : str = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' __lowerCamelCase : int = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' __lowerCamelCase : List[str] = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): """simple docstring""" def _A ( self :str ) -> Union[str, Any]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } ) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"] , ) def _A ( self :Optional[int] , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :Optional[Any]=False ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[Any] = spearmanr(lowerCAmelCase__ , lowerCAmelCase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
656
0
'''simple docstring''' import os from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen, xsplitext from ..table import array_cast from ..utils.py_utils import no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: from .features import FeatureType __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : List[str] = False, False, False @dataclass class A_ : """simple docstring""" a__ = None a__ = True a__ = True a__ = None # Automatically constructed a__ = "dict" a__ = pa.struct({'''bytes''': pa.binary(), '''path''': pa.string()} ) a__ = field(default='''Audio''' , init=lowerCAmelCase__ , repr=lowerCAmelCase__ ) def __call__( self :int ) -> Optional[int]: '''simple docstring''' return self.pa_type def _A ( self :int , lowerCAmelCase__ :int ) -> dict: '''simple docstring''' try: import soundfile as sf # soundfile is a dependency of librosa, needed to decode audio files. except ImportError as err: raise ImportError("To support encoding audio data, please install \'soundfile\'." ) from err if isinstance(_lowerCamelCase , _lowerCamelCase ): return {"bytes": None, "path": value} elif isinstance(_lowerCamelCase , _lowerCamelCase ): return {"bytes": value, "path": None} elif "array" in value: # convert the audio array to wav bytes snake_case_ : Optional[Any] = BytesIO() sf.write(_lowerCamelCase , value["array"] , value["sampling_rate"] , format="wav" ) return {"bytes": buffer.getvalue(), "path": None} elif value.get("path" ) is not None and os.path.isfile(value["path"] ): # we set "bytes": None to not duplicate the data if they're already available locally if value["path"].endswith("pcm" ): # "PCM" only has raw audio bytes if value.get("sampling_rate" ) is None: # At least, If you want to convert "PCM-byte" to "WAV-byte", you have to know sampling rate raise KeyError("To use PCM files, please specify a \'sampling_rate\' in Audio object" ) if value.get("bytes" ): # If we already had PCM-byte, we don`t have to make "read file, make bytes" (just use it!) snake_case_ : int = np.frombuffer(value["bytes"] , dtype=np.intaa ).astype(np.floataa ) / 32_767 else: snake_case_ : Dict = np.memmap(value["path"] , dtype="h" , mode="r" ).astype(np.floataa ) / 32_767 snake_case_ : Optional[int] = BytesIO(bytes() ) sf.write(_lowerCamelCase , _lowerCamelCase , value["sampling_rate"] , format="wav" ) return {"bytes": buffer.getvalue(), "path": None} else: return {"bytes": None, "path": value.get("path" )} elif value.get("bytes" ) is not None or value.get("path" ) is not None: # store the audio bytes, and path is used to infer the audio format using the file extension return {"bytes": value.get("bytes" ), "path": value.get("path" )} else: raise ValueError( F'''An audio sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.''' ) def _A ( self :Tuple , lowerCAmelCase__ :Dict , lowerCAmelCase__ :Optional[Any] = None ) -> dict: '''simple docstring''' if not self.decode: raise RuntimeError("Decoding is disabled for this feature. Please use Audio(decode=True) instead." ) snake_case_, snake_case_ : Optional[Any] = (value["path"], BytesIO(value["bytes"] )) if value["bytes"] is not None else (value["path"], None) if path is None and file is None: raise ValueError(F'''An audio sample should have one of \'path\' or \'bytes\' but both are None in {value}.''' ) try: import librosa import soundfile as sf except ImportError as err: raise ImportError("To support decoding audio files, please install \'librosa\' and \'soundfile\'." ) from err snake_case_ : Tuple = xsplitext(_lowerCamelCase )[1][1:].lower() if path is not None else None if not config.IS_OPUS_SUPPORTED and audio_format == "opus": raise RuntimeError( "Decoding \'opus\' files requires system library \'libsndfile\'>=1.0.31, " "You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. " ) elif not config.IS_MP3_SUPPORTED and audio_format == "mp3": raise RuntimeError( "Decoding \'mp3\' files requires system library \'libsndfile\'>=1.1.0, " "You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. " ) if file is None: snake_case_ : Optional[int] = token_per_repo_id or {} snake_case_ : List[Any] = path.split("::" )[-1] try: snake_case_ : Any = string_to_dict(_lowerCamelCase , config.HUB_DATASETS_URL )["repo_id"] snake_case_ : int = token_per_repo_id[repo_id] except (ValueError, KeyError): snake_case_ : Tuple = None with xopen(_lowerCamelCase , "rb" , use_auth_token=_lowerCamelCase ) as f: snake_case_, snake_case_ : Optional[Any] = sf.read(_lowerCamelCase ) else: snake_case_, snake_case_ : Dict = sf.read(_lowerCamelCase ) snake_case_ : Tuple = array.T if self.mono: snake_case_ : Dict = librosa.to_mono(_lowerCamelCase ) if self.sampling_rate and self.sampling_rate != sampling_rate: snake_case_ : Optional[Any] = librosa.resample(_lowerCamelCase , orig_sr=_lowerCamelCase , target_sr=self.sampling_rate ) snake_case_ : Optional[int] = self.sampling_rate return {"path": path, "array": array, "sampling_rate": sampling_rate} def _A ( self :int ) -> Union["FeatureType", Dict[str, "FeatureType"]]: '''simple docstring''' from .features import Value if self.decode: raise ValueError("Cannot flatten a decoded Audio feature." ) return { "bytes": Value("binary" ), "path": Value("string" ), } def _A ( self :Tuple , lowerCAmelCase__ :Any ) -> pa.StructArray: '''simple docstring''' if pa.types.is_string(storage.type ): snake_case_ : Union[str, Any] = pa.array([None] * len(_lowerCamelCase ) , type=pa.binary() ) snake_case_ : Tuple = pa.StructArray.from_arrays([bytes_array, storage] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): snake_case_ : int = pa.array([None] * len(_lowerCamelCase ) , type=pa.string() ) snake_case_ : Optional[int] = pa.StructArray.from_arrays([storage, path_array] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ) and storage.type.get_all_field_indices("array" ): snake_case_ : Optional[Any] = pa.array([Audio().encode_example(_lowerCamelCase ) if x is not None else None for x in storage.to_pylist()] ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("bytes" ) >= 0: snake_case_ : Union[str, Any] = storage.field("bytes" ) else: snake_case_ : Dict = pa.array([None] * len(_lowerCamelCase ) , type=pa.binary() ) if storage.type.get_field_index("path" ) >= 0: snake_case_ : Optional[int] = storage.field("path" ) else: snake_case_ : Any = pa.array([None] * len(_lowerCamelCase ) , type=pa.string() ) snake_case_ : Dict = pa.StructArray.from_arrays([bytes_array, path_array] , ["bytes", "path"] , mask=storage.is_null() ) return array_cast(_lowerCamelCase , self.pa_type ) def _A ( self :List[Any] , lowerCAmelCase__ :Optional[Any] ) -> pa.StructArray: '''simple docstring''' @no_op_if_value_is_null def path_to_bytes(lowerCAmelCase__ :Dict ): with xopen(_lowerCamelCase , "rb" ) as f: snake_case_ : Optional[int] = f.read() return bytes_ snake_case_ : Any = pa.array( [ (path_to_bytes(x["path"] ) if x["bytes"] is None else x["bytes"]) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) snake_case_ : Optional[int] = pa.array( [os.path.basename(_lowerCamelCase ) if path is not None else None for path in storage.field("path" ).to_pylist()] , type=pa.string() , ) snake_case_ : Optional[Any] = pa.StructArray.from_arrays([bytes_array, path_array] , ["bytes", "path"] , mask=bytes_array.is_null() ) return array_cast(_lowerCamelCase , self.pa_type )
712
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): __lowerCamelCase : Optional[Any] = get_tests_dir('''fixtures/test_sentencepiece.model''') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right __lowerCamelCase : str = 128022 __lowerCamelCase : List[Any] = 128028 @require_sentencepiece class A_ (a_ , unittest.TestCase ): """simple docstring""" a__ = MaMaaaTokenizer a__ = False a__ = False a__ = True def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' super().setUp() snake_case_ : int = ["</s>", "<unk>", "▁This", "▁is", "▁a", "▁t", "est", "\u0120", "<pad>"] snake_case_ : Any = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) snake_case_ : Optional[int] = Path(self.tmpdirname ) save_json(lowerCAmelCase__ , save_dir / VOCAB_FILES_NAMES["vocab_file"] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(lowerCAmelCase__ , save_dir / VOCAB_FILES_NAMES["spm_file"] ) snake_case_ : Union[str, Any] = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def _A ( self :List[Any] , **lowerCAmelCase__ :List[Any] ) -> str: '''simple docstring''' return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def _A ( self :Optional[int] , lowerCAmelCase__ :Any ) -> Optional[int]: '''simple docstring''' return ( "This is a test", "This is a test", ) def _A ( self :List[str] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : str = "</s>" snake_case_ : Union[str, Any] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase__ ) , lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = self.get_tokenizer() snake_case_ : Any = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "</s>" ) self.assertEqual(vocab_keys[1] , "<unk>" ) self.assertEqual(vocab_keys[-1] , "<s>" ) self.assertEqual(len(lowerCAmelCase__ ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip("Skip this test while all models are still to be uploaded." ) def _A ( self :List[Any] ) -> Union[str, Any]: '''simple docstring''' pass def _A ( self :Optional[int] ) -> int: '''simple docstring''' snake_case_ : int = self.get_tokenizer() snake_case_ : List[str] = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowerCAmelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [2, 3, 4, 5, 6] , ) snake_case_ : Any = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(lowerCAmelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) snake_case_ : Any = tokenizer.convert_tokens_to_string(lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , "This is a test" ) @slow def _A ( self :Any ) -> List[Any]: '''simple docstring''' snake_case_ : int = {"input_ids": [[128_022, 110_108, 397, 11, 38_272, 2_247, 124_811, 285, 18_105, 1_586, 207, 7, 39_534, 4_428, 397, 1_019, 18_105, 1_586, 207, 7, 41_337, 16_786, 241, 7, 20_214, 17, 125_690, 10_398, 7, 44_378, 58_069, 68_342, 7_798, 7_343, 11, 299, 33_310, 4, 158, 37_350, 94_077, 4_569, 299, 33_310, 90, 4, 52_840, 290, 4, 31_270, 112, 299, 682, 4, 52_840, 39_953, 14_079, 193, 52_519, 90_894, 17_894, 120_697, 11, 40_445, 551, 17, 1_019, 52_519, 90_894, 17_756, 963, 11, 40_445, 480, 17, 9_792, 1_120, 5_173, 1_393, 6_240, 16_786, 241, 120_996, 28, 1_245, 1_393, 118_240, 11_123, 1_019, 93_612, 2_691, 10_618, 98_058, 120_409, 1_928, 279, 4, 40_683, 367, 178, 207, 1_019, 103, 103_121, 506, 65_296, 5, 2], [128_022, 21_217, 367, 117, 125_450, 128, 719, 7, 7_308, 40, 93_612, 12_669, 1_116, 16_704, 71, 17_785, 3_699, 15_592, 35, 144, 9_584, 241, 11_943, 713, 950, 799, 2_247, 88_427, 150, 149, 118_813, 120_706, 1_019, 106_906, 81_518, 28, 1_224, 22_799, 397, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [128_022, 1_658, 123_311, 5_155, 5_578, 4_722, 279, 14_947, 2_366, 1_120, 1_197, 14, 1_348, 9_232, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase__ , model_name="facebook/m2m100_418M" , revision="c168bae485c864188cf9aa0e4108b0b6934dc91e" , ) @require_torch @require_sentencepiece @require_tokenizers class A_ (unittest.TestCase ): """simple docstring""" a__ = '''facebook/m2m100_418M''' a__ = [ '''In my opinion, there are two levels of response from the French government.''', '''NSA Affair Emphasizes Complete Lack of Debate on Intelligence''', ] a__ = [ '''Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.''', '''L\'affaire NSA souligne l\'absence totale de débat sur le renseignement''', ] # fmt: off a__ = [EN_CODE, 593, 1949, 115781, 4, 71586, 4234, 60633, 126233, 432, 123808, 15592, 1197, 117132, 120618, 5, 2] @classmethod def _A ( cls :str ) -> int: '''simple docstring''' snake_case_ : MaMaaaTokenizer = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en" , tgt_lang="fr" ) snake_case_ : List[str] = 1 return cls def _A ( self :Tuple ) -> Union[str, Any]: '''simple docstring''' self.assertEqual(self.tokenizer.get_lang_id("ar" ) , 128_006 ) self.assertEqual(self.tokenizer.get_lang_id("en" ) , 128_022 ) self.assertEqual(self.tokenizer.get_lang_id("ro" ) , 128_076 ) self.assertEqual(self.tokenizer.get_lang_id("mr" ) , 128_063 ) def _A ( self :Optional[int] ) -> List[str]: '''simple docstring''' snake_case_ : Dict = self.tokenizer.get_vocab() self.assertEqual(len(lowerCAmelCase__ ) , self.tokenizer.vocab_size ) self.assertEqual(vocab["<unk>"] , 3 ) self.assertIn(self.tokenizer.get_lang_token("en" ) , lowerCAmelCase__ ) def _A ( self :Any ) -> Dict: '''simple docstring''' snake_case_ : List[str] = "en" snake_case_ : Dict = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' self.assertIn(lowerCAmelCase__ , self.tokenizer.all_special_ids ) # fmt: off snake_case_ : Dict = [FR_CODE, 5_364, 82, 8_642, 4, 294, 47, 8, 14_028, 136, 3_286, 9_706, 6, 90_797, 6, 144_012, 162, 88_128, 30_061, 5, 2] # fmt: on snake_case_ : List[str] = self.tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) snake_case_ : str = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertNotIn(self.tokenizer.eos_token , lowerCAmelCase__ ) def _A ( self :Tuple ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = tempfile.mkdtemp() snake_case_ : int = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(lowerCAmelCase__ ) snake_case_ : List[str] = MaMaaaTokenizer.from_pretrained(lowerCAmelCase__ ) self.assertDictEqual(new_tok.lang_token_to_id , lowerCAmelCase__ ) @require_torch def _A ( self :Optional[Any] ) -> str: '''simple docstring''' snake_case_ : Union[str, Any] = "en" snake_case_ : Tuple = "fr" snake_case_ : Optional[int] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=lowerCAmelCase__ , return_tensors="pt" ) snake_case_ : Dict = shift_tokens_right( batch["labels"] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: snake_case_ : str = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : List[str] = "mr" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) snake_case_ : int = "zh" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def _A ( self :str ) -> int: '''simple docstring''' snake_case_ : Dict = "mr" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) snake_case_ : Tuple = "zh" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Optional[int] = self.tokenizer._build_translation_inputs("A test" , return_tensors="pt" , src_lang="en" , tgt_lang="ar" ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { # en_XX, A, test, EOS "input_ids": [[128_022, 58, 4_183, 2]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 128_006, } , )
656
0
'''simple docstring''' from __future__ import annotations from math import pi def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> dict[str, float]: """simple docstring""" if (inductance, frequency, reactance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if inductance < 0: raise ValueError("Inductance cannot be negative" ) if frequency < 0: raise ValueError("Frequency cannot be negative" ) if reactance < 0: raise ValueError("Inductive reactance cannot be negative" ) if inductance == 0: return {"inductance": reactance / (2 * pi * frequency)} elif frequency == 0: return {"frequency": reactance / (2 * pi * inductance)} elif reactance == 0: return {"reactance": 2 * pi * frequency * inductance} else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
713
'''simple docstring''' import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py __lowerCamelCase : str = '''.''' # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) __lowerCamelCase : Tuple = [ '''Assert''', '''AssignVariableOp''', '''EmptyTensorList''', '''MergeV2Checkpoints''', '''ReadVariableOp''', '''ResourceGather''', '''RestoreV2''', '''SaveV2''', '''ShardedFilename''', '''StatefulPartitionedCall''', '''StaticRegexFullMatch''', '''VarHandleOp''', ] def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Tuple = SavedModel() snake_case_ : Dict = [] with open(os.path.join(__magic_name__ ,"utils" ,"tf_ops" ,"onnx.json" ) ) as f: snake_case_ : Dict = json.load(__magic_name__ )["opsets"] for i in range(1 ,opset + 1 ): onnx_ops.extend(onnx_opsets[str(__magic_name__ )] ) with open(__magic_name__ ,"rb" ) as f: saved_model.ParseFromString(f.read() ) snake_case_ : Tuple = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want snake_case_ : str = sorted(__magic_name__ ) snake_case_ : Optional[int] = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(__magic_name__ ) if strict and len(__magic_name__ ) > 0: raise Exception(F'''Found the following incompatible ops for the opset {opset}:\n''' + incompatible_ops ) elif len(__magic_name__ ) > 0: print(F'''Found the following incompatible ops for the opset {opset}:''' ) print(*__magic_name__ ,sep="\n" ) else: print(F'''The saved model {saved_model_path} can properly be converted with ONNX.''' ) if __name__ == "__main__": __lowerCamelCase : Any = argparse.ArgumentParser() parser.add_argument('''--saved_model_path''', help='''Path of the saved model to check (the .pb file).''') parser.add_argument( '''--opset''', default=12, type=int, help='''The ONNX opset against which the model has to be tested.''' ) parser.add_argument( '''--framework''', choices=['''onnx'''], default='''onnx''', help='''Frameworks against which to test the saved model.''' ) parser.add_argument( '''--strict''', action='''store_true''', help='''Whether make the checking strict (raise errors) or not (raise warnings)''' ) __lowerCamelCase : Dict = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
656
0
from __future__ import annotations from random import random from typing import Generic, TypeVar __lowerCamelCase : Optional[int] = TypeVar('''KT''') __lowerCamelCase : Tuple = TypeVar('''VT''') class A_ (Generic[KT, VT] ): """simple docstring""" def __init__( self :List[str] , lowerCAmelCase__ :List[Any] = "root" , lowerCAmelCase__ :int = None ) -> Dict: '''simple docstring''' snake_case_ : Optional[int] = key snake_case_ : Optional[Any] = value snake_case_ : str = [] def __repr__( self :int ) -> str: '''simple docstring''' return F'''Node({self.key}: {self.value})''' @property def _A ( self :Optional[Any] ) -> int: '''simple docstring''' return len(self.forward ) class A_ (Generic[KT, VT] ): """simple docstring""" def __init__( self :List[str] , lowerCAmelCase__ :List[Any] = 0.5 , lowerCAmelCase__ :Dict = 16 ) -> int: '''simple docstring''' snake_case_ : Dict = Node[KT, VT]() snake_case_ : Tuple = 0 snake_case_ : List[str] = p snake_case_ : Tuple = max_level def __str__( self :Dict ) -> str: '''simple docstring''' snake_case_ : Dict = list(self ) if len(__A ) == 0: return F'''SkipList(level={self.level})''' snake_case_ : str = max((len(str(__A ) ) for item in items) , default=4 ) snake_case_ : List[Any] = max(__A , 4 ) + 4 snake_case_ : Optional[Any] = self.head snake_case_ : List[str] = [] snake_case_ : Union[str, Any] = node.forward.copy() lines.append(F'''[{node.key}]'''.ljust(__A , "-" ) + "* " * len(__A ) ) lines.append(" " * label_size + "| " * len(__A ) ) while len(node.forward ) != 0: snake_case_ : List[str] = node.forward[0] lines.append( F'''[{node.key}]'''.ljust(__A , "-" ) + " ".join(str(n.key ) if n.key == node.key else "|" for n in forwards ) ) lines.append(" " * label_size + "| " * len(__A ) ) snake_case_ : Any = node.forward lines.append("None".ljust(__A ) + "* " * len(__A ) ) return F'''SkipList(level={self.level})\n''' + "\n".join(__A ) def __iter__( self :Dict ) -> int: '''simple docstring''' snake_case_ : int = self.head while len(node.forward ) != 0: yield node.forward[0].key snake_case_ : int = node.forward[0] def _A ( self :int ) -> int: '''simple docstring''' snake_case_ : List[Any] = 1 while random() < self.p and level < self.max_level: level += 1 return level def _A ( self :Any , lowerCAmelCase__ :List[str] ) -> tuple[Node[KT, VT] | None, list[Node[KT, VT]]]: '''simple docstring''' snake_case_ : int = [] snake_case_ : Union[str, Any] = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: snake_case_ : Optional[Any] = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(__A ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def _A ( self :List[Any] , lowerCAmelCase__ :Optional[int] ) -> int: '''simple docstring''' snake_case_, snake_case_ : Dict = self._locate_node(__A ) if node is not None: for i, update_node in enumerate(__A ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: snake_case_ : Any = node.forward[i] else: snake_case_ : List[Any] = update_node.forward[:i] def _A ( self :Dict , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Tuple ) -> Tuple: '''simple docstring''' snake_case_, snake_case_ : Dict = self._locate_node(__A ) if node is not None: snake_case_ : List[str] = value else: snake_case_ : Optional[Any] = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , __A ): update_vector.append(self.head ) snake_case_ : str = level snake_case_ : int = Node(__A , __A ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(__A ) else: snake_case_ : Union[str, Any] = new_node def _A ( self :Optional[Any] , lowerCAmelCase__ :int ) -> VT | None: '''simple docstring''' snake_case_, snake_case_ : List[str] = self._locate_node(__A ) if node is not None: return node.value return None def __UpperCAmelCase ( )-> List[str]: """simple docstring""" snake_case_ : List[Any] = SkipList() skip_list.insert("Key1" ,3 ) skip_list.insert("Key2" ,12 ) skip_list.insert("Key3" ,41 ) skip_list.insert("Key4" ,-19 ) snake_case_ : Any = skip_list.head snake_case_ : Union[str, Any] = {} while node.level != 0: snake_case_ : Union[str, Any] = node.forward[0] snake_case_ : int = node.value assert len(a__ ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 12 assert all_values["Key3"] == 41 assert all_values["Key4"] == -19 def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : Optional[Any] = SkipList() skip_list.insert("Key1" ,10 ) skip_list.insert("Key1" ,12 ) skip_list.insert("Key5" ,7 ) skip_list.insert("Key7" ,10 ) skip_list.insert("Key10" ,5 ) skip_list.insert("Key7" ,7 ) skip_list.insert("Key5" ,5 ) skip_list.insert("Key10" ,10 ) snake_case_ : Union[str, Any] = skip_list.head snake_case_ : str = {} while node.level != 0: snake_case_ : List[str] = node.forward[0] snake_case_ : Tuple = node.value if len(a__ ) != 4: print() assert len(a__ ) == 4 assert all_values["Key1"] == 12 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 10 def __UpperCAmelCase ( )-> List[Any]: """simple docstring""" snake_case_ : Dict = SkipList() assert skip_list.find("Some key" ) is None def __UpperCAmelCase ( )-> str: """simple docstring""" snake_case_ : Optional[int] = SkipList() skip_list.insert("Key2" ,20 ) assert skip_list.find("Key2" ) == 20 skip_list.insert("Some Key" ,10 ) skip_list.insert("Key2" ,8 ) skip_list.insert("V" ,13 ) assert skip_list.find("Y" ) is None assert skip_list.find("Key2" ) == 8 assert skip_list.find("Some Key" ) == 10 assert skip_list.find("V" ) == 13 def __UpperCAmelCase ( )-> List[Any]: """simple docstring""" snake_case_ : Dict = SkipList() skip_list.delete("Some key" ) assert len(skip_list.head.forward ) == 0 def __UpperCAmelCase ( )-> Union[str, Any]: """simple docstring""" snake_case_ : str = SkipList() skip_list.insert("Key1" ,12 ) skip_list.insert("V" ,13 ) skip_list.insert("X" ,14 ) skip_list.insert("Key2" ,15 ) skip_list.delete("V" ) skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("Key2" ) is None def __UpperCAmelCase ( )-> str: """simple docstring""" snake_case_ : Optional[Any] = SkipList() skip_list.insert("Key1" ,12 ) skip_list.insert("V" ,13 ) skip_list.insert("X" ,14 ) skip_list.insert("Key2" ,15 ) skip_list.delete("V" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) == 14 assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("X" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("Key1" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) == 15 skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) is None def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : int = SkipList() skip_list.insert("Key1" ,12 ) skip_list.insert("V" ,13 ) skip_list.insert("X" ,142 ) skip_list.insert("Key2" ,15 ) skip_list.delete("X" ) def traverse_keys(__magic_name__ ): yield node.key for forward_node in node.forward: yield from traverse_keys(a__ ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def __UpperCAmelCase ( )-> Optional[Any]: """simple docstring""" def is_sorted(__magic_name__ ): return all(next_item >= item for item, next_item in zip(a__ ,lst[1:] ) ) snake_case_ : Dict = SkipList() for i in range(10 ): skip_list.insert(a__ ,a__ ) assert is_sorted(list(a__ ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(a__ ) ) skip_list.insert(-12 ,-12 ) skip_list.insert(77 ,77 ) assert is_sorted(list(a__ ) ) def __UpperCAmelCase ( )-> Dict: """simple docstring""" for _ in range(100 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def __UpperCAmelCase ( )-> str: """simple docstring""" snake_case_ : Tuple = SkipList() skip_list.insert(2 ,"2" ) skip_list.insert(4 ,"4" ) skip_list.insert(6 ,"4" ) skip_list.insert(4 ,"5" ) skip_list.insert(8 ,"4" ) skip_list.insert(9 ,"4" ) skip_list.delete(4 ) print(a__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
714
'''simple docstring''' import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal __lowerCamelCase : Optional[Any] = datasets.utils.logging.get_logger(__name__) __lowerCamelCase : List[str] = ['''names''', '''prefix'''] __lowerCamelCase : int = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] __lowerCamelCase : str = ['''encoding_errors''', '''on_bad_lines'''] __lowerCamelCase : Optional[Any] = ['''date_format'''] @dataclass class A_ (datasets.BuilderConfig ): """simple docstring""" a__ = "," a__ = None a__ = "infer" a__ = None a__ = None a__ = None a__ = None a__ = None a__ = True a__ = None a__ = None a__ = None a__ = None a__ = False a__ = None a__ = None a__ = None a__ = True a__ = True a__ = False a__ = True a__ = None a__ = "." a__ = None a__ = '"' a__ = 0 a__ = None a__ = None a__ = None a__ = None a__ = True a__ = True a__ = 0 a__ = True a__ = False a__ = None a__ = 10000 a__ = None a__ = "strict" a__ = "error" a__ = None def _A ( self :List[str] ) -> Any: '''simple docstring''' if self.delimiter is not None: snake_case_ : Tuple = self.delimiter if self.column_names is not None: snake_case_ : List[Any] = self.column_names @property def _A ( self :Optional[Any] ) -> int: '''simple docstring''' snake_case_ : Optional[int] = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class A_ (datasets.ArrowBasedBuilder ): """simple docstring""" a__ = CsvConfig def _A ( self :Optional[Any] ) -> Optional[Any]: '''simple docstring''' return datasets.DatasetInfo(features=self.config.features ) def _A ( self :Tuple , lowerCAmelCase__ :Dict ) -> List[Any]: '''simple docstring''' if not self.config.data_files: raise ValueError(F'''At least one data file must be specified, but got data_files={self.config.data_files}''' ) snake_case_ : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): snake_case_ : int = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ : List[str] = [files] snake_case_ : Tuple = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] snake_case_ : str = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ : str = [files] snake_case_ : Any = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def _A ( self :List[Any] , lowerCAmelCase__ :pa.Table ) -> pa.Table: '''simple docstring''' if self.config.features is not None: snake_case_ : int = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast snake_case_ : Optional[Any] = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example snake_case_ : Dict = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def _A ( self :Dict , lowerCAmelCase__ :Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str snake_case_ : str = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): snake_case_ : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): snake_case_ : Optional[int] = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(F'''Failed to read file \'{file}\' with error {type(lowerCAmelCase__ )}: {e}''' ) raise
656
0
import pytest __lowerCamelCase : List[str] = '''__dummy_dataset1__''' __lowerCamelCase : List[str] = '''\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/"\nURLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n "tokens": datasets.Sequence(datasets.Value("string")),\n "ner_tags": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n "O",\n "B-PER",\n "I-PER",\n "B-ORG",\n "I-ORG",\n "B-LOC",\n "I-LOC",\n ]\n )\n ),\n "langs": datasets.Sequence(datasets.Value("string")),\n "spans": datasets.Sequence(datasets.Value("string")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, "r", encoding="utf-8") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n''' @pytest.fixture def __UpperCAmelCase ( )-> Optional[Any]: """simple docstring""" return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def __UpperCAmelCase ( )-> List[str]: """simple docstring""" return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : Any = dataset_loading_script_name snake_case_ : int = tmp_path / "datasets" / script_name script_dir.mkdir(parents=UpperCamelCase__ ) snake_case_ : Any = script_dir / F'''{script_name}.py''' with open(UpperCamelCase__ ,"w" ) as f: f.write(UpperCamelCase__ ) return str(UpperCamelCase__ )
715
'''simple docstring''' import json import os import unittest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A_ (a_ , unittest.TestCase ): """simple docstring""" a__ = MgpstrTokenizer a__ = False a__ = {} a__ = False def _A ( self :List[str] ) -> List[str]: '''simple docstring''' super().setUp() # fmt: off snake_case_ : Dict = ["[GO]", "[s]", "0", "1", "2", "3", "4", "5", "6", "7", "8", "9", "a", "b", "c", "d", "e", "f", "g", "h", "i", "j", "k", "l", "m", "n", "o", "p", "q", "r", "s", "t", "u", "v", "w", "x", "y", "z"] # fmt: on snake_case_ : List[str] = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) snake_case_ : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase__ ) + "\n" ) def _A ( self :Optional[Any] , **lowerCAmelCase__ :Optional[Any] ) -> Dict: '''simple docstring''' return MgpstrTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def _A ( self :Dict , lowerCAmelCase__ :Any ) -> str: '''simple docstring''' snake_case_ : Dict = "tester" snake_case_ : Tuple = "tester" return input_text, output_text @unittest.skip("MGP-STR always lower cases letters." ) def _A ( self :Dict ) -> str: '''simple docstring''' pass def _A ( self :Tuple ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[str] = self.get_tokenizers(do_lower_case=lowerCAmelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): snake_case_ : Tuple = "[SPECIAL_TOKEN]" tokenizer.add_special_tokens({"cls_token": special_token} ) snake_case_ : str = tokenizer.encode([special_token] , add_special_tokens=lowerCAmelCase__ ) self.assertEqual(len(lowerCAmelCase__ ) , 1 ) snake_case_ : Tuple = tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) self.assertTrue(special_token not in decoded ) def _A ( self :int ) -> List[str]: '''simple docstring''' snake_case_ : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): snake_case_, snake_case_ : str = self.get_input_output_texts(lowerCAmelCase__ ) snake_case_ : Union[str, Any] = tokenizer.tokenize(lowerCAmelCase__ ) snake_case_ : List[Any] = tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) snake_case_ : Dict = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : List[str] = tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) self.assertNotEqual(len(lowerCAmelCase__ ) , 0 ) snake_case_ : List[str] = tokenizer.decode(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertEqual(text_a.replace(" " , "" ) , lowerCAmelCase__ ) @unittest.skip("MGP-STR tokenizer only handles one sequence." ) def _A ( self :Union[str, Any] ) -> Any: '''simple docstring''' pass @unittest.skip("inputs cannot be pretokenized in MgpstrTokenizer" ) def _A ( self :int ) -> Dict: '''simple docstring''' pass
656
0
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : List[str] = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) __lowerCamelCase : Union[str, Any] = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', f'''encoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', f'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.weight''', f'''encoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.bias''', f'''encoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.weight''', f'''encoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.bias''', f'''encoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.encoder.layers.{i}.norm1.weight''', f'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.norm1.bias''', f'''encoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.weight''', f'''encoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.bias''', f'''encoder.layers.{i}.final_layer_norm.bias''')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', f'''decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', f'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.cross_attn.out_proj.weight''', f'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.cross_attn.out_proj.bias''', f'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.weight''', f'''decoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.bias''', f'''decoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.weight''', f'''decoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.bias''', f'''decoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm1.weight''', f'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm1.bias''', f'''decoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.weight''', f'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.bias''', f'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.weight''', f'''decoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.bias''', f'''decoder.layers.{i}.final_layer_norm.bias''')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qcontent_proj.weight''', f'''decoder.layers.{i}.sa_qcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kcontent_proj.weight''', f'''decoder.layers.{i}.sa_kcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qpos_proj.weight''', f'''decoder.layers.{i}.sa_qpos_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kpos_proj.weight''', f'''decoder.layers.{i}.sa_kpos_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_v_proj.weight''', f'''decoder.layers.{i}.sa_v_proj.weight''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qcontent_proj.weight''', f'''decoder.layers.{i}.ca_qcontent_proj.weight''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kcontent_proj.weight''', f'''decoder.layers.{i}.ca_kcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kpos_proj.weight''', f'''decoder.layers.{i}.ca_kpos_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_v_proj.weight''', f'''decoder.layers.{i}.ca_v_proj.weight''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight''', f'''decoder.layers.{i}.ca_qpos_sine_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qcontent_proj.bias''', f'''decoder.layers.{i}.sa_qcontent_proj.bias''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kcontent_proj.bias''', f'''decoder.layers.{i}.sa_kcontent_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_qpos_proj.bias''', f'''decoder.layers.{i}.sa_qpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_kpos_proj.bias''', f'''decoder.layers.{i}.sa_kpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_v_proj.bias''', f'''decoder.layers.{i}.sa_v_proj.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qcontent_proj.bias''', f'''decoder.layers.{i}.ca_qcontent_proj.bias''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kcontent_proj.bias''', f'''decoder.layers.{i}.ca_kcontent_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_kpos_proj.bias''', f'''decoder.layers.{i}.ca_kpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_v_proj.bias''', f'''decoder.layers.{i}.ca_v_proj.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias''', f'''decoder.layers.{i}.ca_qpos_sine_proj.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : List[Any] = state_dict.pop(_UpperCAmelCase ) snake_case_ : Optional[Any] = val def __UpperCAmelCase ( __magic_name__ )-> Optional[int]: """simple docstring""" snake_case_ : str = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: snake_case_ : List[str] = key.replace("backbone.0.body" ,"backbone.conv_encoder.model" ) snake_case_ : List[str] = value else: snake_case_ : List[Any] = value return new_state_dict def __UpperCAmelCase ( __magic_name__ ,__magic_name__=False )-> Optional[Any]: """simple docstring""" snake_case_ : Tuple = "" if is_panoptic: snake_case_ : Any = "conditional_detr." # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) snake_case_ : Optional[Any] = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) snake_case_ : Any = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case_ : Tuple = in_proj_weight[:256, :] snake_case_ : List[str] = in_proj_bias[:256] snake_case_ : List[Any] = in_proj_weight[256:512, :] snake_case_ : Optional[Any] = in_proj_bias[256:512] snake_case_ : str = in_proj_weight[-256:, :] snake_case_ : str = in_proj_bias[-256:] def __UpperCAmelCase ( )-> Tuple: """simple docstring""" snake_case_ : str = "http://images.cocodataset.org/val2017/000000039769.jpg" snake_case_ : Dict = Image.open(requests.get(_UpperCAmelCase ,stream=_UpperCAmelCase ).raw ) return im @torch.no_grad() def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> int: """simple docstring""" snake_case_ : List[Any] = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: snake_case_ : int = "resnet101" if "dc5" in model_name: snake_case_ : int = True snake_case_ : int = "panoptic" in model_name if is_panoptic: snake_case_ : Any = 250 else: snake_case_ : Tuple = 91 snake_case_ : Any = "huggingface/label-files" snake_case_ : Union[str, Any] = "coco-detection-id2label.json" snake_case_ : Dict = json.load(open(hf_hub_download(_UpperCAmelCase ,_UpperCAmelCase ,repo_type="dataset" ) ,"r" ) ) snake_case_ : Any = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} snake_case_ : int = idalabel snake_case_ : Optional[Any] = {v: k for k, v in idalabel.items()} # load image processor snake_case_ : Union[str, Any] = "coco_panoptic" if is_panoptic else "coco_detection" snake_case_ : Optional[Any] = ConditionalDetrImageProcessor(format=_UpperCAmelCase ) # prepare image snake_case_ : Dict = prepare_img() snake_case_ : str = image_processor(images=_UpperCAmelCase ,return_tensors="pt" ) snake_case_ : List[Any] = encoding["pixel_values"] logger.info(F'''Converting model {model_name}...''' ) # load original model from torch hub snake_case_ : List[Any] = torch.hub.load("DeppMeng/ConditionalDETR" ,_UpperCAmelCase ,pretrained=_UpperCAmelCase ).eval() snake_case_ : Any = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: snake_case_ : Optional[int] = "conditional_detr." + src rename_key(_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ) snake_case_ : int = rename_backbone_keys(_UpperCAmelCase ) # query, key and value matrices need special treatment read_in_q_k_v(_UpperCAmelCase ,is_panoptic=_UpperCAmelCase ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them snake_case_ : List[str] = "conditional_detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("conditional_detr" ) and not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ) ): snake_case_ : Union[str, Any] = state_dict.pop(_UpperCAmelCase ) snake_case_ : Optional[int] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: snake_case_ : str = state_dict.pop(_UpperCAmelCase ) snake_case_ : Optional[int] = val elif key.startswith("bbox_attention" ) or key.startswith("mask_head" ): continue else: snake_case_ : Dict = state_dict.pop(_UpperCAmelCase ) snake_case_ : List[Any] = val else: if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): snake_case_ : Union[str, Any] = state_dict.pop(_UpperCAmelCase ) snake_case_ : Tuple = val # finally, create HuggingFace model and load state dict snake_case_ : Dict = ConditionalDetrForSegmentation(_UpperCAmelCase ) if is_panoptic else ConditionalDetrForObjectDetection(_UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) model.eval() model.push_to_hub(repo_id=_UpperCAmelCase ,organization="DepuMeng" ,commit_message="Add model" ) # verify our conversion snake_case_ : Any = conditional_detr(_UpperCAmelCase ) snake_case_ : Optional[int] = model(_UpperCAmelCase ) assert torch.allclose(outputs.logits ,original_outputs["pred_logits"] ,atol=1E-4 ) assert torch.allclose(outputs.pred_boxes ,original_outputs["pred_boxes"] ,atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks ,original_outputs["pred_masks"] ,atol=1E-4 ) # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) image_processor.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": __lowerCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) __lowerCamelCase : List[Any] = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
716
'''simple docstring''' from __future__ import annotations import math import numpy as np from numpy.linalg import norm def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> float: """simple docstring""" return math.sqrt(sum(pow(a - b ,2 ) for a, b in zip(__magic_name__ ,__magic_name__ ) ) ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> list[list[list[float] | float]]: """simple docstring""" if dataset.ndim != value_array.ndim: snake_case_ : int = ( "Wrong input data's dimensions... " F'''dataset : {dataset.ndim}, value_array : {value_array.ndim}''' ) raise ValueError(__magic_name__ ) try: if dataset.shape[1] != value_array.shape[1]: snake_case_ : Dict = ( "Wrong input data's shape... " F'''dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}''' ) raise ValueError(__magic_name__ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError("Wrong shape" ) if dataset.dtype != value_array.dtype: snake_case_ : Dict = ( "Input data have different datatype... " F'''dataset : {dataset.dtype}, value_array : {value_array.dtype}''' ) raise TypeError(__magic_name__ ) snake_case_ : Optional[int] = [] for value in value_array: snake_case_ : List[str] = euclidean(__magic_name__ ,dataset[0] ) snake_case_ : int = dataset[0].tolist() for dataset_value in dataset[1:]: snake_case_ : Optional[Any] = euclidean(__magic_name__ ,__magic_name__ ) if dist > temp_dist: snake_case_ : Tuple = temp_dist snake_case_ : Optional[int] = dataset_value.tolist() answer.append([vector, dist] ) return answer def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> float: """simple docstring""" return np.dot(__magic_name__ ,__magic_name__ ) / (norm(__magic_name__ ) * norm(__magic_name__ )) if __name__ == "__main__": import doctest doctest.testmod()
656
0
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> list[int]: """simple docstring""" snake_case_ : List[str] = int(__magic_name__ ) # Initialize Result snake_case_ : Optional[Any] = [] # Traverse through all denomination for denomination in reversed(__magic_name__ ): # Find denominations while int(__magic_name__ ) >= int(__magic_name__ ): total_value -= int(__magic_name__ ) answer.append(__magic_name__ ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": __lowerCamelCase : Optional[Any] = [] __lowerCamelCase : List[Any] = '''0''' if ( input('''Do you want to enter your denominations ? (yY/n): ''').strip().lower() == "y" ): __lowerCamelCase : Union[str, Any] = int(input('''Enter the number of denominations you want to add: ''').strip()) for i in range(0, n): denominations.append(int(input(f'''Denomination {i}: ''').strip())) __lowerCamelCase : Dict = input('''Enter the change you want to make in Indian Currency: ''').strip() else: # All denominations of Indian Currency if user does not enter __lowerCamelCase : Optional[int] = [1, 2, 5, 10, 20, 50, 100, 500, 2000] __lowerCamelCase : Dict = input('''Enter the change you want to make: ''').strip() if int(value) == 0 or int(value) < 0: print('''The total value cannot be zero or negative.''') else: print(f'''Following is minimal change for {value}: ''') __lowerCamelCase : Optional[int] = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=''' ''')
717
'''simple docstring''' import fire from utils import calculate_rouge, save_json def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__=None ,**__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : int = [x.strip() for x in open(__magic_name__ ).readlines()] snake_case_ : Optional[int] = [x.strip() for x in open(__magic_name__ ).readlines()][: len(__magic_name__ )] snake_case_ : List[Any] = calculate_rouge(__magic_name__ ,__magic_name__ ,**__magic_name__ ) if save_path is not None: save_json(__magic_name__ ,__magic_name__ ,indent=__magic_name__ ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
656
0
'''simple docstring''' import argparse import json import logging import os import shutil import sys import tempfile import unittest from unittest import mock import torch from accelerate.utils import write_basic_config from transformers.testing_utils import TestCasePlus, get_gpu_count, run_command, slow, torch_device from transformers.utils import is_apex_available logging.basicConfig(level=logging.DEBUG) __lowerCamelCase : Optional[Any] = logging.getLogger() def __UpperCAmelCase ( )-> Tuple: """simple docstring""" snake_case_ : Union[str, Any] = argparse.ArgumentParser() parser.add_argument("-f" ) snake_case_ : int = parser.parse_args() return args.f def __UpperCAmelCase ( __magic_name__ )-> Union[str, Any]: """simple docstring""" snake_case_ : Dict = {} snake_case_ : Union[str, Any] = os.path.join(_lowercase ,"all_results.json" ) if os.path.exists(_lowercase ): with open(_lowercase ,"r" ) as f: snake_case_ : Tuple = json.load(_lowercase ) else: raise ValueError(F'''can\'t find {path}''' ) return results def __UpperCAmelCase ( )-> Optional[int]: """simple docstring""" snake_case_ : str = torch.cuda.is_available() and torch_device == "cuda" return is_using_cuda and is_apex_available() __lowerCamelCase : int = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class A_ (a_ ): """simple docstring""" @classmethod def _A ( cls :List[Any] ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[Any] = tempfile.mkdtemp() snake_case_ : Dict = os.path.join(cls.tmpdir , "default_config.yml" ) write_basic_config(save_location=cls.configPath ) snake_case_ : Tuple = ["accelerate", "launch", "--config_file", cls.configPath] @classmethod def _A ( cls :List[str] ) -> Optional[Any]: '''simple docstring''' shutil.rmtree(cls.tmpdir ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def _A ( self :Optional[int] ) -> List[Any]: '''simple docstring''' snake_case_ : Tuple = self.get_auto_remove_tmp_dir() snake_case_ : Union[str, Any] = F''' {self.examples_dir}/pytorch/text-classification/run_glue_no_trainer.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --seed=42 --checkpointing_steps epoch --with_tracking '''.split() if is_cuda_and_apex_available(): testargs.append("--fp16" ) run_command(self._launch_args + testargs ) snake_case_ : Any = get_results(__A ) self.assertGreaterEqual(result["eval_accuracy"] , 0.7_5 ) self.assertTrue(os.path.exists(os.path.join(__A , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(__A , "glue_no_trainer" ) ) ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def _A ( self :int ) -> List[str]: '''simple docstring''' snake_case_ : Dict = self.get_auto_remove_tmp_dir() snake_case_ : Tuple = F''' {self.examples_dir}/pytorch/language-modeling/run_clm_no_trainer.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --block_size 128 --per_device_train_batch_size 5 --per_device_eval_batch_size 5 --num_train_epochs 2 --output_dir {tmp_dir} --checkpointing_steps epoch --with_tracking '''.split() if torch.cuda.device_count() > 1: # Skipping because there are not enough batches to train the model + would need a drop_last to work. return run_command(self._launch_args + testargs ) snake_case_ : List[str] = get_results(__A ) self.assertLess(result["perplexity"] , 100 ) self.assertTrue(os.path.exists(os.path.join(__A , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(__A , "clm_no_trainer" ) ) ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def _A ( self :int ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = self.get_auto_remove_tmp_dir() snake_case_ : Any = F''' {self.examples_dir}/pytorch/language-modeling/run_mlm_no_trainer.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --num_train_epochs=1 --checkpointing_steps epoch --with_tracking '''.split() run_command(self._launch_args + testargs ) snake_case_ : Any = get_results(__A ) self.assertLess(result["perplexity"] , 42 ) self.assertTrue(os.path.exists(os.path.join(__A , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(__A , "mlm_no_trainer" ) ) ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def _A ( self :Optional[int] ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = 7 if get_gpu_count() > 1 else 2 snake_case_ : Tuple = self.get_auto_remove_tmp_dir() snake_case_ : Union[str, Any] = F''' {self.examples_dir}/pytorch/token-classification/run_ner_no_trainer.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 --checkpointing_steps epoch --with_tracking '''.split() run_command(self._launch_args + testargs ) snake_case_ : Dict = get_results(__A ) self.assertGreaterEqual(result["eval_accuracy"] , 0.7_5 ) self.assertLess(result["train_loss"] , 0.5 ) self.assertTrue(os.path.exists(os.path.join(__A , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(__A , "ner_no_trainer" ) ) ) @unittest.skip(reason="Fix me @muellerzr" ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def _A ( self :Tuple ) -> int: '''simple docstring''' snake_case_ : Union[str, Any] = self.get_auto_remove_tmp_dir() snake_case_ : str = F''' {self.examples_dir}/pytorch/question-answering/run_qa_no_trainer.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --seed=42 --max_train_steps=10 --num_warmup_steps=2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --checkpointing_steps epoch --with_tracking '''.split() run_command(self._launch_args + testargs ) snake_case_ : Tuple = get_results(__A ) # Because we use --version_2_with_negative the testing script uses SQuAD v2 metrics. self.assertGreaterEqual(result["eval_f1"] , 28 ) self.assertGreaterEqual(result["eval_exact"] , 28 ) self.assertTrue(os.path.exists(os.path.join(__A , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(__A , "qa_no_trainer" ) ) ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def _A ( self :List[str] ) -> Tuple: '''simple docstring''' snake_case_ : List[Any] = self.get_auto_remove_tmp_dir() snake_case_ : List[Any] = F''' {self.examples_dir}/pytorch/multiple-choice/run_swag_no_trainer.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/swag/sample.json --validation_file tests/fixtures/tests_samples/swag/sample.json --output_dir {tmp_dir} --max_train_steps=20 --num_warmup_steps=2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --with_tracking '''.split() run_command(self._launch_args + testargs ) snake_case_ : int = get_results(__A ) self.assertGreaterEqual(result["eval_accuracy"] , 0.8 ) self.assertTrue(os.path.exists(os.path.join(__A , "swag_no_trainer" ) ) ) @slow @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def _A ( self :List[Any] ) -> Optional[Any]: '''simple docstring''' snake_case_ : str = self.get_auto_remove_tmp_dir() snake_case_ : Dict = F''' {self.examples_dir}/pytorch/summarization/run_summarization_no_trainer.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --max_train_steps=50 --num_warmup_steps=8 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --checkpointing_steps epoch --with_tracking '''.split() run_command(self._launch_args + testargs ) snake_case_ : Optional[int] = get_results(__A ) self.assertGreaterEqual(result["eval_rouge1"] , 10 ) self.assertGreaterEqual(result["eval_rouge2"] , 2 ) self.assertGreaterEqual(result["eval_rougeL"] , 7 ) self.assertGreaterEqual(result["eval_rougeLsum"] , 7 ) self.assertTrue(os.path.exists(os.path.join(__A , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(__A , "summarization_no_trainer" ) ) ) @slow @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def _A ( self :Any ) -> Union[str, Any]: '''simple docstring''' snake_case_ : int = self.get_auto_remove_tmp_dir() snake_case_ : Union[str, Any] = F''' {self.examples_dir}/pytorch/translation/run_translation_no_trainer.py --model_name_or_path sshleifer/student_marian_en_ro_6_1 --source_lang en --target_lang ro --train_file tests/fixtures/tests_samples/wmt16/sample.json --validation_file tests/fixtures/tests_samples/wmt16/sample.json --output_dir {tmp_dir} --max_train_steps=50 --num_warmup_steps=8 --num_beams=6 --learning_rate=3e-3 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --source_lang en_XX --target_lang ro_RO --checkpointing_steps epoch --with_tracking '''.split() run_command(self._launch_args + testargs ) snake_case_ : Union[str, Any] = get_results(__A ) self.assertGreaterEqual(result["eval_bleu"] , 30 ) self.assertTrue(os.path.exists(os.path.join(__A , "epoch_0" ) ) ) self.assertTrue(os.path.exists(os.path.join(__A , "translation_no_trainer" ) ) ) @slow def _A ( self :Union[str, Any] ) -> Tuple: '''simple docstring''' snake_case_ : List[Any] = logging.StreamHandler(sys.stdout ) logger.addHandler(__A ) snake_case_ : Any = self.get_auto_remove_tmp_dir() snake_case_ : Any = F''' {self.examples_dir}/pytorch/semantic-segmentation/run_semantic_segmentation_no_trainer.py --dataset_name huggingface/semantic-segmentation-test-sample --output_dir {tmp_dir} --max_train_steps=10 --num_warmup_steps=2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --checkpointing_steps epoch '''.split() run_command(self._launch_args + testargs ) snake_case_ : Union[str, Any] = get_results(__A ) self.assertGreaterEqual(result["eval_overall_accuracy"] , 0.1_0 ) @mock.patch.dict(os.environ , {"WANDB_MODE": "offline"} ) def _A ( self :str ) -> Dict: '''simple docstring''' snake_case_ : str = self.get_auto_remove_tmp_dir() snake_case_ : List[str] = F''' {self.examples_dir}/pytorch/image-classification/run_image_classification_no_trainer.py --model_name_or_path google/vit-base-patch16-224-in21k --dataset_name hf-internal-testing/cats_vs_dogs_sample --learning_rate 1e-4 --per_device_train_batch_size 2 --per_device_eval_batch_size 1 --max_train_steps 2 --train_val_split 0.1 --seed 42 --output_dir {tmp_dir} --with_tracking --checkpointing_steps 1 '''.split() if is_cuda_and_apex_available(): testargs.append("--fp16" ) run_command(self._launch_args + testargs ) snake_case_ : Optional[int] = get_results(__A ) # The base model scores a 25% self.assertGreaterEqual(result["eval_accuracy"] , 0.6 ) self.assertTrue(os.path.exists(os.path.join(__A , "step_1" ) ) ) self.assertTrue(os.path.exists(os.path.join(__A , "image_classification_no_trainer" ) ) )
718
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) __lowerCamelCase : Optional[Any] = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', f'''encoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', f'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.weight''', f'''encoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.bias''', f'''encoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.weight''', f'''encoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.bias''', f'''encoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.encoder.layers.{i}.norm1.weight''', f'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.norm1.bias''', f'''encoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.weight''', f'''encoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.bias''', f'''encoder.layers.{i}.final_layer_norm.bias''')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', f'''decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', f'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.cross_attn.out_proj.weight''', f'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.cross_attn.out_proj.bias''', f'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.weight''', f'''decoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.bias''', f'''decoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.weight''', f'''decoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.bias''', f'''decoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm1.weight''', f'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm1.bias''', f'''decoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.weight''', f'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.bias''', f'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.weight''', f'''decoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.bias''', f'''decoder.layers.{i}.final_layer_norm.bias''')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qcontent_proj.weight''', f'''decoder.layers.{i}.sa_qcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kcontent_proj.weight''', f'''decoder.layers.{i}.sa_kcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qpos_proj.weight''', f'''decoder.layers.{i}.sa_qpos_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kpos_proj.weight''', f'''decoder.layers.{i}.sa_kpos_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_v_proj.weight''', f'''decoder.layers.{i}.sa_v_proj.weight''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qcontent_proj.weight''', f'''decoder.layers.{i}.ca_qcontent_proj.weight''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kcontent_proj.weight''', f'''decoder.layers.{i}.ca_kcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kpos_proj.weight''', f'''decoder.layers.{i}.ca_kpos_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_v_proj.weight''', f'''decoder.layers.{i}.ca_v_proj.weight''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight''', f'''decoder.layers.{i}.ca_qpos_sine_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qcontent_proj.bias''', f'''decoder.layers.{i}.sa_qcontent_proj.bias''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kcontent_proj.bias''', f'''decoder.layers.{i}.sa_kcontent_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_qpos_proj.bias''', f'''decoder.layers.{i}.sa_qpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_kpos_proj.bias''', f'''decoder.layers.{i}.sa_kpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_v_proj.bias''', f'''decoder.layers.{i}.sa_v_proj.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qcontent_proj.bias''', f'''decoder.layers.{i}.ca_qcontent_proj.bias''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kcontent_proj.bias''', f'''decoder.layers.{i}.ca_kcontent_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_kpos_proj.bias''', f'''decoder.layers.{i}.ca_kpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_v_proj.bias''', f'''decoder.layers.{i}.ca_v_proj.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias''', f'''decoder.layers.{i}.ca_qpos_sine_proj.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> int: """simple docstring""" snake_case_ : Optional[Any] = state_dict.pop(__magic_name__ ) snake_case_ : Any = val def __UpperCAmelCase ( __magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : Any = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: snake_case_ : Optional[Any] = key.replace("backbone.0.body" ,"backbone.conv_encoder.model" ) snake_case_ : int = value else: snake_case_ : int = value return new_state_dict def __UpperCAmelCase ( __magic_name__ ,__magic_name__=False )-> Optional[int]: """simple docstring""" snake_case_ : str = "" if is_panoptic: snake_case_ : Dict = "conditional_detr." # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) snake_case_ : Any = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) snake_case_ : Optional[int] = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case_ : Tuple = in_proj_weight[:256, :] snake_case_ : List[Any] = in_proj_bias[:256] snake_case_ : Optional[Any] = in_proj_weight[256:512, :] snake_case_ : Optional[int] = in_proj_bias[256:512] snake_case_ : Optional[int] = in_proj_weight[-256:, :] snake_case_ : str = in_proj_bias[-256:] def __UpperCAmelCase ( )-> Optional[Any]: """simple docstring""" snake_case_ : Optional[int] = "http://images.cocodataset.org/val2017/000000039769.jpg" snake_case_ : Optional[Any] = Image.open(requests.get(__magic_name__ ,stream=__magic_name__ ).raw ) return im @torch.no_grad() def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Optional[Any] = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: snake_case_ : Optional[Any] = "resnet101" if "dc5" in model_name: snake_case_ : List[str] = True snake_case_ : Tuple = "panoptic" in model_name if is_panoptic: snake_case_ : List[Any] = 250 else: snake_case_ : Optional[Any] = 91 snake_case_ : Optional[int] = "huggingface/label-files" snake_case_ : Dict = "coco-detection-id2label.json" snake_case_ : List[Any] = json.load(open(hf_hub_download(__magic_name__ ,__magic_name__ ,repo_type="dataset" ) ,"r" ) ) snake_case_ : Optional[int] = {int(__magic_name__ ): v for k, v in idalabel.items()} snake_case_ : int = idalabel snake_case_ : Dict = {v: k for k, v in idalabel.items()} # load image processor snake_case_ : Optional[int] = "coco_panoptic" if is_panoptic else "coco_detection" snake_case_ : str = ConditionalDetrImageProcessor(format=__magic_name__ ) # prepare image snake_case_ : str = prepare_img() snake_case_ : int = image_processor(images=__magic_name__ ,return_tensors="pt" ) snake_case_ : Union[str, Any] = encoding["pixel_values"] logger.info(F'''Converting model {model_name}...''' ) # load original model from torch hub snake_case_ : Union[str, Any] = torch.hub.load("DeppMeng/ConditionalDETR" ,__magic_name__ ,pretrained=__magic_name__ ).eval() snake_case_ : Any = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: snake_case_ : Any = "conditional_detr." + src rename_key(__magic_name__ ,__magic_name__ ,__magic_name__ ) snake_case_ : Tuple = rename_backbone_keys(__magic_name__ ) # query, key and value matrices need special treatment read_in_q_k_v(__magic_name__ ,is_panoptic=__magic_name__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them snake_case_ : int = "conditional_detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("conditional_detr" ) and not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ) ): snake_case_ : Any = state_dict.pop(__magic_name__ ) snake_case_ : Optional[int] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: snake_case_ : Tuple = state_dict.pop(__magic_name__ ) snake_case_ : Any = val elif key.startswith("bbox_attention" ) or key.startswith("mask_head" ): continue else: snake_case_ : Union[str, Any] = state_dict.pop(__magic_name__ ) snake_case_ : List[Any] = val else: if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): snake_case_ : Any = state_dict.pop(__magic_name__ ) snake_case_ : List[Any] = val # finally, create HuggingFace model and load state dict snake_case_ : Optional[int] = ConditionalDetrForSegmentation(__magic_name__ ) if is_panoptic else ConditionalDetrForObjectDetection(__magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() model.push_to_hub(repo_id=__magic_name__ ,organization="DepuMeng" ,commit_message="Add model" ) # verify our conversion snake_case_ : Dict = conditional_detr(__magic_name__ ) snake_case_ : Union[str, Any] = model(__magic_name__ ) assert torch.allclose(outputs.logits ,original_outputs["pred_logits"] ,atol=1E-4 ) assert torch.allclose(outputs.pred_boxes ,original_outputs["pred_boxes"] ,atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks ,original_outputs["pred_masks"] ,atol=1E-4 ) # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) image_processor.save_pretrained(__magic_name__ ) if __name__ == "__main__": __lowerCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) __lowerCamelCase : int = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
656
0
__lowerCamelCase : Optional[Any] = '''ABCDEFGHIJKLMNOPQRSTUVWXYZ''' def __UpperCAmelCase ( )-> str: """simple docstring""" snake_case_ : Dict = input("Enter message: " ) snake_case_ : Union[str, Any] = input("Enter key [alphanumeric]: " ) snake_case_ : Optional[Any] = input("Encrypt/Decrypt [e/d]: " ) if mode.lower().startswith("e" ): snake_case_ : int = """encrypt""" snake_case_ : str = encrypt_message(_lowerCamelCase ,_lowerCamelCase ) elif mode.lower().startswith("d" ): snake_case_ : Optional[Any] = """decrypt""" snake_case_ : Optional[Any] = decrypt_message(_lowerCamelCase ,_lowerCamelCase ) print(F'''\n{mode.title()}ed message:''' ) print(_lowerCamelCase ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" return translate_message(_lowerCamelCase ,_lowerCamelCase ,"encrypt" ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Any: """simple docstring""" return translate_message(_lowerCamelCase ,_lowerCamelCase ,"decrypt" ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Optional[int] = [] snake_case_ : str = 0 snake_case_ : Union[str, Any] = key.upper() for symbol in message: snake_case_ : Dict = LETTERS.find(symbol.upper() ) if num != -1: if mode == "encrypt": num += LETTERS.find(key[key_index] ) elif mode == "decrypt": num -= LETTERS.find(key[key_index] ) num %= len(_lowerCamelCase ) if symbol.isupper(): translated.append(LETTERS[num] ) elif symbol.islower(): translated.append(LETTERS[num].lower() ) key_index += 1 if key_index == len(_lowerCamelCase ): snake_case_ : Optional[int] = 0 else: translated.append(_lowerCamelCase ) return "".join(_lowerCamelCase ) if __name__ == "__main__": main()
719
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Any ) -> str: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _A ( self :List[Any] ) -> List[str]: '''simple docstring''' snake_case_ : Any = 1 snake_case_ : Dict = 3 snake_case_ : Union[str, Any] = (32, 32) snake_case_ : Optional[int] = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) return image @property def _A ( self :Optional[int] ) -> Any: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : List[str] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) return model @property def _A ( self :Dict ) -> Any: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) return model @property def _A ( self :Dict ) -> Optional[int]: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : str = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5_006 , ) return RobertaSeriesModelWithTransformation(lowerCAmelCase__ ) @property def _A ( self :Any ) -> str: '''simple docstring''' def extract(*lowerCAmelCase__ :Any , **lowerCAmelCase__ :List[str] ): class A_ : """simple docstring""" def __init__( self :Optional[int] ) -> List[str]: '''simple docstring''' snake_case_ : str = torch.ones([0] ) def _A ( self :int , lowerCAmelCase__ :List[Any] ) -> Tuple: '''simple docstring''' self.pixel_values.to(lowerCAmelCase__ ) return self return Out() return extract def _A ( self :int ) -> Dict: '''simple docstring''' snake_case_ : str = "cpu" # ensure determinism for the device-dependent torch.Generator snake_case_ : str = self.dummy_cond_unet snake_case_ : Optional[int] = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) snake_case_ : Dict = self.dummy_vae snake_case_ : Dict = self.dummy_text_encoder snake_case_ : Optional[int] = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) snake_case_ : str = 77 snake_case_ : Any = self.dummy_image.to(lowerCAmelCase__ ) snake_case_ : Tuple = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk snake_case_ : Optional[Any] = AltDiffusionImgaImgPipeline( unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , vae=lowerCAmelCase__ , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , feature_extractor=self.dummy_extractor , ) snake_case_ : Union[str, Any] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowerCAmelCase__ ) snake_case_ : Optional[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) snake_case_ : Dict = "A painting of a squirrel eating a burger" snake_case_ : List[str] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) snake_case_ : Dict = alt_pipe( [prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , ) snake_case_ : Any = output.images snake_case_ : List[str] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) snake_case_ : Optional[Any] = alt_pipe( [prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , return_dict=lowerCAmelCase__ , )[0] snake_case_ : Tuple = image[0, -3:, -3:, -1] snake_case_ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) snake_case_ : int = np.array([0.4_4_2_7, 0.3_7_3_1, 0.4_2_4_9, 0.4_9_4_1, 0.4_5_4_6, 0.4_1_4_8, 0.4_1_9_3, 0.4_6_6_6, 0.4_4_9_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5E-3 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def _A ( self :int ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = self.dummy_cond_unet snake_case_ : Union[str, Any] = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) snake_case_ : int = self.dummy_vae snake_case_ : List[Any] = self.dummy_text_encoder snake_case_ : int = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) snake_case_ : int = 77 snake_case_ : Dict = self.dummy_image.to(lowerCAmelCase__ ) # put models in fp16 snake_case_ : Optional[Any] = unet.half() snake_case_ : Tuple = vae.half() snake_case_ : List[str] = bert.half() # make sure here that pndm scheduler skips prk snake_case_ : Optional[int] = AltDiffusionImgaImgPipeline( unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , vae=lowerCAmelCase__ , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , feature_extractor=self.dummy_extractor , ) snake_case_ : List[str] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowerCAmelCase__ ) snake_case_ : Optional[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) snake_case_ : List[Any] = "A painting of a squirrel eating a burger" snake_case_ : str = torch.manual_seed(0 ) snake_case_ : Any = alt_pipe( [prompt] , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def _A ( self :Optional[int] ) -> Any: '''simple docstring''' snake_case_ : Union[str, Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) # resize to resolution that is divisible by 8 but not 16 or 32 snake_case_ : str = init_image.resize((760, 504) ) snake_case_ : Optional[Any] = "BAAI/AltDiffusion" snake_case_ : int = AltDiffusionImgaImgPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() snake_case_ : Tuple = "A fantasy landscape, trending on artstation" snake_case_ : int = torch.manual_seed(0 ) snake_case_ : List[str] = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , strength=0.7_5 , guidance_scale=7.5 , generator=lowerCAmelCase__ , output_type="np" , ) snake_case_ : str = output.images[0] snake_case_ : List[Any] = image[255:258, 383:386, -1] assert image.shape == (504, 760, 3) snake_case_ : Tuple = np.array([0.9_3_5_8, 0.9_3_9_7, 0.9_5_9_9, 0.9_9_0_1, 1.0_0_0_0, 1.0_0_0_0, 0.9_8_8_2, 1.0_0_0_0, 1.0_0_0_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _A ( self :str ) -> Any: '''simple docstring''' snake_case_ : Optional[Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) snake_case_ : List[Any] = init_image.resize((768, 512) ) snake_case_ : Tuple = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy" ) snake_case_ : Any = "BAAI/AltDiffusion" snake_case_ : List[str] = AltDiffusionImgaImgPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() snake_case_ : Tuple = "A fantasy landscape, trending on artstation" snake_case_ : Tuple = torch.manual_seed(0 ) snake_case_ : List[Any] = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , strength=0.7_5 , guidance_scale=7.5 , generator=lowerCAmelCase__ , output_type="np" , ) snake_case_ : Optional[int] = output.images[0] assert image.shape == (512, 768, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1E-2
656
0
'''simple docstring''' import os import random import sys from . import cryptomath_module as cryptomath from . import rabin_miller __lowerCamelCase : List[str] = 3 def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" print("Generating primitive root of p" ) while True: snake_case_ : Tuple = random.randrange(3 ,lowercase_ ) if pow(lowercase_ ,2 ,lowercase_ ) == 1: continue if pow(lowercase_ ,lowercase_ ,lowercase_ ) == 1: continue return g def __UpperCAmelCase ( __magic_name__ )-> tuple[tuple[int, int, int, int], tuple[int, int]]: """simple docstring""" print("Generating prime p..." ) snake_case_ : Union[str, Any] = rabin_miller.generate_large_prime(lowercase_ ) # select large prime number. snake_case_ : Optional[int] = primitive_root(lowercase_ ) # one primitive root on modulo p. snake_case_ : List[Any] = random.randrange(3 ,lowercase_ ) # private_key -> have to be greater than 2 for safety. snake_case_ : Tuple = cryptomath.find_mod_inverse(pow(lowercase_ ,lowercase_ ,lowercase_ ) ,lowercase_ ) snake_case_ : Union[str, Any] = (key_size, e_a, e_a, p) snake_case_ : str = (key_size, d) return public_key, private_key def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> None: """simple docstring""" if os.path.exists(F'''{name}_pubkey.txt''' ) or os.path.exists(F'''{name}_privkey.txt''' ): print("\nWARNING:" ) print( F'''\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n''' "Use a different name or delete these files and re-run this program." ) sys.exit() snake_case_ : Dict = generate_key(lowercase_ ) print(F'''\nWriting public key to file {name}_pubkey.txt...''' ) with open(F'''{name}_pubkey.txt''' ,"w" ) as fo: fo.write(F'''{public_key[0]},{public_key[1]},{public_key[2]},{public_key[3]}''' ) print(F'''Writing private key to file {name}_privkey.txt...''' ) with open(F'''{name}_privkey.txt''' ,"w" ) as fo: fo.write(F'''{private_key[0]},{private_key[1]}''' ) def __UpperCAmelCase ( )-> None: """simple docstring""" print("Making key files..." ) make_key_files("elgamal" ,2048 ) print("Key files generation successful" ) if __name__ == "__main__": main()
720
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. __lowerCamelCase : List[str] = {'''LayoutLMv2Config''', '''LayoutLMv3Config'''} @is_pipeline_test class A_ (unittest.TestCase ): """simple docstring""" a__ = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING a__ = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: a__ = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: a__ = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def _A ( self :Tuple , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :Dict ) -> Any: '''simple docstring''' snake_case_ : Optional[Any] = ZeroShotClassificationPipeline( model=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , candidate_labels=["polics", "health"] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def _A ( self :List[str] , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Any ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = classifier("Who are you voting for in 2020?" , candidate_labels="politics" ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) # No kwarg snake_case_ : List[Any] = classifier("Who are you voting for in 2020?" , ["politics"] ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) snake_case_ : Dict = classifier("Who are you voting for in 2020?" , candidate_labels=["politics"] ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) snake_case_ : int = classifier("Who are you voting for in 2020?" , candidate_labels="politics, public health" ) self.assertEqual( lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ) , 1.0 ) snake_case_ : Optional[int] = classifier("Who are you voting for in 2020?" , candidate_labels=["politics", "public health"] ) self.assertEqual( lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ) , 1.0 ) snake_case_ : str = classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template="This text is about {}" ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) # https://github.com/huggingface/transformers/issues/13846 snake_case_ : Dict = classifier(["I am happy"] , ["positive", "negative"] ) self.assertEqual( lowerCAmelCase__ , [ {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} for i in range(1 ) ] , ) snake_case_ : Tuple = classifier(["I am happy", "I am sad"] , ["positive", "negative"] ) self.assertEqual( lowerCAmelCase__ , [ {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} for i in range(2 ) ] , ) with self.assertRaises(lowerCAmelCase__ ): classifier("" , candidate_labels="politics" ) with self.assertRaises(lowerCAmelCase__ ): classifier(lowerCAmelCase__ , candidate_labels="politics" ) with self.assertRaises(lowerCAmelCase__ ): classifier("Who are you voting for in 2020?" , candidate_labels="" ) with self.assertRaises(lowerCAmelCase__ ): classifier("Who are you voting for in 2020?" , candidate_labels=lowerCAmelCase__ ) with self.assertRaises(lowerCAmelCase__ ): classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template="Not formatting template" , ) with self.assertRaises(lowerCAmelCase__ ): classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template=lowerCAmelCase__ , ) self.run_entailment_id(lowerCAmelCase__ ) def _A ( self :List[Any] , lowerCAmelCase__ :Pipeline ) -> Union[str, Any]: '''simple docstring''' snake_case_ : int = zero_shot_classifier.model.config snake_case_ : Optional[int] = config.labelaid snake_case_ : Tuple = zero_shot_classifier.entailment_id snake_case_ : Optional[Any] = {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2} self.assertEqual(zero_shot_classifier.entailment_id , -1 ) snake_case_ : Tuple = {"entailment": 0, "neutral": 1, "contradiction": 2} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) snake_case_ : str = {"ENTAIL": 0, "NON-ENTAIL": 1} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) snake_case_ : str = {"ENTAIL": 2, "NEUTRAL": 1, "CONTR": 0} self.assertEqual(zero_shot_classifier.entailment_id , 2 ) snake_case_ : List[str] = original_labelaid self.assertEqual(lowerCAmelCase__ , zero_shot_classifier.entailment_id ) @require_torch def _A ( self :Tuple ) -> Any: '''simple docstring''' snake_case_ : List[Any] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="pt" , ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( "Who are you voting for in 2020?" * 100 , candidate_labels=["politics", "public health", "science"] ) @require_torch def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="pt" , ) snake_case_ : int = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.3_3_3, 0.3_3_3, 0.3_3_3], } , ) @require_tf def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : List[str] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="tf" , ) snake_case_ : Optional[int] = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.3_3_3, 0.3_3_3, 0.3_3_3], } , ) @slow @require_torch def _A ( self :Union[str, Any] ) -> int: '''simple docstring''' snake_case_ : int = pipeline("zero-shot-classification" , model="roberta-large-mnli" , framework="pt" ) snake_case_ : str = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.9_7_6, 0.0_1_5, 0.0_0_9], } , ) snake_case_ : Optional[int] = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data." , candidate_labels=["machine learning", "statistics", "translation", "vision"] , multi_label=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], } , ) @slow @require_tf def _A ( self :List[str] ) -> str: '''simple docstring''' snake_case_ : int = pipeline("zero-shot-classification" , model="roberta-large-mnli" , framework="tf" ) snake_case_ : Optional[Any] = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.9_7_6, 0.0_1_5, 0.0_0_9], } , ) snake_case_ : Tuple = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data." , candidate_labels=["machine learning", "statistics", "translation", "vision"] , multi_label=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], } , )
656
0
'''simple docstring''' import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __lowerCamelCase : int = '''platform''' import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class A_ : """simple docstring""" a__ = PegasusConfig a__ = {} a__ = 'gelu' def __init__( self :Dict , lowerCAmelCase__ :int , lowerCAmelCase__ :int=13 , lowerCAmelCase__ :str=7 , lowerCAmelCase__ :Dict=True , lowerCAmelCase__ :Any=False , lowerCAmelCase__ :Dict=99 , lowerCAmelCase__ :int=32 , lowerCAmelCase__ :Dict=5 , lowerCAmelCase__ :List[str]=4 , lowerCAmelCase__ :Any=37 , lowerCAmelCase__ :List[Any]=0.1 , lowerCAmelCase__ :List[str]=0.1 , lowerCAmelCase__ :Any=20 , lowerCAmelCase__ :Union[str, Any]=2 , lowerCAmelCase__ :Dict=1 , lowerCAmelCase__ :Tuple=0 , ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = parent snake_case_ : int = batch_size snake_case_ : List[Any] = seq_length snake_case_ : int = is_training snake_case_ : Optional[int] = use_labels snake_case_ : List[str] = vocab_size snake_case_ : int = hidden_size snake_case_ : List[Any] = num_hidden_layers snake_case_ : Tuple = num_attention_heads snake_case_ : Any = intermediate_size snake_case_ : Optional[Any] = hidden_dropout_prob snake_case_ : Optional[int] = attention_probs_dropout_prob snake_case_ : Union[str, Any] = max_position_embeddings snake_case_ : Union[str, Any] = eos_token_id snake_case_ : int = pad_token_id snake_case_ : Optional[int] = bos_token_id def _A ( self :Dict ) -> List[Any]: '''simple docstring''' snake_case_ : List[str] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ).clip(3 , self.vocab_size ) snake_case_ : Any = np.expand_dims(np.array([self.eos_token_id] * self.batch_size ) , 1 ) snake_case_ : int = np.concatenate([input_ids, eos_tensor] , axis=1 ) snake_case_ : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ : Dict = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) snake_case_ : List[Any] = prepare_pegasus_inputs_dict(_a , _a , _a ) return config, inputs_dict def _A ( self :Tuple , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :str ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = 20 snake_case_ : List[str] = model_class_name(_a ) snake_case_ : List[str] = model.encode(inputs_dict["input_ids"] ) snake_case_, snake_case_ : List[str] = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) snake_case_ : List[Any] = model.init_cache(decoder_input_ids.shape[0] , _a , _a ) snake_case_ : Any = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="i4" ) snake_case_ : Dict = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) snake_case_ : str = model.decode( decoder_input_ids[:, :-1] , _a , decoder_attention_mask=_a , past_key_values=_a , decoder_position_ids=_a , ) snake_case_ : Optional[Any] = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) snake_case_ : Optional[int] = model.decode( decoder_input_ids[:, -1:] , _a , decoder_attention_mask=_a , past_key_values=outputs_cache.past_key_values , decoder_position_ids=_a , ) snake_case_ : str = model.decode(_a , _a ) snake_case_ : Dict = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''' ) def _A ( self :Union[str, Any] , lowerCAmelCase__ :List[str] , lowerCAmelCase__ :str , lowerCAmelCase__ :Optional[int] ) -> Optional[int]: '''simple docstring''' snake_case_ : List[Any] = 20 snake_case_ : List[str] = model_class_name(_a ) snake_case_ : Any = model.encode(inputs_dict["input_ids"] ) snake_case_, snake_case_ : str = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) snake_case_ : List[Any] = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) snake_case_ : Any = model.init_cache(decoder_input_ids.shape[0] , _a , _a ) snake_case_ : List[str] = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) snake_case_ : Any = model.decode( decoder_input_ids[:, :-1] , _a , decoder_attention_mask=_a , past_key_values=_a , decoder_position_ids=_a , ) snake_case_ : str = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) snake_case_ : int = model.decode( decoder_input_ids[:, -1:] , _a , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=_a , decoder_position_ids=_a , ) snake_case_ : Tuple = model.decode(_a , _a , decoder_attention_mask=_a ) snake_case_ : Union[str, Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''' ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__=None ,__magic_name__=None ,)-> Dict: """simple docstring""" if attention_mask is None: snake_case_ : int = np.not_equal(lowerCAmelCase__ ,config.pad_token_id ).astype(np.inta ) if decoder_attention_mask is None: snake_case_ : str = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape ,dtype=np.inta ), np.not_equal(decoder_input_ids[:, 1:] ,config.pad_token_id ).astype(np.inta ), ] ,axis=-1 ,) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class A_ (__SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" a__ = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) a__ = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () a__ = True a__ = False a__ = False a__ = False def _A ( self :Any ) -> Union[str, Any]: '''simple docstring''' snake_case_ : int = FlaxPegasusModelTester(self ) snake_case_ : List[str] = ConfigTester(self , config_class=_a ) def _A ( self :Dict ) -> List[Any]: '''simple docstring''' self.config_tester.run_common_tests() def _A ( self :int ) -> str: '''simple docstring''' snake_case_, snake_case_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(_a , _a , _a ) def _A ( self :int ) -> Optional[Any]: '''simple docstring''' snake_case_, snake_case_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(_a , _a , _a ) def _A ( self :Any ) -> int: '''simple docstring''' snake_case_, snake_case_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): snake_case_ : int = self._prepare_for_class(_a , _a ) snake_case_ : Union[str, Any] = model_class(_a ) @jax.jit def encode_jitted(lowerCAmelCase__ :Dict , lowerCAmelCase__ :Tuple=None , **lowerCAmelCase__ :List[str] ): return model.encode(input_ids=_a , attention_mask=_a ) with self.subTest("JIT Enabled" ): snake_case_ : Union[str, Any] = encode_jitted(**_a ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): snake_case_ : Tuple = encode_jitted(**_a ).to_tuple() self.assertEqual(len(_a ) , len(_a ) ) for jitted_output, output in zip(_a , _a ): self.assertEqual(jitted_output.shape , output.shape ) def _A ( self :Optional[int] ) -> List[str]: '''simple docstring''' snake_case_, snake_case_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): snake_case_ : Optional[Any] = model_class(_a ) snake_case_ : Dict = model.encode(inputs_dict["input_ids"] , inputs_dict["attention_mask"] ) snake_case_ : Optional[int] = { "decoder_input_ids": inputs_dict["decoder_input_ids"], "decoder_attention_mask": inputs_dict["decoder_attention_mask"], "encoder_outputs": encoder_outputs, } @jax.jit def decode_jitted(lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :List[str] , lowerCAmelCase__ :Optional[Any] ): return model.decode( decoder_input_ids=_a , decoder_attention_mask=_a , encoder_outputs=_a , ) with self.subTest("JIT Enabled" ): snake_case_ : Any = decode_jitted(**_a ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): snake_case_ : Union[str, Any] = decode_jitted(**_a ).to_tuple() self.assertEqual(len(_a ) , len(_a ) ) for jitted_output, output in zip(_a , _a ): self.assertEqual(jitted_output.shape , output.shape ) @slow def _A ( self :List[str] ) -> Union[str, Any]: '''simple docstring''' for model_class_name in self.all_model_classes: snake_case_ : List[Any] = model_class_name.from_pretrained("google/pegasus-large" , from_pt=_a ) snake_case_ : int = np.ones((1, 1) ) snake_case_ : List[str] = model(_a ) self.assertIsNotNone(_a ) @slow def _A ( self :int ) -> str: '''simple docstring''' snake_case_ : Optional[Any] = FlaxPegasusForConditionalGeneration.from_pretrained("google/pegasus-xsum" ) snake_case_ : Tuple = PegasusTokenizer.from_pretrained("google/pegasus-xsum" ) snake_case_ : Tuple = [ " PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.", " The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning \'Oh I think you\'re nominated\'\", said Dappy.\"And I was like \'Oh yeah, which one?\' And now we\'ve got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it\'s best of us to come down and mingle with everyone and say hello to the cameras. And now we find we\'ve got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn\'t be too disappointed if they didn\'t win this time around.\"At the end of the day we\'re grateful to be where we are in our careers.\"If it don\'t happen then it don\'t happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers\' All These Things That I\'ve Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year\'s Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" ", ] snake_case_ : Tuple = [ "California\'s largest electricity provider has turned off power to hundreds of thousands of customers.", "Pop group N-Dubz have revealed they were surprised to get four nominations for this year\'s Mobo Awards.", ] snake_case_ : List[str] = tokenizer(_a , return_tensors="np" , truncation=_a , max_length=512 , padding=_a ) snake_case_ : Optional[int] = model.generate(**_a , num_beams=2 ).sequences snake_case_ : Union[str, Any] = tokenizer.batch_decode(_a , skip_special_tokens=_a ) assert tgt_text == decoded
721
'''simple docstring''' import argparse import pathlib import fairseq import torch from fairseq.models.roberta import RobertaModel as FairseqRobertaModel from fairseq.modules import TransformerSentenceEncoderLayer from packaging import version from transformers import XLMRobertaConfig, XLMRobertaXLForMaskedLM, XLMRobertaXLForSequenceClassification from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.models.roberta.modeling_roberta import RobertaAttention from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse('''1.0.0a'''): raise Exception('''requires fairseq >= 1.0.0a''') logging.set_verbosity_info() __lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) __lowerCamelCase : Union[str, Any] = '''Hello world! cécé herlolip''' def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : str = FairseqRobertaModel.from_pretrained(__magic_name__ ) roberta.eval() # disable dropout snake_case_ : Dict = roberta.model.encoder.sentence_encoder snake_case_ : List[str] = XLMRobertaConfig( vocab_size=roberta_sent_encoder.embed_tokens.num_embeddings ,hidden_size=roberta.cfg.model.encoder_embed_dim ,num_hidden_layers=roberta.cfg.model.encoder_layers ,num_attention_heads=roberta.cfg.model.encoder_attention_heads ,intermediate_size=roberta.cfg.model.encoder_ffn_embed_dim ,max_position_embeddings=514 ,type_vocab_size=1 ,layer_norm_eps=1E-5 ,) if classification_head: snake_case_ : List[str] = roberta.model.classification_heads["mnli"].out_proj.weight.shape[0] print("Our RoBERTa config:" ,__magic_name__ ) snake_case_ : List[str] = XLMRobertaXLForSequenceClassification(__magic_name__ ) if classification_head else XLMRobertaXLForMaskedLM(__magic_name__ ) model.eval() # Now let's copy all the weights. # Embeddings snake_case_ : List[Any] = roberta_sent_encoder.embed_tokens.weight snake_case_ : int = roberta_sent_encoder.embed_positions.weight snake_case_ : Union[str, Any] = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c RoBERTa doesn't use them. snake_case_ : Union[str, Any] = roberta_sent_encoder.layer_norm.weight snake_case_ : str = roberta_sent_encoder.layer_norm.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer snake_case_ : BertLayer = model.roberta.encoder.layer[i] snake_case_ : TransformerSentenceEncoderLayer = roberta_sent_encoder.layers[i] snake_case_ : RobertaAttention = layer.attention snake_case_ : Dict = roberta_layer.self_attn_layer_norm.weight snake_case_ : Dict = roberta_layer.self_attn_layer_norm.bias # self attention snake_case_ : BertSelfAttention = layer.attention.self assert ( roberta_layer.self_attn.k_proj.weight.data.shape == roberta_layer.self_attn.q_proj.weight.data.shape == roberta_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ) snake_case_ : Dict = roberta_layer.self_attn.q_proj.weight snake_case_ : Any = roberta_layer.self_attn.q_proj.bias snake_case_ : Optional[Any] = roberta_layer.self_attn.k_proj.weight snake_case_ : Optional[Any] = roberta_layer.self_attn.k_proj.bias snake_case_ : Optional[int] = roberta_layer.self_attn.v_proj.weight snake_case_ : Any = roberta_layer.self_attn.v_proj.bias # self-attention output snake_case_ : BertSelfOutput = layer.attention.output assert self_output.dense.weight.shape == roberta_layer.self_attn.out_proj.weight.shape snake_case_ : List[str] = roberta_layer.self_attn.out_proj.weight snake_case_ : Optional[int] = roberta_layer.self_attn.out_proj.bias # this one is final layer norm snake_case_ : int = roberta_layer.final_layer_norm.weight snake_case_ : Union[str, Any] = roberta_layer.final_layer_norm.bias # intermediate snake_case_ : BertIntermediate = layer.intermediate assert intermediate.dense.weight.shape == roberta_layer.fca.weight.shape snake_case_ : List[str] = roberta_layer.fca.weight snake_case_ : List[Any] = roberta_layer.fca.bias # output snake_case_ : BertOutput = layer.output assert bert_output.dense.weight.shape == roberta_layer.fca.weight.shape snake_case_ : Any = roberta_layer.fca.weight snake_case_ : Any = roberta_layer.fca.bias # end of layer if classification_head: snake_case_ : int = roberta.model.classification_heads["mnli"].dense.weight snake_case_ : Union[str, Any] = roberta.model.classification_heads["mnli"].dense.bias snake_case_ : Tuple = roberta.model.classification_heads["mnli"].out_proj.weight snake_case_ : str = roberta.model.classification_heads["mnli"].out_proj.bias else: # LM Head snake_case_ : Optional[Any] = roberta.model.encoder.lm_head.dense.weight snake_case_ : int = roberta.model.encoder.lm_head.dense.bias snake_case_ : Optional[Any] = roberta.model.encoder.lm_head.layer_norm.weight snake_case_ : Optional[int] = roberta.model.encoder.lm_head.layer_norm.bias snake_case_ : int = roberta.model.encoder.lm_head.weight snake_case_ : List[str] = roberta.model.encoder.lm_head.bias # Let's check that we get the same results. snake_case_ : torch.Tensor = roberta.encode(__magic_name__ ).unsqueeze(0 ) # batch of size 1 snake_case_ : Union[str, Any] = model(__magic_name__ )[0] if classification_head: snake_case_ : Optional[Any] = roberta.model.classification_heads["mnli"](roberta.extract_features(__magic_name__ ) ) else: snake_case_ : List[str] = roberta.model(__magic_name__ )[0] print(our_output.shape ,their_output.shape ) snake_case_ : str = torch.max(torch.abs(our_output - their_output ) ).item() print(F'''max_absolute_diff = {max_absolute_diff}''' ) # ~ 1e-7 snake_case_ : Any = torch.allclose(__magic_name__ ,__magic_name__ ,atol=1E-3 ) print("Do both models output the same tensors?" ,"🔥" if success else "💩" ) if not success: raise Exception("Something went wRoNg" ) pathlib.Path(__magic_name__ ).mkdir(parents=__magic_name__ ,exist_ok=__magic_name__ ) print(F'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(__magic_name__ ) if __name__ == "__main__": __lowerCamelCase : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--roberta_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--classification_head''', action='''store_true''', help='''Whether to convert a final classification head.''' ) __lowerCamelCase : Tuple = parser.parse_args() convert_xlm_roberta_xl_checkpoint_to_pytorch( args.roberta_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
656
0
'''simple docstring''' import os from distutils.util import strtobool def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" for e in env_keys: snake_case_ : Union[str, Any] = int(os.environ.get(_lowerCamelCase ,-1 ) ) if val >= 0: return val return default def __UpperCAmelCase ( __magic_name__ ,__magic_name__=False )-> Optional[Any]: """simple docstring""" snake_case_ : int = os.environ.get(_lowerCamelCase ,str(_lowerCamelCase ) ) return strtobool(_lowerCamelCase ) == 1 # As its name indicates `strtobool` actually returns an int... def __UpperCAmelCase ( __magic_name__ ,__magic_name__="no" )-> List[str]: """simple docstring""" snake_case_ : Optional[Any] = os.environ.get(_lowerCamelCase ,str(_lowerCamelCase ) ) return value
700
'''simple docstring''' import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def __UpperCAmelCase ( __magic_name__ ,__magic_name__=() ,__magic_name__=None ,__magic_name__="no" ,__magic_name__="29500" )-> Optional[int]: """simple docstring""" snake_case_ : str = False snake_case_ : int = False if any(key.startswith("KAGGLE" ) for key in os.environ.keys() ): snake_case_ : Any = True elif "IPython" in sys.modules: snake_case_ : Union[str, Any] = "google.colab" in str(sys.modules["IPython"].get_ipython() ) try: snake_case_ : Any = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F'''Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.''' ) if (in_colab or in_kaggle) and (os.environ.get("TPU_NAME" ,__magic_name__ ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( "To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside " "your training function. Restart your notebook and make sure no cells initializes an " "`Accelerator`." ) if num_processes is None: snake_case_ : Tuple = 8 snake_case_ : Optional[int] = PrepareForLaunch(__magic_name__ ,distributed_type="TPU" ) print(F'''Launching a training on {num_processes} TPU cores.''' ) xmp.spawn(__magic_name__ ,args=__magic_name__ ,nprocs=__magic_name__ ,start_method="fork" ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print("Launching training on one GPU." ) else: print("Launching training on one CPU." ) function(*__magic_name__ ) else: if num_processes is None: raise ValueError( "You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call." ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( "To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized " "inside your training function. Restart your notebook and make sure no cells initializes an " "`Accelerator`." ) if torch.cuda.is_initialized(): raise ValueError( "To launch a multi-GPU training from your notebook, you need to avoid running any instruction " "using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA " "function." ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=__magic_name__ ,master_addr="127.0.01" ,master_port=__magic_name__ ,mixed_precision=__magic_name__ ): snake_case_ : Optional[int] = PrepareForLaunch(__magic_name__ ,distributed_type="MULTI_GPU" ) print(F'''Launching training on {num_processes} GPUs.''' ) try: start_processes(__magic_name__ ,args=__magic_name__ ,nprocs=__magic_name__ ,start_method="fork" ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( "CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. " "This likely stems from an outside import causing issues once the `notebook_launcher()` is called. " "Please review your imports and test them when running the `notebook_launcher()` to identify " "which one is problematic." ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): snake_case_ : Any = "1" print("Launching training on MPS." ) elif torch.cuda.is_available(): print("Launching training on one GPU." ) else: print("Launching training on CPU." ) function(*__magic_name__ ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__=() ,__magic_name__=2 )-> Dict: """simple docstring""" from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=__magic_name__ ,master_addr="127.0.01" ,master_port="29500" ,accelerate_mixed_precision="no" ,accelerate_debug_rdv_file=tmp_file.name ,accelerate_use_cpu="yes" ,): snake_case_ : Any = PrepareForLaunch(__magic_name__ ,debug=__magic_name__ ) start_processes(__magic_name__ ,args=__magic_name__ ,nprocs=__magic_name__ ,start_method="fork" )
656
0
'''simple docstring''' from typing import Any import numpy as np def __UpperCAmelCase ( __magic_name__ )-> Any: """simple docstring""" return np.array_equal(__magic_name__ ,matrix.conjugate().T ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> int: """simple docstring""" snake_case_ : Optional[Any] = v.conjugate().T snake_case_ : int = v_star.dot(__magic_name__ ) assert isinstance(__magic_name__ ,np.ndarray ) return (v_star_dot.dot(__magic_name__ )) / (v_star.dot(__magic_name__ )) def __UpperCAmelCase ( )-> Dict: """simple docstring""" snake_case_ : Any = np.array([[2, 2 + 1j, 4], [2 - 1j, 3, 1j], [4, -1j, 1]] ) snake_case_ : List[str] = np.array([[1], [2], [3]] ) assert is_hermitian(__magic_name__ ), F'''{a} is not hermitian.''' print(rayleigh_quotient(__magic_name__ ,__magic_name__ ) ) snake_case_ : Union[str, Any] = np.array([[1, 2, 4], [2, 3, -1], [4, -1, 1]] ) assert is_hermitian(__magic_name__ ), F'''{a} is not hermitian.''' assert rayleigh_quotient(__magic_name__ ,__magic_name__ ) == float(3 ) if __name__ == "__main__": import doctest doctest.testmod() tests()
701
'''simple docstring''' from collections import deque from math import floor from random import random from time import time class A_ : """simple docstring""" def __init__( self :Dict ) -> List[str]: '''simple docstring''' snake_case_ : int = {} def _A ( self :Any , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Optional[Any]=1 ) -> Any: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: snake_case_ : Optional[int] = [[w, v]] if not self.graph.get(lowerCAmelCase__ ): snake_case_ : Dict = [] def _A ( self :List[Any] ) -> Optional[int]: '''simple docstring''' return list(self.graph ) def _A ( self :str , lowerCAmelCase__ :Any , lowerCAmelCase__ :int ) -> List[Any]: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCAmelCase__ ) def _A ( self :List[str] , lowerCAmelCase__ :Optional[Any]=-2 , lowerCAmelCase__ :str=-1 ) -> str: '''simple docstring''' if s == d: return [] snake_case_ : str = [] snake_case_ : Optional[int] = [] if s == -2: snake_case_ : List[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Dict = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : str = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCAmelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCAmelCase__ ) != 0: snake_case_ : Union[str, Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return visited def _A ( self :Tuple , lowerCAmelCase__ :int=-1 ) -> int: '''simple docstring''' if c == -1: snake_case_ : Any = floor(random() * 10_000 ) + 10 for i in range(lowerCAmelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): snake_case_ : Optional[Any] = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCAmelCase__ , lowerCAmelCase__ , 1 ) def _A ( self :Tuple , lowerCAmelCase__ :Dict=-2 ) -> Dict: '''simple docstring''' snake_case_ : Union[str, Any] = deque() snake_case_ : Optional[Any] = [] if s == -2: snake_case_ : Tuple = list(self.graph )[0] d.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) while d: snake_case_ : Optional[int] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def _A ( self :List[str] , lowerCAmelCase__ :str ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = 0 for x in self.graph: for y in self.graph[x]: if y[1] == u: count += 1 return count def _A ( self :Any , lowerCAmelCase__ :int ) -> Optional[Any]: '''simple docstring''' return len(self.graph[u] ) def _A ( self :Tuple , lowerCAmelCase__ :List[str]=-2 ) -> Optional[Any]: '''simple docstring''' snake_case_ : str = [] snake_case_ : str = [] if s == -2: snake_case_ : Optional[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : int = s snake_case_ : Optional[int] = [] while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : List[str] = node[1] break # check if all the children are visited if s == ss: sorted_nodes.append(stack.pop() ) if len(lowerCAmelCase__ ) != 0: snake_case_ : int = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Union[str, Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return sorted_nodes def _A ( self :Dict ) -> Any: '''simple docstring''' snake_case_ : Dict = [] snake_case_ : Any = [] snake_case_ : str = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Optional[int] = -2 snake_case_ : Any = [] snake_case_ : List[Any] = s snake_case_ : int = False snake_case_ : Optional[int] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Any = len(lowerCAmelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Optional[int] = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[Any] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : str = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : List[str] = s snake_case_ : Optional[int] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return list(lowerCAmelCase__ ) def _A ( self :Tuple ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = [] snake_case_ : Tuple = [] snake_case_ : List[str] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : str = -2 snake_case_ : List[str] = [] snake_case_ : List[Any] = s snake_case_ : List[str] = False snake_case_ : Dict = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Any = len(lowerCAmelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Tuple = True if len(lowerCAmelCase__ ) != 0: snake_case_ : List[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : int = s snake_case_ : Union[str, Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return False def _A ( self :Optional[int] , lowerCAmelCase__ :Optional[int]=-2 , lowerCAmelCase__ :Tuple=-1 ) -> str: '''simple docstring''' snake_case_ : Optional[int] = time() self.dfs(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : Optional[Any] = time() return end - begin def _A ( self :Any , lowerCAmelCase__ :Tuple=-2 ) -> Optional[Any]: '''simple docstring''' snake_case_ : Any = time() self.bfs(lowerCAmelCase__ ) snake_case_ : Any = time() return end - begin class A_ : """simple docstring""" def __init__( self :Tuple ) -> List[str]: '''simple docstring''' snake_case_ : Optional[Any] = {} def _A ( self :str , lowerCAmelCase__ :Dict , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Union[str, Any]=1 ) -> str: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): # if there already is a edge if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: # if u does not exist snake_case_ : str = [[w, v]] # add the other way if self.graph.get(lowerCAmelCase__ ): # if there already is a edge if self.graph[v].count([w, u] ) == 0: self.graph[v].append([w, u] ) else: # if u does not exist snake_case_ : List[str] = [[w, u]] def _A ( self :Dict , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Optional[Any] ) -> Optional[Any]: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCAmelCase__ ) # the other way round if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[v]: if _[1] == u: self.graph[v].remove(lowerCAmelCase__ ) def _A ( self :Optional[Any] , lowerCAmelCase__ :Optional[Any]=-2 , lowerCAmelCase__ :Optional[int]=-1 ) -> int: '''simple docstring''' if s == d: return [] snake_case_ : Any = [] snake_case_ : Dict = [] if s == -2: snake_case_ : Optional[int] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[str] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCAmelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : str = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return visited def _A ( self :Optional[int] , lowerCAmelCase__ :str=-1 ) -> List[Any]: '''simple docstring''' if c == -1: snake_case_ : Optional[int] = floor(random() * 10_000 ) + 10 for i in range(lowerCAmelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): snake_case_ : str = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCAmelCase__ , lowerCAmelCase__ , 1 ) def _A ( self :Any , lowerCAmelCase__ :Optional[Any]=-2 ) -> List[Any]: '''simple docstring''' snake_case_ : List[str] = deque() snake_case_ : Optional[Any] = [] if s == -2: snake_case_ : List[Any] = list(self.graph )[0] d.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) while d: snake_case_ : Optional[int] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def _A ( self :str , lowerCAmelCase__ :Union[str, Any] ) -> Optional[Any]: '''simple docstring''' return len(self.graph[u] ) def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : Any = [] snake_case_ : Optional[Any] = [] snake_case_ : Optional[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = -2 snake_case_ : Optional[int] = [] snake_case_ : Tuple = s snake_case_ : Optional[Any] = False snake_case_ : Optional[int] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : Optional[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Optional[int] = len(lowerCAmelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Tuple = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[int] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : List[Any] = s snake_case_ : Dict = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return list(lowerCAmelCase__ ) def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Optional[Any] = [] snake_case_ : int = [] snake_case_ : List[str] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = -2 snake_case_ : int = [] snake_case_ : int = s snake_case_ : Optional[Any] = False snake_case_ : List[Any] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : Union[str, Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Tuple = len(lowerCAmelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Optional[Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[Any] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Tuple = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : Union[str, Any] = s snake_case_ : Tuple = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return False def _A ( self :Any ) -> Tuple: '''simple docstring''' return list(self.graph ) def _A ( self :Optional[Any] , lowerCAmelCase__ :Tuple=-2 , lowerCAmelCase__ :Optional[int]=-1 ) -> str: '''simple docstring''' snake_case_ : List[str] = time() self.dfs(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : List[Any] = time() return end - begin def _A ( self :Union[str, Any] , lowerCAmelCase__ :List[Any]=-2 ) -> int: '''simple docstring''' snake_case_ : List[str] = time() self.bfs(lowerCAmelCase__ ) snake_case_ : Tuple = time() return end - begin
656
0
'''simple docstring''' from PIL import Image def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Image: """simple docstring""" snake_case_ : List[str] = (259 * (level + 255)) / (255 * (259 - level)) def contrast(__magic_name__ ) -> int: return int(128 + factor * (c - 128) ) return img.point(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": # Load image with Image.open('''image_data/lena.jpg''') as img: # Change contrast to 170 __lowerCamelCase : List[Any] = change_contrast(img, 170) cont_img.save('''image_data/lena_high_contrast.png''', format='''png''')
702
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __lowerCamelCase : List[str] = re.compile(R'''\s+''') def __UpperCAmelCase ( __magic_name__ )-> Union[str, Any]: """simple docstring""" return {"hash": hashlib.mda(re.sub(__magic_name__ ,"" ,example["content"] ).encode("utf-8" ) ).hexdigest()} def __UpperCAmelCase ( __magic_name__ )-> str: """simple docstring""" snake_case_ : Optional[Any] = [len(__magic_name__ ) for line in example["content"].splitlines()] return {"line_mean": np.mean(__magic_name__ ), "line_max": max(__magic_name__ )} def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" snake_case_ : Optional[int] = np.mean([c.isalnum() for c in example["content"]] ) return {"alpha_frac": alpha_frac} def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" if example["hash"] in uniques: uniques.remove(example["hash"] ) return True else: return False def __UpperCAmelCase ( __magic_name__ ,__magic_name__=5 )-> Tuple: """simple docstring""" snake_case_ : List[str] = ["auto-generated", "autogenerated", "automatically generated"] snake_case_ : Optional[Any] = example["content"].splitlines() for _, line in zip(range(__magic_name__ ) ,__magic_name__ ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def __UpperCAmelCase ( __magic_name__ ,__magic_name__=5 ,__magic_name__=0.05 )-> Optional[Any]: """simple docstring""" snake_case_ : str = ["unit tests", "test file", "configuration file"] snake_case_ : int = example["content"].splitlines() snake_case_ : Optional[Any] = 0 snake_case_ : Any = 0 # first test for _, line in zip(range(__magic_name__ ) ,__magic_name__ ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test snake_case_ : Tuple = example["content"].count("\n" ) snake_case_ : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count("config" ) count_test += line.lower().count("test" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def __UpperCAmelCase ( __magic_name__ )-> str: """simple docstring""" snake_case_ : List[Any] = ["def ", "class ", "for ", "while "] snake_case_ : Optional[Any] = example["content"].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def __UpperCAmelCase ( __magic_name__ ,__magic_name__=4 )-> Optional[int]: """simple docstring""" snake_case_ : Tuple = example["content"].splitlines() snake_case_ : Tuple = 0 for line in lines: counter += line.lower().count("=" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : Tuple = tokenizer(example["content"] ,truncation=__magic_name__ )["input_ids"] snake_case_ : int = len(example["content"] ) / len(__magic_name__ ) return {"ratio": ratio} def __UpperCAmelCase ( __magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : Union[str, Any] = {} results.update(get_hash(__magic_name__ ) ) results.update(line_stats(__magic_name__ ) ) results.update(alpha_stats(__magic_name__ ) ) results.update(char_token_ratio(__magic_name__ ) ) results.update(is_autogenerated(__magic_name__ ) ) results.update(is_config_or_test(__magic_name__ ) ) results.update(has_no_keywords(__magic_name__ ) ) results.update(has_few_assignments(__magic_name__ ) ) return results def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" if not check_uniques(__magic_name__ ,__magic_name__ ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def __UpperCAmelCase ( __magic_name__ )-> Dict: """simple docstring""" with open(__magic_name__ ,"rb" ) as f_in: with gzip.open(str(__magic_name__ ) + ".gz" ,"wb" ,compresslevel=6 ) as f_out: shutil.copyfileobj(__magic_name__ ,__magic_name__ ) os.unlink(__magic_name__ ) # Settings __lowerCamelCase : List[Any] = HfArgumentParser(PreprocessingArguments) __lowerCamelCase : str = parser.parse_args() if args.num_workers is None: __lowerCamelCase : List[Any] = multiprocessing.cpu_count() __lowerCamelCase : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __lowerCamelCase : Any = time.time() __lowerCamelCase : str = load_dataset(args.dataset_name, split='''train''') print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __lowerCamelCase : List[str] = time.time() __lowerCamelCase : Any = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __lowerCamelCase : Any = set(ds.unique('''hash''')) __lowerCamelCase : Optional[int] = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __lowerCamelCase : List[str] = time.time() __lowerCamelCase : Tuple = ds.filter(filter, fn_kwargs={'''uniques''': uniques, '''args''': args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __lowerCamelCase : List[str] = time.time() __lowerCamelCase , __lowerCamelCase : Tuple = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __lowerCamelCase : List[Any] = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / '''duplicate_clusters.json''', '''w''') as f: json.dump(duplicate_clusters, f) __lowerCamelCase : List[str] = output_dir / '''data''' data_dir.mkdir(exist_ok=True) __lowerCamelCase : int = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __lowerCamelCase : Union[str, Any] = str(data_dir / f'''file-{file_number+1:012}.json''') __lowerCamelCase : List[Any] = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
656
0
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __lowerCamelCase : Optional[Any] = logging.get_logger(__name__) __lowerCamelCase : Optional[Any] = { '''ut/deta''': '''https://huggingface.co/ut/deta/resolve/main/config.json''', } class A_ (a_ ): """simple docstring""" a__ = '''deta''' a__ = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self :Union[str, Any] , lowerCAmelCase__ :List[Any]=None , lowerCAmelCase__ :Optional[int]=900 , lowerCAmelCase__ :Any=2_048 , lowerCAmelCase__ :Any=6 , lowerCAmelCase__ :Tuple=2_048 , lowerCAmelCase__ :Optional[int]=8 , lowerCAmelCase__ :List[str]=6 , lowerCAmelCase__ :Optional[int]=1_024 , lowerCAmelCase__ :Optional[int]=8 , lowerCAmelCase__ :int=0.0 , lowerCAmelCase__ :str=True , lowerCAmelCase__ :Optional[Any]="relu" , lowerCAmelCase__ :Optional[Any]=256 , lowerCAmelCase__ :Tuple=0.1 , lowerCAmelCase__ :Union[str, Any]=0.0 , lowerCAmelCase__ :List[Any]=0.0 , lowerCAmelCase__ :int=0.0_2 , lowerCAmelCase__ :str=1.0 , lowerCAmelCase__ :Dict=True , lowerCAmelCase__ :Any=False , lowerCAmelCase__ :Dict="sine" , lowerCAmelCase__ :Optional[Any]=5 , lowerCAmelCase__ :int=4 , lowerCAmelCase__ :List[Any]=4 , lowerCAmelCase__ :Optional[int]=True , lowerCAmelCase__ :Union[str, Any]=300 , lowerCAmelCase__ :List[Any]=True , lowerCAmelCase__ :Optional[int]=True , lowerCAmelCase__ :int=1 , lowerCAmelCase__ :List[Any]=5 , lowerCAmelCase__ :Dict=2 , lowerCAmelCase__ :int=1 , lowerCAmelCase__ :List[Any]=1 , lowerCAmelCase__ :Optional[Any]=5 , lowerCAmelCase__ :Any=2 , lowerCAmelCase__ :Optional[int]=0.1 , lowerCAmelCase__ :Any=0.2_5 , **lowerCAmelCase__ :str , ) -> Any: '''simple docstring''' if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) snake_case_ : Optional[Any] = CONFIG_MAPPING['resnet'](out_features=["stage2", "stage3", "stage4"] ) else: if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ : Dict = backbone_config.pop("model_type" ) snake_case_ : Union[str, Any] = CONFIG_MAPPING[backbone_model_type] snake_case_ : str = config_class.from_dict(lowerCAmelCase__ ) snake_case_ : int = backbone_config snake_case_ : Dict = num_queries snake_case_ : List[Any] = max_position_embeddings snake_case_ : Tuple = d_model snake_case_ : int = encoder_ffn_dim snake_case_ : Optional[Any] = encoder_layers snake_case_ : str = encoder_attention_heads snake_case_ : List[str] = decoder_ffn_dim snake_case_ : Optional[int] = decoder_layers snake_case_ : List[str] = decoder_attention_heads snake_case_ : List[str] = dropout snake_case_ : List[Any] = attention_dropout snake_case_ : str = activation_dropout snake_case_ : Union[str, Any] = activation_function snake_case_ : List[str] = init_std snake_case_ : Optional[Any] = init_xavier_std snake_case_ : Optional[Any] = encoder_layerdrop snake_case_ : Union[str, Any] = auxiliary_loss snake_case_ : List[str] = position_embedding_type # deformable attributes snake_case_ : Dict = num_feature_levels snake_case_ : Optional[Any] = encoder_n_points snake_case_ : Union[str, Any] = decoder_n_points snake_case_ : str = two_stage snake_case_ : int = two_stage_num_proposals snake_case_ : int = with_box_refine snake_case_ : Optional[int] = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError("If two_stage is True, with_box_refine must be True." ) # Hungarian matcher snake_case_ : Dict = class_cost snake_case_ : Tuple = bbox_cost snake_case_ : Any = giou_cost # Loss coefficients snake_case_ : List[Any] = mask_loss_coefficient snake_case_ : Dict = dice_loss_coefficient snake_case_ : Tuple = bbox_loss_coefficient snake_case_ : str = giou_loss_coefficient snake_case_ : str = eos_coefficient snake_case_ : str = focal_alpha super().__init__(is_encoder_decoder=lowerCAmelCase__ , **lowerCAmelCase__ ) @property def _A ( self :Optional[Any] ) -> Union[str, Any]: '''simple docstring''' return self.encoder_attention_heads @property def _A ( self :Optional[Any] ) -> List[str]: '''simple docstring''' return self.d_model def _A ( self :Any ) -> Any: '''simple docstring''' snake_case_ : Union[str, Any] = copy.deepcopy(self.__dict__ ) snake_case_ : Any = self.backbone_config.to_dict() snake_case_ : Union[str, Any] = self.__class__.model_type return output
703
'''simple docstring''' import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :str ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Union[str, Any] = torch.nn.Linear(10 , 10 ) snake_case_ : Dict = torch.optim.SGD(model.parameters() , 0.1 ) snake_case_ : Tuple = Accelerator() snake_case_ : Optional[Any] = accelerator.prepare(lowerCAmelCase__ ) try: pickle.loads(pickle.dumps(lowerCAmelCase__ ) ) except Exception as e: self.fail(F'''Accelerated optimizer pickling failed with {e}''' ) AcceleratorState._reset_state()
656
0
import requests def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Dict: """simple docstring""" snake_case_ : str = {'Content-Type': 'application/json'} snake_case_ : int = requests.post(__lowercase ,json={"text": message_body} ,headers=__lowercase ) if response.status_code != 200: snake_case_ : List[Any] = ( 'Request to slack returned an error ' F'''{response.status_code}, the response is:\n{response.text}''' ) raise ValueError(__lowercase ) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message('''<YOUR MESSAGE BODY>''', '''<SLACK CHANNEL URL>''')
704
'''simple docstring''' import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __lowerCamelCase : Any = '''src/transformers''' # This is to make sure the transformers module imported is the one in the repo. __lowerCamelCase : List[str] = direct_transformers_import(PATH_TO_TRANSFORMERS) __lowerCamelCase : Optional[Any] = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __lowerCamelCase : Union[str, Any] = re.compile(R'''\[(.+?)\]\((https://huggingface\.co/.+?)\)''') __lowerCamelCase : Any = { '''DecisionTransformerConfig''', '''EncoderDecoderConfig''', '''MusicgenConfig''', '''RagConfig''', '''SpeechEncoderDecoderConfig''', '''TimmBackboneConfig''', '''VisionEncoderDecoderConfig''', '''VisionTextDualEncoderConfig''', '''LlamaConfig''', } def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : Tuple = None # source code of `config_class` snake_case_ : List[Any] = inspect.getsource(__magic_name__ ) snake_case_ : List[str] = _re_checkpoint.findall(__magic_name__ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith("/" ): snake_case_ : Optional[Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link snake_case_ : str = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: snake_case_ : Dict = ckpt_name break return checkpoint def __UpperCAmelCase ( )-> Dict: """simple docstring""" snake_case_ : Optional[int] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue snake_case_ : str = get_checkpoint_from_config_class(__magic_name__ ) snake_case_ : Union[str, Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(__magic_name__ ) if len(__magic_name__ ) > 0: snake_case_ : Tuple = "\n".join(sorted(__magic_name__ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
656
0
'''simple docstring''' import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class A_ : """simple docstring""" def __init__( self :Union[str, Any] , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Optional[Any]=13 , lowerCAmelCase__ :List[Any]=30 , lowerCAmelCase__ :Any=2 , lowerCAmelCase__ :Optional[int]=3 , lowerCAmelCase__ :Optional[Any]=True , lowerCAmelCase__ :Dict=True , lowerCAmelCase__ :Optional[Any]=32 , lowerCAmelCase__ :Optional[Any]=5 , lowerCAmelCase__ :List[str]=4 , lowerCAmelCase__ :Tuple=37 , lowerCAmelCase__ :Optional[int]="gelu" , lowerCAmelCase__ :int=0.1 , lowerCAmelCase__ :List[Any]=0.1 , lowerCAmelCase__ :int=10 , lowerCAmelCase__ :Any=0.0_2 , lowerCAmelCase__ :int=3 , lowerCAmelCase__ :Optional[Any]=None , lowerCAmelCase__ :Optional[int]=2 , ) -> str: '''simple docstring''' snake_case_ : Optional[int] = parent snake_case_ : int = batch_size snake_case_ : Any = image_size snake_case_ : List[Any] = patch_size snake_case_ : Tuple = num_channels snake_case_ : Optional[int] = is_training snake_case_ : Any = use_labels snake_case_ : Optional[Any] = hidden_size snake_case_ : Any = num_hidden_layers snake_case_ : Any = num_attention_heads snake_case_ : Tuple = intermediate_size snake_case_ : Dict = hidden_act snake_case_ : Any = hidden_dropout_prob snake_case_ : List[str] = attention_probs_dropout_prob snake_case_ : Tuple = type_sequence_label_size snake_case_ : Any = initializer_range snake_case_ : List[str] = scope snake_case_ : Union[str, Any] = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) snake_case_ : Union[str, Any] = (image_size // patch_size) ** 2 snake_case_ : Dict = num_patches + 2 def _A ( self :Any ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ : int = None if self.use_labels: snake_case_ : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ : List[str] = self.get_config() return config, pixel_values, labels def _A ( self :Tuple ) -> Any: '''simple docstring''' return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_A , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _A ( self :str , lowerCAmelCase__ :List[str] , lowerCAmelCase__ :str , lowerCAmelCase__ :Tuple ) -> Dict: '''simple docstring''' snake_case_ : Optional[Any] = DeiTModel(config=_A ) model.to(_A ) model.eval() snake_case_ : Tuple = model(_A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _A ( self :Optional[int] , lowerCAmelCase__ :Dict , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Any ) -> List[Any]: '''simple docstring''' snake_case_ : Dict = DeiTForMaskedImageModeling(config=_A ) model.to(_A ) model.eval() snake_case_ : List[Any] = model(_A ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images snake_case_ : Union[str, Any] = 1 snake_case_ : List[Any] = DeiTForMaskedImageModeling(_A ) model.to(_A ) model.eval() snake_case_ : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case_ : Dict = model(_A ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _A ( self :int , lowerCAmelCase__ :Dict , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :int ) -> int: '''simple docstring''' snake_case_ : str = self.type_sequence_label_size snake_case_ : Optional[int] = DeiTForImageClassification(_A ) model.to(_A ) model.eval() snake_case_ : Union[str, Any] = model(_A , labels=_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images snake_case_ : int = 1 snake_case_ : List[str] = DeiTForImageClassification(_A ) model.to(_A ) model.eval() snake_case_ : Optional[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case_ : Optional[Any] = model(_A , labels=_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _A ( self :Optional[int] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Union[str, Any] = self.prepare_config_and_inputs() ( ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ) : List[str] = config_and_inputs snake_case_ : Optional[int] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class A_ (__lowercase , __lowercase , unittest.TestCase ): """simple docstring""" a__ = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) a__ = ( { '''feature-extraction''': DeiTModel, '''image-classification''': (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) a__ = False a__ = False a__ = False def _A ( self :int ) -> Optional[int]: '''simple docstring''' snake_case_ : List[str] = DeiTModelTester(self ) snake_case_ : Tuple = ConfigTester(self , config_class=_A , has_text_modality=_A , hidden_size=37 ) def _A ( self :Dict ) -> List[str]: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="DeiT does not use inputs_embeds" ) def _A ( self :List[Any] ) -> Optional[int]: '''simple docstring''' pass def _A ( self :Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_, snake_case_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ : List[Any] = model_class(_A ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case_ : Union[str, Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_A , nn.Linear ) ) def _A ( self :Union[str, Any] ) -> Tuple: '''simple docstring''' snake_case_, snake_case_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ : Any = model_class(_A ) snake_case_ : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ : List[str] = [*signature.parameters.keys()] snake_case_ : int = ["pixel_values"] self.assertListEqual(arg_names[:1] , _A ) def _A ( self :Dict ) -> Optional[Any]: '''simple docstring''' snake_case_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_A ) def _A ( self :List[str] ) -> str: '''simple docstring''' snake_case_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_A ) def _A ( self :Optional[int] ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_A ) def _A ( self :Optional[Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Any=False ) -> List[str]: '''simple docstring''' snake_case_ : Any = super()._prepare_for_class(_A , _A , return_labels=_A ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def _A ( self :str ) -> Dict: '''simple docstring''' if not self.model_tester.is_training: return snake_case_, snake_case_ : int = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ : Optional[int] = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(_A ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue snake_case_ : List[Any] = model_class(_A ) model.to(_A ) model.train() snake_case_ : Tuple = self._prepare_for_class(_A , _A , return_labels=_A ) snake_case_ : Optional[int] = model(**_A ).loss loss.backward() def _A ( self :str ) -> Union[str, Any]: '''simple docstring''' snake_case_, snake_case_ : str = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return snake_case_ : int = False snake_case_ : Dict = True for model_class in self.all_model_classes: if model_class in get_values(_A ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue snake_case_ : Tuple = model_class(_A ) model.gradient_checkpointing_enable() model.to(_A ) model.train() snake_case_ : List[str] = self._prepare_for_class(_A , _A , return_labels=_A ) snake_case_ : Optional[int] = model(**_A ).loss loss.backward() def _A ( self :List[str] ) -> int: '''simple docstring''' snake_case_, snake_case_ : int = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ : Optional[Any] = [ {"title": "multi_label_classification", "num_labels": 2, "dtype": torch.float}, {"title": "single_label_classification", "num_labels": 1, "dtype": torch.long}, {"title": "regression", "num_labels": 1, "dtype": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(_A ), *get_values(_A ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F'''Testing {model_class} with {problem_type['title']}''' ): snake_case_ : Union[str, Any] = problem_type["title"] snake_case_ : List[str] = problem_type["num_labels"] snake_case_ : str = model_class(_A ) model.to(_A ) model.train() snake_case_ : Dict = self._prepare_for_class(_A , _A , return_labels=_A ) if problem_type["num_labels"] > 1: snake_case_ : List[Any] = inputs["labels"].unsqueeze(1 ).repeat(1 , problem_type["num_labels"] ) snake_case_ : List[Any] = inputs["labels"].to(problem_type["dtype"] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=_A ) as warning_list: snake_case_ : Union[str, Any] = model(**_A ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F'''Something is going wrong in the regression problem: intercepted {w.message}''' ) loss.backward() @slow def _A ( self :List[Any] ) -> Union[str, Any]: '''simple docstring''' for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ : Optional[Any] = DeiTModel.from_pretrained(_A ) self.assertIsNotNone(_A ) def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : str = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class A_ (unittest.TestCase ): """simple docstring""" @cached_property def _A ( self :int ) -> str: '''simple docstring''' return ( DeiTImageProcessor.from_pretrained("facebook/deit-base-distilled-patch16-224" ) if is_vision_available() else None ) @slow def _A ( self :Any ) -> List[str]: '''simple docstring''' snake_case_ : str = DeiTForImageClassificationWithTeacher.from_pretrained("facebook/deit-base-distilled-patch16-224" ).to( _A ) snake_case_ : List[Any] = self.default_image_processor snake_case_ : Any = prepare_img() snake_case_ : Tuple = image_processor(images=_A , return_tensors="pt" ).to(_A ) # forward pass with torch.no_grad(): snake_case_ : List[str] = model(**_A ) # verify the logits snake_case_ : Tuple = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , _A ) snake_case_ : str = torch.tensor([-1.0_2_6_6, 0.1_9_1_2, -1.2_8_6_1] ).to(_A ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _A , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def _A ( self :int ) -> Dict: '''simple docstring''' snake_case_ : Optional[Any] = DeiTModel.from_pretrained( "facebook/deit-base-distilled-patch16-224" , torch_dtype=torch.floataa , device_map="auto" ) snake_case_ : Optional[int] = self.default_image_processor snake_case_ : Union[str, Any] = prepare_img() snake_case_ : Optional[Any] = image_processor(images=_A , return_tensors="pt" ) snake_case_ : Dict = inputs.pixel_values.to(_A ) # forward pass to make sure inference works in fp16 with torch.no_grad(): snake_case_ : List[str] = model(_A )
705
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCamelCase : List[str] = logging.get_logger(__name__) __lowerCamelCase : int = { '''microsoft/cvt-13''': '''https://huggingface.co/microsoft/cvt-13/resolve/main/config.json''', # See all Cvt models at https://huggingface.co/models?filter=cvt } class A_ (a_ ): """simple docstring""" a__ = '''cvt''' def __init__( self :List[Any] , lowerCAmelCase__ :Optional[int]=3 , lowerCAmelCase__ :Any=[7, 3, 3] , lowerCAmelCase__ :Dict=[4, 2, 2] , lowerCAmelCase__ :Union[str, Any]=[2, 1, 1] , lowerCAmelCase__ :Any=[64, 192, 384] , lowerCAmelCase__ :List[str]=[1, 3, 6] , lowerCAmelCase__ :str=[1, 2, 10] , lowerCAmelCase__ :Any=[4.0, 4.0, 4.0] , lowerCAmelCase__ :int=[0.0, 0.0, 0.0] , lowerCAmelCase__ :Optional[Any]=[0.0, 0.0, 0.0] , lowerCAmelCase__ :Dict=[0.0, 0.0, 0.1] , lowerCAmelCase__ :List[Any]=[True, True, True] , lowerCAmelCase__ :List[Any]=[False, False, True] , lowerCAmelCase__ :Dict=["dw_bn", "dw_bn", "dw_bn"] , lowerCAmelCase__ :Any=[3, 3, 3] , lowerCAmelCase__ :Tuple=[1, 1, 1] , lowerCAmelCase__ :Optional[int]=[2, 2, 2] , lowerCAmelCase__ :Union[str, Any]=[1, 1, 1] , lowerCAmelCase__ :Any=[1, 1, 1] , lowerCAmelCase__ :List[str]=0.0_2 , lowerCAmelCase__ :Dict=1E-1_2 , **lowerCAmelCase__ :Optional[Any] , ) -> str: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) snake_case_ : int = num_channels snake_case_ : int = patch_sizes snake_case_ : Optional[Any] = patch_stride snake_case_ : Dict = patch_padding snake_case_ : Tuple = embed_dim snake_case_ : Optional[int] = num_heads snake_case_ : Union[str, Any] = depth snake_case_ : Optional[int] = mlp_ratio snake_case_ : Tuple = attention_drop_rate snake_case_ : str = drop_rate snake_case_ : Tuple = drop_path_rate snake_case_ : Any = qkv_bias snake_case_ : Union[str, Any] = cls_token snake_case_ : int = qkv_projection_method snake_case_ : Any = kernel_qkv snake_case_ : Union[str, Any] = padding_kv snake_case_ : str = stride_kv snake_case_ : Dict = padding_q snake_case_ : Tuple = stride_q snake_case_ : Any = initializer_range snake_case_ : Any = layer_norm_eps
656
0
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ ,__magic_name__ = 0 )-> Optional[int]: """simple docstring""" snake_case_ : Dict = length or len(lowerCAmelCase__ ) snake_case_ : Optional[int] = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: snake_case_, snake_case_ : List[str] = list_data[i + 1], list_data[i] snake_case_ : Optional[Any] = True return list_data if not swapped else bubble_sort(lowerCAmelCase__ ,length - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
706
'''simple docstring''' import sacrebleu as scb from packaging import version from sacrebleu import TER import datasets __lowerCamelCase : str = '''\ @inproceedings{snover-etal-2006-study, title = "A Study of Translation Edit Rate with Targeted Human Annotation", author = "Snover, Matthew and Dorr, Bonnie and Schwartz, Rich and Micciulla, Linnea and Makhoul, John", booktitle = "Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers", month = aug # " 8-12", year = "2006", address = "Cambridge, Massachusetts, USA", publisher = "Association for Machine Translation in the Americas", url = "https://aclanthology.org/2006.amta-papers.25", pages = "223--231", } @inproceedings{post-2018-call, title = "A Call for Clarity in Reporting {BLEU} Scores", author = "Post, Matt", booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers", month = oct, year = "2018", address = "Belgium, Brussels", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W18-6319", pages = "186--191", } ''' __lowerCamelCase : Dict = '''\ TER (Translation Edit Rate, also called Translation Error Rate) is a metric to quantify the edit operations that a hypothesis requires to match a reference translation. We use the implementation that is already present in sacrebleu (https://github.com/mjpost/sacreBLEU#ter), which in turn is inspired by the TERCOM implementation, which can be found here: https://github.com/jhclark/tercom. The implementation here is slightly different from sacrebleu in terms of the required input format. The length of the references and hypotheses lists need to be the same, so you may need to transpose your references compared to sacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534 See the README.md file at https://github.com/mjpost/sacreBLEU#ter for more information. ''' __lowerCamelCase : int = ''' Produces TER scores alongside the number of edits and reference length. Args: predictions (list of str): The system stream (a sequence of segments). references (list of list of str): A list of one or more reference streams (each a sequence of segments). normalized (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. ignore_punct (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. support_zh_ja_chars (boolean): If `True`, tokenization/normalization supports processing of Chinese characters, as well as Japanese Kanji, Hiragana, Katakana, and Phonetic Extensions of Katakana. Only applies if `normalized = True`. Defaults to `False`. case_sensitive (boolean): If `False`, makes all predictions and references lowercase to ignore differences in case. Defaults to `False`. Returns: \'score\' (float): TER score (num_edits / sum_ref_lengths * 100) \'num_edits\' (int): The cumulative number of edits \'ref_length\' (float): The cumulative average reference length Examples: Example 1: >>> predictions = ["does this sentence match??", ... "what about this sentence?", ... "What did the TER metric user say to the developer?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"], ... ["Your jokes are...", "...TERrible"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {\'score\': 150.0, \'num_edits\': 15, \'ref_length\': 10.0} Example 2: >>> predictions = ["does this sentence match??", ... "what about this sentence?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {\'score\': 62.5, \'num_edits\': 5, \'ref_length\': 8.0} Example 3: >>> predictions = ["does this sentence match??", ... "what about this sentence?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... normalized=True, ... case_sensitive=True) >>> print(results) {\'score\': 57.14285714285714, \'num_edits\': 6, \'ref_length\': 10.5} Example 4: >>> predictions = ["does this sentence match??", ... "what about this sentence?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {\'score\': 0.0, \'num_edits\': 0, \'ref_length\': 8.0} Example 5: >>> predictions = ["does this sentence match??", ... "what about this sentence?", ... "What did the TER metric user say to the developer?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"], ... ["Your jokes are...", "...TERrible"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {\'score\': 100.0, \'num_edits\': 10, \'ref_length\': 10.0} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): """simple docstring""" def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' if version.parse(scb.__version__ ) < version.parse("1.4.12" ): raise ImportWarning( "To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n" "You can install it with `pip install \"sacrebleu>=1.4.12\"`." ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="http://www.cs.umd.edu/~snover/tercom/" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Sequence(datasets.Value("string" , id="sequence" ) , id="references" ), } ) , codebase_urls=["https://github.com/mjpost/sacreBLEU#ter"] , reference_urls=[ "https://github.com/jhclark/tercom", ] , ) def _A ( self :Any , lowerCAmelCase__ :Any , lowerCAmelCase__ :str , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , ) -> Optional[Any]: '''simple docstring''' snake_case_ : List[str] = len(references[0] ) if any(len(lowerCAmelCase__ ) != references_per_prediction for refs in references ): raise ValueError("Sacrebleu requires the same number of references for each prediction" ) snake_case_ : List[str] = [[refs[i] for refs in references] for i in range(lowerCAmelCase__ )] snake_case_ : List[str] = TER( normalized=lowerCAmelCase__ , no_punct=lowerCAmelCase__ , asian_support=lowerCAmelCase__ , case_sensitive=lowerCAmelCase__ , ) snake_case_ : Any = sb_ter.corpus_score(lowerCAmelCase__ , lowerCAmelCase__ ) return {"score": output.score, "num_edits": output.num_edits, "ref_length": output.ref_length}
656
0
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" if p < 2: raise ValueError("p should not be less than 2!" ) elif p == 2: return True snake_case_ : Tuple = 4 snake_case_ : Union[str, Any] = (1 << p) - 1 for _ in range(p - 2 ): snake_case_ : Tuple = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
707
'''simple docstring''' from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : Any = { "repo_name": ["test_repo1", "test_repo2", "test_repo3"], "path": ["test_1.py", "test_2.py", "unit_test.py"], "content": ["a " * 20, "a " * 30, "b " * 7], } snake_case_ : int = Dataset.from_dict(__magic_name__ ) return dataset class A_ (a_ ): """simple docstring""" def _A ( self :List[str] ) -> str: '''simple docstring''' snake_case_ : Union[str, Any] = get_dataset() snake_case_ : Optional[int] = make_duplicate_clusters(lowerCAmelCase__ , 0.8_5 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case_ : Optional[int] = get_dataset() snake_case_, snake_case_ : List[Any] = deduplicate_dataset(lowerCAmelCase__ ) self.assertEqual(len(lowerCAmelCase__ ) , 2 ) print(lowerCAmelCase__ ) self.assertEqual(duplicate_clusters[0][0]["copies"] , 2 ) self.assertEqual(duplicate_clusters[0][0]["is_extreme"] , lowerCAmelCase__ )
656
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCamelCase : Optional[int] = logging.get_logger(__name__) __lowerCamelCase : Optional[int] = { '''facebook/data2vec-vision-base-ft''': ( '''https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json''' ), } class A_ (a_ ): """simple docstring""" a__ = '''data2vec-vision''' def __init__( self :List[str] , lowerCAmelCase__ :List[Any]=768 , lowerCAmelCase__ :List[Any]=12 , lowerCAmelCase__ :Dict=12 , lowerCAmelCase__ :Tuple=3_072 , lowerCAmelCase__ :int="gelu" , lowerCAmelCase__ :Tuple=0.0 , lowerCAmelCase__ :Dict=0.0 , lowerCAmelCase__ :Optional[int]=0.0_2 , lowerCAmelCase__ :Dict=1E-1_2 , lowerCAmelCase__ :Optional[Any]=224 , lowerCAmelCase__ :Any=16 , lowerCAmelCase__ :Tuple=3 , lowerCAmelCase__ :Dict=False , lowerCAmelCase__ :str=False , lowerCAmelCase__ :List[Any]=False , lowerCAmelCase__ :Union[str, Any]=False , lowerCAmelCase__ :List[Any]=0.1 , lowerCAmelCase__ :List[str]=0.1 , lowerCAmelCase__ :Optional[int]=True , lowerCAmelCase__ :Tuple=[3, 5, 7, 11] , lowerCAmelCase__ :Any=[1, 2, 3, 6] , lowerCAmelCase__ :Optional[int]=True , lowerCAmelCase__ :List[str]=0.4 , lowerCAmelCase__ :Optional[int]=256 , lowerCAmelCase__ :int=1 , lowerCAmelCase__ :str=False , lowerCAmelCase__ :int=255 , **lowerCAmelCase__ :List[Any] , ) -> Union[str, Any]: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) snake_case_ : int = hidden_size snake_case_ : int = num_hidden_layers snake_case_ : List[Any] = num_attention_heads snake_case_ : List[Any] = intermediate_size snake_case_ : List[str] = hidden_act snake_case_ : Dict = hidden_dropout_prob snake_case_ : Any = attention_probs_dropout_prob snake_case_ : List[str] = initializer_range snake_case_ : Union[str, Any] = layer_norm_eps snake_case_ : Any = image_size snake_case_ : Any = patch_size snake_case_ : List[Any] = num_channels snake_case_ : Union[str, Any] = use_mask_token snake_case_ : Any = use_absolute_position_embeddings snake_case_ : Tuple = use_relative_position_bias snake_case_ : Union[str, Any] = use_shared_relative_position_bias snake_case_ : Dict = layer_scale_init_value snake_case_ : Union[str, Any] = drop_path_rate snake_case_ : List[Any] = use_mean_pooling # decode head attributes (semantic segmentation) snake_case_ : List[Any] = out_indices snake_case_ : Optional[Any] = pool_scales # auxiliary head attributes (semantic segmentation) snake_case_ : Optional[Any] = use_auxiliary_head snake_case_ : List[Any] = auxiliary_loss_weight snake_case_ : Optional[Any] = auxiliary_channels snake_case_ : Union[str, Any] = auxiliary_num_convs snake_case_ : List[str] = auxiliary_concat_input snake_case_ : Dict = semantic_loss_ignore_index class A_ (a_ ): """simple docstring""" a__ = version.parse('''1.11''' ) @property def _A ( self :Optional[int] ) -> Optional[Any]: '''simple docstring''' return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def _A ( self :List[str] ) -> List[str]: '''simple docstring''' return 1E-4
708
'''simple docstring''' from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable __lowerCamelCase : Dict = { '''configuration_gpt_neox_japanese''': ['''GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoXJapaneseConfig'''], '''tokenization_gpt_neox_japanese''': ['''GPTNeoXJapaneseTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ '''GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoXJapaneseForCausalLM''', '''GPTNeoXJapaneseLayer''', '''GPTNeoXJapaneseModel''', '''GPTNeoXJapanesePreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neox_japanese import GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXJapaneseConfig from .tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox_japanese import ( GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseLayer, GPTNeoXJapaneseModel, GPTNeoXJapanesePreTrainedModel, ) else: import sys __lowerCamelCase : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
656
0
'''simple docstring''' import math def __UpperCAmelCase ( )-> None: """simple docstring""" snake_case_ : Optional[Any] = input("Enter message: " ) snake_case_ : Dict = int(input(F'''Enter key [2-{len(_A ) - 1}]: ''' ) ) snake_case_ : Any = input("Encryption/Decryption [e/d]: " ) if mode.lower().startswith("e" ): snake_case_ : Union[str, Any] = encrypt_message(_A ,_A ) elif mode.lower().startswith("d" ): snake_case_ : int = decrypt_message(_A ,_A ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(F'''Output:\n{text + '|'}''' ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> str: """simple docstring""" snake_case_ : str = [""] * key for col in range(_A ): snake_case_ : str = col while pointer < len(_A ): cipher_text[col] += message[pointer] pointer += key return "".join(_A ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> str: """simple docstring""" snake_case_ : str = math.ceil(len(_A ) / key ) snake_case_ : List[str] = key snake_case_ : Union[str, Any] = (num_cols * num_rows) - len(_A ) snake_case_ : Tuple = [""] * num_cols snake_case_ : Any = 0 snake_case_ : str = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): snake_case_ : List[str] = 0 row += 1 return "".join(_A ) if __name__ == "__main__": import doctest doctest.testmod() main()
709
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ )-> list[int]: """simple docstring""" if length <= 0 or not isinstance(__magic_name__ ,__magic_name__ ): raise ValueError("Length must be a positive integer." ) return [n * (2 * n - 1) for n in range(__magic_name__ )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=10))
656
0
'''simple docstring''' __lowerCamelCase : Any = { 'Pillow': 'Pillow<10.0.0', 'accelerate': 'accelerate>=0.20.3', 'av': 'av==9.2.0', 'beautifulsoup4': 'beautifulsoup4', 'black': 'black~=23.1', 'codecarbon': 'codecarbon==1.2.0', 'cookiecutter': 'cookiecutter==1.7.3', 'dataclasses': 'dataclasses', 'datasets': 'datasets!=2.5.0', 'decord': 'decord==0.6.0', 'deepspeed': 'deepspeed>=0.9.3', 'diffusers': 'diffusers', 'dill': 'dill<0.3.5', 'evaluate': 'evaluate>=0.2.0', 'fairscale': 'fairscale>0.3', 'faiss-cpu': 'faiss-cpu', 'fastapi': 'fastapi', 'filelock': 'filelock', 'flax': 'flax>=0.4.1,<=0.7.0', 'ftfy': 'ftfy', 'fugashi': 'fugashi>=1.0', 'GitPython': 'GitPython<3.1.19', 'hf-doc-builder': 'hf-doc-builder>=0.3.0', 'huggingface-hub': 'huggingface-hub>=0.14.1,<1.0', 'importlib_metadata': 'importlib_metadata', 'ipadic': 'ipadic>=1.0.0,<2.0', 'isort': 'isort>=5.5.4', 'jax': 'jax>=0.2.8,!=0.3.2,<=0.4.13', 'jaxlib': 'jaxlib>=0.1.65,<=0.4.13', 'jieba': 'jieba', 'kenlm': 'kenlm', 'keras-nlp': 'keras-nlp>=0.3.1', 'librosa': 'librosa', 'nltk': 'nltk', 'natten': 'natten>=0.14.6', 'numpy': 'numpy>=1.17', 'onnxconverter-common': 'onnxconverter-common', 'onnxruntime-tools': 'onnxruntime-tools>=1.4.2', 'onnxruntime': 'onnxruntime>=1.4.0', 'opencv-python': 'opencv-python', 'optuna': 'optuna', 'optax': 'optax>=0.0.8,<=0.1.4', 'packaging': 'packaging>=20.0', 'parameterized': 'parameterized', 'phonemizer': 'phonemizer', 'protobuf': 'protobuf', 'psutil': 'psutil', 'pyyaml': 'pyyaml>=5.1', 'pydantic': 'pydantic<2', 'pytest': 'pytest>=7.2.0', 'pytest-timeout': 'pytest-timeout', 'pytest-xdist': 'pytest-xdist', 'python': 'python>=3.8.0', 'ray[tune]': 'ray[tune]', 'regex': 'regex!=2019.12.17', 'requests': 'requests', 'rhoknp': 'rhoknp>=1.1.0,<1.3.1', 'rjieba': 'rjieba', 'rouge-score': 'rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1', 'ruff': 'ruff>=0.0.241,<=0.0.259', 'sacrebleu': 'sacrebleu>=1.4.12,<2.0.0', 'sacremoses': 'sacremoses', 'safetensors': 'safetensors>=0.3.1', 'sagemaker': 'sagemaker>=2.31.0', 'scikit-learn': 'scikit-learn', 'sentencepiece': 'sentencepiece>=0.1.91,!=0.1.92', 'sigopt': 'sigopt', 'starlette': 'starlette', 'sudachipy': 'sudachipy>=0.6.6', 'sudachidict_core': 'sudachidict_core>=20220729', 'tensorflow-cpu': 'tensorflow-cpu>=2.6,<2.14', 'tensorflow': 'tensorflow>=2.6,<2.14', 'tensorflow-text': 'tensorflow-text<2.14', 'tf2onnx': 'tf2onnx', 'timeout-decorator': 'timeout-decorator', 'timm': 'timm', 'tokenizers': 'tokenizers>=0.11.1,!=0.11.3,<0.14', 'torch': 'torch>=1.9,!=1.12.0', 'torchaudio': 'torchaudio', 'torchvision': 'torchvision', 'pyctcdecode': 'pyctcdecode>=0.4.0', 'tqdm': 'tqdm>=4.27', 'unidic': 'unidic>=1.0.2', 'unidic_lite': 'unidic_lite>=1.0.7', 'urllib3': 'urllib3<2.0.0', 'uvicorn': 'uvicorn', }
710
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def __UpperCAmelCase ( __magic_name__=None )-> List[str]: """simple docstring""" if subparsers is not None: snake_case_ : List[str] = subparsers.add_parser("test" ) else: snake_case_ : List[Any] = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" ,default=__magic_name__ ,help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) ,) if subparsers is not None: parser.set_defaults(func=__magic_name__ ) return parser def __UpperCAmelCase ( __magic_name__ )-> Tuple: """simple docstring""" snake_case_ : Optional[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: snake_case_ : str = script_name else: snake_case_ : Any = F'''--config_file={args.config_file} {script_name}''' snake_case_ : Union[str, Any] = ["accelerate-launch"] + test_args.split() snake_case_ : Optional[int] = execute_subprocess_async(__magic_name__ ,env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : Dict = test_command_parser() snake_case_ : Dict = parser.parse_args() test_command(__magic_name__ ) if __name__ == "__main__": main()
656
0
'''simple docstring''' import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel __lowerCamelCase : str = HfApi() __lowerCamelCase : List[str] = {} # fmt: off __lowerCamelCase : Optional[Any] = torch.tensor([ -0.7_515, -1.6_883, 0.2_420, 0.0_300, 0.6_347, 1.3_433, -1.1_743, -3.7_467, 1.2_342, -2.2_485, 0.4_636, 0.8_076, -0.7_991, 0.3_969, 0.8_498, 0.9_189, -1.8_887, -3.3_522, 0.7_639, 0.2_040, 0.6_271, -2.7_148, -1.6_316, 3.0_839, 0.3_186, 0.2_721, -0.9_759, -1.2_461, 2.6_257, 1.3_557 ]) __lowerCamelCase : Dict = torch.tensor([ -2.3_639, -2.5_344, 0.0_054, -0.6_674, 1.5_990, 1.0_158, 0.3_124, -2.1_436, 1.8_795, -2.5_429, -0.1_566, -0.3_973, 1.2_490, 2.6_447, 1.2_283, -0.5_208, -2.8_154, -3.5_119, 2.3_838, 1.2_033, 1.7_201, -2.1_256, -1.4_576, 2.7_948, 2.4_204, -0.9_752, -1.2_546, 0.8_027, 3.2_758, 3.1_365 ]) __lowerCamelCase : Union[str, Any] = torch.tensor([ -0.6_531, -0.6_891, -0.3_172, -0.5_375, -0.9_140, -0.5_367, -0.1_175, -0.7_869, -0.3_808, -0.4_513, -0.2_098, -0.0_083, 0.3_183, 0.5_140, 0.2_247, -0.1_304, -0.1_302, -0.2_802, -0.2_084, -0.2_025, -0.4_967, -0.4_873, -0.0_861, 0.6_925, 0.0_250, 0.1_290, -0.1_543, 0.6_316, 1.0_460, 1.4_943 ]) __lowerCamelCase : str = torch.tensor([ 0.0_911, 0.1_107, 0.0_182, 0.0_435, -0.0_805, -0.0_608, 0.0_381, 0.2_172, -0.0_280, 0.1_327, -0.0_299, -0.0_255, -0.0_050, -0.1_170, -0.1_046, 0.0_309, 0.1_367, 0.1_728, -0.0_533, -0.0_748, -0.0_534, 0.1_624, 0.0_384, -0.1_805, -0.0_707, 0.0_642, 0.0_220, -0.0_134, -0.1_333, -0.1_505 ]) __lowerCamelCase : Optional[Any] = torch.tensor([ 0.1_321, 0.1_337, 0.0_440, 0.0_622, -0.0_591, -0.0_370, 0.0_503, 0.2_133, -0.0_177, 0.1_415, -0.0_116, -0.0_112, 0.0_044, -0.0_980, -0.0_789, 0.0_395, 0.1_502, 0.1_785, -0.0_488, -0.0_514, -0.0_404, 0.1_539, 0.0_454, -0.1_559, -0.0_665, 0.0_659, 0.0_383, -0.0_005, -0.1_266, -0.1_386 ]) __lowerCamelCase : List[Any] = torch.tensor([ 0.1_154, 0.1_218, 0.0_307, 0.0_526, -0.0_711, -0.0_541, 0.0_366, 0.2_078, -0.0_267, 0.1_317, -0.0_226, -0.0_193, -0.0_014, -0.1_055, -0.0_902, 0.0_330, 0.1_391, 0.1_709, -0.0_562, -0.0_693, -0.0_560, 0.1_482, 0.0_381, -0.1_683, -0.0_681, 0.0_661, 0.0_331, -0.0_046, -0.1_268, -0.1_431 ]) __lowerCamelCase : Optional[int] = torch.tensor([ 0.1_192, 0.1_240, 0.0_414, 0.0_606, -0.0_557, -0.0_412, 0.0_430, 0.2_042, -0.0_200, 0.1_385, -0.0_115, -0.0_132, 0.0_017, -0.0_965, -0.0_802, 0.0_398, 0.1_433, 0.1_747, -0.0_458, -0.0_533, -0.0_407, 0.1_545, 0.0_419, -0.1_574, -0.0_645, 0.0_626, 0.0_341, -0.0_010, -0.1_199, -0.1_390 ]) __lowerCamelCase : Tuple = torch.tensor([ 0.1_075, 0.1_074, 0.0_205, 0.0_431, -0.0_774, -0.0_607, 0.0_298, 0.2_042, -0.0_320, 0.1_267, -0.0_281, -0.0_250, -0.0_064, -0.1_091, -0.0_946, 0.0_290, 0.1_328, 0.1_650, -0.0_580, -0.0_738, -0.0_586, 0.1_440, 0.0_337, -0.1_746, -0.0_712, 0.0_605, 0.0_250, -0.0_099, -0.1_316, -0.1_473 ]) __lowerCamelCase : Any = torch.tensor([ -1.4_572, -2.0_481, -0.0_414, -0.6_005, 1.4_136, 0.5_848, 0.4_028, -2.7_330, 1.2_212, -2.1_228, 0.2_155, 0.4_039, 0.7_662, 2.0_535, 0.7_477, -0.3_243, -2.1_758, -2.7_648, 1.6_947, 0.7_026, 1.2_338, -1.6_078, -0.8_682, 2.2_810, 1.8_574, -0.5_718, -0.5_586, -0.0_186, 2.3_415, 2.1_251]) __lowerCamelCase : Union[str, Any] = torch.tensor([ -1.3_690, -1.9_720, -0.4_090, -0.6_966, 1.4_660, 0.9_938, -0.1_385, -2.7_324, 0.7_736, -1.8_917, 0.2_923, 0.4_293, 0.1_693, 1.4_112, 1.1_887, -0.3_181, -2.2_160, -2.6_381, 1.3_170, 0.8_163, 0.9_240, -1.6_544, -0.6_099, 2.5_259, 1.6_430, -0.9_090, -0.9_392, -0.0_126, 2.4_268, 2.3_266 ]) __lowerCamelCase : Tuple = torch.tensor([ -1.3_525, -1.9_628, -0.3_956, -0.6_860, 1.4_664, 1.0_014, -0.1_259, -2.7_212, 0.7_772, -1.8_811, 0.2_996, 0.4_388, 0.1_704, 1.4_029, 1.1_701, -0.3_027, -2.2_053, -2.6_287, 1.3_350, 0.8_131, 0.9_274, -1.6_292, -0.6_098, 2.5_131, 1.6_505, -0.8_958, -0.9_298, -0.0_151, 2.4_257, 2.3_355 ]) __lowerCamelCase : Dict = torch.tensor([ -2.0_585, -2.7_897, -0.2_850, -0.8_940, 1.9_052, 0.5_702, 0.6_345, -3.8_959, 1.5_932, -3.2_319, 0.1_974, 0.0_287, 1.7_566, 2.6_543, 0.8_387, -0.5_351, -3.2_736, -4.3_375, 2.9_029, 1.6_390, 1.4_640, -2.1_701, -1.9_013, 2.9_341, 3.4_981, -0.6_255, -1.1_644, -0.1_591, 3.7_097, 3.2_066 ]) __lowerCamelCase : Tuple = torch.tensor([ -2.3_139, -2.5_594, -0.0_197, -0.6_785, 1.7_001, 1.1_606, 0.3_075, -2.1_740, 1.8_071, -2.5_630, -0.0_926, -0.3_811, 1.2_116, 2.6_246, 1.2_731, -0.5_398, -2.8_153, -3.6_140, 2.3_893, 1.3_262, 1.6_258, -2.1_856, -1.3_267, 2.8_395, 2.3_779, -1.0_623, -1.2_468, 0.8_959, 3.3_367, 3.2_243 ]) __lowerCamelCase : List[str] = torch.tensor([ -2.0_628, -2.7_667, -0.2_089, -0.8_263, 2.0_539, 0.5_992, 0.6_495, -3.8_336, 1.6_025, -3.2_817, 0.1_721, -0.0_633, 1.7_516, 2.7_039, 0.8_100, -0.5_908, -3.2_113, -4.4_343, 2.9_257, 1.3_632, 1.5_562, -2.1_489, -1.9_894, 3.0_560, 3.3_396, -0.7_328, -1.0_417, 0.0_383, 3.7_093, 3.2_343 ]) __lowerCamelCase : Union[str, Any] = torch.tensor([ -1.4_574, -2.0_569, -0.0_473, -0.6_117, 1.4_018, 0.5_769, 0.4_129, -2.7_344, 1.2_241, -2.1_397, 0.2_000, 0.3_937, 0.7_616, 2.0_453, 0.7_324, -0.3_391, -2.1_746, -2.7_744, 1.6_963, 0.6_921, 1.2_187, -1.6_172, -0.8_877, 2.2_439, 1.8_471, -0.5_839, -0.5_605, -0.0_464, 2.3_250, 2.1_219 ]) # fmt: on __lowerCamelCase : Any = api.list_models(filter='''diffusers''') for mod in models: if "google" in mod.author or mod.modelId == "CompVis/ldm-celebahq-256": __lowerCamelCase : Union[str, Any] = '''/home/patrick/google_checkpoints/''' + mod.modelId.split('''/''')[-1] print(f'''Started running {mod.modelId}!!!''') if mod.modelId.startswith('''CompVis'''): __lowerCamelCase : List[str] = UNetaDModel.from_pretrained(local_checkpoint, subfolder='''unet''') else: __lowerCamelCase : List[str] = UNetaDModel.from_pretrained(local_checkpoint) torch.manual_seed(0) random.seed(0) __lowerCamelCase : int = torch.randn(1, model.config.in_channels, model.config.sample_size, model.config.sample_size) __lowerCamelCase : Optional[int] = torch.tensor([10] * noise.shape[0]) with torch.no_grad(): __lowerCamelCase : Any = model(noise, time_step).sample assert torch.allclose( logits[0, 0, 0, :30], results['''_'''.join('''_'''.join(mod.modelId.split('''/''')).split('''-'''))], atol=1E-3 ) print(f'''{mod.modelId} has passed successfully!!!''')
711
'''simple docstring''' from scipy.stats import spearmanr import datasets __lowerCamelCase : str = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' __lowerCamelCase : int = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' __lowerCamelCase : List[str] = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): """simple docstring""" def _A ( self :str ) -> Union[str, Any]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } ) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"] , ) def _A ( self :Optional[int] , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :Optional[Any]=False ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[Any] = spearmanr(lowerCAmelCase__ , lowerCAmelCase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
656
0
'''simple docstring''' from math import sqrt def __UpperCAmelCase ( __magic_name__ )-> bool: """simple docstring""" assert isinstance(__snake_case ,__snake_case ) and ( number >= 0 ), "'number' must been an int and positive" snake_case_ : Optional[int] = True # 0 and 1 are none primes. if number <= 1: snake_case_ : Dict = False for divisor in range(2 ,int(round(sqrt(__snake_case ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: snake_case_ : Union[str, Any] = False break # precondition assert isinstance(__snake_case ,__snake_case ), "'status' must been from type bool" return status def __UpperCAmelCase ( __magic_name__ )-> Dict: """simple docstring""" assert isinstance(__snake_case ,__snake_case ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N snake_case_ : Any = list(range(2 ,n + 1 ) ) snake_case_ : Union[str, Any] = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(__snake_case ) ): for j in range(i + 1 ,len(__snake_case ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): snake_case_ : str = 0 # filters actual prime numbers. snake_case_ : str = [x for x in begin_list if x != 0] # precondition assert isinstance(__snake_case ,__snake_case ), "'ans' must been from type list" return ans def __UpperCAmelCase ( __magic_name__ )-> List[str]: """simple docstring""" assert isinstance(__snake_case ,__snake_case ) and (n > 2), "'N' must been an int and > 2" snake_case_ : Union[str, Any] = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 ,n + 1 ): if is_prime(__snake_case ): ans.append(__snake_case ) # precondition assert isinstance(__snake_case ,__snake_case ), "'ans' must been from type list" return ans def __UpperCAmelCase ( __magic_name__ )-> Optional[int]: """simple docstring""" assert isinstance(__snake_case ,__snake_case ) and number >= 0, "'number' must been an int and >= 0" snake_case_ : Any = [] # this list will be returns of the function. # potential prime number factors. snake_case_ : Tuple = 2 snake_case_ : List[str] = number if number == 0 or number == 1: ans.append(__snake_case ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(__snake_case ): while quotient != 1: if is_prime(__snake_case ) and (quotient % factor == 0): ans.append(__snake_case ) quotient /= factor else: factor += 1 else: ans.append(__snake_case ) # precondition assert isinstance(__snake_case ,__snake_case ), "'ans' must been from type list" return ans def __UpperCAmelCase ( __magic_name__ )-> Optional[Any]: """simple docstring""" assert isinstance(__snake_case ,__snake_case ) and ( number >= 0 ), "'number' bust been an int and >= 0" snake_case_ : Optional[int] = 0 # prime factorization of 'number' snake_case_ : Optional[Any] = prime_factorization(__snake_case ) snake_case_ : List[str] = max(__snake_case ) # precondition assert isinstance(__snake_case ,__snake_case ), "'ans' must been from type int" return ans def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" assert isinstance(__snake_case ,__snake_case ) and ( number >= 0 ), "'number' bust been an int and >= 0" snake_case_ : Tuple = 0 # prime factorization of 'number' snake_case_ : Dict = prime_factorization(__snake_case ) snake_case_ : Dict = min(__snake_case ) # precondition assert isinstance(__snake_case ,__snake_case ), "'ans' must been from type int" return ans def __UpperCAmelCase ( __magic_name__ )-> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case ,__snake_case ), "'number' must been an int" assert isinstance(number % 2 == 0 ,__snake_case ), "compare bust been from type bool" return number % 2 == 0 def __UpperCAmelCase ( __magic_name__ )-> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case ,__snake_case ), "'number' must been an int" assert isinstance(number % 2 != 0 ,__snake_case ), "compare bust been from type bool" return number % 2 != 0 def __UpperCAmelCase ( __magic_name__ )-> str: """simple docstring""" assert ( isinstance(__snake_case ,__snake_case ) and (number > 2) and is_even(__snake_case ) ), "'number' must been an int, even and > 2" snake_case_ : int = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' snake_case_ : List[str] = get_prime_numbers(__snake_case ) snake_case_ : List[Any] = len(__snake_case ) # run variable for while-loops. snake_case_ : Any = 0 snake_case_ : str = None # exit variable. for break up the loops snake_case_ : str = True while i < len_pn and loop: snake_case_ : List[str] = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: snake_case_ : Dict = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(__snake_case ,__snake_case ) and (len(__snake_case ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> str: """simple docstring""" assert ( isinstance(__snake_case ,__snake_case ) and isinstance(__snake_case ,__snake_case ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." snake_case_ : Tuple = 0 while numbera != 0: snake_case_ : str = numbera % numbera snake_case_ : List[str] = numbera snake_case_ : Optional[int] = rest # precondition assert isinstance(__snake_case ,__snake_case ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" assert ( isinstance(__snake_case ,__snake_case ) and isinstance(__snake_case ,__snake_case ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." snake_case_ : int = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' snake_case_ : Union[str, Any] = prime_factorization(__snake_case ) snake_case_ : Optional[Any] = prime_factorization(__snake_case ) elif numbera == 1 or numbera == 1: snake_case_ : List[Any] = [] snake_case_ : Tuple = [] snake_case_ : List[Any] = max(__snake_case ,__snake_case ) snake_case_ : List[str] = 0 snake_case_ : List[Any] = 0 snake_case_ : List[Any] = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: snake_case_ : Optional[Any] = prime_fac_a.count(__snake_case ) snake_case_ : Any = prime_fac_a.count(__snake_case ) for _ in range(max(__snake_case ,__snake_case ) ): ans *= n else: snake_case_ : Optional[int] = prime_fac_a.count(__snake_case ) for _ in range(__snake_case ): ans *= n done.append(__snake_case ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: snake_case_ : Union[str, Any] = prime_fac_a.count(__snake_case ) for _ in range(__snake_case ): ans *= n done.append(__snake_case ) # precondition assert isinstance(__snake_case ,__snake_case ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def __UpperCAmelCase ( __magic_name__ )-> Tuple: """simple docstring""" assert isinstance(__snake_case ,__snake_case ) and (n >= 0), "'number' must been a positive int" snake_case_ : Union[str, Any] = 0 snake_case_ : List[Any] = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(__snake_case ): ans += 1 # precondition assert isinstance(__snake_case ,__snake_case ) and is_prime( __snake_case ), "'ans' must been a prime number and from type int" return ans def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" assert ( is_prime(__snake_case ) and is_prime(__snake_case ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" snake_case_ : Optional[Any] = p_number_a + 1 # jump to the next number snake_case_ : Union[str, Any] = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(__snake_case ): number += 1 while number < p_number_a: ans.append(__snake_case ) number += 1 # fetch the next prime number. while not is_prime(__snake_case ): number += 1 # precondition assert ( isinstance(__snake_case ,__snake_case ) and ans[0] != p_number_a and ans[len(__snake_case ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" assert isinstance(__snake_case ,__snake_case ) and (n >= 1), "'n' must been int and >= 1" snake_case_ : str = [] # will be returned. for divisor in range(1 ,n + 1 ): if n % divisor == 0: ans.append(__snake_case ) # precondition assert ans[0] == 1 and ans[len(__snake_case ) - 1] == n, "Error in function getDivisiors(...)" return ans def __UpperCAmelCase ( __magic_name__ )-> str: """simple docstring""" assert isinstance(__snake_case ,__snake_case ) and ( number > 1 ), "'number' must been an int and >= 1" snake_case_ : List[str] = get_divisors(__snake_case ) # precondition assert ( isinstance(__snake_case ,__snake_case ) and (divisors[0] == 1) and (divisors[len(__snake_case ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Optional[Any]: """simple docstring""" assert ( isinstance(__snake_case ,__snake_case ) and isinstance(__snake_case ,__snake_case ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. snake_case_ : str = gcd(abs(__snake_case ) ,abs(__snake_case ) ) # precondition assert ( isinstance(__snake_case ,__snake_case ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def __UpperCAmelCase ( __magic_name__ )-> Optional[Any]: """simple docstring""" assert isinstance(__snake_case ,__snake_case ) and (n >= 0), "'n' must been a int and >= 0" snake_case_ : List[Any] = 1 # this will be return. for factor in range(1 ,n + 1 ): ans *= factor return ans def __UpperCAmelCase ( __magic_name__ )-> Union[str, Any]: """simple docstring""" assert isinstance(__snake_case ,__snake_case ) and (n >= 0), "'n' must been an int and >= 0" snake_case_ : Optional[Any] = 0 snake_case_ : Optional[int] = 1 snake_case_ : List[str] = 1 # this will be return for _ in range(n - 1 ): snake_case_ : Union[str, Any] = ans ans += fiba snake_case_ : List[Any] = tmp return ans
712
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): __lowerCamelCase : Optional[Any] = get_tests_dir('''fixtures/test_sentencepiece.model''') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right __lowerCamelCase : str = 128022 __lowerCamelCase : List[Any] = 128028 @require_sentencepiece class A_ (a_ , unittest.TestCase ): """simple docstring""" a__ = MaMaaaTokenizer a__ = False a__ = False a__ = True def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' super().setUp() snake_case_ : int = ["</s>", "<unk>", "▁This", "▁is", "▁a", "▁t", "est", "\u0120", "<pad>"] snake_case_ : Any = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) snake_case_ : Optional[int] = Path(self.tmpdirname ) save_json(lowerCAmelCase__ , save_dir / VOCAB_FILES_NAMES["vocab_file"] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(lowerCAmelCase__ , save_dir / VOCAB_FILES_NAMES["spm_file"] ) snake_case_ : Union[str, Any] = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def _A ( self :List[Any] , **lowerCAmelCase__ :List[Any] ) -> str: '''simple docstring''' return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def _A ( self :Optional[int] , lowerCAmelCase__ :Any ) -> Optional[int]: '''simple docstring''' return ( "This is a test", "This is a test", ) def _A ( self :List[str] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : str = "</s>" snake_case_ : Union[str, Any] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase__ ) , lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = self.get_tokenizer() snake_case_ : Any = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "</s>" ) self.assertEqual(vocab_keys[1] , "<unk>" ) self.assertEqual(vocab_keys[-1] , "<s>" ) self.assertEqual(len(lowerCAmelCase__ ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip("Skip this test while all models are still to be uploaded." ) def _A ( self :List[Any] ) -> Union[str, Any]: '''simple docstring''' pass def _A ( self :Optional[int] ) -> int: '''simple docstring''' snake_case_ : int = self.get_tokenizer() snake_case_ : List[str] = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowerCAmelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [2, 3, 4, 5, 6] , ) snake_case_ : Any = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(lowerCAmelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) snake_case_ : Any = tokenizer.convert_tokens_to_string(lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , "This is a test" ) @slow def _A ( self :Any ) -> List[Any]: '''simple docstring''' snake_case_ : int = {"input_ids": [[128_022, 110_108, 397, 11, 38_272, 2_247, 124_811, 285, 18_105, 1_586, 207, 7, 39_534, 4_428, 397, 1_019, 18_105, 1_586, 207, 7, 41_337, 16_786, 241, 7, 20_214, 17, 125_690, 10_398, 7, 44_378, 58_069, 68_342, 7_798, 7_343, 11, 299, 33_310, 4, 158, 37_350, 94_077, 4_569, 299, 33_310, 90, 4, 52_840, 290, 4, 31_270, 112, 299, 682, 4, 52_840, 39_953, 14_079, 193, 52_519, 90_894, 17_894, 120_697, 11, 40_445, 551, 17, 1_019, 52_519, 90_894, 17_756, 963, 11, 40_445, 480, 17, 9_792, 1_120, 5_173, 1_393, 6_240, 16_786, 241, 120_996, 28, 1_245, 1_393, 118_240, 11_123, 1_019, 93_612, 2_691, 10_618, 98_058, 120_409, 1_928, 279, 4, 40_683, 367, 178, 207, 1_019, 103, 103_121, 506, 65_296, 5, 2], [128_022, 21_217, 367, 117, 125_450, 128, 719, 7, 7_308, 40, 93_612, 12_669, 1_116, 16_704, 71, 17_785, 3_699, 15_592, 35, 144, 9_584, 241, 11_943, 713, 950, 799, 2_247, 88_427, 150, 149, 118_813, 120_706, 1_019, 106_906, 81_518, 28, 1_224, 22_799, 397, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [128_022, 1_658, 123_311, 5_155, 5_578, 4_722, 279, 14_947, 2_366, 1_120, 1_197, 14, 1_348, 9_232, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase__ , model_name="facebook/m2m100_418M" , revision="c168bae485c864188cf9aa0e4108b0b6934dc91e" , ) @require_torch @require_sentencepiece @require_tokenizers class A_ (unittest.TestCase ): """simple docstring""" a__ = '''facebook/m2m100_418M''' a__ = [ '''In my opinion, there are two levels of response from the French government.''', '''NSA Affair Emphasizes Complete Lack of Debate on Intelligence''', ] a__ = [ '''Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.''', '''L\'affaire NSA souligne l\'absence totale de débat sur le renseignement''', ] # fmt: off a__ = [EN_CODE, 593, 1949, 115781, 4, 71586, 4234, 60633, 126233, 432, 123808, 15592, 1197, 117132, 120618, 5, 2] @classmethod def _A ( cls :str ) -> int: '''simple docstring''' snake_case_ : MaMaaaTokenizer = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en" , tgt_lang="fr" ) snake_case_ : List[str] = 1 return cls def _A ( self :Tuple ) -> Union[str, Any]: '''simple docstring''' self.assertEqual(self.tokenizer.get_lang_id("ar" ) , 128_006 ) self.assertEqual(self.tokenizer.get_lang_id("en" ) , 128_022 ) self.assertEqual(self.tokenizer.get_lang_id("ro" ) , 128_076 ) self.assertEqual(self.tokenizer.get_lang_id("mr" ) , 128_063 ) def _A ( self :Optional[int] ) -> List[str]: '''simple docstring''' snake_case_ : Dict = self.tokenizer.get_vocab() self.assertEqual(len(lowerCAmelCase__ ) , self.tokenizer.vocab_size ) self.assertEqual(vocab["<unk>"] , 3 ) self.assertIn(self.tokenizer.get_lang_token("en" ) , lowerCAmelCase__ ) def _A ( self :Any ) -> Dict: '''simple docstring''' snake_case_ : List[str] = "en" snake_case_ : Dict = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' self.assertIn(lowerCAmelCase__ , self.tokenizer.all_special_ids ) # fmt: off snake_case_ : Dict = [FR_CODE, 5_364, 82, 8_642, 4, 294, 47, 8, 14_028, 136, 3_286, 9_706, 6, 90_797, 6, 144_012, 162, 88_128, 30_061, 5, 2] # fmt: on snake_case_ : List[str] = self.tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) snake_case_ : str = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertNotIn(self.tokenizer.eos_token , lowerCAmelCase__ ) def _A ( self :Tuple ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = tempfile.mkdtemp() snake_case_ : int = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(lowerCAmelCase__ ) snake_case_ : List[str] = MaMaaaTokenizer.from_pretrained(lowerCAmelCase__ ) self.assertDictEqual(new_tok.lang_token_to_id , lowerCAmelCase__ ) @require_torch def _A ( self :Optional[Any] ) -> str: '''simple docstring''' snake_case_ : Union[str, Any] = "en" snake_case_ : Tuple = "fr" snake_case_ : Optional[int] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=lowerCAmelCase__ , return_tensors="pt" ) snake_case_ : Dict = shift_tokens_right( batch["labels"] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: snake_case_ : str = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : List[str] = "mr" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) snake_case_ : int = "zh" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def _A ( self :str ) -> int: '''simple docstring''' snake_case_ : Dict = "mr" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) snake_case_ : Tuple = "zh" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Optional[int] = self.tokenizer._build_translation_inputs("A test" , return_tensors="pt" , src_lang="en" , tgt_lang="ar" ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { # en_XX, A, test, EOS "input_ids": [[128_022, 58, 4_183, 2]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 128_006, } , )
656
0
'''simple docstring''' import json import os import re import sys import urllib.request import requests from bsa import BeautifulSoup __lowerCamelCase : Dict = { '''User-Agent''': '''Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36''' ''' (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582''' } def __UpperCAmelCase ( __magic_name__ = "dhaka" ,__magic_name__ = 5 )-> Union[str, Any]: """simple docstring""" snake_case_ : Union[str, Any] = min(lowerCamelCase_ ,50 ) # Prevent abuse! snake_case_ : Union[str, Any] = { """q""": query, """tbm""": """isch""", """hl""": """en""", """ijn""": """0""", } snake_case_ : Union[str, Any] = requests.get("https://www.google.com/search" ,params=lowerCamelCase_ ,headers=lowerCamelCase_ ) snake_case_ : List[Any] = BeautifulSoup(html.text ,"html.parser" ) snake_case_ : Tuple = """""".join( re.findall(r"AF_initDataCallback\(([^<]+)\);" ,str(soup.select("script" ) ) ) ) snake_case_ : Optional[Any] = json.dumps(lowerCamelCase_ ) snake_case_ : Tuple = json.loads(lowerCamelCase_ ) snake_case_ : Optional[Any] = re.findall( r"\[\"GRID_STATE0\",null,\[\[1,\[0,\".*?\",(.*),\"All\"," ,lowerCamelCase_ ,) if not matched_google_image_data: return 0 snake_case_ : int = re.sub( r"\[\"(https\:\/\/encrypted-tbn0\.gstatic\.com\/images\?.*?)\",\d+,\d+\]" ,"" ,str(lowerCamelCase_ ) ,) snake_case_ : Any = re.findall( r"(?:'|,),\[\"(https:|http.*?)\",\d+,\d+\]" ,lowerCamelCase_ ,) for index, fixed_full_res_image in enumerate(lowerCamelCase_ ): if index >= max_images: return index snake_case_ : int = bytes(lowerCamelCase_ ,"ascii" ).decode( "unicode-escape" ) snake_case_ : Dict = bytes(lowerCamelCase_ ,"ascii" ).decode( "unicode-escape" ) snake_case_ : Union[str, Any] = urllib.request.build_opener() snake_case_ : List[str] = [ ( """User-Agent""", """Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36""" """ (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582""", ) ] urllib.request.install_opener(lowerCamelCase_ ) snake_case_ : Optional[int] = F'''query_{query.replace(' ' ,'_' )}''' if not os.path.exists(lowerCamelCase_ ): os.makedirs(lowerCamelCase_ ) urllib.request.urlretrieve( # noqa: S310 lowerCamelCase_ ,F'''{path_name}/original_size_img_{index}.jpg''' ) return index if __name__ == "__main__": try: __lowerCamelCase : List[Any] = download_images_from_google_query(sys.argv[1]) print(f'''{image_count} images were downloaded to disk.''') except IndexError: print('''Please provide a search term.''') raise
713
'''simple docstring''' import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py __lowerCamelCase : str = '''.''' # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) __lowerCamelCase : Tuple = [ '''Assert''', '''AssignVariableOp''', '''EmptyTensorList''', '''MergeV2Checkpoints''', '''ReadVariableOp''', '''ResourceGather''', '''RestoreV2''', '''SaveV2''', '''ShardedFilename''', '''StatefulPartitionedCall''', '''StaticRegexFullMatch''', '''VarHandleOp''', ] def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Tuple = SavedModel() snake_case_ : Dict = [] with open(os.path.join(__magic_name__ ,"utils" ,"tf_ops" ,"onnx.json" ) ) as f: snake_case_ : Dict = json.load(__magic_name__ )["opsets"] for i in range(1 ,opset + 1 ): onnx_ops.extend(onnx_opsets[str(__magic_name__ )] ) with open(__magic_name__ ,"rb" ) as f: saved_model.ParseFromString(f.read() ) snake_case_ : Tuple = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want snake_case_ : str = sorted(__magic_name__ ) snake_case_ : Optional[int] = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(__magic_name__ ) if strict and len(__magic_name__ ) > 0: raise Exception(F'''Found the following incompatible ops for the opset {opset}:\n''' + incompatible_ops ) elif len(__magic_name__ ) > 0: print(F'''Found the following incompatible ops for the opset {opset}:''' ) print(*__magic_name__ ,sep="\n" ) else: print(F'''The saved model {saved_model_path} can properly be converted with ONNX.''' ) if __name__ == "__main__": __lowerCamelCase : Any = argparse.ArgumentParser() parser.add_argument('''--saved_model_path''', help='''Path of the saved model to check (the .pb file).''') parser.add_argument( '''--opset''', default=12, type=int, help='''The ONNX opset against which the model has to be tested.''' ) parser.add_argument( '''--framework''', choices=['''onnx'''], default='''onnx''', help='''Frameworks against which to test the saved model.''' ) parser.add_argument( '''--strict''', action='''store_true''', help='''Whether make the checking strict (raise errors) or not (raise warnings)''' ) __lowerCamelCase : Dict = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
656
0
from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import ( BackboneOutput, BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import ( add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging, replace_return_docstrings, ) from ...utils.backbone_utils import BackboneMixin from .configuration_resnet import ResNetConfig __lowerCamelCase : List[str] = logging.get_logger(__name__) # General docstring __lowerCamelCase : List[str] = '''ResNetConfig''' # Base docstring __lowerCamelCase : List[Any] = '''microsoft/resnet-50''' __lowerCamelCase : Union[str, Any] = [1, 2048, 7, 7] # Image classification docstring __lowerCamelCase : List[str] = '''microsoft/resnet-50''' __lowerCamelCase : int = '''tiger cat''' __lowerCamelCase : Any = [ '''microsoft/resnet-50''', # See all resnet models at https://huggingface.co/models?filter=resnet ] class A_ (nn.Module ): """simple docstring""" def __init__( self :Any , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :int = 3 , lowerCAmelCase__ :int = 1 , lowerCAmelCase__ :str = "relu" ) -> Union[str, Any]: '''simple docstring''' super().__init__() snake_case_ : Dict = nn.Convad( UpperCamelCase_ , UpperCamelCase_ , kernel_size=UpperCamelCase_ , stride=UpperCamelCase_ , padding=kernel_size // 2 , bias=UpperCamelCase_ ) snake_case_ : Tuple = nn.BatchNormad(UpperCamelCase_ ) snake_case_ : List[str] = ACTaFN[activation] if activation is not None else nn.Identity() def _A ( self :Optional[Any] , lowerCAmelCase__ :Tensor ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = self.convolution(UpperCamelCase_ ) snake_case_ : Tuple = self.normalization(UpperCamelCase_ ) snake_case_ : Dict = self.activation(UpperCamelCase_ ) return hidden_state class A_ (nn.Module ): """simple docstring""" def __init__( self :Union[str, Any] , lowerCAmelCase__ :ResNetConfig ) -> Optional[int]: '''simple docstring''' super().__init__() snake_case_ : Tuple = ResNetConvLayer( config.num_channels , config.embedding_size , kernel_size=7 , stride=2 , activation=config.hidden_act ) snake_case_ : Tuple = nn.MaxPoolad(kernel_size=3 , stride=2 , padding=1 ) snake_case_ : Union[str, Any] = config.num_channels def _A ( self :str , lowerCAmelCase__ :Tensor ) -> Dict: '''simple docstring''' snake_case_ : Any = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) snake_case_ : Optional[int] = self.embedder(UpperCamelCase_ ) snake_case_ : int = self.pooler(UpperCamelCase_ ) return embedding class A_ (nn.Module ): """simple docstring""" def __init__( self :Tuple , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :int = 2 ) -> Optional[Any]: '''simple docstring''' super().__init__() snake_case_ : str = nn.Convad(UpperCamelCase_ , UpperCamelCase_ , kernel_size=1 , stride=UpperCamelCase_ , bias=UpperCamelCase_ ) snake_case_ : str = nn.BatchNormad(UpperCamelCase_ ) def _A ( self :int , lowerCAmelCase__ :Tensor ) -> Any: '''simple docstring''' snake_case_ : Tuple = self.convolution(UpperCamelCase_ ) snake_case_ : Optional[Any] = self.normalization(UpperCamelCase_ ) return hidden_state class A_ (nn.Module ): """simple docstring""" def __init__( self :Optional[Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :int = 1 , lowerCAmelCase__ :str = "relu" ) -> Any: '''simple docstring''' super().__init__() snake_case_ : Optional[Any] = in_channels != out_channels or stride != 1 snake_case_ : List[str] = ( ResNetShortCut(UpperCamelCase_ , UpperCamelCase_ , stride=UpperCamelCase_ ) if should_apply_shortcut else nn.Identity() ) snake_case_ : Optional[Any] = nn.Sequential( ResNetConvLayer(UpperCamelCase_ , UpperCamelCase_ , stride=UpperCamelCase_ ) , ResNetConvLayer(UpperCamelCase_ , UpperCamelCase_ , activation=UpperCamelCase_ ) , ) snake_case_ : Any = ACTaFN[activation] def _A ( self :str , lowerCAmelCase__ :Tuple ) -> int: '''simple docstring''' snake_case_ : Union[str, Any] = hidden_state snake_case_ : str = self.layer(UpperCamelCase_ ) snake_case_ : Dict = self.shortcut(UpperCamelCase_ ) hidden_state += residual snake_case_ : List[str] = self.activation(UpperCamelCase_ ) return hidden_state class A_ (nn.Module ): """simple docstring""" def __init__( self :Optional[Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :int = 1 , lowerCAmelCase__ :str = "relu" , lowerCAmelCase__ :int = 4 ) -> Optional[Any]: '''simple docstring''' super().__init__() snake_case_ : Optional[int] = in_channels != out_channels or stride != 1 snake_case_ : Optional[Any] = out_channels // reduction snake_case_ : str = ( ResNetShortCut(UpperCamelCase_ , UpperCamelCase_ , stride=UpperCamelCase_ ) if should_apply_shortcut else nn.Identity() ) snake_case_ : List[Any] = nn.Sequential( ResNetConvLayer(UpperCamelCase_ , UpperCamelCase_ , kernel_size=1 ) , ResNetConvLayer(UpperCamelCase_ , UpperCamelCase_ , stride=UpperCamelCase_ ) , ResNetConvLayer(UpperCamelCase_ , UpperCamelCase_ , kernel_size=1 , activation=UpperCamelCase_ ) , ) snake_case_ : List[str] = ACTaFN[activation] def _A ( self :int , lowerCAmelCase__ :List[str] ) -> int: '''simple docstring''' snake_case_ : Tuple = hidden_state snake_case_ : Optional[int] = self.layer(UpperCamelCase_ ) snake_case_ : Tuple = self.shortcut(UpperCamelCase_ ) hidden_state += residual snake_case_ : str = self.activation(UpperCamelCase_ ) return hidden_state class A_ (nn.Module ): """simple docstring""" def __init__( self :Union[str, Any] , lowerCAmelCase__ :ResNetConfig , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :int = 2 , lowerCAmelCase__ :int = 2 , ) -> Union[str, Any]: '''simple docstring''' super().__init__() snake_case_ : Any = ResNetBottleNeckLayer if config.layer_type == "bottleneck" else ResNetBasicLayer snake_case_ : List[Any] = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer(UpperCamelCase_ , UpperCamelCase_ , stride=UpperCamelCase_ , activation=config.hidden_act ) , *[layer(UpperCamelCase_ , UpperCamelCase_ , activation=config.hidden_act ) for _ in range(depth - 1 )] , ) def _A ( self :Any , lowerCAmelCase__ :Tensor ) -> int: '''simple docstring''' snake_case_ : Union[str, Any] = input for layer in self.layers: snake_case_ : List[Any] = layer(UpperCamelCase_ ) return hidden_state class A_ (nn.Module ): """simple docstring""" def __init__( self :Any , lowerCAmelCase__ :ResNetConfig ) -> List[str]: '''simple docstring''' super().__init__() snake_case_ : Any = nn.ModuleList([] ) # based on `downsample_in_first_stage` the first layer of the first stage may or may not downsample the input self.stages.append( ResNetStage( UpperCamelCase_ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) snake_case_ : str = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(UpperCamelCase_ , config.depths[1:] ): self.stages.append(ResNetStage(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , depth=UpperCamelCase_ ) ) def _A ( self :Tuple , lowerCAmelCase__ :Tensor , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = True ) -> str: '''simple docstring''' snake_case_ : Optional[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: snake_case_ : Any = hidden_states + (hidden_state,) snake_case_ : Any = stage_module(UpperCamelCase_ ) if output_hidden_states: snake_case_ : Any = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention( last_hidden_state=UpperCamelCase_ , hidden_states=UpperCamelCase_ , ) class A_ (__lowerCamelCase ): """simple docstring""" a__ = ResNetConfig a__ = 'resnet' a__ = 'pixel_values' a__ = True def _A ( self :List[str] , lowerCAmelCase__ :Any ) -> List[str]: '''simple docstring''' if isinstance(UpperCamelCase_ , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode="fan_out" , nonlinearity="relu" ) elif isinstance(UpperCamelCase_ , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def _A ( self :Union[str, Any] , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :List[str]=False ) -> str: '''simple docstring''' if isinstance(UpperCamelCase_ , UpperCamelCase_ ): snake_case_ : List[str] = value __lowerCamelCase : Dict = R'''\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it\n as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`ResNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n''' __lowerCamelCase : str = R'''\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConvNextImageProcessor.__call__`] for details.\n\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n''' @add_start_docstrings( '''The bare ResNet model outputting raw features without any specific head on top.''' , __lowerCamelCase , ) class A_ (__lowerCamelCase ): """simple docstring""" def __init__( self :int , lowerCAmelCase__ :Optional[Any] ) -> List[str]: '''simple docstring''' super().__init__(UpperCamelCase_ ) snake_case_ : str = config snake_case_ : Optional[int] = ResNetEmbeddings(UpperCamelCase_ ) snake_case_ : List[Any] = ResNetEncoder(UpperCamelCase_ ) snake_case_ : List[str] = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCamelCase_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCamelCase_ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def _A ( self :Dict , lowerCAmelCase__ :Tensor , lowerCAmelCase__ :Optional[bool] = None , lowerCAmelCase__ :Optional[bool] = None ) -> Dict: '''simple docstring''' snake_case_ : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) snake_case_ : Dict = return_dict if return_dict is not None else self.config.use_return_dict snake_case_ : str = self.embedder(UpperCamelCase_ ) snake_case_ : int = self.encoder( UpperCamelCase_ , output_hidden_states=UpperCamelCase_ , return_dict=UpperCamelCase_ ) snake_case_ : str = encoder_outputs[0] snake_case_ : Optional[int] = self.pooler(UpperCamelCase_ ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=UpperCamelCase_ , pooler_output=UpperCamelCase_ , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( '''\n ResNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n ''' , __lowerCamelCase , ) class A_ (__lowerCamelCase ): """simple docstring""" def __init__( self :str , lowerCAmelCase__ :int ) -> List[Any]: '''simple docstring''' super().__init__(UpperCamelCase_ ) snake_case_ : int = config.num_labels snake_case_ : Tuple = ResNetModel(UpperCamelCase_ ) # classification head snake_case_ : Any = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCamelCase_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCamelCase_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def _A ( self :Optional[Any] , lowerCAmelCase__ :Optional[torch.FloatTensor] = None , lowerCAmelCase__ :Optional[torch.LongTensor] = None , lowerCAmelCase__ :Optional[bool] = None , lowerCAmelCase__ :Optional[bool] = None , ) -> int: '''simple docstring''' snake_case_ : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict snake_case_ : Optional[Any] = self.resnet(UpperCamelCase_ , output_hidden_states=UpperCamelCase_ , return_dict=UpperCamelCase_ ) snake_case_ : Optional[Any] = outputs.pooler_output if return_dict else outputs[1] snake_case_ : Any = self.classifier(UpperCamelCase_ ) snake_case_ : Optional[int] = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: snake_case_ : Optional[int] = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): snake_case_ : int = "single_label_classification" else: snake_case_ : Union[str, Any] = "multi_label_classification" if self.config.problem_type == "regression": snake_case_ : Tuple = MSELoss() if self.num_labels == 1: snake_case_ : List[Any] = loss_fct(logits.squeeze() , labels.squeeze() ) else: snake_case_ : Dict = loss_fct(UpperCamelCase_ , UpperCamelCase_ ) elif self.config.problem_type == "single_label_classification": snake_case_ : Optional[Any] = CrossEntropyLoss() snake_case_ : Optional[int] = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": snake_case_ : int = BCEWithLogitsLoss() snake_case_ : Optional[int] = loss_fct(UpperCamelCase_ , UpperCamelCase_ ) if not return_dict: snake_case_ : Dict = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=UpperCamelCase_ , logits=UpperCamelCase_ , hidden_states=outputs.hidden_states ) @add_start_docstrings( '''\n ResNet backbone, to be used with frameworks like DETR and MaskFormer.\n ''' , __lowerCamelCase , ) class A_ (__lowerCamelCase , __lowerCamelCase ): """simple docstring""" def __init__( self :Dict , lowerCAmelCase__ :Optional[Any] ) -> List[Any]: '''simple docstring''' super().__init__(UpperCamelCase_ ) super()._init_backbone(UpperCamelCase_ ) snake_case_ : Tuple = [config.embedding_size] + config.hidden_sizes snake_case_ : Any = ResNetEmbeddings(UpperCamelCase_ ) snake_case_ : List[Any] = ResNetEncoder(UpperCamelCase_ ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCamelCase_ ) @replace_return_docstrings(output_type=UpperCamelCase_ , config_class=_CONFIG_FOR_DOC ) def _A ( self :Dict , lowerCAmelCase__ :Tensor , lowerCAmelCase__ :Optional[bool] = None , lowerCAmelCase__ :Optional[bool] = None ) -> List[str]: '''simple docstring''' snake_case_ : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict snake_case_ : Any = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) snake_case_ : str = self.embedder(UpperCamelCase_ ) snake_case_ : Optional[Any] = self.encoder(UpperCamelCase_ , output_hidden_states=UpperCamelCase_ , return_dict=UpperCamelCase_ ) snake_case_ : str = outputs.hidden_states snake_case_ : Optional[int] = () for idx, stage in enumerate(self.stage_names ): if stage in self.out_features: feature_maps += (hidden_states[idx],) if not return_dict: snake_case_ : Dict = (feature_maps,) if output_hidden_states: output += (outputs.hidden_states,) return output return BackboneOutput( feature_maps=UpperCamelCase_ , hidden_states=outputs.hidden_states if output_hidden_states else None , attentions=UpperCamelCase_ , )
714
'''simple docstring''' import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal __lowerCamelCase : Optional[Any] = datasets.utils.logging.get_logger(__name__) __lowerCamelCase : List[str] = ['''names''', '''prefix'''] __lowerCamelCase : int = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] __lowerCamelCase : str = ['''encoding_errors''', '''on_bad_lines'''] __lowerCamelCase : Optional[Any] = ['''date_format'''] @dataclass class A_ (datasets.BuilderConfig ): """simple docstring""" a__ = "," a__ = None a__ = "infer" a__ = None a__ = None a__ = None a__ = None a__ = None a__ = True a__ = None a__ = None a__ = None a__ = None a__ = False a__ = None a__ = None a__ = None a__ = True a__ = True a__ = False a__ = True a__ = None a__ = "." a__ = None a__ = '"' a__ = 0 a__ = None a__ = None a__ = None a__ = None a__ = True a__ = True a__ = 0 a__ = True a__ = False a__ = None a__ = 10000 a__ = None a__ = "strict" a__ = "error" a__ = None def _A ( self :List[str] ) -> Any: '''simple docstring''' if self.delimiter is not None: snake_case_ : Tuple = self.delimiter if self.column_names is not None: snake_case_ : List[Any] = self.column_names @property def _A ( self :Optional[Any] ) -> int: '''simple docstring''' snake_case_ : Optional[int] = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class A_ (datasets.ArrowBasedBuilder ): """simple docstring""" a__ = CsvConfig def _A ( self :Optional[Any] ) -> Optional[Any]: '''simple docstring''' return datasets.DatasetInfo(features=self.config.features ) def _A ( self :Tuple , lowerCAmelCase__ :Dict ) -> List[Any]: '''simple docstring''' if not self.config.data_files: raise ValueError(F'''At least one data file must be specified, but got data_files={self.config.data_files}''' ) snake_case_ : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): snake_case_ : int = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ : List[str] = [files] snake_case_ : Tuple = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] snake_case_ : str = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ : str = [files] snake_case_ : Any = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def _A ( self :List[Any] , lowerCAmelCase__ :pa.Table ) -> pa.Table: '''simple docstring''' if self.config.features is not None: snake_case_ : int = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast snake_case_ : Optional[Any] = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example snake_case_ : Dict = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def _A ( self :Dict , lowerCAmelCase__ :Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str snake_case_ : str = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): snake_case_ : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): snake_case_ : Optional[int] = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(F'''Failed to read file \'{file}\' with error {type(lowerCAmelCase__ )}: {e}''' ) raise
656
0
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class A_ (unittest.TestCase ): """simple docstring""" def __init__( self :List[Any] , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :List[Any]=7 , lowerCAmelCase__ :Dict=3 , lowerCAmelCase__ :Union[str, Any]=30 , lowerCAmelCase__ :Optional[Any]=400 , lowerCAmelCase__ :Any=True , lowerCAmelCase__ :List[Any]=None , lowerCAmelCase__ :Tuple=True , lowerCAmelCase__ :str=[0.5, 0.5, 0.5] , lowerCAmelCase__ :List[Any]=[0.5, 0.5, 0.5] , lowerCAmelCase__ :Any=True , lowerCAmelCase__ :Union[str, Any]=1 / 255 , lowerCAmelCase__ :List[str]=True , ) -> Optional[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1_333} snake_case_ : str = parent snake_case_ : int = batch_size snake_case_ : Optional[int] = num_channels snake_case_ : Tuple = min_resolution snake_case_ : Dict = max_resolution snake_case_ : Union[str, Any] = do_resize snake_case_ : Dict = size snake_case_ : Any = do_normalize snake_case_ : str = image_mean snake_case_ : Dict = image_std snake_case_ : Optional[Any] = do_rescale snake_case_ : Union[str, Any] = rescale_factor snake_case_ : List[Any] = do_pad def _A ( self :Union[str, Any] ) -> str: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def _A ( self :Tuple , lowerCAmelCase__ :List[str] , lowerCAmelCase__ :Optional[int]=False ) -> Tuple: '''simple docstring''' if not batched: snake_case_ : Optional[int] = image_inputs[0] if isinstance(lowerCAmelCase__ , Image.Image ): snake_case_ : Union[str, Any] = image.size else: snake_case_ : Optional[Any] = image.shape[1], image.shape[2] if w < h: snake_case_ : Union[str, Any] = int(self.size["shortest_edge"] * h / w ) snake_case_ : Optional[int] = self.size["shortest_edge"] elif w > h: snake_case_ : Tuple = self.size["shortest_edge"] snake_case_ : Dict = int(self.size["shortest_edge"] * w / h ) else: snake_case_ : List[str] = self.size["shortest_edge"] snake_case_ : List[str] = self.size["shortest_edge"] else: snake_case_ : Dict = [] for image in image_inputs: snake_case_ : List[str] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) snake_case_ : Optional[Any] = max(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : item[0] )[0] snake_case_ : Optional[Any] = max(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class A_ (_UpperCamelCase , unittest.TestCase ): """simple docstring""" a__ = ConditionalDetrImageProcessor if is_vision_available() else None def _A ( self :str ) -> Any: '''simple docstring''' snake_case_ : str = ConditionalDetrImageProcessingTester(self ) @property def _A ( self :Tuple ) -> Tuple: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def _A ( self :List[Any] ) -> Any: '''simple docstring''' snake_case_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCAmelCase__ , "image_mean" ) ) self.assertTrue(hasattr(lowerCAmelCase__ , "image_std" ) ) self.assertTrue(hasattr(lowerCAmelCase__ , "do_normalize" ) ) self.assertTrue(hasattr(lowerCAmelCase__ , "do_resize" ) ) self.assertTrue(hasattr(lowerCAmelCase__ , "size" ) ) def _A ( self :Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1_333} ) self.assertEqual(image_processor.do_pad , lowerCAmelCase__ ) snake_case_ : Tuple = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCAmelCase__ ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCAmelCase__ ) def _A ( self :Optional[int] ) -> Optional[int]: '''simple docstring''' pass def _A ( self :Optional[Any] ) -> List[str]: '''simple docstring''' snake_case_ : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , Image.Image ) # Test not batched input snake_case_ : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ : int = self.image_processor_tester.get_expected_values(lowerCAmelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : str = self.image_processor_tester.get_expected_values(lowerCAmelCase__ , batched=lowerCAmelCase__ ) snake_case_ : Dict = image_processing(lowerCAmelCase__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _A ( self :Any ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , numpify=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , np.ndarray ) # Test not batched input snake_case_ : Tuple = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ : List[str] = self.image_processor_tester.get_expected_values(lowerCAmelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : List[Any] = image_processing(lowerCAmelCase__ , return_tensors="pt" ).pixel_values snake_case_ : Optional[int] = self.image_processor_tester.get_expected_values(lowerCAmelCase__ , batched=lowerCAmelCase__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _A ( self :List[str] ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , torchify=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , torch.Tensor ) # Test not batched input snake_case_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ : Any = self.image_processor_tester.get_expected_values(lowerCAmelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : Optional[Any] = image_processing(lowerCAmelCase__ , return_tensors="pt" ).pixel_values snake_case_ : Optional[int] = self.image_processor_tester.get_expected_values(lowerCAmelCase__ , batched=lowerCAmelCase__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def _A ( self :int ) -> int: '''simple docstring''' snake_case_ : int = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: snake_case_ : str = json.loads(f.read() ) snake_case_ : Dict = {"image_id": 39_769, "annotations": target} # encode them snake_case_ : Union[str, Any] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) snake_case_ : List[str] = image_processing(images=lowerCAmelCase__ , annotations=lowerCAmelCase__ , return_tensors="pt" ) # verify pixel values snake_case_ : Union[str, Any] = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCAmelCase__ ) snake_case_ : int = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCAmelCase__ , atol=1E-4 ) ) # verify area snake_case_ : Any = torch.tensor([5_887.9_600, 11_250.2_061, 489_353.8_438, 837_122.7_500, 147_967.5_156, 165_732.3_438] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCAmelCase__ ) ) # verify boxes snake_case_ : Union[str, Any] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCAmelCase__ ) snake_case_ : List[Any] = torch.tensor([0.5_5_0_3, 0.2_7_6_5, 0.0_6_0_4, 0.2_2_1_5] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCAmelCase__ , atol=1E-3 ) ) # verify image_id snake_case_ : str = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCAmelCase__ ) ) # verify is_crowd snake_case_ : Optional[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCAmelCase__ ) ) # verify class_labels snake_case_ : str = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCAmelCase__ ) ) # verify orig_size snake_case_ : str = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCAmelCase__ ) ) # verify size snake_case_ : Optional[Any] = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCAmelCase__ ) ) @slow def _A ( self :Dict ) -> List[str]: '''simple docstring''' snake_case_ : Any = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: snake_case_ : str = json.loads(f.read() ) snake_case_ : Tuple = {"file_name": "000000039769.png", "image_id": 39_769, "segments_info": target} snake_case_ : str = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them snake_case_ : Union[str, Any] = ConditionalDetrImageProcessor(format="coco_panoptic" ) snake_case_ : Dict = image_processing(images=lowerCAmelCase__ , annotations=lowerCAmelCase__ , masks_path=lowerCAmelCase__ , return_tensors="pt" ) # verify pixel values snake_case_ : Optional[int] = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCAmelCase__ ) snake_case_ : Union[str, Any] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCAmelCase__ , atol=1E-4 ) ) # verify area snake_case_ : Tuple = torch.tensor([147_979.6_875, 165_527.0_469, 484_638.5_938, 11_292.9_375, 5_879.6_562, 7_634.1_147] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCAmelCase__ ) ) # verify boxes snake_case_ : Optional[Any] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCAmelCase__ ) snake_case_ : List[str] = torch.tensor([0.2_6_2_5, 0.5_4_3_7, 0.4_6_8_8, 0.8_6_2_5] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCAmelCase__ , atol=1E-3 ) ) # verify image_id snake_case_ : int = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCAmelCase__ ) ) # verify is_crowd snake_case_ : int = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCAmelCase__ ) ) # verify class_labels snake_case_ : Union[str, Any] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCAmelCase__ ) ) # verify masks snake_case_ : Union[str, Any] = 822_873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCAmelCase__ ) # verify orig_size snake_case_ : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCAmelCase__ ) ) # verify size snake_case_ : str = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCAmelCase__ ) )
715
'''simple docstring''' import json import os import unittest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A_ (a_ , unittest.TestCase ): """simple docstring""" a__ = MgpstrTokenizer a__ = False a__ = {} a__ = False def _A ( self :List[str] ) -> List[str]: '''simple docstring''' super().setUp() # fmt: off snake_case_ : Dict = ["[GO]", "[s]", "0", "1", "2", "3", "4", "5", "6", "7", "8", "9", "a", "b", "c", "d", "e", "f", "g", "h", "i", "j", "k", "l", "m", "n", "o", "p", "q", "r", "s", "t", "u", "v", "w", "x", "y", "z"] # fmt: on snake_case_ : List[str] = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) snake_case_ : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase__ ) + "\n" ) def _A ( self :Optional[Any] , **lowerCAmelCase__ :Optional[Any] ) -> Dict: '''simple docstring''' return MgpstrTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def _A ( self :Dict , lowerCAmelCase__ :Any ) -> str: '''simple docstring''' snake_case_ : Dict = "tester" snake_case_ : Tuple = "tester" return input_text, output_text @unittest.skip("MGP-STR always lower cases letters." ) def _A ( self :Dict ) -> str: '''simple docstring''' pass def _A ( self :Tuple ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[str] = self.get_tokenizers(do_lower_case=lowerCAmelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): snake_case_ : Tuple = "[SPECIAL_TOKEN]" tokenizer.add_special_tokens({"cls_token": special_token} ) snake_case_ : str = tokenizer.encode([special_token] , add_special_tokens=lowerCAmelCase__ ) self.assertEqual(len(lowerCAmelCase__ ) , 1 ) snake_case_ : Tuple = tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) self.assertTrue(special_token not in decoded ) def _A ( self :int ) -> List[str]: '''simple docstring''' snake_case_ : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): snake_case_, snake_case_ : str = self.get_input_output_texts(lowerCAmelCase__ ) snake_case_ : Union[str, Any] = tokenizer.tokenize(lowerCAmelCase__ ) snake_case_ : List[Any] = tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) snake_case_ : Dict = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : List[str] = tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) self.assertNotEqual(len(lowerCAmelCase__ ) , 0 ) snake_case_ : List[str] = tokenizer.decode(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertEqual(text_a.replace(" " , "" ) , lowerCAmelCase__ ) @unittest.skip("MGP-STR tokenizer only handles one sequence." ) def _A ( self :Union[str, Any] ) -> Any: '''simple docstring''' pass @unittest.skip("inputs cannot be pretokenized in MgpstrTokenizer" ) def _A ( self :int ) -> Dict: '''simple docstring''' pass
656
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available __lowerCamelCase : List[str] = { '''configuration_ernie''': ['''ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ErnieConfig''', '''ErnieOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Optional[int] = [ '''ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ErnieForCausalLM''', '''ErnieForMaskedLM''', '''ErnieForMultipleChoice''', '''ErnieForNextSentencePrediction''', '''ErnieForPreTraining''', '''ErnieForQuestionAnswering''', '''ErnieForSequenceClassification''', '''ErnieForTokenClassification''', '''ErnieModel''', '''ErniePreTrainedModel''', ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys __lowerCamelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
716
'''simple docstring''' from __future__ import annotations import math import numpy as np from numpy.linalg import norm def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> float: """simple docstring""" return math.sqrt(sum(pow(a - b ,2 ) for a, b in zip(__magic_name__ ,__magic_name__ ) ) ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> list[list[list[float] | float]]: """simple docstring""" if dataset.ndim != value_array.ndim: snake_case_ : int = ( "Wrong input data's dimensions... " F'''dataset : {dataset.ndim}, value_array : {value_array.ndim}''' ) raise ValueError(__magic_name__ ) try: if dataset.shape[1] != value_array.shape[1]: snake_case_ : Dict = ( "Wrong input data's shape... " F'''dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}''' ) raise ValueError(__magic_name__ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError("Wrong shape" ) if dataset.dtype != value_array.dtype: snake_case_ : Dict = ( "Input data have different datatype... " F'''dataset : {dataset.dtype}, value_array : {value_array.dtype}''' ) raise TypeError(__magic_name__ ) snake_case_ : Optional[int] = [] for value in value_array: snake_case_ : List[str] = euclidean(__magic_name__ ,dataset[0] ) snake_case_ : int = dataset[0].tolist() for dataset_value in dataset[1:]: snake_case_ : Optional[Any] = euclidean(__magic_name__ ,__magic_name__ ) if dist > temp_dist: snake_case_ : Tuple = temp_dist snake_case_ : Optional[int] = dataset_value.tolist() answer.append([vector, dist] ) return answer def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> float: """simple docstring""" return np.dot(__magic_name__ ,__magic_name__ ) / (norm(__magic_name__ ) * norm(__magic_name__ )) if __name__ == "__main__": import doctest doctest.testmod()
656
0
'''simple docstring''' from collections import defaultdict from graphs.minimum_spanning_tree_prims import prisms_algorithm as mst def __UpperCAmelCase ( )-> List[str]: """simple docstring""" snake_case_, snake_case_ : str = 9, 14 # noqa: F841 snake_case_ : Tuple = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4], [6, 5, 2], [3, 5, 14], [3, 4, 9], [5, 4, 10], [1, 7, 11], ] snake_case_ : int = defaultdict(snake_case__ ) for nodea, nodea, cost in edges: adjancency[nodea].append([nodea, cost] ) adjancency[nodea].append([nodea, cost] ) snake_case_ : int = mst(snake_case__ ) snake_case_ : Optional[Any] = [ [7, 6, 1], [2, 8, 2], [6, 5, 2], [0, 1, 4], [2, 5, 4], [2, 3, 7], [0, 7, 8], [3, 4, 9], ] for answer in expected: snake_case_ : str = tuple(answer[:2] ) snake_case_ : Union[str, Any] = tuple(edge[::-1] ) assert edge in result or reverse in result
717
'''simple docstring''' import fire from utils import calculate_rouge, save_json def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__=None ,**__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : int = [x.strip() for x in open(__magic_name__ ).readlines()] snake_case_ : Optional[int] = [x.strip() for x in open(__magic_name__ ).readlines()][: len(__magic_name__ )] snake_case_ : List[Any] = calculate_rouge(__magic_name__ ,__magic_name__ ,**__magic_name__ ) if save_path is not None: save_json(__magic_name__ ,__magic_name__ ,indent=__magic_name__ ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
656
0
'''simple docstring''' import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : Any = logging.get_logger(__name__) def __UpperCAmelCase ( __magic_name__ )-> Union[str, Any]: """simple docstring""" snake_case_ : int = torch.load(__magic_name__ ,map_location="cpu" ) if "model" in sd.keys(): snake_case_ : Union[str, Any] = torch.load(__magic_name__ ,map_location="cpu" )['''model'''] # pop unnecessary weights snake_case_ : List[str] = [ '''decoder.version''', '''decoder.output_projection.weight''', ] for key in keys_to_delete: if key in sd: sd.pop(__magic_name__ ) snake_case_ : Any = { '''decoder.project_in_dim.weight''': '''decoder.project_in.weight''', '''decoder.project_out_dim.weight''': '''decoder.project_out.weight''', '''decoder.layer_norm.weight''': '''decoder.final_layer_norm.weight''', '''decoder.layer_norm.bias''': '''decoder.final_layer_norm.bias''', } for old_key, new_key in keys_to_rename.items(): if old_key in sd: snake_case_ : Dict = sd.pop(__magic_name__ ) snake_case_ : Optional[Any] = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: snake_case_ : Tuple = sd[key] # We split QKV in separate Q,K,V snake_case_ : Optional[Any] = key.replace(".qkv_proj." ,".q_proj." ) snake_case_ : Optional[Any] = key.replace(".qkv_proj." ,".k_proj." ) snake_case_ : Any = key.replace(".qkv_proj." ,".v_proj." ) snake_case_ : Dict = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 snake_case_ : Union[str, Any] = torch.split(__magic_name__ ,depth // 3 ,dim=0 ) snake_case_ : Optional[int] = q snake_case_ : str = k snake_case_ : List[str] = v del sd[key] return sd @torch.no_grad() def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__=None )-> Dict: """simple docstring""" snake_case_ : Any = load_checkpoint(__magic_name__ ) if config is not None: snake_case_ : str = OPTConfig.from_pretrained(__magic_name__ ) else: snake_case_ : Optional[int] = OPTConfig() snake_case_ : Dict = OPTModel(__magic_name__ ).half().eval() model.load_state_dict(__magic_name__ ) # Check results Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) if __name__ == "__main__": __lowerCamelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fairseq_path''', type=str, help=( '''path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:''' ''' https://huggingface.co/models?other=opt_metasq''' ), ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--hf_config''', default=None, type=str, help='''Define HF config.''') __lowerCamelCase : List[Any] = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
718
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) __lowerCamelCase : Optional[Any] = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', f'''encoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', f'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.weight''', f'''encoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.bias''', f'''encoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.weight''', f'''encoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.bias''', f'''encoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.encoder.layers.{i}.norm1.weight''', f'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.norm1.bias''', f'''encoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.weight''', f'''encoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.bias''', f'''encoder.layers.{i}.final_layer_norm.bias''')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', f'''decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', f'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.cross_attn.out_proj.weight''', f'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.cross_attn.out_proj.bias''', f'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.weight''', f'''decoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.bias''', f'''decoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.weight''', f'''decoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.bias''', f'''decoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm1.weight''', f'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm1.bias''', f'''decoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.weight''', f'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.bias''', f'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.weight''', f'''decoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.bias''', f'''decoder.layers.{i}.final_layer_norm.bias''')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qcontent_proj.weight''', f'''decoder.layers.{i}.sa_qcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kcontent_proj.weight''', f'''decoder.layers.{i}.sa_kcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qpos_proj.weight''', f'''decoder.layers.{i}.sa_qpos_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kpos_proj.weight''', f'''decoder.layers.{i}.sa_kpos_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_v_proj.weight''', f'''decoder.layers.{i}.sa_v_proj.weight''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qcontent_proj.weight''', f'''decoder.layers.{i}.ca_qcontent_proj.weight''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kcontent_proj.weight''', f'''decoder.layers.{i}.ca_kcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kpos_proj.weight''', f'''decoder.layers.{i}.ca_kpos_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_v_proj.weight''', f'''decoder.layers.{i}.ca_v_proj.weight''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight''', f'''decoder.layers.{i}.ca_qpos_sine_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qcontent_proj.bias''', f'''decoder.layers.{i}.sa_qcontent_proj.bias''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kcontent_proj.bias''', f'''decoder.layers.{i}.sa_kcontent_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_qpos_proj.bias''', f'''decoder.layers.{i}.sa_qpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_kpos_proj.bias''', f'''decoder.layers.{i}.sa_kpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_v_proj.bias''', f'''decoder.layers.{i}.sa_v_proj.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qcontent_proj.bias''', f'''decoder.layers.{i}.ca_qcontent_proj.bias''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kcontent_proj.bias''', f'''decoder.layers.{i}.ca_kcontent_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_kpos_proj.bias''', f'''decoder.layers.{i}.ca_kpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_v_proj.bias''', f'''decoder.layers.{i}.ca_v_proj.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias''', f'''decoder.layers.{i}.ca_qpos_sine_proj.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> int: """simple docstring""" snake_case_ : Optional[Any] = state_dict.pop(__magic_name__ ) snake_case_ : Any = val def __UpperCAmelCase ( __magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : Any = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: snake_case_ : Optional[Any] = key.replace("backbone.0.body" ,"backbone.conv_encoder.model" ) snake_case_ : int = value else: snake_case_ : int = value return new_state_dict def __UpperCAmelCase ( __magic_name__ ,__magic_name__=False )-> Optional[int]: """simple docstring""" snake_case_ : str = "" if is_panoptic: snake_case_ : Dict = "conditional_detr." # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) snake_case_ : Any = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) snake_case_ : Optional[int] = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case_ : Tuple = in_proj_weight[:256, :] snake_case_ : List[Any] = in_proj_bias[:256] snake_case_ : Optional[Any] = in_proj_weight[256:512, :] snake_case_ : Optional[int] = in_proj_bias[256:512] snake_case_ : Optional[int] = in_proj_weight[-256:, :] snake_case_ : str = in_proj_bias[-256:] def __UpperCAmelCase ( )-> Optional[Any]: """simple docstring""" snake_case_ : Optional[int] = "http://images.cocodataset.org/val2017/000000039769.jpg" snake_case_ : Optional[Any] = Image.open(requests.get(__magic_name__ ,stream=__magic_name__ ).raw ) return im @torch.no_grad() def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Optional[Any] = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: snake_case_ : Optional[Any] = "resnet101" if "dc5" in model_name: snake_case_ : List[str] = True snake_case_ : Tuple = "panoptic" in model_name if is_panoptic: snake_case_ : List[Any] = 250 else: snake_case_ : Optional[Any] = 91 snake_case_ : Optional[int] = "huggingface/label-files" snake_case_ : Dict = "coco-detection-id2label.json" snake_case_ : List[Any] = json.load(open(hf_hub_download(__magic_name__ ,__magic_name__ ,repo_type="dataset" ) ,"r" ) ) snake_case_ : Optional[int] = {int(__magic_name__ ): v for k, v in idalabel.items()} snake_case_ : int = idalabel snake_case_ : Dict = {v: k for k, v in idalabel.items()} # load image processor snake_case_ : Optional[int] = "coco_panoptic" if is_panoptic else "coco_detection" snake_case_ : str = ConditionalDetrImageProcessor(format=__magic_name__ ) # prepare image snake_case_ : str = prepare_img() snake_case_ : int = image_processor(images=__magic_name__ ,return_tensors="pt" ) snake_case_ : Union[str, Any] = encoding["pixel_values"] logger.info(F'''Converting model {model_name}...''' ) # load original model from torch hub snake_case_ : Union[str, Any] = torch.hub.load("DeppMeng/ConditionalDETR" ,__magic_name__ ,pretrained=__magic_name__ ).eval() snake_case_ : Any = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: snake_case_ : Any = "conditional_detr." + src rename_key(__magic_name__ ,__magic_name__ ,__magic_name__ ) snake_case_ : Tuple = rename_backbone_keys(__magic_name__ ) # query, key and value matrices need special treatment read_in_q_k_v(__magic_name__ ,is_panoptic=__magic_name__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them snake_case_ : int = "conditional_detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("conditional_detr" ) and not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ) ): snake_case_ : Any = state_dict.pop(__magic_name__ ) snake_case_ : Optional[int] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: snake_case_ : Tuple = state_dict.pop(__magic_name__ ) snake_case_ : Any = val elif key.startswith("bbox_attention" ) or key.startswith("mask_head" ): continue else: snake_case_ : Union[str, Any] = state_dict.pop(__magic_name__ ) snake_case_ : List[Any] = val else: if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): snake_case_ : Any = state_dict.pop(__magic_name__ ) snake_case_ : List[Any] = val # finally, create HuggingFace model and load state dict snake_case_ : Optional[int] = ConditionalDetrForSegmentation(__magic_name__ ) if is_panoptic else ConditionalDetrForObjectDetection(__magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() model.push_to_hub(repo_id=__magic_name__ ,organization="DepuMeng" ,commit_message="Add model" ) # verify our conversion snake_case_ : Dict = conditional_detr(__magic_name__ ) snake_case_ : Union[str, Any] = model(__magic_name__ ) assert torch.allclose(outputs.logits ,original_outputs["pred_logits"] ,atol=1E-4 ) assert torch.allclose(outputs.pred_boxes ,original_outputs["pred_boxes"] ,atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks ,original_outputs["pred_masks"] ,atol=1E-4 ) # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) image_processor.save_pretrained(__magic_name__ ) if __name__ == "__main__": __lowerCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) __lowerCamelCase : int = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
656
0
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCamelCase : List[str] = logging.get_logger(__name__) __lowerCamelCase : Any = { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json""", } class A_ (SCREAMING_SNAKE_CASE__ ): """simple docstring""" a__ = '''mvp''' a__ = ['''past_key_values'''] a__ = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self :Dict , lowerCAmelCase__ :int=50_267 , lowerCAmelCase__ :Dict=1_024 , lowerCAmelCase__ :str=12 , lowerCAmelCase__ :str=4_096 , lowerCAmelCase__ :Union[str, Any]=16 , lowerCAmelCase__ :Union[str, Any]=12 , lowerCAmelCase__ :List[str]=4_096 , lowerCAmelCase__ :int=16 , lowerCAmelCase__ :Union[str, Any]=0.0 , lowerCAmelCase__ :Optional[int]=0.0 , lowerCAmelCase__ :Any="gelu" , lowerCAmelCase__ :Optional[Any]=1_024 , lowerCAmelCase__ :Union[str, Any]=0.1 , lowerCAmelCase__ :str=0.0 , lowerCAmelCase__ :Dict=0.0 , lowerCAmelCase__ :int=0.0_2 , lowerCAmelCase__ :Optional[Any]=0.0 , lowerCAmelCase__ :str=False , lowerCAmelCase__ :Any=True , lowerCAmelCase__ :List[str]=1 , lowerCAmelCase__ :List[str]=0 , lowerCAmelCase__ :Optional[Any]=2 , lowerCAmelCase__ :Optional[Any]=True , lowerCAmelCase__ :Optional[Any]=2 , lowerCAmelCase__ :Union[str, Any]=2 , lowerCAmelCase__ :Dict=False , lowerCAmelCase__ :Optional[Any]=100 , lowerCAmelCase__ :Optional[Any]=800 , **lowerCAmelCase__ :Union[str, Any] , ) -> Optional[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = vocab_size snake_case_ : List[str] = max_position_embeddings snake_case_ : Dict = d_model snake_case_ : List[str] = encoder_ffn_dim snake_case_ : Dict = encoder_layers snake_case_ : Union[str, Any] = encoder_attention_heads snake_case_ : Any = decoder_ffn_dim snake_case_ : Union[str, Any] = decoder_layers snake_case_ : Union[str, Any] = decoder_attention_heads snake_case_ : Dict = dropout snake_case_ : Tuple = attention_dropout snake_case_ : int = activation_dropout snake_case_ : str = activation_function snake_case_ : str = init_std snake_case_ : Any = encoder_layerdrop snake_case_ : int = decoder_layerdrop snake_case_ : Union[str, Any] = classifier_dropout snake_case_ : Optional[int] = use_cache snake_case_ : List[str] = encoder_layers snake_case_ : List[str] = scale_embedding # scale factor will be sqrt(d_model) if True snake_case_ : str = use_prompt snake_case_ : Optional[int] = prompt_length snake_case_ : Optional[int] = prompt_mid_dim super().__init__( pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ , is_encoder_decoder=snake_case__ , decoder_start_token_id=snake_case__ , forced_eos_token_id=snake_case__ , **snake_case__ , ) if self.forced_bos_token_id is None and kwargs.get("force_bos_token_to_be_generated" , snake_case__ ): snake_case_ : Optional[int] = self.bos_token_id warnings.warn( F'''Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ''' "The config can simply be saved and uploaded again to be fixed." )
719
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Any ) -> str: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _A ( self :List[Any] ) -> List[str]: '''simple docstring''' snake_case_ : Any = 1 snake_case_ : Dict = 3 snake_case_ : Union[str, Any] = (32, 32) snake_case_ : Optional[int] = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) return image @property def _A ( self :Optional[int] ) -> Any: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : List[str] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) return model @property def _A ( self :Dict ) -> Any: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) return model @property def _A ( self :Dict ) -> Optional[int]: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : str = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5_006 , ) return RobertaSeriesModelWithTransformation(lowerCAmelCase__ ) @property def _A ( self :Any ) -> str: '''simple docstring''' def extract(*lowerCAmelCase__ :Any , **lowerCAmelCase__ :List[str] ): class A_ : """simple docstring""" def __init__( self :Optional[int] ) -> List[str]: '''simple docstring''' snake_case_ : str = torch.ones([0] ) def _A ( self :int , lowerCAmelCase__ :List[Any] ) -> Tuple: '''simple docstring''' self.pixel_values.to(lowerCAmelCase__ ) return self return Out() return extract def _A ( self :int ) -> Dict: '''simple docstring''' snake_case_ : str = "cpu" # ensure determinism for the device-dependent torch.Generator snake_case_ : str = self.dummy_cond_unet snake_case_ : Optional[int] = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) snake_case_ : Dict = self.dummy_vae snake_case_ : Dict = self.dummy_text_encoder snake_case_ : Optional[int] = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) snake_case_ : str = 77 snake_case_ : Any = self.dummy_image.to(lowerCAmelCase__ ) snake_case_ : Tuple = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk snake_case_ : Optional[Any] = AltDiffusionImgaImgPipeline( unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , vae=lowerCAmelCase__ , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , feature_extractor=self.dummy_extractor , ) snake_case_ : Union[str, Any] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowerCAmelCase__ ) snake_case_ : Optional[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) snake_case_ : Dict = "A painting of a squirrel eating a burger" snake_case_ : List[str] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) snake_case_ : Dict = alt_pipe( [prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , ) snake_case_ : Any = output.images snake_case_ : List[str] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) snake_case_ : Optional[Any] = alt_pipe( [prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , return_dict=lowerCAmelCase__ , )[0] snake_case_ : Tuple = image[0, -3:, -3:, -1] snake_case_ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) snake_case_ : int = np.array([0.4_4_2_7, 0.3_7_3_1, 0.4_2_4_9, 0.4_9_4_1, 0.4_5_4_6, 0.4_1_4_8, 0.4_1_9_3, 0.4_6_6_6, 0.4_4_9_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5E-3 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def _A ( self :int ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = self.dummy_cond_unet snake_case_ : Union[str, Any] = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) snake_case_ : int = self.dummy_vae snake_case_ : List[Any] = self.dummy_text_encoder snake_case_ : int = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) snake_case_ : int = 77 snake_case_ : Dict = self.dummy_image.to(lowerCAmelCase__ ) # put models in fp16 snake_case_ : Optional[Any] = unet.half() snake_case_ : Tuple = vae.half() snake_case_ : List[str] = bert.half() # make sure here that pndm scheduler skips prk snake_case_ : Optional[int] = AltDiffusionImgaImgPipeline( unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , vae=lowerCAmelCase__ , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , feature_extractor=self.dummy_extractor , ) snake_case_ : List[str] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowerCAmelCase__ ) snake_case_ : Optional[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) snake_case_ : List[Any] = "A painting of a squirrel eating a burger" snake_case_ : str = torch.manual_seed(0 ) snake_case_ : Any = alt_pipe( [prompt] , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def _A ( self :Optional[int] ) -> Any: '''simple docstring''' snake_case_ : Union[str, Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) # resize to resolution that is divisible by 8 but not 16 or 32 snake_case_ : str = init_image.resize((760, 504) ) snake_case_ : Optional[Any] = "BAAI/AltDiffusion" snake_case_ : int = AltDiffusionImgaImgPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() snake_case_ : Tuple = "A fantasy landscape, trending on artstation" snake_case_ : int = torch.manual_seed(0 ) snake_case_ : List[str] = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , strength=0.7_5 , guidance_scale=7.5 , generator=lowerCAmelCase__ , output_type="np" , ) snake_case_ : str = output.images[0] snake_case_ : List[Any] = image[255:258, 383:386, -1] assert image.shape == (504, 760, 3) snake_case_ : Tuple = np.array([0.9_3_5_8, 0.9_3_9_7, 0.9_5_9_9, 0.9_9_0_1, 1.0_0_0_0, 1.0_0_0_0, 0.9_8_8_2, 1.0_0_0_0, 1.0_0_0_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _A ( self :str ) -> Any: '''simple docstring''' snake_case_ : Optional[Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) snake_case_ : List[Any] = init_image.resize((768, 512) ) snake_case_ : Tuple = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy" ) snake_case_ : Any = "BAAI/AltDiffusion" snake_case_ : List[str] = AltDiffusionImgaImgPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() snake_case_ : Tuple = "A fantasy landscape, trending on artstation" snake_case_ : Tuple = torch.manual_seed(0 ) snake_case_ : List[Any] = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , strength=0.7_5 , guidance_scale=7.5 , generator=lowerCAmelCase__ , output_type="np" , ) snake_case_ : Optional[int] = output.images[0] assert image.shape == (512, 768, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1E-2
656
0
'''simple docstring''' import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() __lowerCamelCase : Dict = logging.get_logger(__name__) __lowerCamelCase : List[str] = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } __lowerCamelCase : str = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ )-> Dict: """simple docstring""" for attribute in key.split("." ): snake_case_ : Tuple = getattr(__a ,__a ) if weight_type is not None: snake_case_ : List[str] = getattr(__a ,__a ).shape else: snake_case_ : Dict = hf_pointer.shape assert hf_shape == value.shape, ( F'''Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": snake_case_ : Union[str, Any] = value elif weight_type == "weight_g": snake_case_ : Tuple = value elif weight_type == "weight_v": snake_case_ : Dict = value elif weight_type == "bias": snake_case_ : Dict = value else: snake_case_ : Union[str, Any] = value logger.info(F'''{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.''' ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> str: """simple docstring""" snake_case_ : str = [] snake_case_ : str = fairseq_model.state_dict() snake_case_ : Dict = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight snake_case_ : Optional[int] = None for name, value in fairseq_dict.items(): snake_case_ : Optional[int] = False if "conv_layers" in name: load_conv_layer( __a ,__a ,__a ,__a ,hf_model.config.feat_extract_norm == "group" ,) snake_case_ : Optional[int] = True elif name.split("." )[0] == "proj": snake_case_ : Optional[int] = fairseq_model.proj snake_case_ : Optional[Any] = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: snake_case_ : List[str] = True if "*" in mapped_key: snake_case_ : List[Any] = name.split(__a )[0].split("." )[-2] snake_case_ : Any = mapped_key.replace("*" ,__a ) if "weight_g" in name: snake_case_ : Dict = "weight_g" elif "weight_v" in name: snake_case_ : Any = "weight_v" elif "bias" in name: snake_case_ : Optional[int] = "bias" elif "weight" in name: snake_case_ : str = "weight" else: snake_case_ : Union[str, Any] = None set_recursively(__a ,__a ,__a ,__a ,__a ) continue if not is_used: unused_weights.append(__a ) logger.warning(F'''Unused weights: {unused_weights}''' ) return proj_weight def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" snake_case_ : Optional[Any] = full_name.split("conv_layers." )[-1] snake_case_ : Optional[Any] = name.split("." ) snake_case_ : List[Any] = int(items[0] ) snake_case_ : Tuple = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) snake_case_ : str = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) snake_case_ : Dict = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) snake_case_ : int = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) snake_case_ : Optional[int] = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(__a ) def __UpperCAmelCase ( __magic_name__ )-> Tuple: """simple docstring""" snake_case_, snake_case_ : List[Any] = emb.weight.shape snake_case_ : List[Any] = nn.Linear(__a ,__a ,bias=__a ) snake_case_ : Optional[int] = emb.weight.data return lin_layer def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" with open(__a ,"r" ,encoding="utf-8" ) as f: snake_case_ : Tuple = f.readlines() snake_case_ : Union[str, Any] = [line.split(" " )[0] for line in lines] snake_case_ : Optional[Any] = len(__a ) snake_case_ : Optional[int] = { "<s>": 0, "<pad>": 1, "</s>": 2, "<unk>": 3, } vocab_dict.update(dict(zip(__a ,range(4 ,num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,)-> Tuple: """simple docstring""" snake_case_ : Union[str, Any] = WavaVecaConfig.from_pretrained(__a ) snake_case_ : Tuple = SpeechaTextaConfig.from_pretrained( __a ,vocab_size=__a ,decoder_layers=__a ,do_stable_layer_norm=__a ) snake_case_ : Optional[Any] = WavaVecaFeatureExtractor( feature_size=1 ,sampling_rate=1_6000 ,padding_value=0 ,do_normalize=__a ,return_attention_mask=__a ,) snake_case_, snake_case_, snake_case_ : Tuple = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] ,arg_overrides={"data": "/".join(dict_path.split("/" )[:-1] )} ) snake_case_ : List[Any] = model[0].eval() # set weights for wav2vec2 encoder snake_case_ : Any = WavaVecaModel(__a ) snake_case_ : Union[str, Any] = recursively_load_weights_wavaveca(model.encoder ,__a ) snake_case_ : Tuple = SpeechaTextaForCausalLM(__a ) snake_case_, snake_case_ : List[str] = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() ,strict=__a ) # set output linear layer unexpected_keys.remove("embed_out" ) snake_case_ : Optional[Any] = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(F'''The following keys are missing when loading the decoder weights: {missing_keys}''' ) logger.warning(F'''The following keys are unexpected when loading the decoder weights: {unexpected_keys}''' ) snake_case_ : Tuple = SpeechEncoderDecoderModel(encoder=__a ,decoder=__a ) snake_case_ : Any = False # add projection layer snake_case_ : Tuple = nn.Parameter(projection_layer.weight ) snake_case_ : Dict = nn.Parameter(projection_layer.bias ) snake_case_ : Optional[Any] = create_vocab_dict(__a ) with open(os.path.join(__a ,"vocab.json" ) ,"w" ) as fp: json.dump(__a ,__a ) snake_case_ : Any = SpeechaTextaTokenizer(os.path.join(__a ,"vocab.json" ) ) tokenizer.save_pretrained(__a ) snake_case_ : int = hf_wavavec.config.to_dict() snake_case_ : List[str] = tokenizer.pad_token_id snake_case_ : List[str] = tokenizer.bos_token_id snake_case_ : List[Any] = tokenizer.eos_token_id snake_case_ : str = "speech_to_text_2" snake_case_ : Optional[int] = "wav2vec2" snake_case_ : Dict = SpeechEncoderDecoderConfig.from_dict(__a ) hf_wavavec.save_pretrained(__a ) feature_extractor.save_pretrained(__a ) if __name__ == "__main__": __lowerCamelCase : Optional[Any] = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument( '''--encoder_config_path''', default='''facebook/wav2vec2-large-lv60''', type=str, help='''Path to hf encoder wav2vec2 checkpoint config''', ) parser.add_argument( '''--decoder_config_path''', default='''facebook/s2t-small-mustc-en-fr-st''', type=str, help='''Path to hf decoder s2t checkpoint config''', ) parser.add_argument('''--vocab_size''', default=10224, type=int, help='''Vocab size of decoder''') parser.add_argument('''--num_decoder_layers''', default=7, type=int, help='''Number of decoder layers''') __lowerCamelCase : List[str] = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
720
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. __lowerCamelCase : List[str] = {'''LayoutLMv2Config''', '''LayoutLMv3Config'''} @is_pipeline_test class A_ (unittest.TestCase ): """simple docstring""" a__ = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING a__ = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: a__ = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: a__ = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def _A ( self :Tuple , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :Dict ) -> Any: '''simple docstring''' snake_case_ : Optional[Any] = ZeroShotClassificationPipeline( model=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , candidate_labels=["polics", "health"] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def _A ( self :List[str] , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Any ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = classifier("Who are you voting for in 2020?" , candidate_labels="politics" ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) # No kwarg snake_case_ : List[Any] = classifier("Who are you voting for in 2020?" , ["politics"] ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) snake_case_ : Dict = classifier("Who are you voting for in 2020?" , candidate_labels=["politics"] ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) snake_case_ : int = classifier("Who are you voting for in 2020?" , candidate_labels="politics, public health" ) self.assertEqual( lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ) , 1.0 ) snake_case_ : Optional[int] = classifier("Who are you voting for in 2020?" , candidate_labels=["politics", "public health"] ) self.assertEqual( lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ) , 1.0 ) snake_case_ : str = classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template="This text is about {}" ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) # https://github.com/huggingface/transformers/issues/13846 snake_case_ : Dict = classifier(["I am happy"] , ["positive", "negative"] ) self.assertEqual( lowerCAmelCase__ , [ {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} for i in range(1 ) ] , ) snake_case_ : Tuple = classifier(["I am happy", "I am sad"] , ["positive", "negative"] ) self.assertEqual( lowerCAmelCase__ , [ {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} for i in range(2 ) ] , ) with self.assertRaises(lowerCAmelCase__ ): classifier("" , candidate_labels="politics" ) with self.assertRaises(lowerCAmelCase__ ): classifier(lowerCAmelCase__ , candidate_labels="politics" ) with self.assertRaises(lowerCAmelCase__ ): classifier("Who are you voting for in 2020?" , candidate_labels="" ) with self.assertRaises(lowerCAmelCase__ ): classifier("Who are you voting for in 2020?" , candidate_labels=lowerCAmelCase__ ) with self.assertRaises(lowerCAmelCase__ ): classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template="Not formatting template" , ) with self.assertRaises(lowerCAmelCase__ ): classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template=lowerCAmelCase__ , ) self.run_entailment_id(lowerCAmelCase__ ) def _A ( self :List[Any] , lowerCAmelCase__ :Pipeline ) -> Union[str, Any]: '''simple docstring''' snake_case_ : int = zero_shot_classifier.model.config snake_case_ : Optional[int] = config.labelaid snake_case_ : Tuple = zero_shot_classifier.entailment_id snake_case_ : Optional[Any] = {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2} self.assertEqual(zero_shot_classifier.entailment_id , -1 ) snake_case_ : Tuple = {"entailment": 0, "neutral": 1, "contradiction": 2} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) snake_case_ : str = {"ENTAIL": 0, "NON-ENTAIL": 1} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) snake_case_ : str = {"ENTAIL": 2, "NEUTRAL": 1, "CONTR": 0} self.assertEqual(zero_shot_classifier.entailment_id , 2 ) snake_case_ : List[str] = original_labelaid self.assertEqual(lowerCAmelCase__ , zero_shot_classifier.entailment_id ) @require_torch def _A ( self :Tuple ) -> Any: '''simple docstring''' snake_case_ : List[Any] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="pt" , ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( "Who are you voting for in 2020?" * 100 , candidate_labels=["politics", "public health", "science"] ) @require_torch def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="pt" , ) snake_case_ : int = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.3_3_3, 0.3_3_3, 0.3_3_3], } , ) @require_tf def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : List[str] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="tf" , ) snake_case_ : Optional[int] = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.3_3_3, 0.3_3_3, 0.3_3_3], } , ) @slow @require_torch def _A ( self :Union[str, Any] ) -> int: '''simple docstring''' snake_case_ : int = pipeline("zero-shot-classification" , model="roberta-large-mnli" , framework="pt" ) snake_case_ : str = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.9_7_6, 0.0_1_5, 0.0_0_9], } , ) snake_case_ : Optional[int] = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data." , candidate_labels=["machine learning", "statistics", "translation", "vision"] , multi_label=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], } , ) @slow @require_tf def _A ( self :List[str] ) -> str: '''simple docstring''' snake_case_ : int = pipeline("zero-shot-classification" , model="roberta-large-mnli" , framework="tf" ) snake_case_ : Optional[Any] = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.9_7_6, 0.0_1_5, 0.0_0_9], } , ) snake_case_ : Tuple = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data." , candidate_labels=["machine learning", "statistics", "translation", "vision"] , multi_label=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], } , )
656
0
'''simple docstring''' import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class A_ (pl.LightningModule ): """simple docstring""" def __init__( self :Tuple , lowerCAmelCase__ :Any ) -> Union[str, Any]: '''simple docstring''' super().__init__() snake_case_ : Optional[int] = model snake_case_ : Optional[int] = 2 snake_case_ : List[Any] = nn.Linear(self.model.config.hidden_size , self.num_labels ) def _A ( self :Optional[int] ) -> Any: '''simple docstring''' pass def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : List[str] = LongformerModel.from_pretrained(__magic_name__ ) snake_case_ : Union[str, Any] = LightningModel(__magic_name__ ) snake_case_ : Optional[Any] = torch.load(__magic_name__ ,map_location=torch.device("cpu" ) ) lightning_model.load_state_dict(ckpt["state_dict"] ) # init longformer question answering model snake_case_ : Union[str, Any] = LongformerForQuestionAnswering.from_pretrained(__magic_name__ ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(__magic_name__ ) print(F'''Conversion successful. Model saved under {pytorch_dump_folder_path}''' ) if __name__ == "__main__": __lowerCamelCase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--longformer_model''', default=None, type=str, required=True, help='''model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.''', ) parser.add_argument( '''--longformer_question_answering_ckpt_path''', default=None, type=str, required=True, help='''Path the official PyTorch Lightning Checkpoint.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __lowerCamelCase : Tuple = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
721
'''simple docstring''' import argparse import pathlib import fairseq import torch from fairseq.models.roberta import RobertaModel as FairseqRobertaModel from fairseq.modules import TransformerSentenceEncoderLayer from packaging import version from transformers import XLMRobertaConfig, XLMRobertaXLForMaskedLM, XLMRobertaXLForSequenceClassification from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.models.roberta.modeling_roberta import RobertaAttention from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse('''1.0.0a'''): raise Exception('''requires fairseq >= 1.0.0a''') logging.set_verbosity_info() __lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) __lowerCamelCase : Union[str, Any] = '''Hello world! cécé herlolip''' def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : str = FairseqRobertaModel.from_pretrained(__magic_name__ ) roberta.eval() # disable dropout snake_case_ : Dict = roberta.model.encoder.sentence_encoder snake_case_ : List[str] = XLMRobertaConfig( vocab_size=roberta_sent_encoder.embed_tokens.num_embeddings ,hidden_size=roberta.cfg.model.encoder_embed_dim ,num_hidden_layers=roberta.cfg.model.encoder_layers ,num_attention_heads=roberta.cfg.model.encoder_attention_heads ,intermediate_size=roberta.cfg.model.encoder_ffn_embed_dim ,max_position_embeddings=514 ,type_vocab_size=1 ,layer_norm_eps=1E-5 ,) if classification_head: snake_case_ : List[str] = roberta.model.classification_heads["mnli"].out_proj.weight.shape[0] print("Our RoBERTa config:" ,__magic_name__ ) snake_case_ : List[str] = XLMRobertaXLForSequenceClassification(__magic_name__ ) if classification_head else XLMRobertaXLForMaskedLM(__magic_name__ ) model.eval() # Now let's copy all the weights. # Embeddings snake_case_ : List[Any] = roberta_sent_encoder.embed_tokens.weight snake_case_ : int = roberta_sent_encoder.embed_positions.weight snake_case_ : Union[str, Any] = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c RoBERTa doesn't use them. snake_case_ : Union[str, Any] = roberta_sent_encoder.layer_norm.weight snake_case_ : str = roberta_sent_encoder.layer_norm.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer snake_case_ : BertLayer = model.roberta.encoder.layer[i] snake_case_ : TransformerSentenceEncoderLayer = roberta_sent_encoder.layers[i] snake_case_ : RobertaAttention = layer.attention snake_case_ : Dict = roberta_layer.self_attn_layer_norm.weight snake_case_ : Dict = roberta_layer.self_attn_layer_norm.bias # self attention snake_case_ : BertSelfAttention = layer.attention.self assert ( roberta_layer.self_attn.k_proj.weight.data.shape == roberta_layer.self_attn.q_proj.weight.data.shape == roberta_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ) snake_case_ : Dict = roberta_layer.self_attn.q_proj.weight snake_case_ : Any = roberta_layer.self_attn.q_proj.bias snake_case_ : Optional[Any] = roberta_layer.self_attn.k_proj.weight snake_case_ : Optional[Any] = roberta_layer.self_attn.k_proj.bias snake_case_ : Optional[int] = roberta_layer.self_attn.v_proj.weight snake_case_ : Any = roberta_layer.self_attn.v_proj.bias # self-attention output snake_case_ : BertSelfOutput = layer.attention.output assert self_output.dense.weight.shape == roberta_layer.self_attn.out_proj.weight.shape snake_case_ : List[str] = roberta_layer.self_attn.out_proj.weight snake_case_ : Optional[int] = roberta_layer.self_attn.out_proj.bias # this one is final layer norm snake_case_ : int = roberta_layer.final_layer_norm.weight snake_case_ : Union[str, Any] = roberta_layer.final_layer_norm.bias # intermediate snake_case_ : BertIntermediate = layer.intermediate assert intermediate.dense.weight.shape == roberta_layer.fca.weight.shape snake_case_ : List[str] = roberta_layer.fca.weight snake_case_ : List[Any] = roberta_layer.fca.bias # output snake_case_ : BertOutput = layer.output assert bert_output.dense.weight.shape == roberta_layer.fca.weight.shape snake_case_ : Any = roberta_layer.fca.weight snake_case_ : Any = roberta_layer.fca.bias # end of layer if classification_head: snake_case_ : int = roberta.model.classification_heads["mnli"].dense.weight snake_case_ : Union[str, Any] = roberta.model.classification_heads["mnli"].dense.bias snake_case_ : Tuple = roberta.model.classification_heads["mnli"].out_proj.weight snake_case_ : str = roberta.model.classification_heads["mnli"].out_proj.bias else: # LM Head snake_case_ : Optional[Any] = roberta.model.encoder.lm_head.dense.weight snake_case_ : int = roberta.model.encoder.lm_head.dense.bias snake_case_ : Optional[Any] = roberta.model.encoder.lm_head.layer_norm.weight snake_case_ : Optional[int] = roberta.model.encoder.lm_head.layer_norm.bias snake_case_ : int = roberta.model.encoder.lm_head.weight snake_case_ : List[str] = roberta.model.encoder.lm_head.bias # Let's check that we get the same results. snake_case_ : torch.Tensor = roberta.encode(__magic_name__ ).unsqueeze(0 ) # batch of size 1 snake_case_ : Union[str, Any] = model(__magic_name__ )[0] if classification_head: snake_case_ : Optional[Any] = roberta.model.classification_heads["mnli"](roberta.extract_features(__magic_name__ ) ) else: snake_case_ : List[str] = roberta.model(__magic_name__ )[0] print(our_output.shape ,their_output.shape ) snake_case_ : str = torch.max(torch.abs(our_output - their_output ) ).item() print(F'''max_absolute_diff = {max_absolute_diff}''' ) # ~ 1e-7 snake_case_ : Any = torch.allclose(__magic_name__ ,__magic_name__ ,atol=1E-3 ) print("Do both models output the same tensors?" ,"🔥" if success else "💩" ) if not success: raise Exception("Something went wRoNg" ) pathlib.Path(__magic_name__ ).mkdir(parents=__magic_name__ ,exist_ok=__magic_name__ ) print(F'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(__magic_name__ ) if __name__ == "__main__": __lowerCamelCase : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--roberta_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--classification_head''', action='''store_true''', help='''Whether to convert a final classification head.''' ) __lowerCamelCase : Tuple = parser.parse_args() convert_xlm_roberta_xl_checkpoint_to_pytorch( args.roberta_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
656
0
'''simple docstring''' from collections import Counter from timeit import timeit def __UpperCAmelCase ( __magic_name__ = "" ,)-> bool: """simple docstring""" return sum(c % 2 for c in Counter(input_str.replace(" " ,"" ).lower() ).values() ) < 2 def __UpperCAmelCase ( __magic_name__ = "" )-> bool: """simple docstring""" if len(__magic_name__ ) == 0: return True snake_case_ : Union[str, Any] = input_str.replace(" " ,"" ).lower() # character_freq_dict: Stores the frequency of every character in the input string snake_case_ : dict[str, int] = {} for character in lower_case_input_str: snake_case_ : Dict = character_freq_dict.get(__magic_name__ ,0 ) + 1 snake_case_ : str = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def __UpperCAmelCase ( __magic_name__ = "" )-> None: """simple docstring""" print("\nFor string = " ,__magic_name__ ,":" ) print( "> can_string_be_rearranged_as_palindrome_counter()" ,"\tans =" ,can_string_be_rearranged_as_palindrome_counter(__magic_name__ ) ,"\ttime =" ,timeit( "z.can_string_be_rearranged_as_palindrome_counter(z.check_str)" ,setup="import __main__ as z" ,) ,"seconds" ,) print( "> can_string_be_rearranged_as_palindrome()" ,"\tans =" ,can_string_be_rearranged_as_palindrome(__magic_name__ ) ,"\ttime =" ,timeit( "z.can_string_be_rearranged_as_palindrome(z.check_str)" ,setup="import __main__ as z" ,) ,"seconds" ,) if __name__ == "__main__": __lowerCamelCase : str = input( '''Enter string to determine if it can be rearranged as a palindrome or not: ''' ).strip() benchmark(check_str) __lowerCamelCase : List[str] = can_string_be_rearranged_as_palindrome_counter(check_str) print(f'''{check_str} can {"" if status else "not "}be rearranged as a palindrome''')
700
'''simple docstring''' import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def __UpperCAmelCase ( __magic_name__ ,__magic_name__=() ,__magic_name__=None ,__magic_name__="no" ,__magic_name__="29500" )-> Optional[int]: """simple docstring""" snake_case_ : str = False snake_case_ : int = False if any(key.startswith("KAGGLE" ) for key in os.environ.keys() ): snake_case_ : Any = True elif "IPython" in sys.modules: snake_case_ : Union[str, Any] = "google.colab" in str(sys.modules["IPython"].get_ipython() ) try: snake_case_ : Any = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F'''Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.''' ) if (in_colab or in_kaggle) and (os.environ.get("TPU_NAME" ,__magic_name__ ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( "To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside " "your training function. Restart your notebook and make sure no cells initializes an " "`Accelerator`." ) if num_processes is None: snake_case_ : Tuple = 8 snake_case_ : Optional[int] = PrepareForLaunch(__magic_name__ ,distributed_type="TPU" ) print(F'''Launching a training on {num_processes} TPU cores.''' ) xmp.spawn(__magic_name__ ,args=__magic_name__ ,nprocs=__magic_name__ ,start_method="fork" ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print("Launching training on one GPU." ) else: print("Launching training on one CPU." ) function(*__magic_name__ ) else: if num_processes is None: raise ValueError( "You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call." ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( "To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized " "inside your training function. Restart your notebook and make sure no cells initializes an " "`Accelerator`." ) if torch.cuda.is_initialized(): raise ValueError( "To launch a multi-GPU training from your notebook, you need to avoid running any instruction " "using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA " "function." ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=__magic_name__ ,master_addr="127.0.01" ,master_port=__magic_name__ ,mixed_precision=__magic_name__ ): snake_case_ : Optional[int] = PrepareForLaunch(__magic_name__ ,distributed_type="MULTI_GPU" ) print(F'''Launching training on {num_processes} GPUs.''' ) try: start_processes(__magic_name__ ,args=__magic_name__ ,nprocs=__magic_name__ ,start_method="fork" ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( "CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. " "This likely stems from an outside import causing issues once the `notebook_launcher()` is called. " "Please review your imports and test them when running the `notebook_launcher()` to identify " "which one is problematic." ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): snake_case_ : Any = "1" print("Launching training on MPS." ) elif torch.cuda.is_available(): print("Launching training on one GPU." ) else: print("Launching training on CPU." ) function(*__magic_name__ ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__=() ,__magic_name__=2 )-> Dict: """simple docstring""" from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=__magic_name__ ,master_addr="127.0.01" ,master_port="29500" ,accelerate_mixed_precision="no" ,accelerate_debug_rdv_file=tmp_file.name ,accelerate_use_cpu="yes" ,): snake_case_ : Any = PrepareForLaunch(__magic_name__ ,debug=__magic_name__ ) start_processes(__magic_name__ ,args=__magic_name__ ,nprocs=__magic_name__ ,start_method="fork" )
656
0
'''simple docstring''' from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def __UpperCAmelCase ( __magic_name__ )-> str: """simple docstring""" if not is_accelerate_available(): return method snake_case_ : Tuple = version.parse(accelerate.__version__ ).base_version if version.parse(_lowerCAmelCase ) < version.parse("0.17.0" ): return method def wrapper(self ,*__magic_name__ ,**__magic_name__ ): if hasattr(self ,"_hf_hook" ) and hasattr(self._hf_hook ,"pre_forward" ): self._hf_hook.pre_forward(self ) return method(self ,*_lowerCAmelCase ,**_lowerCAmelCase ) return wrapper
701
'''simple docstring''' from collections import deque from math import floor from random import random from time import time class A_ : """simple docstring""" def __init__( self :Dict ) -> List[str]: '''simple docstring''' snake_case_ : int = {} def _A ( self :Any , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Optional[Any]=1 ) -> Any: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: snake_case_ : Optional[int] = [[w, v]] if not self.graph.get(lowerCAmelCase__ ): snake_case_ : Dict = [] def _A ( self :List[Any] ) -> Optional[int]: '''simple docstring''' return list(self.graph ) def _A ( self :str , lowerCAmelCase__ :Any , lowerCAmelCase__ :int ) -> List[Any]: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCAmelCase__ ) def _A ( self :List[str] , lowerCAmelCase__ :Optional[Any]=-2 , lowerCAmelCase__ :str=-1 ) -> str: '''simple docstring''' if s == d: return [] snake_case_ : str = [] snake_case_ : Optional[int] = [] if s == -2: snake_case_ : List[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Dict = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : str = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCAmelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCAmelCase__ ) != 0: snake_case_ : Union[str, Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return visited def _A ( self :Tuple , lowerCAmelCase__ :int=-1 ) -> int: '''simple docstring''' if c == -1: snake_case_ : Any = floor(random() * 10_000 ) + 10 for i in range(lowerCAmelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): snake_case_ : Optional[Any] = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCAmelCase__ , lowerCAmelCase__ , 1 ) def _A ( self :Tuple , lowerCAmelCase__ :Dict=-2 ) -> Dict: '''simple docstring''' snake_case_ : Union[str, Any] = deque() snake_case_ : Optional[Any] = [] if s == -2: snake_case_ : Tuple = list(self.graph )[0] d.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) while d: snake_case_ : Optional[int] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def _A ( self :List[str] , lowerCAmelCase__ :str ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = 0 for x in self.graph: for y in self.graph[x]: if y[1] == u: count += 1 return count def _A ( self :Any , lowerCAmelCase__ :int ) -> Optional[Any]: '''simple docstring''' return len(self.graph[u] ) def _A ( self :Tuple , lowerCAmelCase__ :List[str]=-2 ) -> Optional[Any]: '''simple docstring''' snake_case_ : str = [] snake_case_ : str = [] if s == -2: snake_case_ : Optional[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : int = s snake_case_ : Optional[int] = [] while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : List[str] = node[1] break # check if all the children are visited if s == ss: sorted_nodes.append(stack.pop() ) if len(lowerCAmelCase__ ) != 0: snake_case_ : int = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Union[str, Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return sorted_nodes def _A ( self :Dict ) -> Any: '''simple docstring''' snake_case_ : Dict = [] snake_case_ : Any = [] snake_case_ : str = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Optional[int] = -2 snake_case_ : Any = [] snake_case_ : List[Any] = s snake_case_ : int = False snake_case_ : Optional[int] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Any = len(lowerCAmelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Optional[int] = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[Any] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : str = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : List[str] = s snake_case_ : Optional[int] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return list(lowerCAmelCase__ ) def _A ( self :Tuple ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = [] snake_case_ : Tuple = [] snake_case_ : List[str] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : str = -2 snake_case_ : List[str] = [] snake_case_ : List[Any] = s snake_case_ : List[str] = False snake_case_ : Dict = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Any = len(lowerCAmelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Tuple = True if len(lowerCAmelCase__ ) != 0: snake_case_ : List[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : int = s snake_case_ : Union[str, Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return False def _A ( self :Optional[int] , lowerCAmelCase__ :Optional[int]=-2 , lowerCAmelCase__ :Tuple=-1 ) -> str: '''simple docstring''' snake_case_ : Optional[int] = time() self.dfs(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : Optional[Any] = time() return end - begin def _A ( self :Any , lowerCAmelCase__ :Tuple=-2 ) -> Optional[Any]: '''simple docstring''' snake_case_ : Any = time() self.bfs(lowerCAmelCase__ ) snake_case_ : Any = time() return end - begin class A_ : """simple docstring""" def __init__( self :Tuple ) -> List[str]: '''simple docstring''' snake_case_ : Optional[Any] = {} def _A ( self :str , lowerCAmelCase__ :Dict , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Union[str, Any]=1 ) -> str: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): # if there already is a edge if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: # if u does not exist snake_case_ : str = [[w, v]] # add the other way if self.graph.get(lowerCAmelCase__ ): # if there already is a edge if self.graph[v].count([w, u] ) == 0: self.graph[v].append([w, u] ) else: # if u does not exist snake_case_ : List[str] = [[w, u]] def _A ( self :Dict , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Optional[Any] ) -> Optional[Any]: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCAmelCase__ ) # the other way round if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[v]: if _[1] == u: self.graph[v].remove(lowerCAmelCase__ ) def _A ( self :Optional[Any] , lowerCAmelCase__ :Optional[Any]=-2 , lowerCAmelCase__ :Optional[int]=-1 ) -> int: '''simple docstring''' if s == d: return [] snake_case_ : Any = [] snake_case_ : Dict = [] if s == -2: snake_case_ : Optional[int] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[str] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCAmelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : str = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return visited def _A ( self :Optional[int] , lowerCAmelCase__ :str=-1 ) -> List[Any]: '''simple docstring''' if c == -1: snake_case_ : Optional[int] = floor(random() * 10_000 ) + 10 for i in range(lowerCAmelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): snake_case_ : str = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCAmelCase__ , lowerCAmelCase__ , 1 ) def _A ( self :Any , lowerCAmelCase__ :Optional[Any]=-2 ) -> List[Any]: '''simple docstring''' snake_case_ : List[str] = deque() snake_case_ : Optional[Any] = [] if s == -2: snake_case_ : List[Any] = list(self.graph )[0] d.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) while d: snake_case_ : Optional[int] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def _A ( self :str , lowerCAmelCase__ :Union[str, Any] ) -> Optional[Any]: '''simple docstring''' return len(self.graph[u] ) def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : Any = [] snake_case_ : Optional[Any] = [] snake_case_ : Optional[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = -2 snake_case_ : Optional[int] = [] snake_case_ : Tuple = s snake_case_ : Optional[Any] = False snake_case_ : Optional[int] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : Optional[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Optional[int] = len(lowerCAmelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Tuple = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[int] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : List[Any] = s snake_case_ : Dict = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return list(lowerCAmelCase__ ) def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Optional[Any] = [] snake_case_ : int = [] snake_case_ : List[str] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = -2 snake_case_ : int = [] snake_case_ : int = s snake_case_ : Optional[Any] = False snake_case_ : List[Any] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : Union[str, Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Tuple = len(lowerCAmelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Optional[Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[Any] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Tuple = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : Union[str, Any] = s snake_case_ : Tuple = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return False def _A ( self :Any ) -> Tuple: '''simple docstring''' return list(self.graph ) def _A ( self :Optional[Any] , lowerCAmelCase__ :Tuple=-2 , lowerCAmelCase__ :Optional[int]=-1 ) -> str: '''simple docstring''' snake_case_ : List[str] = time() self.dfs(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : List[Any] = time() return end - begin def _A ( self :Union[str, Any] , lowerCAmelCase__ :List[Any]=-2 ) -> int: '''simple docstring''' snake_case_ : List[str] = time() self.bfs(lowerCAmelCase__ ) snake_case_ : Tuple = time() return end - begin
656
0
'''simple docstring''' import flax.linen as nn import jax import jax.numpy as jnp class A_ (nn.Module ): """simple docstring""" a__ = 42 a__ = jnp.floataa def _A ( self :Optional[int] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self :Any , lowerCAmelCase__ :int ) -> Optional[Any]: '''simple docstring''' snake_case_ : Optional[Any] = hidden_states.shape snake_case_ : str = jax.image.resize( __lowerCamelCase , shape=(batch, height * 2, width * 2, channels) , method="nearest" , ) snake_case_ : Tuple = self.conv(__lowerCamelCase ) return hidden_states class A_ (nn.Module ): """simple docstring""" a__ = 42 a__ = jnp.floataa def _A ( self :Any ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self :Optional[Any] , lowerCAmelCase__ :Union[str, Any] ) -> int: '''simple docstring''' snake_case_ : int = self.conv(__lowerCamelCase ) return hidden_states class A_ (nn.Module ): """simple docstring""" a__ = 42 a__ = None a__ = 0.0 a__ = None a__ = jnp.floataa def _A ( self :List[Any] ) -> Any: '''simple docstring''' snake_case_ : Optional[Any] = self.in_channels if self.out_channels is None else self.out_channels snake_case_ : List[Any] = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) snake_case_ : int = nn.Conv( __lowerCamelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) snake_case_ : List[str] = nn.Dense(__lowerCamelCase , dtype=self.dtype ) snake_case_ : Dict = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) snake_case_ : Dict = nn.Dropout(self.dropout_prob ) snake_case_ : Any = nn.Conv( __lowerCamelCase , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) snake_case_ : Dict = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut snake_case_ : Union[str, Any] = None if use_nin_shortcut: snake_case_ : Dict = nn.Conv( __lowerCamelCase , kernel_size=(1, 1) , strides=(1, 1) , padding="VALID" , dtype=self.dtype , ) def __call__( self :Tuple , lowerCAmelCase__ :int , lowerCAmelCase__ :Any , lowerCAmelCase__ :Tuple=True ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[Any] = hidden_states snake_case_ : Any = self.norma(__lowerCamelCase ) snake_case_ : Dict = nn.swish(__lowerCamelCase ) snake_case_ : Tuple = self.conva(__lowerCamelCase ) snake_case_ : List[str] = self.time_emb_proj(nn.swish(__lowerCamelCase ) ) snake_case_ : Optional[int] = jnp.expand_dims(jnp.expand_dims(__lowerCamelCase , 1 ) , 1 ) snake_case_ : Dict = hidden_states + temb snake_case_ : Optional[Any] = self.norma(__lowerCamelCase ) snake_case_ : Optional[int] = nn.swish(__lowerCamelCase ) snake_case_ : Dict = self.dropout(__lowerCamelCase , __lowerCamelCase ) snake_case_ : Dict = self.conva(__lowerCamelCase ) if self.conv_shortcut is not None: snake_case_ : List[str] = self.conv_shortcut(__lowerCamelCase ) return hidden_states + residual
702
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __lowerCamelCase : List[str] = re.compile(R'''\s+''') def __UpperCAmelCase ( __magic_name__ )-> Union[str, Any]: """simple docstring""" return {"hash": hashlib.mda(re.sub(__magic_name__ ,"" ,example["content"] ).encode("utf-8" ) ).hexdigest()} def __UpperCAmelCase ( __magic_name__ )-> str: """simple docstring""" snake_case_ : Optional[Any] = [len(__magic_name__ ) for line in example["content"].splitlines()] return {"line_mean": np.mean(__magic_name__ ), "line_max": max(__magic_name__ )} def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" snake_case_ : Optional[int] = np.mean([c.isalnum() for c in example["content"]] ) return {"alpha_frac": alpha_frac} def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" if example["hash"] in uniques: uniques.remove(example["hash"] ) return True else: return False def __UpperCAmelCase ( __magic_name__ ,__magic_name__=5 )-> Tuple: """simple docstring""" snake_case_ : List[str] = ["auto-generated", "autogenerated", "automatically generated"] snake_case_ : Optional[Any] = example["content"].splitlines() for _, line in zip(range(__magic_name__ ) ,__magic_name__ ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def __UpperCAmelCase ( __magic_name__ ,__magic_name__=5 ,__magic_name__=0.05 )-> Optional[Any]: """simple docstring""" snake_case_ : str = ["unit tests", "test file", "configuration file"] snake_case_ : int = example["content"].splitlines() snake_case_ : Optional[Any] = 0 snake_case_ : Any = 0 # first test for _, line in zip(range(__magic_name__ ) ,__magic_name__ ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test snake_case_ : Tuple = example["content"].count("\n" ) snake_case_ : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count("config" ) count_test += line.lower().count("test" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def __UpperCAmelCase ( __magic_name__ )-> str: """simple docstring""" snake_case_ : List[Any] = ["def ", "class ", "for ", "while "] snake_case_ : Optional[Any] = example["content"].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def __UpperCAmelCase ( __magic_name__ ,__magic_name__=4 )-> Optional[int]: """simple docstring""" snake_case_ : Tuple = example["content"].splitlines() snake_case_ : Tuple = 0 for line in lines: counter += line.lower().count("=" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : Tuple = tokenizer(example["content"] ,truncation=__magic_name__ )["input_ids"] snake_case_ : int = len(example["content"] ) / len(__magic_name__ ) return {"ratio": ratio} def __UpperCAmelCase ( __magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : Union[str, Any] = {} results.update(get_hash(__magic_name__ ) ) results.update(line_stats(__magic_name__ ) ) results.update(alpha_stats(__magic_name__ ) ) results.update(char_token_ratio(__magic_name__ ) ) results.update(is_autogenerated(__magic_name__ ) ) results.update(is_config_or_test(__magic_name__ ) ) results.update(has_no_keywords(__magic_name__ ) ) results.update(has_few_assignments(__magic_name__ ) ) return results def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" if not check_uniques(__magic_name__ ,__magic_name__ ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def __UpperCAmelCase ( __magic_name__ )-> Dict: """simple docstring""" with open(__magic_name__ ,"rb" ) as f_in: with gzip.open(str(__magic_name__ ) + ".gz" ,"wb" ,compresslevel=6 ) as f_out: shutil.copyfileobj(__magic_name__ ,__magic_name__ ) os.unlink(__magic_name__ ) # Settings __lowerCamelCase : List[Any] = HfArgumentParser(PreprocessingArguments) __lowerCamelCase : str = parser.parse_args() if args.num_workers is None: __lowerCamelCase : List[Any] = multiprocessing.cpu_count() __lowerCamelCase : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __lowerCamelCase : Any = time.time() __lowerCamelCase : str = load_dataset(args.dataset_name, split='''train''') print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __lowerCamelCase : List[str] = time.time() __lowerCamelCase : Any = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __lowerCamelCase : Any = set(ds.unique('''hash''')) __lowerCamelCase : Optional[int] = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __lowerCamelCase : List[str] = time.time() __lowerCamelCase : Tuple = ds.filter(filter, fn_kwargs={'''uniques''': uniques, '''args''': args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __lowerCamelCase : List[str] = time.time() __lowerCamelCase , __lowerCamelCase : Tuple = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __lowerCamelCase : List[Any] = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / '''duplicate_clusters.json''', '''w''') as f: json.dump(duplicate_clusters, f) __lowerCamelCase : List[str] = output_dir / '''data''' data_dir.mkdir(exist_ok=True) __lowerCamelCase : int = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __lowerCamelCase : Union[str, Any] = str(data_dir / f'''file-{file_number+1:012}.json''') __lowerCamelCase : List[Any] = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
656
0
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class A_ (unittest.TestCase ): """simple docstring""" @slow def _A ( self :List[str] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[str] = XLMRobertaModel.from_pretrained("xlm-roberta-base" ) snake_case_ : Union[str, Any] = torch.tensor([[0, 581, 10_269, 83, 99_942, 136, 60_742, 23, 70, 80_583, 18_276, 2]] ) # The dog is cute and lives in the garden house snake_case_ : List[Any] = torch.Size((1, 12, 768) ) # batch_size, sequence_length, embedding_vector_dim snake_case_ : Any = torch.tensor( [[-0.0_1_0_1, 0.1_2_1_8, -0.0_8_0_3, 0.0_8_0_1, 0.1_3_2_7, 0.0_7_7_6, -0.1_2_1_5, 0.2_3_8_3, 0.3_3_3_8, 0.3_1_0_6, 0.0_3_0_0, 0.0_2_5_2]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): snake_case_ : str = model(_lowercase )["last_hidden_state"].detach() self.assertEqual(output.shape , _lowercase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , _lowercase , atol=1E-3 ) ) @slow def _A ( self :Any ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = XLMRobertaModel.from_pretrained("xlm-roberta-large" ) snake_case_ : int = torch.tensor([[0, 581, 10_269, 83, 99_942, 136, 60_742, 23, 70, 80_583, 18_276, 2]] ) # The dog is cute and lives in the garden house snake_case_ : Tuple = torch.Size((1, 12, 1_024) ) # batch_size, sequence_length, embedding_vector_dim snake_case_ : Union[str, Any] = torch.tensor( [[-0.0_6_9_9, -0.0_3_1_8, 0.0_7_0_5, -0.1_2_4_1, 0.0_9_9_9, -0.0_5_2_0, 0.1_0_0_4, -0.1_8_3_8, -0.4_7_0_4, 0.1_4_3_7, 0.0_8_2_1, 0.0_1_2_6]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): snake_case_ : Any = model(_lowercase )["last_hidden_state"].detach() self.assertEqual(output.shape , _lowercase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , _lowercase , atol=1E-3 ) )
703
'''simple docstring''' import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :str ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Union[str, Any] = torch.nn.Linear(10 , 10 ) snake_case_ : Dict = torch.optim.SGD(model.parameters() , 0.1 ) snake_case_ : Tuple = Accelerator() snake_case_ : Optional[Any] = accelerator.prepare(lowerCAmelCase__ ) try: pickle.loads(pickle.dumps(lowerCAmelCase__ ) ) except Exception as e: self.fail(F'''Accelerated optimizer pickling failed with {e}''' ) AcceleratorState._reset_state()
656
0
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class A_ (unittest.TestCase ): """simple docstring""" def __init__( self :List[str] , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :str=7 , lowerCAmelCase__ :List[Any]=3 , lowerCAmelCase__ :Union[str, Any]=30 , lowerCAmelCase__ :Optional[int]=400 , lowerCAmelCase__ :Tuple=True , lowerCAmelCase__ :str=None , lowerCAmelCase__ :List[Any]=True , lowerCAmelCase__ :Union[str, Any]=1 / 255 , lowerCAmelCase__ :int=True , lowerCAmelCase__ :List[str]=[0.5, 0.5, 0.5] , lowerCAmelCase__ :Optional[Any]=[0.5, 0.5, 0.5] , lowerCAmelCase__ :Tuple=True , ) -> List[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1_333} snake_case_ : Optional[Any] = parent snake_case_ : List[Any] = batch_size snake_case_ : str = num_channels snake_case_ : Optional[int] = min_resolution snake_case_ : List[str] = max_resolution snake_case_ : str = do_resize snake_case_ : Tuple = size snake_case_ : Tuple = do_rescale snake_case_ : List[str] = rescale_factor snake_case_ : Optional[int] = do_normalize snake_case_ : List[Any] = image_mean snake_case_ : List[Any] = image_std snake_case_ : Optional[Any] = do_pad def _A ( self :int ) -> Optional[Any]: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def _A ( self :List[Any] , lowerCAmelCase__ :Any , lowerCAmelCase__ :str=False ) -> str: '''simple docstring''' if not batched: snake_case_ : Tuple = image_inputs[0] if isinstance(lowercase_ , Image.Image ): snake_case_ : Any = image.size else: snake_case_ : str = image.shape[1], image.shape[2] if w < h: snake_case_ : Optional[int] = int(self.size["shortest_edge"] * h / w ) snake_case_ : Optional[Any] = self.size["shortest_edge"] elif w > h: snake_case_ : Union[str, Any] = self.size["shortest_edge"] snake_case_ : Union[str, Any] = int(self.size["shortest_edge"] * w / h ) else: snake_case_ : Tuple = self.size["shortest_edge"] snake_case_ : str = self.size["shortest_edge"] else: snake_case_ : str = [] for image in image_inputs: snake_case_ : Optional[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) snake_case_ : Optional[int] = max(lowercase_ , key=lambda lowerCAmelCase__ : item[0] )[0] snake_case_ : List[str] = max(lowercase_ , key=lambda lowerCAmelCase__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class A_ (snake_case__ , unittest.TestCase ): """simple docstring""" a__ = DetrImageProcessor if is_vision_available() else None def _A ( self :Optional[Any] ) -> Dict: '''simple docstring''' snake_case_ : Optional[int] = DetrImageProcessingTester(self ) @property def _A ( self :Tuple ) -> Dict: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def _A ( self :List[Any] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase_ , "image_mean" ) ) self.assertTrue(hasattr(lowercase_ , "image_std" ) ) self.assertTrue(hasattr(lowercase_ , "do_normalize" ) ) self.assertTrue(hasattr(lowercase_ , "do_rescale" ) ) self.assertTrue(hasattr(lowercase_ , "rescale_factor" ) ) self.assertTrue(hasattr(lowercase_ , "do_resize" ) ) self.assertTrue(hasattr(lowercase_ , "size" ) ) self.assertTrue(hasattr(lowercase_ , "do_pad" ) ) def _A ( self :Optional[Any] ) -> Optional[Any]: '''simple docstring''' snake_case_ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1_333} ) self.assertEqual(image_processor.do_pad , lowercase_ ) snake_case_ : int = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowercase_ ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowercase_ ) def _A ( self :Dict ) -> Union[str, Any]: '''simple docstring''' pass def _A ( self :Any ) -> Dict: '''simple docstring''' snake_case_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase_ ) for image in image_inputs: self.assertIsInstance(lowercase_ , Image.Image ) # Test not batched input snake_case_ : List[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ : Dict = self.image_processor_tester.get_expected_values(lowercase_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : List[Any] = self.image_processor_tester.get_expected_values(lowercase_ , batched=lowercase_ ) snake_case_ : Dict = image_processing(lowercase_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _A ( self :Union[str, Any] ) -> str: '''simple docstring''' snake_case_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase_ , numpify=lowercase_ ) for image in image_inputs: self.assertIsInstance(lowercase_ , np.ndarray ) # Test not batched input snake_case_ : Dict = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ : Any = self.image_processor_tester.get_expected_values(lowercase_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : Tuple = image_processing(lowercase_ , return_tensors="pt" ).pixel_values snake_case_ : int = self.image_processor_tester.get_expected_values(lowercase_ , batched=lowercase_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _A ( self :List[str] ) -> List[str]: '''simple docstring''' snake_case_ : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase_ , torchify=lowercase_ ) for image in image_inputs: self.assertIsInstance(lowercase_ , torch.Tensor ) # Test not batched input snake_case_ : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values snake_case_ : Optional[Any] = self.image_processor_tester.get_expected_values(lowercase_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : List[Any] = image_processing(lowercase_ , return_tensors="pt" ).pixel_values snake_case_ : Tuple = self.image_processor_tester.get_expected_values(lowercase_ , batched=lowercase_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def _A ( self :List[Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: snake_case_ : List[str] = json.loads(f.read() ) snake_case_ : Optional[Any] = {"image_id": 39_769, "annotations": target} # encode them snake_case_ : str = DetrImageProcessor.from_pretrained("facebook/detr-resnet-50" ) snake_case_ : int = image_processing(images=lowercase_ , annotations=lowercase_ , return_tensors="pt" ) # verify pixel values snake_case_ : List[str] = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding["pixel_values"].shape , lowercase_ ) snake_case_ : Optional[int] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowercase_ , atol=1E-4 ) ) # verify area snake_case_ : List[Any] = torch.tensor([5_887.9_600, 11_250.2_061, 489_353.8_438, 837_122.7_500, 147_967.5_156, 165_732.3_438] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowercase_ ) ) # verify boxes snake_case_ : Any = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowercase_ ) snake_case_ : Optional[Any] = torch.tensor([0.5_5_0_3, 0.2_7_6_5, 0.0_6_0_4, 0.2_2_1_5] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowercase_ , atol=1E-3 ) ) # verify image_id snake_case_ : List[str] = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowercase_ ) ) # verify is_crowd snake_case_ : Optional[int] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowercase_ ) ) # verify class_labels snake_case_ : Tuple = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowercase_ ) ) # verify orig_size snake_case_ : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowercase_ ) ) # verify size snake_case_ : Any = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowercase_ ) ) @slow def _A ( self :List[str] ) -> List[Any]: '''simple docstring''' snake_case_ : int = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: snake_case_ : Optional[Any] = json.loads(f.read() ) snake_case_ : List[Any] = {"file_name": "000000039769.png", "image_id": 39_769, "segments_info": target} snake_case_ : Union[str, Any] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them snake_case_ : Any = DetrImageProcessor.from_pretrained("facebook/detr-resnet-50-panoptic" ) snake_case_ : Optional[Any] = image_processing(images=lowercase_ , annotations=lowercase_ , masks_path=lowercase_ , return_tensors="pt" ) # verify pixel values snake_case_ : Any = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding["pixel_values"].shape , lowercase_ ) snake_case_ : Optional[int] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowercase_ , atol=1E-4 ) ) # verify area snake_case_ : Optional[Any] = torch.tensor([147_979.6_875, 165_527.0_469, 484_638.5_938, 11_292.9_375, 5_879.6_562, 7_634.1_147] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowercase_ ) ) # verify boxes snake_case_ : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowercase_ ) snake_case_ : Tuple = torch.tensor([0.2_6_2_5, 0.5_4_3_7, 0.4_6_8_8, 0.8_6_2_5] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowercase_ , atol=1E-3 ) ) # verify image_id snake_case_ : List[str] = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowercase_ ) ) # verify is_crowd snake_case_ : Union[str, Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowercase_ ) ) # verify class_labels snake_case_ : Tuple = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowercase_ ) ) # verify masks snake_case_ : Any = 822_873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowercase_ ) # verify orig_size snake_case_ : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowercase_ ) ) # verify size snake_case_ : Optional[int] = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowercase_ ) )
704
'''simple docstring''' import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __lowerCamelCase : Any = '''src/transformers''' # This is to make sure the transformers module imported is the one in the repo. __lowerCamelCase : List[str] = direct_transformers_import(PATH_TO_TRANSFORMERS) __lowerCamelCase : Optional[Any] = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __lowerCamelCase : Union[str, Any] = re.compile(R'''\[(.+?)\]\((https://huggingface\.co/.+?)\)''') __lowerCamelCase : Any = { '''DecisionTransformerConfig''', '''EncoderDecoderConfig''', '''MusicgenConfig''', '''RagConfig''', '''SpeechEncoderDecoderConfig''', '''TimmBackboneConfig''', '''VisionEncoderDecoderConfig''', '''VisionTextDualEncoderConfig''', '''LlamaConfig''', } def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : Tuple = None # source code of `config_class` snake_case_ : List[Any] = inspect.getsource(__magic_name__ ) snake_case_ : List[str] = _re_checkpoint.findall(__magic_name__ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith("/" ): snake_case_ : Optional[Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link snake_case_ : str = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: snake_case_ : Dict = ckpt_name break return checkpoint def __UpperCAmelCase ( )-> Dict: """simple docstring""" snake_case_ : Optional[int] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue snake_case_ : str = get_checkpoint_from_config_class(__magic_name__ ) snake_case_ : Union[str, Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(__magic_name__ ) if len(__magic_name__ ) > 0: snake_case_ : Tuple = "\n".join(sorted(__magic_name__ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
656
0
'''simple docstring''' import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : int = logging.get_logger(__name__) __lowerCamelCase : List[Any] = "https://openaipublic.azureedge.net/jukebox/models/" __lowerCamelCase : Optional[Any] = { "jukebox-1b-lyrics": [ "5b/vqvae.pth.tar", "5b/prior_level_0.pth.tar", "5b/prior_level_1.pth.tar", "1b_lyrics/prior_level_2.pth.tar", ], "jukebox-5b-lyrics": [ "5b/vqvae.pth.tar", "5b/prior_level_0.pth.tar", "5b/prior_level_1.pth.tar", "5b_lyrics/prior_level_2.pth.tar", ], } def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" if key.endswith(".model.1.bias" ) and len(key.split("." ) ) > 10: snake_case_ : List[str] = key.replace(".model.1.bias" ,".conv1d_1.bias" ) elif key.endswith(".model.1.weight" ) and len(key.split("." ) ) > 10: snake_case_ : str = key.replace(".model.1.weight" ,".conv1d_1.weight" ) elif key.endswith(".model.3.bias" ) and len(key.split("." ) ) > 10: snake_case_ : List[Any] = key.replace(".model.3.bias" ,".conv1d_2.bias" ) elif key.endswith(".model.3.weight" ) and len(key.split("." ) ) > 10: snake_case_ : Any = key.replace(".model.3.weight" ,".conv1d_2.weight" ) if "conditioner_blocks.0." in key: snake_case_ : str = key.replace("conditioner_blocks.0" ,"conditioner_blocks" ) if "prime_prior" in key: snake_case_ : Any = key.replace("prime_prior" ,"encoder" ) if ".emb." in key and "total" not in key and "absolute" not in key and "relative" not in key: snake_case_ : int = key.replace(".emb." ,"." ) if key.endswith("k" ): # replace vqvae.X.k with vqvae.X.codebook return key.replace(".k" ,".codebook" ) if "y_emb." in key: return key.replace("y_emb." ,"metadata_embedding." ) if "x_emb.emb." in key: snake_case_ : Optional[int] = key.replace("0.x_emb.emb" ,"embed_tokens" ) if "prime_state_ln" in key: return key.replace("prime_state_ln" ,"encoder.final_layer_norm" ) if ".ln" in key: return key.replace(".ln" ,".layer_norm" ) if "_ln" in key: return key.replace("_ln" ,"_layer_norm" ) if "prime_state_proj" in key: return key.replace("prime_state_proj" ,"encoder.proj_in" ) if "prime_x_out" in key: return key.replace("prime_x_out" ,"encoder.lm_head" ) if "prior.x_out" in key: return key.replace("x_out" ,"fc_proj_out" ) if "x_emb" in key: return key.replace("x_emb" ,"embed_tokens" ) return key def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ )-> int: """simple docstring""" snake_case_ : Tuple = {} import re snake_case_ : Optional[int] = re.compile(r"encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)" ) snake_case_ : Dict = re.compile( r"encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) snake_case_ : str = re.compile(r"encoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)" ) snake_case_ : List[Any] = re.compile(r"decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)" ) snake_case_ : List[Any] = re.compile( r"decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) snake_case_ : str = re.compile(r"decoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)" ) snake_case_ : Any = re.compile(r"conditioner_blocks.(\d*).cond.model.(\d*).(\d).(bias|weight)" ) snake_case_ : Optional[Any] = re.compile( r"conditioner_blocks.(\d*).cond.model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) snake_case_ : Any = re.compile(r"conditioner_blocks.(\d*).cond.model.(\d*).(bias|weight)" ) for original_key, value in state_dict.items(): # rename vqvae.encoder keys if re_encoder_block_conv_in.fullmatch(__magic_name__ ): snake_case_ : Any = re_encoder_block_conv_in.match(__magic_name__ ) snake_case_ : Optional[Any] = regex_match.groups() snake_case_ : List[str] = int(groups[2] ) * 2 + int(groups[3] ) snake_case_ : List[str] = F'''encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.{groups[-1]}''' snake_case_ : int = re_encoder_block_conv_in.sub(__magic_name__ ,__magic_name__ ) elif re_encoder_block_resnet.fullmatch(__magic_name__ ): snake_case_ : Dict = re_encoder_block_resnet.match(__magic_name__ ) snake_case_ : Dict = regex_match.groups() snake_case_ : List[Any] = int(groups[2] ) * 2 + int(groups[3] ) snake_case_ : Tuple = {"1": 1, "3": 2}[groups[-2]] snake_case_ : Tuple = F'''encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.''' snake_case_ : Union[str, Any] = F'''resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}''' snake_case_ : Dict = prefix + resnet_block snake_case_ : List[str] = re_encoder_block_resnet.sub(__magic_name__ ,__magic_name__ ) elif re_encoder_block_proj_out.fullmatch(__magic_name__ ): snake_case_ : List[str] = re_encoder_block_proj_out.match(__magic_name__ ) snake_case_ : Optional[int] = regex_match.groups() snake_case_ : List[str] = F'''encoders.{groups[0]}.level_blocks.{groups[1]}.proj_out.{groups[-1]}''' snake_case_ : Union[str, Any] = re_encoder_block_proj_out.sub(__magic_name__ ,__magic_name__ ) # rename vqvae.decoder keys elif re_decoder_block_conv_out.fullmatch(__magic_name__ ): snake_case_ : List[Any] = re_decoder_block_conv_out.match(__magic_name__ ) snake_case_ : int = regex_match.groups() snake_case_ : Tuple = int(groups[2] ) * 2 + int(groups[3] ) - 2 snake_case_ : int = F'''decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.{groups[-1]}''' snake_case_ : Dict = re_decoder_block_conv_out.sub(__magic_name__ ,__magic_name__ ) elif re_decoder_block_resnet.fullmatch(__magic_name__ ): snake_case_ : Dict = re_decoder_block_resnet.match(__magic_name__ ) snake_case_ : List[Any] = regex_match.groups() snake_case_ : Optional[int] = int(groups[2] ) * 2 + int(groups[3] ) - 2 snake_case_ : List[str] = {"1": 1, "3": 2}[groups[-2]] snake_case_ : Union[str, Any] = F'''decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.''' snake_case_ : Union[str, Any] = F'''resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}''' snake_case_ : Dict = prefix + resnet_block snake_case_ : Optional[int] = re_decoder_block_resnet.sub(__magic_name__ ,__magic_name__ ) elif re_decoder_block_proj_in.fullmatch(__magic_name__ ): snake_case_ : Optional[Any] = re_decoder_block_proj_in.match(__magic_name__ ) snake_case_ : Dict = regex_match.groups() snake_case_ : List[Any] = F'''decoders.{groups[0]}.level_blocks.{groups[1]}.proj_in.{groups[-1]}''' snake_case_ : Tuple = re_decoder_block_proj_in.sub(__magic_name__ ,__magic_name__ ) # rename prior cond.model to upsampler.upsample_block and resnet elif re_prior_cond_conv_out.fullmatch(__magic_name__ ): snake_case_ : str = re_prior_cond_conv_out.match(__magic_name__ ) snake_case_ : List[Any] = regex_match.groups() snake_case_ : Optional[Any] = int(groups[1] ) * 2 + int(groups[2] ) - 2 snake_case_ : List[Any] = F'''conditioner_blocks.upsampler.upsample_block.{block_index}.{groups[-1]}''' snake_case_ : List[str] = re_prior_cond_conv_out.sub(__magic_name__ ,__magic_name__ ) elif re_prior_cond_resnet.fullmatch(__magic_name__ ): snake_case_ : List[Any] = re_prior_cond_resnet.match(__magic_name__ ) snake_case_ : int = regex_match.groups() snake_case_ : Dict = int(groups[1] ) * 2 + int(groups[2] ) - 2 snake_case_ : Any = {"1": 1, "3": 2}[groups[-2]] snake_case_ : List[str] = F'''conditioner_blocks.upsampler.upsample_block.{block_index}.''' snake_case_ : Union[str, Any] = F'''resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}''' snake_case_ : List[Any] = prefix + resnet_block snake_case_ : Tuple = re_prior_cond_resnet.sub(__magic_name__ ,__magic_name__ ) elif re_prior_cond_proj_in.fullmatch(__magic_name__ ): snake_case_ : List[Any] = re_prior_cond_proj_in.match(__magic_name__ ) snake_case_ : List[str] = regex_match.groups() snake_case_ : Optional[Any] = F'''conditioner_blocks.upsampler.proj_in.{groups[-1]}''' snake_case_ : Dict = re_prior_cond_proj_in.sub(__magic_name__ ,__magic_name__ ) # keep original key else: snake_case_ : Union[str, Any] = original_key snake_case_ : List[Any] = replace_key(__magic_name__ ) if F'''{key_prefix}.{key}''' not in model_state_dict or key is None: print(F'''failed converting {original_key} to {key}, does not match''' ) # handle missmatched shape elif value.shape != model_state_dict[F'''{key_prefix}.{key}'''].shape: snake_case_ : Dict = model_state_dict[F'''{key_prefix}.{key}'''] print(F'''{original_key}-> {key} : \nshape {val.shape} and { value.shape}, do not match''' ) snake_case_ : List[str] = original_key snake_case_ : str = original_key snake_case_ : Union[str, Any] = value return new_dict @torch.no_grad() def __UpperCAmelCase ( __magic_name__=None ,__magic_name__=None )-> Any: """simple docstring""" for file in MODEL_MAPPING[model_name]: if not os.path.isfile(F'''{pytorch_dump_folder_path}/{file.split('/' )[-1]}''' ): snake_case_ : Tuple = requests.get(F'''{PREFIX}{file}''' ,allow_redirects=__magic_name__ ) os.makedirs(F'''{pytorch_dump_folder_path}/''' ,exist_ok=__magic_name__ ) open(F'''{pytorch_dump_folder_path}/{file.split('/' )[-1]}''' ,"wb" ).write(r.content ) snake_case_ : int = MODEL_MAPPING[model_name.split("/" )[-1]] snake_case_ : List[str] = JukeboxConfig.from_pretrained(__magic_name__ ) snake_case_ : Tuple = JukeboxModel(__magic_name__ ) snake_case_ : Dict = [] snake_case_ : Optional[int] = {} for i, dict_name in enumerate(__magic_name__ ): snake_case_ : int = torch.load(F'''{pytorch_dump_folder_path}/{dict_name.split('/' )[-1]}''' )["model"] snake_case_ : List[Any] = {} for k in old_dic.keys(): if k.endswith(".b" ): snake_case_ : Any = old_dic[k] elif k.endswith(".w" ): snake_case_ : int = old_dic[k] elif "level_2" not in dict_name and "cond.model." in k: snake_case_ : Union[str, Any] = old_dic[k] else: snake_case_ : Dict = old_dic[k] snake_case_ : Optional[int] = "vqvae" if i == 0 else F'''priors.{3 - i}''' snake_case_ : Union[str, Any] = fix_jukebox_keys(__magic_name__ ,model.state_dict() ,__magic_name__ ,__magic_name__ ) weight_dict.append(__magic_name__ ) snake_case_ : Dict = weight_dict.pop(0 ) model.vqvae.load_state_dict(__magic_name__ ) for i in range(len(__magic_name__ ) ): model.priors[i].load_state_dict(weight_dict[2 - i] ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) with open(F'''{pytorch_dump_folder_path}/mapping.json''' ,"w" ) as txtfile: json.dump(__magic_name__ ,__magic_name__ ) print(F'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__magic_name__ ) return weight_dict if __name__ == "__main__": __lowerCamelCase : str = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''jukebox-5b-lyrics''', type=str, help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''jukebox-5b-lyrics-converted''', type=str, help='''Path to the output PyTorch model directory.''', ) __lowerCamelCase : Dict = parser.parse_args() convert_openai_checkpoint(args.model_name, args.pytorch_dump_folder_path)
705
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCamelCase : List[str] = logging.get_logger(__name__) __lowerCamelCase : int = { '''microsoft/cvt-13''': '''https://huggingface.co/microsoft/cvt-13/resolve/main/config.json''', # See all Cvt models at https://huggingface.co/models?filter=cvt } class A_ (a_ ): """simple docstring""" a__ = '''cvt''' def __init__( self :List[Any] , lowerCAmelCase__ :Optional[int]=3 , lowerCAmelCase__ :Any=[7, 3, 3] , lowerCAmelCase__ :Dict=[4, 2, 2] , lowerCAmelCase__ :Union[str, Any]=[2, 1, 1] , lowerCAmelCase__ :Any=[64, 192, 384] , lowerCAmelCase__ :List[str]=[1, 3, 6] , lowerCAmelCase__ :str=[1, 2, 10] , lowerCAmelCase__ :Any=[4.0, 4.0, 4.0] , lowerCAmelCase__ :int=[0.0, 0.0, 0.0] , lowerCAmelCase__ :Optional[Any]=[0.0, 0.0, 0.0] , lowerCAmelCase__ :Dict=[0.0, 0.0, 0.1] , lowerCAmelCase__ :List[Any]=[True, True, True] , lowerCAmelCase__ :List[Any]=[False, False, True] , lowerCAmelCase__ :Dict=["dw_bn", "dw_bn", "dw_bn"] , lowerCAmelCase__ :Any=[3, 3, 3] , lowerCAmelCase__ :Tuple=[1, 1, 1] , lowerCAmelCase__ :Optional[int]=[2, 2, 2] , lowerCAmelCase__ :Union[str, Any]=[1, 1, 1] , lowerCAmelCase__ :Any=[1, 1, 1] , lowerCAmelCase__ :List[str]=0.0_2 , lowerCAmelCase__ :Dict=1E-1_2 , **lowerCAmelCase__ :Optional[Any] , ) -> str: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) snake_case_ : int = num_channels snake_case_ : int = patch_sizes snake_case_ : Optional[Any] = patch_stride snake_case_ : Dict = patch_padding snake_case_ : Tuple = embed_dim snake_case_ : Optional[int] = num_heads snake_case_ : Union[str, Any] = depth snake_case_ : Optional[int] = mlp_ratio snake_case_ : Tuple = attention_drop_rate snake_case_ : str = drop_rate snake_case_ : Tuple = drop_path_rate snake_case_ : Any = qkv_bias snake_case_ : Union[str, Any] = cls_token snake_case_ : int = qkv_projection_method snake_case_ : Any = kernel_qkv snake_case_ : Union[str, Any] = padding_kv snake_case_ : str = stride_kv snake_case_ : Dict = padding_q snake_case_ : Tuple = stride_q snake_case_ : Any = initializer_range snake_case_ : Any = layer_norm_eps
656
0
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) __lowerCamelCase : int = logging.getLogger(__name__) __lowerCamelCase : Union[str, Any] = 'Hello world! cécé herlolip' __lowerCamelCase : Union[str, Any] = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : List[str] = BertAbsConfig( temp_dir="." ,finetune_bert=_lowerCamelCase ,large=_lowerCamelCase ,share_emb=_lowerCamelCase ,use_bert_emb=_lowerCamelCase ,encoder="bert" ,max_pos=512 ,enc_layers=6 ,enc_hidden_size=512 ,enc_heads=8 ,enc_ff_size=512 ,enc_dropout=0.2 ,dec_layers=6 ,dec_hidden_size=768 ,dec_heads=8 ,dec_ff_size=2048 ,dec_dropout=0.2 ,) snake_case_ : Tuple = torch.load(_lowerCamelCase ,lambda __magic_name__ ,__magic_name__ : storage ) snake_case_ : int = AbsSummarizer(_lowerCamelCase ,torch.device("cpu" ) ,_lowerCamelCase ) original.eval() snake_case_ : Any = BertAbsSummarizer(_lowerCamelCase ,torch.device("cpu" ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info("convert the model" ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info("Make sure that the models' outputs are identical" ) snake_case_ : Any = BertTokenizer.from_pretrained("bert-base-uncased" ) # prepare the model inputs snake_case_ : Tuple = tokenizer.encode("This is sample éàalj'-." ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(_lowerCamelCase )) ) snake_case_ : Optional[Any] = torch.tensor(_lowerCamelCase ).unsqueeze(0 ) snake_case_ : Optional[Any] = tokenizer.encode("This is sample 3 éàalj'-." ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(_lowerCamelCase )) ) snake_case_ : Optional[int] = torch.tensor(_lowerCamelCase ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass snake_case_ : Any = encoder_input_ids snake_case_ : Union[str, Any] = decoder_input_ids snake_case_ : Tuple = None snake_case_ : Optional[int] = None snake_case_ : int = None snake_case_ : Any = None snake_case_ : Optional[int] = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical snake_case_ : List[str] = original(_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase )[0] snake_case_ : Dict = original.generator(_lowerCamelCase ) snake_case_ : str = new_model( _lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase )[0] snake_case_ : Optional[int] = new_model.generator(_lowerCamelCase ) snake_case_ : Optional[Any] = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print("Maximum absolute difference beween weights: {:.2f}".format(_lowerCamelCase ) ) snake_case_ : Union[str, Any] = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print("Maximum absolute difference beween weights: {:.2f}".format(_lowerCamelCase ) ) snake_case_ : Any = torch.allclose(_lowerCamelCase ,_lowerCamelCase ,atol=1E-3 ) if are_identical: logging.info("all weights are equal up to 1e-3" ) else: raise ValueError("the weights are different. The new model is likely different from the original one." ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info("saving the model's state dictionary" ) torch.save( new_model.state_dict() ,"./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin" ) if __name__ == "__main__": __lowerCamelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) __lowerCamelCase : Optional[Any] = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
706
'''simple docstring''' import sacrebleu as scb from packaging import version from sacrebleu import TER import datasets __lowerCamelCase : str = '''\ @inproceedings{snover-etal-2006-study, title = "A Study of Translation Edit Rate with Targeted Human Annotation", author = "Snover, Matthew and Dorr, Bonnie and Schwartz, Rich and Micciulla, Linnea and Makhoul, John", booktitle = "Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers", month = aug # " 8-12", year = "2006", address = "Cambridge, Massachusetts, USA", publisher = "Association for Machine Translation in the Americas", url = "https://aclanthology.org/2006.amta-papers.25", pages = "223--231", } @inproceedings{post-2018-call, title = "A Call for Clarity in Reporting {BLEU} Scores", author = "Post, Matt", booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers", month = oct, year = "2018", address = "Belgium, Brussels", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W18-6319", pages = "186--191", } ''' __lowerCamelCase : Dict = '''\ TER (Translation Edit Rate, also called Translation Error Rate) is a metric to quantify the edit operations that a hypothesis requires to match a reference translation. We use the implementation that is already present in sacrebleu (https://github.com/mjpost/sacreBLEU#ter), which in turn is inspired by the TERCOM implementation, which can be found here: https://github.com/jhclark/tercom. The implementation here is slightly different from sacrebleu in terms of the required input format. The length of the references and hypotheses lists need to be the same, so you may need to transpose your references compared to sacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534 See the README.md file at https://github.com/mjpost/sacreBLEU#ter for more information. ''' __lowerCamelCase : int = ''' Produces TER scores alongside the number of edits and reference length. Args: predictions (list of str): The system stream (a sequence of segments). references (list of list of str): A list of one or more reference streams (each a sequence of segments). normalized (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. ignore_punct (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. support_zh_ja_chars (boolean): If `True`, tokenization/normalization supports processing of Chinese characters, as well as Japanese Kanji, Hiragana, Katakana, and Phonetic Extensions of Katakana. Only applies if `normalized = True`. Defaults to `False`. case_sensitive (boolean): If `False`, makes all predictions and references lowercase to ignore differences in case. Defaults to `False`. Returns: \'score\' (float): TER score (num_edits / sum_ref_lengths * 100) \'num_edits\' (int): The cumulative number of edits \'ref_length\' (float): The cumulative average reference length Examples: Example 1: >>> predictions = ["does this sentence match??", ... "what about this sentence?", ... "What did the TER metric user say to the developer?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"], ... ["Your jokes are...", "...TERrible"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {\'score\': 150.0, \'num_edits\': 15, \'ref_length\': 10.0} Example 2: >>> predictions = ["does this sentence match??", ... "what about this sentence?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {\'score\': 62.5, \'num_edits\': 5, \'ref_length\': 8.0} Example 3: >>> predictions = ["does this sentence match??", ... "what about this sentence?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... normalized=True, ... case_sensitive=True) >>> print(results) {\'score\': 57.14285714285714, \'num_edits\': 6, \'ref_length\': 10.5} Example 4: >>> predictions = ["does this sentence match??", ... "what about this sentence?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {\'score\': 0.0, \'num_edits\': 0, \'ref_length\': 8.0} Example 5: >>> predictions = ["does this sentence match??", ... "what about this sentence?", ... "What did the TER metric user say to the developer?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"], ... ["Your jokes are...", "...TERrible"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {\'score\': 100.0, \'num_edits\': 10, \'ref_length\': 10.0} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): """simple docstring""" def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' if version.parse(scb.__version__ ) < version.parse("1.4.12" ): raise ImportWarning( "To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n" "You can install it with `pip install \"sacrebleu>=1.4.12\"`." ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="http://www.cs.umd.edu/~snover/tercom/" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Sequence(datasets.Value("string" , id="sequence" ) , id="references" ), } ) , codebase_urls=["https://github.com/mjpost/sacreBLEU#ter"] , reference_urls=[ "https://github.com/jhclark/tercom", ] , ) def _A ( self :Any , lowerCAmelCase__ :Any , lowerCAmelCase__ :str , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , ) -> Optional[Any]: '''simple docstring''' snake_case_ : List[str] = len(references[0] ) if any(len(lowerCAmelCase__ ) != references_per_prediction for refs in references ): raise ValueError("Sacrebleu requires the same number of references for each prediction" ) snake_case_ : List[str] = [[refs[i] for refs in references] for i in range(lowerCAmelCase__ )] snake_case_ : List[str] = TER( normalized=lowerCAmelCase__ , no_punct=lowerCAmelCase__ , asian_support=lowerCAmelCase__ , case_sensitive=lowerCAmelCase__ , ) snake_case_ : Any = sb_ter.corpus_score(lowerCAmelCase__ , lowerCAmelCase__ ) return {"score": output.score, "num_edits": output.num_edits, "ref_length": output.ref_length}
656
0
'''simple docstring''' import argparse import os import transformers from .convert_slow_tokenizer import SLOW_TO_FAST_CONVERTERS from .utils import logging logging.set_verbosity_info() __lowerCamelCase : Dict = logging.get_logger(__name__) __lowerCamelCase : List[str] = {name: getattr(transformers, name + '''Fast''') for name in SLOW_TO_FAST_CONVERTERS} def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ )-> Any: """simple docstring""" if tokenizer_name is not None and tokenizer_name not in TOKENIZER_CLASSES: raise ValueError(F'''Unrecognized tokenizer name, should be one of {list(TOKENIZER_CLASSES.keys() )}.''' ) if tokenizer_name is None: snake_case_ : Dict = TOKENIZER_CLASSES else: snake_case_ : List[str] = {tokenizer_name: getattr(_SCREAMING_SNAKE_CASE ,tokenizer_name + "Fast" )} logger.info(F'''Loading tokenizer classes: {tokenizer_names}''' ) for tokenizer_name in tokenizer_names: snake_case_ : Dict = TOKENIZER_CLASSES[tokenizer_name] snake_case_ : int = True if checkpoint_name is None: snake_case_ : str = list(tokenizer_class.max_model_input_sizes.keys() ) else: snake_case_ : int = [checkpoint_name] logger.info(F'''For tokenizer {tokenizer_class.__class__.__name__} loading checkpoints: {checkpoint_names}''' ) for checkpoint in checkpoint_names: logger.info(F'''Loading {tokenizer_class.__class__.__name__} {checkpoint}''' ) # Load tokenizer snake_case_ : Dict = tokenizer_class.from_pretrained(_SCREAMING_SNAKE_CASE ,force_download=_SCREAMING_SNAKE_CASE ) # Save fast tokenizer logger.info(F'''Save fast tokenizer to {dump_path} with prefix {checkpoint} add_prefix {add_prefix}''' ) # For organization names we create sub-directories if "/" in checkpoint: snake_case_, snake_case_ : int = checkpoint.split("/" ) snake_case_ : Optional[int] = os.path.join(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) elif add_prefix: snake_case_ : Union[str, Any] = checkpoint snake_case_ : Optional[Any] = dump_path else: snake_case_ : str = None snake_case_ : str = dump_path logger.info(F'''=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}''' ) if checkpoint in list(tokenizer.pretrained_vocab_files_map.values() )[0]: snake_case_ : Any = list(tokenizer.pretrained_vocab_files_map.values() )[0][checkpoint] snake_case_ : Optional[int] = file_path.split(_SCREAMING_SNAKE_CASE )[-1][0] if next_char == "/": snake_case_ : Optional[Any] = os.path.join(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) snake_case_ : Tuple = None logger.info(F'''=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}''' ) snake_case_ : Union[str, Any] = tokenizer.save_pretrained( _SCREAMING_SNAKE_CASE ,legacy_format=_SCREAMING_SNAKE_CASE ,filename_prefix=_SCREAMING_SNAKE_CASE ) logger.info(F'''=> File names {file_names}''' ) for file_name in file_names: if not file_name.endswith("tokenizer.json" ): os.remove(_SCREAMING_SNAKE_CASE ) logger.info(F'''=> removing {file_name}''' ) if __name__ == "__main__": __lowerCamelCase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--dump_path''', default=None, type=str, required=True, help='''Path to output generated fast tokenizer files.''' ) parser.add_argument( '''--tokenizer_name''', default=None, type=str, help=( f'''Optional tokenizer type selected in the list of {list(TOKENIZER_CLASSES.keys())}. If not given, will ''' '''download and convert all the checkpoints from AWS.''' ), ) parser.add_argument( '''--checkpoint_name''', default=None, type=str, help='''Optional checkpoint name. If not given, will download and convert the canonical checkpoints from AWS.''', ) parser.add_argument( '''--force_download''', action='''store_true''', help='''Re-download checkpoints.''', ) __lowerCamelCase : Any = parser.parse_args() convert_slow_checkpoint_to_fast(args.tokenizer_name, args.checkpoint_name, args.dump_path, args.force_download)
707
'''simple docstring''' from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : Any = { "repo_name": ["test_repo1", "test_repo2", "test_repo3"], "path": ["test_1.py", "test_2.py", "unit_test.py"], "content": ["a " * 20, "a " * 30, "b " * 7], } snake_case_ : int = Dataset.from_dict(__magic_name__ ) return dataset class A_ (a_ ): """simple docstring""" def _A ( self :List[str] ) -> str: '''simple docstring''' snake_case_ : Union[str, Any] = get_dataset() snake_case_ : Optional[int] = make_duplicate_clusters(lowerCAmelCase__ , 0.8_5 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case_ : Optional[int] = get_dataset() snake_case_, snake_case_ : List[Any] = deduplicate_dataset(lowerCAmelCase__ ) self.assertEqual(len(lowerCAmelCase__ ) , 2 ) print(lowerCAmelCase__ ) self.assertEqual(duplicate_clusters[0][0]["copies"] , 2 ) self.assertEqual(duplicate_clusters[0][0]["is_extreme"] , lowerCAmelCase__ )
656
0
'''simple docstring''' import os from bleurt import score # From: git+https://github.com/google-research/bleurt.git import datasets __lowerCamelCase : List[str] = datasets.logging.get_logger(__name__) __lowerCamelCase : List[str] = '''\ @inproceedings{bleurt, title={BLEURT: Learning Robust Metrics for Text Generation}, author={Thibault Sellam and Dipanjan Das and Ankur P. Parikh}, booktitle={ACL}, year={2020}, url={https://arxiv.org/abs/2004.04696} } ''' __lowerCamelCase : Any = '''\ BLEURT a learnt evaluation metric for Natural Language Generation. It is built using multiple phases of transfer learning starting from a pretrained BERT model (Devlin et al. 2018) and then employing another pre-training phrase using synthetic data. Finally it is trained on WMT human annotations. You may run BLEURT out-of-the-box or fine-tune it for your specific application (the latter is expected to perform better). See the project\'s README at https://github.com/google-research/bleurt#readme for more information. ''' __lowerCamelCase : str = ''' BLEURT score. Args: `predictions` (list of str): prediction/candidate sentences `references` (list of str): reference sentences `checkpoint` BLEURT checkpoint. Will default to BLEURT-tiny if None. Returns: \'scores\': List of scores. Examples: >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> bleurt = datasets.load_metric("bleurt") >>> results = bleurt.compute(predictions=predictions, references=references) >>> print([round(v, 2) for v in results["scores"]]) [1.03, 1.04] ''' __lowerCamelCase : int = { '''bleurt-tiny-128''': '''https://storage.googleapis.com/bleurt-oss/bleurt-tiny-128.zip''', '''bleurt-tiny-512''': '''https://storage.googleapis.com/bleurt-oss/bleurt-tiny-512.zip''', '''bleurt-base-128''': '''https://storage.googleapis.com/bleurt-oss/bleurt-base-128.zip''', '''bleurt-base-512''': '''https://storage.googleapis.com/bleurt-oss/bleurt-base-512.zip''', '''bleurt-large-128''': '''https://storage.googleapis.com/bleurt-oss/bleurt-large-128.zip''', '''bleurt-large-512''': '''https://storage.googleapis.com/bleurt-oss/bleurt-large-512.zip''', '''BLEURT-20-D3''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D3.zip''', '''BLEURT-20-D6''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D6.zip''', '''BLEURT-20-D12''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D12.zip''', '''BLEURT-20''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20.zip''', } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): """simple docstring""" def _A ( self :Optional[int] ) -> Dict: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://github.com/google-research/bleurt" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/google-research/bleurt"] , reference_urls=["https://github.com/google-research/bleurt", "https://arxiv.org/abs/2004.04696"] , ) def _A ( self :int , lowerCAmelCase__ :Union[str, Any] ) -> Optional[int]: '''simple docstring''' if self.config_name == "default": logger.warning( "Using default BLEURT-Base checkpoint for sequence maximum length 128. " "You can use a bigger model for better results with e.g.: datasets.load_metric('bleurt', 'bleurt-large-512')." ) snake_case_ : Optional[Any] = """bleurt-base-128""" if self.config_name.lower() in CHECKPOINT_URLS: snake_case_ : Any = self.config_name.lower() elif self.config_name.upper() in CHECKPOINT_URLS: snake_case_ : Union[str, Any] = self.config_name.upper() else: raise KeyError( F'''{self.config_name} model not found. You should supply the name of a model checkpoint for bleurt in {CHECKPOINT_URLS.keys()}''' ) # download the model checkpoint specified by self.config_name and set up the scorer snake_case_ : Union[str, Any] = dl_manager.download_and_extract(CHECKPOINT_URLS[checkpoint_name] ) snake_case_ : str = score.BleurtScorer(os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) ) def _A ( self :Dict , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Any ) -> int: '''simple docstring''' snake_case_ : Optional[Any] = self.scorer.score(references=lowerCAmelCase__ , candidates=lowerCAmelCase__ ) return {"scores": scores}
708
'''simple docstring''' from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable __lowerCamelCase : Dict = { '''configuration_gpt_neox_japanese''': ['''GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoXJapaneseConfig'''], '''tokenization_gpt_neox_japanese''': ['''GPTNeoXJapaneseTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ '''GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoXJapaneseForCausalLM''', '''GPTNeoXJapaneseLayer''', '''GPTNeoXJapaneseModel''', '''GPTNeoXJapanesePreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neox_japanese import GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXJapaneseConfig from .tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox_japanese import ( GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseLayer, GPTNeoXJapaneseModel, GPTNeoXJapanesePreTrainedModel, ) else: import sys __lowerCamelCase : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
656
0
'''simple docstring''' import string def __UpperCAmelCase ( __magic_name__ )-> Optional[Any]: """simple docstring""" for key in range(len(string.ascii_uppercase ) ): snake_case_ : List[Any] = '' for symbol in message: if symbol in string.ascii_uppercase: snake_case_ : Any = string.ascii_uppercase.find(lowerCAmelCase__ ) snake_case_ : Optional[Any] = num - key if num < 0: snake_case_ : Any = num + len(string.ascii_uppercase ) snake_case_ : List[Any] = translated + string.ascii_uppercase[num] else: snake_case_ : int = translated + symbol print(F'''Decryption using Key #{key}: {translated}''' ) def __UpperCAmelCase ( )-> str: """simple docstring""" snake_case_ : Dict = input("Encrypted message: " ) snake_case_ : List[str] = message.upper() decrypt(lowerCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
709
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ )-> list[int]: """simple docstring""" if length <= 0 or not isinstance(__magic_name__ ,__magic_name__ ): raise ValueError("Length must be a positive integer." ) return [n * (2 * n - 1) for n in range(__magic_name__ )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=10))
656
0
'''simple docstring''' from unittest.mock import patch import pyspark from datasets.packaged_modules.spark.spark import ( Spark, SparkExamplesIterable, _generate_iterable_examples, ) from ..utils import ( require_dill_gt_0_3_2, require_not_windows, ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : str = [] for part_id in partition_order: snake_case_ : Dict = df.where(F'''SPARK_PARTITION_ID() = {part_id}''' ).collect() for row_idx, row in enumerate(__magic_name__ ): expected_row_ids_and_row_dicts.append((F'''{part_id}_{row_idx}''', row.asDict()) ) return expected_row_ids_and_row_dicts @require_not_windows @require_dill_gt_0_3_2 def __UpperCAmelCase ( )-> Tuple: """simple docstring""" snake_case_ : Dict = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() snake_case_ : List[Any] = spark.range(100 ).repartition(1 ) snake_case_ : Tuple = Spark(__magic_name__ ) # The id ints will be converted to Pyarrow int64s, so each row will be 8 bytes. Setting a max_shard_size of 16 means # that each partition can hold 2 rows. spark_builder._repartition_df_if_needed(max_shard_size=16 ) # Given that the dataframe has 100 rows and each partition has 2 rows, we expect 50 partitions. assert spark_builder.df.rdd.getNumPartitions() == 50 @require_not_windows @require_dill_gt_0_3_2 def __UpperCAmelCase ( )-> Optional[int]: """simple docstring""" snake_case_ : Union[str, Any] = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() snake_case_ : Dict = spark.range(10 ).repartition(2 ) snake_case_ : Union[str, Any] = [1, 0] snake_case_ : Optional[Any] = _generate_iterable_examples(__magic_name__ ,__magic_name__ ) # Reverse the partitions. snake_case_ : Dict = _get_expected_row_ids_and_row_dicts_for_partition_order(__magic_name__ ,__magic_name__ ) for i, (row_id, row_dict) in enumerate(generate_fn() ): snake_case_, snake_case_ : List[Any] = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def __UpperCAmelCase ( )-> Optional[Any]: """simple docstring""" snake_case_ : Optional[int] = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() snake_case_ : Optional[Any] = spark.range(10 ).repartition(1 ) snake_case_ : str = SparkExamplesIterable(__magic_name__ ) assert it.n_shards == 1 for i, (row_id, row_dict) in enumerate(__magic_name__ ): assert row_id == F'''0_{i}''' assert row_dict == {"id": i} @require_not_windows @require_dill_gt_0_3_2 def __UpperCAmelCase ( )-> Union[str, Any]: """simple docstring""" snake_case_ : Any = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() snake_case_ : int = spark.range(30 ).repartition(3 ) # Mock the generator so that shuffle reverses the partition indices. with patch("numpy.random.Generator" ) as generator_mock: snake_case_ : List[str] = lambda __magic_name__ : x.reverse() snake_case_ : Dict = _get_expected_row_ids_and_row_dicts_for_partition_order(__magic_name__ ,[2, 1, 0] ) snake_case_ : Union[str, Any] = SparkExamplesIterable(__magic_name__ ).shuffle_data_sources(__magic_name__ ) assert shuffled_it.n_shards == 3 for i, (row_id, row_dict) in enumerate(__magic_name__ ): snake_case_, snake_case_ : Tuple = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def __UpperCAmelCase ( )-> Dict: """simple docstring""" snake_case_ : str = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() snake_case_ : Optional[Any] = spark.range(20 ).repartition(4 ) # Partitions 0 and 2 snake_case_ : List[Any] = SparkExamplesIterable(__magic_name__ ).shard_data_sources(worker_id=0 ,num_workers=2 ) assert shard_it_a.n_shards == 2 snake_case_ : List[str] = _get_expected_row_ids_and_row_dicts_for_partition_order(__magic_name__ ,[0, 2] ) for i, (row_id, row_dict) in enumerate(__magic_name__ ): snake_case_, snake_case_ : List[Any] = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict # Partitions 1 and 3 snake_case_ : List[str] = SparkExamplesIterable(__magic_name__ ).shard_data_sources(worker_id=1 ,num_workers=2 ) assert shard_it_a.n_shards == 2 snake_case_ : Union[str, Any] = _get_expected_row_ids_and_row_dicts_for_partition_order(__magic_name__ ,[1, 3] ) for i, (row_id, row_dict) in enumerate(__magic_name__ ): snake_case_, snake_case_ : int = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def __UpperCAmelCase ( )-> Dict: """simple docstring""" snake_case_ : List[Any] = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() snake_case_ : List[str] = spark.range(100 ).repartition(1 ) snake_case_ : Optional[Any] = Spark(__magic_name__ ) # Choose a small max_shard_size for maximum partitioning. spark_builder._repartition_df_if_needed(max_shard_size=1 ) # The new number of partitions should not be greater than the number of rows. assert spark_builder.df.rdd.getNumPartitions() == 100
710
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def __UpperCAmelCase ( __magic_name__=None )-> List[str]: """simple docstring""" if subparsers is not None: snake_case_ : List[str] = subparsers.add_parser("test" ) else: snake_case_ : List[Any] = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" ,default=__magic_name__ ,help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) ,) if subparsers is not None: parser.set_defaults(func=__magic_name__ ) return parser def __UpperCAmelCase ( __magic_name__ )-> Tuple: """simple docstring""" snake_case_ : Optional[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: snake_case_ : str = script_name else: snake_case_ : Any = F'''--config_file={args.config_file} {script_name}''' snake_case_ : Union[str, Any] = ["accelerate-launch"] + test_args.split() snake_case_ : Optional[int] = execute_subprocess_async(__magic_name__ ,env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : Dict = test_command_parser() snake_case_ : Dict = parser.parse_args() test_command(__magic_name__ ) if __name__ == "__main__": main()
656
0
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : Optional[int] = logging.get_logger(__name__) __lowerCamelCase : Optional[int] = torch.device('''cpu''') def __UpperCAmelCase ( )-> Optional[int]: """simple docstring""" snake_case_ : Union[str, Any] = "http://images.cocodataset.org/val2017/000000039769.jpg" snake_case_ : Any = Image.open(requests.get(__magic_name__ ,stream=__magic_name__ ).raw ) return im def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.17_03E00, 2.11_07E00, -2.08_11E00, 8.86_85E-01, 2.43_60E-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.96_36E-01, 2.34_78E-01, -1.69_63E00, -1.73_81E00, -8.63_37E-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.27_68E-01, -4.74_29E-01, -1.08_97E00, -1.02_48E00, 3.55_23E-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.53_30E-01, 2.42_11E-01, -6.01_85E-01, -8.27_89E-01, -6.04_46E-02] ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Union[str, Any] = dct.pop(__magic_name__ ) snake_case_ : Union[str, Any] = val def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : Dict = [] for k in state_dict.keys(): snake_case_ : str = k if ".pwconv" in k: snake_case_ : List[str] = k_new.replace(".pwconv" ,".point_wise_conv" ) if ".dwconv" in k: snake_case_ : Dict = k_new.replace(".dwconv" ,".depth_wise_conv" ) if ".Proj." in k: snake_case_ : Tuple = k_new.replace(".Proj." ,".proj." ) if "patch_embed" in k_new: snake_case_ : List[Any] = k_new.replace("patch_embed" ,"swiftformer.patch_embed.patch_embedding" ) if "network" in k_new: snake_case_ : int = k_new.split("." ) if ls[2].isdigit(): snake_case_ : Dict = "swiftformer.encoder.network." + ls[1] + ".blocks." + ls[2] + "." + ".".join(ls[3:] ) else: snake_case_ : Dict = k_new.replace("network" ,"swiftformer.encoder.network" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> int: """simple docstring""" snake_case_ : Any = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size snake_case_ : Optional[Any] = 1000 snake_case_ : Tuple = "huggingface/label-files" snake_case_ : Union[str, Any] = "imagenet-1k-id2label.json" snake_case_ : List[str] = json.load(open(hf_hub_download(__magic_name__ ,__magic_name__ ,repo_type="dataset" ) ,"r" ) ) snake_case_ : List[Any] = {int(__magic_name__ ): v for k, v in idalabel.items()} snake_case_ : Any = idalabel snake_case_ : Tuple = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": snake_case_ : Optional[Any] = [3, 3, 6, 4] snake_case_ : Any = [48, 56, 112, 220] elif swiftformer_name == "swiftformer_s": snake_case_ : int = [3, 3, 9, 6] snake_case_ : int = [48, 64, 168, 224] elif swiftformer_name == "swiftformer_l1": snake_case_ : Optional[int] = [4, 3, 10, 5] snake_case_ : Dict = [48, 96, 192, 384] elif swiftformer_name == "swiftformer_l3": snake_case_ : Union[str, Any] = [4, 4, 12, 6] snake_case_ : List[str] = [64, 128, 320, 512] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("https" ): snake_case_ : Optional[int] = torch.hub.load_state_dict_from_url(__magic_name__ ,map_location="cpu" ,check_hash=__magic_name__ ) else: snake_case_ : Optional[int] = torch.load(__magic_name__ ,map_location="cpu" ) snake_case_ : List[str] = checkpoint snake_case_ : int = create_rename_keys(__magic_name__ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(__magic_name__ ,__magic_name__ ,__magic_name__ ) # load HuggingFace model snake_case_ : Tuple = SwiftFormerForImageClassification(__magic_name__ ).eval() hf_model.load_state_dict(__magic_name__ ) # prepare test inputs snake_case_ : Tuple = prepare_img() snake_case_ : int = ViTImageProcessor.from_pretrained("preprocessor_config" ) snake_case_ : List[Any] = processor(images=__magic_name__ ,return_tensors="pt" ) # compare outputs from both models snake_case_ : Optional[Any] = get_expected_output(__magic_name__ ) snake_case_ : Any = hf_model(inputs["pixel_values"] ).logits assert hf_logits.shape == torch.Size([1, 1000] ) assert torch.allclose(hf_logits[0, 0:5] ,__magic_name__ ,atol=1E-3 ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) print(F'''Saving model {swiftformer_name} to {pytorch_dump_folder_path}''' ) hf_model.save_pretrained(__magic_name__ ) if __name__ == "__main__": __lowerCamelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--swiftformer_name''', default='''swiftformer_xs''', choices=['''swiftformer_xs''', '''swiftformer_s''', '''swiftformer_l1''', '''swiftformer_l3'''], type=str, help='''Name of the SwiftFormer model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''./converted_outputs/''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--original_ckpt''', default=None, type=str, help='''Path to the original model checkpoint.''') __lowerCamelCase : Dict = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
711
'''simple docstring''' from scipy.stats import spearmanr import datasets __lowerCamelCase : str = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' __lowerCamelCase : int = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' __lowerCamelCase : List[str] = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): """simple docstring""" def _A ( self :str ) -> Union[str, Any]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } ) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"] , ) def _A ( self :Optional[int] , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :Optional[Any]=False ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[Any] = spearmanr(lowerCAmelCase__ , lowerCAmelCase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
656
0
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" if not grid or not grid[0]: raise TypeError("The grid does not contain the appropriate information" ) for cell_n in range(1 ,len(grid[0] ) ): grid[0][cell_n] += grid[0][cell_n - 1] snake_case_ : Union[str, Any] = grid[0] for row_n in range(1 ,len(__lowerCAmelCase ) ): snake_case_ : Dict = grid[row_n] snake_case_ : List[Any] = fill_row(__lowerCAmelCase ,__lowerCAmelCase ) snake_case_ : List[Any] = grid[row_n] return grid[-1][-1] def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> list: """simple docstring""" current_row[0] += row_above[0] for cell_n in range(1 ,len(__lowerCAmelCase ) ): current_row[cell_n] += min(current_row[cell_n - 1] ,row_above[cell_n] ) return current_row if __name__ == "__main__": import doctest doctest.testmod()
712
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): __lowerCamelCase : Optional[Any] = get_tests_dir('''fixtures/test_sentencepiece.model''') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right __lowerCamelCase : str = 128022 __lowerCamelCase : List[Any] = 128028 @require_sentencepiece class A_ (a_ , unittest.TestCase ): """simple docstring""" a__ = MaMaaaTokenizer a__ = False a__ = False a__ = True def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' super().setUp() snake_case_ : int = ["</s>", "<unk>", "▁This", "▁is", "▁a", "▁t", "est", "\u0120", "<pad>"] snake_case_ : Any = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) snake_case_ : Optional[int] = Path(self.tmpdirname ) save_json(lowerCAmelCase__ , save_dir / VOCAB_FILES_NAMES["vocab_file"] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(lowerCAmelCase__ , save_dir / VOCAB_FILES_NAMES["spm_file"] ) snake_case_ : Union[str, Any] = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def _A ( self :List[Any] , **lowerCAmelCase__ :List[Any] ) -> str: '''simple docstring''' return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def _A ( self :Optional[int] , lowerCAmelCase__ :Any ) -> Optional[int]: '''simple docstring''' return ( "This is a test", "This is a test", ) def _A ( self :List[str] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : str = "</s>" snake_case_ : Union[str, Any] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase__ ) , lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = self.get_tokenizer() snake_case_ : Any = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "</s>" ) self.assertEqual(vocab_keys[1] , "<unk>" ) self.assertEqual(vocab_keys[-1] , "<s>" ) self.assertEqual(len(lowerCAmelCase__ ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip("Skip this test while all models are still to be uploaded." ) def _A ( self :List[Any] ) -> Union[str, Any]: '''simple docstring''' pass def _A ( self :Optional[int] ) -> int: '''simple docstring''' snake_case_ : int = self.get_tokenizer() snake_case_ : List[str] = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowerCAmelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [2, 3, 4, 5, 6] , ) snake_case_ : Any = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(lowerCAmelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) snake_case_ : Any = tokenizer.convert_tokens_to_string(lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , "This is a test" ) @slow def _A ( self :Any ) -> List[Any]: '''simple docstring''' snake_case_ : int = {"input_ids": [[128_022, 110_108, 397, 11, 38_272, 2_247, 124_811, 285, 18_105, 1_586, 207, 7, 39_534, 4_428, 397, 1_019, 18_105, 1_586, 207, 7, 41_337, 16_786, 241, 7, 20_214, 17, 125_690, 10_398, 7, 44_378, 58_069, 68_342, 7_798, 7_343, 11, 299, 33_310, 4, 158, 37_350, 94_077, 4_569, 299, 33_310, 90, 4, 52_840, 290, 4, 31_270, 112, 299, 682, 4, 52_840, 39_953, 14_079, 193, 52_519, 90_894, 17_894, 120_697, 11, 40_445, 551, 17, 1_019, 52_519, 90_894, 17_756, 963, 11, 40_445, 480, 17, 9_792, 1_120, 5_173, 1_393, 6_240, 16_786, 241, 120_996, 28, 1_245, 1_393, 118_240, 11_123, 1_019, 93_612, 2_691, 10_618, 98_058, 120_409, 1_928, 279, 4, 40_683, 367, 178, 207, 1_019, 103, 103_121, 506, 65_296, 5, 2], [128_022, 21_217, 367, 117, 125_450, 128, 719, 7, 7_308, 40, 93_612, 12_669, 1_116, 16_704, 71, 17_785, 3_699, 15_592, 35, 144, 9_584, 241, 11_943, 713, 950, 799, 2_247, 88_427, 150, 149, 118_813, 120_706, 1_019, 106_906, 81_518, 28, 1_224, 22_799, 397, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [128_022, 1_658, 123_311, 5_155, 5_578, 4_722, 279, 14_947, 2_366, 1_120, 1_197, 14, 1_348, 9_232, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase__ , model_name="facebook/m2m100_418M" , revision="c168bae485c864188cf9aa0e4108b0b6934dc91e" , ) @require_torch @require_sentencepiece @require_tokenizers class A_ (unittest.TestCase ): """simple docstring""" a__ = '''facebook/m2m100_418M''' a__ = [ '''In my opinion, there are two levels of response from the French government.''', '''NSA Affair Emphasizes Complete Lack of Debate on Intelligence''', ] a__ = [ '''Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.''', '''L\'affaire NSA souligne l\'absence totale de débat sur le renseignement''', ] # fmt: off a__ = [EN_CODE, 593, 1949, 115781, 4, 71586, 4234, 60633, 126233, 432, 123808, 15592, 1197, 117132, 120618, 5, 2] @classmethod def _A ( cls :str ) -> int: '''simple docstring''' snake_case_ : MaMaaaTokenizer = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en" , tgt_lang="fr" ) snake_case_ : List[str] = 1 return cls def _A ( self :Tuple ) -> Union[str, Any]: '''simple docstring''' self.assertEqual(self.tokenizer.get_lang_id("ar" ) , 128_006 ) self.assertEqual(self.tokenizer.get_lang_id("en" ) , 128_022 ) self.assertEqual(self.tokenizer.get_lang_id("ro" ) , 128_076 ) self.assertEqual(self.tokenizer.get_lang_id("mr" ) , 128_063 ) def _A ( self :Optional[int] ) -> List[str]: '''simple docstring''' snake_case_ : Dict = self.tokenizer.get_vocab() self.assertEqual(len(lowerCAmelCase__ ) , self.tokenizer.vocab_size ) self.assertEqual(vocab["<unk>"] , 3 ) self.assertIn(self.tokenizer.get_lang_token("en" ) , lowerCAmelCase__ ) def _A ( self :Any ) -> Dict: '''simple docstring''' snake_case_ : List[str] = "en" snake_case_ : Dict = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' self.assertIn(lowerCAmelCase__ , self.tokenizer.all_special_ids ) # fmt: off snake_case_ : Dict = [FR_CODE, 5_364, 82, 8_642, 4, 294, 47, 8, 14_028, 136, 3_286, 9_706, 6, 90_797, 6, 144_012, 162, 88_128, 30_061, 5, 2] # fmt: on snake_case_ : List[str] = self.tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) snake_case_ : str = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertNotIn(self.tokenizer.eos_token , lowerCAmelCase__ ) def _A ( self :Tuple ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = tempfile.mkdtemp() snake_case_ : int = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(lowerCAmelCase__ ) snake_case_ : List[str] = MaMaaaTokenizer.from_pretrained(lowerCAmelCase__ ) self.assertDictEqual(new_tok.lang_token_to_id , lowerCAmelCase__ ) @require_torch def _A ( self :Optional[Any] ) -> str: '''simple docstring''' snake_case_ : Union[str, Any] = "en" snake_case_ : Tuple = "fr" snake_case_ : Optional[int] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=lowerCAmelCase__ , return_tensors="pt" ) snake_case_ : Dict = shift_tokens_right( batch["labels"] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: snake_case_ : str = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : List[str] = "mr" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) snake_case_ : int = "zh" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def _A ( self :str ) -> int: '''simple docstring''' snake_case_ : Dict = "mr" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) snake_case_ : Tuple = "zh" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Optional[int] = self.tokenizer._build_translation_inputs("A test" , return_tensors="pt" , src_lang="en" , tgt_lang="ar" ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { # en_XX, A, test, EOS "input_ids": [[128_022, 58, 4_183, 2]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 128_006, } , )
656
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __lowerCamelCase : Tuple = logging.get_logger(__name__) class A_ (_UpperCamelCase ): """simple docstring""" a__ = ['''pixel_values'''] def __init__( self :List[Any] , lowerCAmelCase__ :List[str] = True , lowerCAmelCase__ :List[str] = None , lowerCAmelCase__ :Optional[Any] = None , lowerCAmelCase__ :Union[str, Any] = PILImageResampling.BILINEAR , lowerCAmelCase__ :List[str] = True , lowerCAmelCase__ :int = 1 / 255 , lowerCAmelCase__ :Optional[Any] = True , lowerCAmelCase__ :Tuple = None , lowerCAmelCase__ :Union[str, Any] = None , **lowerCAmelCase__ :List[str] , ) -> List[Any]: '''simple docstring''' super().__init__(**_UpperCAmelCase ) snake_case_ : str = size if size is not None else {'''shortest_edge''': 384} snake_case_ : str = get_size_dict(_UpperCAmelCase , default_to_square=_UpperCAmelCase ) snake_case_ : Optional[Any] = do_resize snake_case_ : Optional[int] = size # Default value set here for backwards compatibility where the value in config is None snake_case_ : str = crop_pct if crop_pct is not None else 224 / 256 snake_case_ : Any = resample snake_case_ : int = do_rescale snake_case_ : int = rescale_factor snake_case_ : int = do_normalize snake_case_ : Dict = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN snake_case_ : int = image_std if image_std is not None else IMAGENET_STANDARD_STD def _A ( self :List[Any] , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :int , lowerCAmelCase__ :Optional[int] = PILImageResampling.BICUBIC , lowerCAmelCase__ :int = None , **lowerCAmelCase__ :Optional[Any] , ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[int] = get_size_dict(_UpperCAmelCase , default_to_square=_UpperCAmelCase ) if "shortest_edge" not in size: raise ValueError(F'''Size dictionary must contain \'shortest_edge\' key. Got {size.keys()}''' ) snake_case_ : Any = size['''shortest_edge'''] if shortest_edge < 384: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct snake_case_ : Optional[int] = int(shortest_edge / crop_pct ) snake_case_ : Optional[Any] = get_resize_output_image_size(_UpperCAmelCase , size=_UpperCAmelCase , default_to_square=_UpperCAmelCase ) snake_case_ : Optional[Any] = resize(image=_UpperCAmelCase , size=_UpperCAmelCase , resample=_UpperCAmelCase , data_format=_UpperCAmelCase , **_UpperCAmelCase ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=_UpperCAmelCase , size=(shortest_edge, shortest_edge) , data_format=_UpperCAmelCase , **_UpperCAmelCase ) else: # warping (no cropping) when evaluated at 384 or larger return resize( _UpperCAmelCase , size=(shortest_edge, shortest_edge) , resample=_UpperCAmelCase , data_format=_UpperCAmelCase , **_UpperCAmelCase ) def _A ( self :Optional[Any] , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :List[Any] = None , **lowerCAmelCase__ :List[Any] , ) -> int: '''simple docstring''' return rescale(_UpperCAmelCase , scale=_UpperCAmelCase , data_format=_UpperCAmelCase , **_UpperCAmelCase ) def _A ( self :Optional[int] , lowerCAmelCase__ :int , lowerCAmelCase__ :Any , lowerCAmelCase__ :int , lowerCAmelCase__ :Dict = None , **lowerCAmelCase__ :Union[str, Any] , ) -> Tuple: '''simple docstring''' return normalize(_UpperCAmelCase , mean=_UpperCAmelCase , std=_UpperCAmelCase , data_format=_UpperCAmelCase , **_UpperCAmelCase ) def _A ( self :Any , lowerCAmelCase__ :str , lowerCAmelCase__ :Optional[Any] = None , lowerCAmelCase__ :Tuple = None , lowerCAmelCase__ :int = None , lowerCAmelCase__ :int = None , lowerCAmelCase__ :Optional[int] = None , lowerCAmelCase__ :int = None , lowerCAmelCase__ :Any = None , lowerCAmelCase__ :Dict = None , lowerCAmelCase__ :str = None , lowerCAmelCase__ :Optional[Any] = None , lowerCAmelCase__ :int = ChannelDimension.FIRST , **lowerCAmelCase__ :List[str] , ) -> Any: '''simple docstring''' snake_case_ : str = do_resize if do_resize is not None else self.do_resize snake_case_ : List[Any] = crop_pct if crop_pct is not None else self.crop_pct snake_case_ : Tuple = resample if resample is not None else self.resample snake_case_ : List[str] = do_rescale if do_rescale is not None else self.do_rescale snake_case_ : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ : Dict = do_normalize if do_normalize is not None else self.do_normalize snake_case_ : Any = image_mean if image_mean is not None else self.image_mean snake_case_ : int = image_std if image_std is not None else self.image_std snake_case_ : List[str] = size if size is not None else self.size snake_case_ : Optional[int] = get_size_dict(_UpperCAmelCase , default_to_square=_UpperCAmelCase ) snake_case_ : Optional[int] = make_list_of_images(_UpperCAmelCase ) if not valid_images(_UpperCAmelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_resize and size["shortest_edge"] < 384 and crop_pct is None: raise ValueError("crop_pct must be specified if size < 384." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. snake_case_ : Tuple = [to_numpy_array(_UpperCAmelCase ) for image in images] if do_resize: snake_case_ : Dict = [self.resize(image=_UpperCAmelCase , size=_UpperCAmelCase , crop_pct=_UpperCAmelCase , resample=_UpperCAmelCase ) for image in images] if do_rescale: snake_case_ : str = [self.rescale(image=_UpperCAmelCase , scale=_UpperCAmelCase ) for image in images] if do_normalize: snake_case_ : Dict = [self.normalize(image=_UpperCAmelCase , mean=_UpperCAmelCase , std=_UpperCAmelCase ) for image in images] snake_case_ : str = [to_channel_dimension_format(_UpperCAmelCase , _UpperCAmelCase ) for image in images] snake_case_ : Dict = {'''pixel_values''': images} return BatchFeature(data=_UpperCAmelCase , tensor_type=_UpperCAmelCase )
713
'''simple docstring''' import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py __lowerCamelCase : str = '''.''' # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) __lowerCamelCase : Tuple = [ '''Assert''', '''AssignVariableOp''', '''EmptyTensorList''', '''MergeV2Checkpoints''', '''ReadVariableOp''', '''ResourceGather''', '''RestoreV2''', '''SaveV2''', '''ShardedFilename''', '''StatefulPartitionedCall''', '''StaticRegexFullMatch''', '''VarHandleOp''', ] def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Tuple = SavedModel() snake_case_ : Dict = [] with open(os.path.join(__magic_name__ ,"utils" ,"tf_ops" ,"onnx.json" ) ) as f: snake_case_ : Dict = json.load(__magic_name__ )["opsets"] for i in range(1 ,opset + 1 ): onnx_ops.extend(onnx_opsets[str(__magic_name__ )] ) with open(__magic_name__ ,"rb" ) as f: saved_model.ParseFromString(f.read() ) snake_case_ : Tuple = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want snake_case_ : str = sorted(__magic_name__ ) snake_case_ : Optional[int] = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(__magic_name__ ) if strict and len(__magic_name__ ) > 0: raise Exception(F'''Found the following incompatible ops for the opset {opset}:\n''' + incompatible_ops ) elif len(__magic_name__ ) > 0: print(F'''Found the following incompatible ops for the opset {opset}:''' ) print(*__magic_name__ ,sep="\n" ) else: print(F'''The saved model {saved_model_path} can properly be converted with ONNX.''' ) if __name__ == "__main__": __lowerCamelCase : Any = argparse.ArgumentParser() parser.add_argument('''--saved_model_path''', help='''Path of the saved model to check (the .pb file).''') parser.add_argument( '''--opset''', default=12, type=int, help='''The ONNX opset against which the model has to be tested.''' ) parser.add_argument( '''--framework''', choices=['''onnx'''], default='''onnx''', help='''Frameworks against which to test the saved model.''' ) parser.add_argument( '''--strict''', action='''store_true''', help='''Whether make the checking strict (raise errors) or not (raise warnings)''' ) __lowerCamelCase : Dict = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
656
0
def __UpperCAmelCase ( __magic_name__ = 10 ,__magic_name__ = 1000 ,__magic_name__ = True )-> str: """simple docstring""" assert ( isinstance(UpperCAmelCase__ ,UpperCAmelCase__ ) and isinstance(UpperCAmelCase__ ,UpperCAmelCase__ ) and isinstance(UpperCAmelCase__ ,UpperCAmelCase__ ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError("Invalid value for min_val or max_val (min_value < max_value)" ) return min_val if option else max_val def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" return int((number_a + number_a) / 2 ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Optional[Any]: """simple docstring""" assert ( isinstance(UpperCAmelCase__ ,UpperCAmelCase__ ) and isinstance(UpperCAmelCase__ ,UpperCAmelCase__ ) and isinstance(UpperCAmelCase__ ,UpperCAmelCase__ ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError("argument value for lower and higher must be(lower > higher)" ) if not lower < to_guess < higher: raise ValueError( "guess value must be within the range of lower and higher value" ) def answer(__magic_name__ ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print("started..." ) snake_case_ : Any = lower snake_case_ : Optional[Any] = higher snake_case_ : str = [] while True: snake_case_ : List[str] = get_avg(UpperCAmelCase__ ,UpperCAmelCase__ ) last_numbers.append(UpperCAmelCase__ ) if answer(UpperCAmelCase__ ) == "low": snake_case_ : str = number elif answer(UpperCAmelCase__ ) == "high": snake_case_ : Optional[Any] = number else: break print(F'''guess the number : {last_numbers[-1]}''' ) print(F'''details : {last_numbers!s}''' ) def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : List[Any] = int(input("Enter lower value : " ).strip() ) snake_case_ : Dict = int(input("Enter high value : " ).strip() ) snake_case_ : Optional[Any] = int(input("Enter value to guess : " ).strip() ) guess_the_number(UpperCAmelCase__ ,UpperCAmelCase__ ,UpperCAmelCase__ ) if __name__ == "__main__": main()
714
'''simple docstring''' import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal __lowerCamelCase : Optional[Any] = datasets.utils.logging.get_logger(__name__) __lowerCamelCase : List[str] = ['''names''', '''prefix'''] __lowerCamelCase : int = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] __lowerCamelCase : str = ['''encoding_errors''', '''on_bad_lines'''] __lowerCamelCase : Optional[Any] = ['''date_format'''] @dataclass class A_ (datasets.BuilderConfig ): """simple docstring""" a__ = "," a__ = None a__ = "infer" a__ = None a__ = None a__ = None a__ = None a__ = None a__ = True a__ = None a__ = None a__ = None a__ = None a__ = False a__ = None a__ = None a__ = None a__ = True a__ = True a__ = False a__ = True a__ = None a__ = "." a__ = None a__ = '"' a__ = 0 a__ = None a__ = None a__ = None a__ = None a__ = True a__ = True a__ = 0 a__ = True a__ = False a__ = None a__ = 10000 a__ = None a__ = "strict" a__ = "error" a__ = None def _A ( self :List[str] ) -> Any: '''simple docstring''' if self.delimiter is not None: snake_case_ : Tuple = self.delimiter if self.column_names is not None: snake_case_ : List[Any] = self.column_names @property def _A ( self :Optional[Any] ) -> int: '''simple docstring''' snake_case_ : Optional[int] = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class A_ (datasets.ArrowBasedBuilder ): """simple docstring""" a__ = CsvConfig def _A ( self :Optional[Any] ) -> Optional[Any]: '''simple docstring''' return datasets.DatasetInfo(features=self.config.features ) def _A ( self :Tuple , lowerCAmelCase__ :Dict ) -> List[Any]: '''simple docstring''' if not self.config.data_files: raise ValueError(F'''At least one data file must be specified, but got data_files={self.config.data_files}''' ) snake_case_ : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): snake_case_ : int = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ : List[str] = [files] snake_case_ : Tuple = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] snake_case_ : str = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ : str = [files] snake_case_ : Any = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def _A ( self :List[Any] , lowerCAmelCase__ :pa.Table ) -> pa.Table: '''simple docstring''' if self.config.features is not None: snake_case_ : int = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast snake_case_ : Optional[Any] = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example snake_case_ : Dict = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def _A ( self :Dict , lowerCAmelCase__ :Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str snake_case_ : str = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): snake_case_ : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): snake_case_ : Optional[int] = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(F'''Failed to read file \'{file}\' with error {type(lowerCAmelCase__ )}: {e}''' ) raise
656
0
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.generation import DisjunctiveConstraint @require_torch class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Optional[Any] ) -> str: '''simple docstring''' snake_case_ : Tuple = [[1, 2, 4], [1, 2, 3, 4]] snake_case_ : Union[str, Any] = DisjunctiveConstraint(__UpperCamelCase ) self.assertTrue(isinstance(dc.token_ids , __UpperCamelCase ) ) with self.assertRaises(__UpperCamelCase ): DisjunctiveConstraint(torch.LongTensor([[1, 2, 4], [1, 2, 3]] ) ) with self.assertRaises(__UpperCamelCase ): DisjunctiveConstraint([torch.LongTensor([1, 2, 4] ), torch.LongTensor([1, 2, 3, 4, 5] )] ) def _A ( self :List[Any] ) -> List[str]: '''simple docstring''' snake_case_ : Tuple = [[1, 2], [1, 2, 3, 4]] with self.assertRaises(__UpperCamelCase ): DisjunctiveConstraint(__UpperCamelCase ) # fails here def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : List[str] = [[1, 2, 3], [1, 2, 4]] snake_case_ : Dict = DisjunctiveConstraint(__UpperCamelCase ) snake_case_, snake_case_, snake_case_ : Tuple = dc.update(1 ) snake_case_ : List[Any] = stepped is True and completed is False and reset is False self.assertTrue(__UpperCamelCase ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) snake_case_, snake_case_, snake_case_ : Optional[Any] = dc.update(2 ) snake_case_ : str = stepped is True and completed is False and reset is False self.assertTrue(__UpperCamelCase ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) snake_case_, snake_case_, snake_case_ : Tuple = dc.update(3 ) snake_case_ : Tuple = stepped is True and completed is True and reset is False self.assertTrue(__UpperCamelCase ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 3] ) def _A ( self :Tuple ) -> int: '''simple docstring''' snake_case_ : List[Any] = [[1, 2, 3], [1, 2, 4, 5], [1, 2, 5]] snake_case_ : Union[str, Any] = DisjunctiveConstraint(__UpperCamelCase ) snake_case_, snake_case_, snake_case_ : int = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) snake_case_, snake_case_, snake_case_ : List[str] = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) snake_case_, snake_case_, snake_case_ : Union[str, Any] = dc.update(4 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2, 4] ) snake_case_, snake_case_, snake_case_ : str = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 4, 5] ) dc.reset() snake_case_, snake_case_, snake_case_ : str = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 3 ) self.assertTrue(dc.current_seq == [1] ) snake_case_, snake_case_, snake_case_ : str = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 2 ) self.assertTrue(dc.current_seq == [1, 2] ) snake_case_, snake_case_, snake_case_ : str = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.remaining() == 0 ) self.assertTrue(dc.current_seq == [1, 2, 5] )
715
'''simple docstring''' import json import os import unittest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A_ (a_ , unittest.TestCase ): """simple docstring""" a__ = MgpstrTokenizer a__ = False a__ = {} a__ = False def _A ( self :List[str] ) -> List[str]: '''simple docstring''' super().setUp() # fmt: off snake_case_ : Dict = ["[GO]", "[s]", "0", "1", "2", "3", "4", "5", "6", "7", "8", "9", "a", "b", "c", "d", "e", "f", "g", "h", "i", "j", "k", "l", "m", "n", "o", "p", "q", "r", "s", "t", "u", "v", "w", "x", "y", "z"] # fmt: on snake_case_ : List[str] = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) snake_case_ : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase__ ) + "\n" ) def _A ( self :Optional[Any] , **lowerCAmelCase__ :Optional[Any] ) -> Dict: '''simple docstring''' return MgpstrTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def _A ( self :Dict , lowerCAmelCase__ :Any ) -> str: '''simple docstring''' snake_case_ : Dict = "tester" snake_case_ : Tuple = "tester" return input_text, output_text @unittest.skip("MGP-STR always lower cases letters." ) def _A ( self :Dict ) -> str: '''simple docstring''' pass def _A ( self :Tuple ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[str] = self.get_tokenizers(do_lower_case=lowerCAmelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): snake_case_ : Tuple = "[SPECIAL_TOKEN]" tokenizer.add_special_tokens({"cls_token": special_token} ) snake_case_ : str = tokenizer.encode([special_token] , add_special_tokens=lowerCAmelCase__ ) self.assertEqual(len(lowerCAmelCase__ ) , 1 ) snake_case_ : Tuple = tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) self.assertTrue(special_token not in decoded ) def _A ( self :int ) -> List[str]: '''simple docstring''' snake_case_ : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): snake_case_, snake_case_ : str = self.get_input_output_texts(lowerCAmelCase__ ) snake_case_ : Union[str, Any] = tokenizer.tokenize(lowerCAmelCase__ ) snake_case_ : List[Any] = tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) snake_case_ : Dict = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : List[str] = tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) self.assertNotEqual(len(lowerCAmelCase__ ) , 0 ) snake_case_ : List[str] = tokenizer.decode(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertEqual(text_a.replace(" " , "" ) , lowerCAmelCase__ ) @unittest.skip("MGP-STR tokenizer only handles one sequence." ) def _A ( self :Union[str, Any] ) -> Any: '''simple docstring''' pass @unittest.skip("inputs cannot be pretokenized in MgpstrTokenizer" ) def _A ( self :int ) -> Dict: '''simple docstring''' pass
656
0
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ )-> Dict: """simple docstring""" snake_case_, snake_case_ : Optional[Any] = [], [] while len(_UpperCamelCase ) > 1: snake_case_, snake_case_ : str = min(_UpperCamelCase ), max(_UpperCamelCase ) start.append(_UpperCamelCase ) end.append(_UpperCamelCase ) collection.remove(_UpperCamelCase ) collection.remove(_UpperCamelCase ) end.reverse() return start + collection + end if __name__ == "__main__": __lowerCamelCase : str = input('''Enter numbers separated by a comma:\n''').strip() __lowerCamelCase : Tuple = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
716
'''simple docstring''' from __future__ import annotations import math import numpy as np from numpy.linalg import norm def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> float: """simple docstring""" return math.sqrt(sum(pow(a - b ,2 ) for a, b in zip(__magic_name__ ,__magic_name__ ) ) ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> list[list[list[float] | float]]: """simple docstring""" if dataset.ndim != value_array.ndim: snake_case_ : int = ( "Wrong input data's dimensions... " F'''dataset : {dataset.ndim}, value_array : {value_array.ndim}''' ) raise ValueError(__magic_name__ ) try: if dataset.shape[1] != value_array.shape[1]: snake_case_ : Dict = ( "Wrong input data's shape... " F'''dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}''' ) raise ValueError(__magic_name__ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError("Wrong shape" ) if dataset.dtype != value_array.dtype: snake_case_ : Dict = ( "Input data have different datatype... " F'''dataset : {dataset.dtype}, value_array : {value_array.dtype}''' ) raise TypeError(__magic_name__ ) snake_case_ : Optional[int] = [] for value in value_array: snake_case_ : List[str] = euclidean(__magic_name__ ,dataset[0] ) snake_case_ : int = dataset[0].tolist() for dataset_value in dataset[1:]: snake_case_ : Optional[Any] = euclidean(__magic_name__ ,__magic_name__ ) if dist > temp_dist: snake_case_ : Tuple = temp_dist snake_case_ : Optional[int] = dataset_value.tolist() answer.append([vector, dist] ) return answer def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> float: """simple docstring""" return np.dot(__magic_name__ ,__magic_name__ ) / (norm(__magic_name__ ) * norm(__magic_name__ )) if __name__ == "__main__": import doctest doctest.testmod()
656
0
'''simple docstring''' import os import sys import unittest __lowerCamelCase : Any = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, '''utils''')) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path __lowerCamelCase : int = os.path.join(git_repo_path, '''src''', '''transformers''') __lowerCamelCase : str = """ {0} = None """ __lowerCamelCase : Union[str, Any] = """ class {0}(metaclass=DummyObject): _backends = {1} def __init__(self, *args, **kwargs): requires_backends(self, {1}) """ __lowerCamelCase : Optional[Any] = """ def {0}(*args, **kwargs): requires_backends({0}, {1}) """ class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Any ) -> Optional[int]: '''simple docstring''' snake_case_ : Any = find_backend(" _import_structure[\"models.albert\"].append(\"AlbertTokenizerFast\")" ) self.assertIsNone(lowerCAmelCase__ ) snake_case_ : Dict = find_backend(" if not is_tokenizers_available():" ) self.assertEqual(lowerCAmelCase__ , "tokenizers" ) snake_case_ : int = find_backend(" if not is_tensorflow_text_available():" ) self.assertEqual(lowerCAmelCase__ , "tensorflow_text" ) snake_case_ : int = find_backend(" if not (is_sentencepiece_available() and is_tokenizers_available()):" ) self.assertEqual(lowerCAmelCase__ , "sentencepiece_and_tokenizers" ) snake_case_ : Optional[Any] = find_backend( " if not (is_sentencepiece_available() and is_tensorflow_text_available()):" ) self.assertEqual(lowerCAmelCase__ , "sentencepiece_and_tensorflow_text" ) snake_case_ : int = find_backend( " if not (is_sentencepiece_available() and is_tokenizers_available() and is_vision_available()):" ) self.assertEqual(lowerCAmelCase__ , "sentencepiece_and_tokenizers_and_vision" ) def _A ( self :List[Any] ) -> List[str]: '''simple docstring''' snake_case_ : Dict = read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn("torch" , lowerCAmelCase__ ) self.assertIn("tensorflow_text" , lowerCAmelCase__ ) self.assertIn("sentencepiece_and_tokenizers" , lowerCAmelCase__ ) # Likewise, we can't assert on the exact content of a key self.assertIn("BertModel" , objects["torch"] ) self.assertIn("TFBertModel" , objects["tf"] ) self.assertIn("FlaxBertModel" , objects["flax"] ) self.assertIn("BertModel" , objects["torch"] ) self.assertIn("TFBertTokenizer" , objects["tensorflow_text"] ) self.assertIn("convert_slow_tokenizer" , objects["sentencepiece_and_tokenizers"] ) def _A ( self :int ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[str] = create_dummy_object("CONSTANT" , "\'torch\'" ) self.assertEqual(lowerCAmelCase__ , "\nCONSTANT = None\n" ) snake_case_ : List[str] = create_dummy_object("function" , "\'torch\'" ) self.assertEqual( lowerCAmelCase__ , "\ndef function(*args, **kwargs):\n requires_backends(function, \'torch\')\n" ) snake_case_ : Dict = "\nclass FakeClass(metaclass=DummyObject):\n _backends = \'torch\'\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, \'torch\')\n" snake_case_ : Any = create_dummy_object("FakeClass" , "\'torch\'" ) self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ ) def _A ( self :Any ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Any = "# This file is autogenerated by the command `make fix-copies`, do not edit.\nfrom ..utils import DummyObject, requires_backends\n\n\nCONSTANT = None\n\n\ndef function(*args, **kwargs):\n requires_backends(function, [\"torch\"])\n\n\nclass FakeClass(metaclass=DummyObject):\n _backends = [\"torch\"]\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, [\"torch\"])\n" snake_case_ : List[str] = create_dummy_files({"torch": ["CONSTANT", "function", "FakeClass"]} ) self.assertEqual(dummy_files["torch"] , lowerCAmelCase__ )
717
'''simple docstring''' import fire from utils import calculate_rouge, save_json def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__=None ,**__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : int = [x.strip() for x in open(__magic_name__ ).readlines()] snake_case_ : Optional[int] = [x.strip() for x in open(__magic_name__ ).readlines()][: len(__magic_name__ )] snake_case_ : List[Any] = calculate_rouge(__magic_name__ ,__magic_name__ ,**__magic_name__ ) if save_path is not None: save_json(__magic_name__ ,__magic_name__ ,indent=__magic_name__ ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
656
0
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __lowerCamelCase : Tuple = logging.get_logger(__name__) __lowerCamelCase : Union[str, Any] = {"vocab_file": "sentencepiece.bpe.model"} __lowerCamelCase : List[Any] = { "vocab_file": { "moussaKam/mbarthez": "https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez": "https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez-orangesum-title": ( "https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model" ), }, } __lowerCamelCase : int = { "moussaKam/mbarthez": 1024, "moussaKam/barthez": 1024, "moussaKam/barthez-orangesum-title": 1024, } __lowerCamelCase : int = "▁" class A_ (__lowerCAmelCase ): """simple docstring""" a__ = VOCAB_FILES_NAMES a__ = PRETRAINED_VOCAB_FILES_MAP a__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ = ["input_ids", "attention_mask"] def __init__( self :Union[str, Any] , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Tuple="<s>" , lowerCAmelCase__ :Tuple="</s>" , lowerCAmelCase__ :int="</s>" , lowerCAmelCase__ :Optional[Any]="<s>" , lowerCAmelCase__ :List[str]="<unk>" , lowerCAmelCase__ :str="<pad>" , lowerCAmelCase__ :int="<mask>" , lowerCAmelCase__ :Optional[Dict[str, Any]] = None , **lowerCAmelCase__ :Any , ) -> None: '''simple docstring''' snake_case_ : str = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else mask_token snake_case_ : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase__ , ) snake_case_ : Any = vocab_file snake_case_ : str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowerCamelCase__ ) ) snake_case_ : Any = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} snake_case_ : Tuple = len(self.sp_model ) - 1 snake_case_ : str = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def _A ( self :List[str] , lowerCAmelCase__ :List[int] , lowerCAmelCase__ :Optional[List[int]] = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] snake_case_ : List[str] = [self.cls_token_id] snake_case_ : Optional[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _A ( self :Dict , lowerCAmelCase__ :List[int] , lowerCAmelCase__ :Optional[List[int]] = None , lowerCAmelCase__ :bool = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase__ )) + [1] return [1] + ([0] * len(lowerCamelCase__ )) + [1, 1] + ([0] * len(lowerCamelCase__ )) + [1] def _A ( self :List[str] , lowerCAmelCase__ :List[int] , lowerCAmelCase__ :Optional[List[int]] = None ) -> List[int]: '''simple docstring''' snake_case_ : int = [self.sep_token_id] snake_case_ : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def _A ( self :Union[str, Any] ) -> int: '''simple docstring''' return len(self.sp_model ) def _A ( self :List[str] ) -> Optional[int]: '''simple docstring''' snake_case_ : Tuple = {self.convert_ids_to_tokens(lowerCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _A ( self :int , lowerCAmelCase__ :str ) -> List[str]: '''simple docstring''' return self.sp_model.encode(lowerCamelCase__ , out_type=lowerCamelCase__ ) def _A ( self :Union[str, Any] , lowerCAmelCase__ :List[str] ) -> Union[str, Any]: '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] snake_case_ : List[str] = self.sp_model.PieceToId(lowerCamelCase__ ) return spm_id if spm_id else self.unk_token_id def _A ( self :Dict , lowerCAmelCase__ :int ) -> List[str]: '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(lowerCamelCase__ ) def _A ( self :str , lowerCAmelCase__ :Tuple ) -> List[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = [] snake_case_ : Any = '''''' snake_case_ : str = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowerCamelCase__ ) + token snake_case_ : str = True snake_case_ : Tuple = [] else: current_sub_tokens.append(lowerCamelCase__ ) snake_case_ : Any = False out_string += self.sp_model.decode(lowerCamelCase__ ) return out_string.strip() def __getstate__( self :Tuple ) -> Dict: '''simple docstring''' snake_case_ : str = self.__dict__.copy() snake_case_ : int = None return state def __setstate__( self :Tuple , lowerCAmelCase__ :Any ) -> str: '''simple docstring''' snake_case_ : List[Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): snake_case_ : Optional[Any] = {} snake_case_ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _A ( self :Optional[int] , lowerCAmelCase__ :str , lowerCAmelCase__ :Optional[str] = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(lowerCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return snake_case_ : Any = os.path.join( lowerCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase__ , "wb" ) as fi: snake_case_ : int = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase__ ) return (out_vocab_file,)
718
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) __lowerCamelCase : Optional[Any] = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', f'''encoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', f'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.weight''', f'''encoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.bias''', f'''encoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.weight''', f'''encoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.bias''', f'''encoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.encoder.layers.{i}.norm1.weight''', f'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.norm1.bias''', f'''encoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.weight''', f'''encoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.bias''', f'''encoder.layers.{i}.final_layer_norm.bias''')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', f'''decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', f'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.cross_attn.out_proj.weight''', f'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.cross_attn.out_proj.bias''', f'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.weight''', f'''decoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.bias''', f'''decoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.weight''', f'''decoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.bias''', f'''decoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm1.weight''', f'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm1.bias''', f'''decoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.weight''', f'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.bias''', f'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.weight''', f'''decoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.bias''', f'''decoder.layers.{i}.final_layer_norm.bias''')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qcontent_proj.weight''', f'''decoder.layers.{i}.sa_qcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kcontent_proj.weight''', f'''decoder.layers.{i}.sa_kcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qpos_proj.weight''', f'''decoder.layers.{i}.sa_qpos_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kpos_proj.weight''', f'''decoder.layers.{i}.sa_kpos_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_v_proj.weight''', f'''decoder.layers.{i}.sa_v_proj.weight''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qcontent_proj.weight''', f'''decoder.layers.{i}.ca_qcontent_proj.weight''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kcontent_proj.weight''', f'''decoder.layers.{i}.ca_kcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kpos_proj.weight''', f'''decoder.layers.{i}.ca_kpos_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_v_proj.weight''', f'''decoder.layers.{i}.ca_v_proj.weight''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight''', f'''decoder.layers.{i}.ca_qpos_sine_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qcontent_proj.bias''', f'''decoder.layers.{i}.sa_qcontent_proj.bias''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kcontent_proj.bias''', f'''decoder.layers.{i}.sa_kcontent_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_qpos_proj.bias''', f'''decoder.layers.{i}.sa_qpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_kpos_proj.bias''', f'''decoder.layers.{i}.sa_kpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_v_proj.bias''', f'''decoder.layers.{i}.sa_v_proj.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qcontent_proj.bias''', f'''decoder.layers.{i}.ca_qcontent_proj.bias''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kcontent_proj.bias''', f'''decoder.layers.{i}.ca_kcontent_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_kpos_proj.bias''', f'''decoder.layers.{i}.ca_kpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_v_proj.bias''', f'''decoder.layers.{i}.ca_v_proj.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias''', f'''decoder.layers.{i}.ca_qpos_sine_proj.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> int: """simple docstring""" snake_case_ : Optional[Any] = state_dict.pop(__magic_name__ ) snake_case_ : Any = val def __UpperCAmelCase ( __magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : Any = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: snake_case_ : Optional[Any] = key.replace("backbone.0.body" ,"backbone.conv_encoder.model" ) snake_case_ : int = value else: snake_case_ : int = value return new_state_dict def __UpperCAmelCase ( __magic_name__ ,__magic_name__=False )-> Optional[int]: """simple docstring""" snake_case_ : str = "" if is_panoptic: snake_case_ : Dict = "conditional_detr." # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) snake_case_ : Any = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) snake_case_ : Optional[int] = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case_ : Tuple = in_proj_weight[:256, :] snake_case_ : List[Any] = in_proj_bias[:256] snake_case_ : Optional[Any] = in_proj_weight[256:512, :] snake_case_ : Optional[int] = in_proj_bias[256:512] snake_case_ : Optional[int] = in_proj_weight[-256:, :] snake_case_ : str = in_proj_bias[-256:] def __UpperCAmelCase ( )-> Optional[Any]: """simple docstring""" snake_case_ : Optional[int] = "http://images.cocodataset.org/val2017/000000039769.jpg" snake_case_ : Optional[Any] = Image.open(requests.get(__magic_name__ ,stream=__magic_name__ ).raw ) return im @torch.no_grad() def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Optional[Any] = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: snake_case_ : Optional[Any] = "resnet101" if "dc5" in model_name: snake_case_ : List[str] = True snake_case_ : Tuple = "panoptic" in model_name if is_panoptic: snake_case_ : List[Any] = 250 else: snake_case_ : Optional[Any] = 91 snake_case_ : Optional[int] = "huggingface/label-files" snake_case_ : Dict = "coco-detection-id2label.json" snake_case_ : List[Any] = json.load(open(hf_hub_download(__magic_name__ ,__magic_name__ ,repo_type="dataset" ) ,"r" ) ) snake_case_ : Optional[int] = {int(__magic_name__ ): v for k, v in idalabel.items()} snake_case_ : int = idalabel snake_case_ : Dict = {v: k for k, v in idalabel.items()} # load image processor snake_case_ : Optional[int] = "coco_panoptic" if is_panoptic else "coco_detection" snake_case_ : str = ConditionalDetrImageProcessor(format=__magic_name__ ) # prepare image snake_case_ : str = prepare_img() snake_case_ : int = image_processor(images=__magic_name__ ,return_tensors="pt" ) snake_case_ : Union[str, Any] = encoding["pixel_values"] logger.info(F'''Converting model {model_name}...''' ) # load original model from torch hub snake_case_ : Union[str, Any] = torch.hub.load("DeppMeng/ConditionalDETR" ,__magic_name__ ,pretrained=__magic_name__ ).eval() snake_case_ : Any = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: snake_case_ : Any = "conditional_detr." + src rename_key(__magic_name__ ,__magic_name__ ,__magic_name__ ) snake_case_ : Tuple = rename_backbone_keys(__magic_name__ ) # query, key and value matrices need special treatment read_in_q_k_v(__magic_name__ ,is_panoptic=__magic_name__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them snake_case_ : int = "conditional_detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("conditional_detr" ) and not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ) ): snake_case_ : Any = state_dict.pop(__magic_name__ ) snake_case_ : Optional[int] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: snake_case_ : Tuple = state_dict.pop(__magic_name__ ) snake_case_ : Any = val elif key.startswith("bbox_attention" ) or key.startswith("mask_head" ): continue else: snake_case_ : Union[str, Any] = state_dict.pop(__magic_name__ ) snake_case_ : List[Any] = val else: if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): snake_case_ : Any = state_dict.pop(__magic_name__ ) snake_case_ : List[Any] = val # finally, create HuggingFace model and load state dict snake_case_ : Optional[int] = ConditionalDetrForSegmentation(__magic_name__ ) if is_panoptic else ConditionalDetrForObjectDetection(__magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() model.push_to_hub(repo_id=__magic_name__ ,organization="DepuMeng" ,commit_message="Add model" ) # verify our conversion snake_case_ : Dict = conditional_detr(__magic_name__ ) snake_case_ : Union[str, Any] = model(__magic_name__ ) assert torch.allclose(outputs.logits ,original_outputs["pred_logits"] ,atol=1E-4 ) assert torch.allclose(outputs.pred_boxes ,original_outputs["pred_boxes"] ,atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks ,original_outputs["pred_masks"] ,atol=1E-4 ) # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) image_processor.save_pretrained(__magic_name__ ) if __name__ == "__main__": __lowerCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) __lowerCamelCase : int = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
656
0
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class A_ : """simple docstring""" def __init__( self :Tuple , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Any=2 , lowerCAmelCase__ :Optional[int]=True , lowerCAmelCase__ :Union[str, Any]=False , lowerCAmelCase__ :Optional[int]=10 , lowerCAmelCase__ :Tuple=3 , lowerCAmelCase__ :int=32 * 4 , lowerCAmelCase__ :Optional[Any]=32 * 6 , lowerCAmelCase__ :Tuple=4 , lowerCAmelCase__ :Union[str, Any]=32 , ) -> Any: '''simple docstring''' snake_case_ : List[str] = parent snake_case_ : str = batch_size snake_case_ : List[str] = is_training snake_case_ : Tuple = use_auxiliary_loss snake_case_ : Optional[Any] = num_queries snake_case_ : Optional[int] = num_channels snake_case_ : Union[str, Any] = min_size snake_case_ : Any = max_size snake_case_ : Any = num_labels snake_case_ : Union[str, Any] = mask_feature_size def _A ( self :Tuple ) -> Dict: '''simple docstring''' snake_case_ : Any = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( lowerCAmelCase__ ) snake_case_ : Dict = torch.ones([self.batch_size, self.min_size, self.max_size] , device=lowerCAmelCase__ ) snake_case_ : int = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=lowerCAmelCase__ ) > 0.5 ).float() snake_case_ : List[Any] = (torch.rand((self.batch_size, self.num_labels) , device=lowerCAmelCase__ ) > 0.5).long() snake_case_ : Optional[int] = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def _A ( self :Tuple ) -> Any: '''simple docstring''' return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=128 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def _A ( self :Optional[int] ) -> List[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = self.prepare_config_and_inputs() snake_case_ : str = {'''pixel_values''': pixel_values, '''pixel_mask''': pixel_mask} return config, inputs_dict def _A ( self :Union[str, Any] , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :List[Any] ) -> List[str]: '''simple docstring''' snake_case_ : Tuple = output.encoder_hidden_states snake_case_ : str = output.pixel_decoder_hidden_states snake_case_ : Tuple = output.transformer_decoder_hidden_states self.parent.assertTrue(len(lowerCAmelCase__ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(lowerCAmelCase__ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(lowerCAmelCase__ ) , config.decoder_config.decoder_layers ) def _A ( self :Any , lowerCAmelCase__ :int , lowerCAmelCase__ :str , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Optional[int]=False ) -> Optional[Any]: '''simple docstring''' with torch.no_grad(): snake_case_ : str = MaskFormerModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case_ : Dict = model(pixel_values=lowerCAmelCase__ , pixel_mask=lowerCAmelCase__ ) snake_case_ : int = model(lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(lowerCAmelCase__ , lowerCAmelCase__ ) def _A ( self :str , lowerCAmelCase__ :Dict , lowerCAmelCase__ :int , lowerCAmelCase__ :Any , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Tuple ) -> List[Any]: '''simple docstring''' snake_case_ : List[Any] = MaskFormerForInstanceSegmentation(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() def comm_check_on_output(lowerCAmelCase__ :int ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): snake_case_ : str = model(pixel_values=lowerCAmelCase__ , pixel_mask=lowerCAmelCase__ ) snake_case_ : Union[str, Any] = model(lowerCAmelCase__ ) comm_check_on_output(lowerCAmelCase__ ) snake_case_ : List[str] = model( pixel_values=lowerCAmelCase__ , pixel_mask=lowerCAmelCase__ , mask_labels=lowerCAmelCase__ , class_labels=lowerCAmelCase__ ) comm_check_on_output(lowerCAmelCase__ ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class A_ (UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase ): """simple docstring""" a__ = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () a__ = ( {'''feature-extraction''': MaskFormerModel, '''image-segmentation''': MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) a__ = False a__ = False a__ = False a__ = False def _A ( self :List[str] ) -> List[str]: '''simple docstring''' snake_case_ : Tuple = MaskFormerModelTester(self ) snake_case_ : Dict = ConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> List[Any]: '''simple docstring''' self.config_tester.run_common_tests() def _A ( self :Optional[Any] ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(lowerCAmelCase__ , **lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ ) def _A ( self :str ) -> List[Any]: '''simple docstring''' snake_case_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*lowerCAmelCase__ ) @unittest.skip(reason="MaskFormer does not use inputs_embeds" ) def _A ( self :Optional[int] ) -> Optional[int]: '''simple docstring''' pass @unittest.skip(reason="MaskFormer does not have a get_input_embeddings method" ) def _A ( self :Optional[Any] ) -> Union[str, Any]: '''simple docstring''' pass @unittest.skip(reason="MaskFormer is not a generative model" ) def _A ( self :Union[str, Any] ) -> int: '''simple docstring''' pass @unittest.skip(reason="MaskFormer does not use token embeddings" ) def _A ( self :Optional[int] ) -> Optional[Any]: '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason="MaskFormer has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`" ) def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' pass def _A ( self :Dict ) -> Optional[Any]: '''simple docstring''' snake_case_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ : List[Any] = model_class(lowerCAmelCase__ ) snake_case_ : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ : List[str] = [*signature.parameters.keys()] snake_case_ : int = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) @slow def _A ( self :str ) -> List[Any]: '''simple docstring''' for model_name in ["facebook/maskformer-swin-small-coco"]: snake_case_ : Any = MaskFormerModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def _A ( self :Optional[int] ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = (self.model_tester.min_size,) * 2 snake_case_ : int = { '''pixel_values''': torch.randn((2, 3, *size) , device=lowerCAmelCase__ ), '''mask_labels''': torch.randn((2, 10, *size) , device=lowerCAmelCase__ ), '''class_labels''': torch.zeros(2 , 10 , device=lowerCAmelCase__ ).long(), } snake_case_ : Any = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(lowerCAmelCase__ ) snake_case_ : str = model(**lowerCAmelCase__ ) self.assertTrue(outputs.loss is not None ) def _A ( self :Optional[int] ) -> Optional[int]: '''simple docstring''' snake_case_ : int = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(lowerCAmelCase__ , **lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ ) def _A ( self :Tuple ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ : str = model_class(lowerCAmelCase__ ).to(lowerCAmelCase__ ) snake_case_ : Dict = model(**lowerCAmelCase__ , output_attentions=lowerCAmelCase__ ) self.assertTrue(outputs.attentions is not None ) def _A ( self :Tuple ) -> List[str]: '''simple docstring''' if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss snake_case_ : Union[str, Any] = self.all_model_classes[1] snake_case_ : Dict = self.model_tester.prepare_config_and_inputs() snake_case_ : List[Any] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.train() snake_case_ : Union[str, Any] = model(lowerCAmelCase__ , mask_labels=lowerCAmelCase__ , class_labels=lowerCAmelCase__ ).loss loss.backward() def _A ( self :Tuple ) -> Optional[int]: '''simple docstring''' snake_case_ : Optional[int] = self.all_model_classes[1] snake_case_ : Dict = self.model_tester.prepare_config_and_inputs() snake_case_ : Dict = True snake_case_ : Optional[int] = True snake_case_ : Union[str, Any] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.train() snake_case_ : int = model(lowerCAmelCase__ , mask_labels=lowerCAmelCase__ , class_labels=lowerCAmelCase__ ) snake_case_ : Optional[int] = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() snake_case_ : List[str] = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't snake_case_ : List[str] = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() snake_case_ : Any = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=lowerCAmelCase__ ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __lowerCamelCase : int = 1E-4 def __UpperCAmelCase ( )-> List[Any]: """simple docstring""" snake_case_ : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_vision @slow class A_ (unittest.TestCase ): """simple docstring""" @cached_property def _A ( self :Union[str, Any] ) -> List[Any]: '''simple docstring''' return ( MaskFormerImageProcessor.from_pretrained("facebook/maskformer-swin-small-coco" ) if is_vision_available() else None ) def _A ( self :List[str] ) -> Tuple: '''simple docstring''' snake_case_ : str = MaskFormerModel.from_pretrained("facebook/maskformer-swin-small-coco" ).to(lowerCAmelCase__ ) snake_case_ : Any = self.default_image_processor snake_case_ : Optional[Any] = prepare_img() snake_case_ : Dict = image_processor(lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) snake_case_ : str = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(lowerCAmelCase__ , (1, 3, 800, 1_088) ) with torch.no_grad(): snake_case_ : List[Any] = model(**lowerCAmelCase__ ) snake_case_ : List[Any] = torch.tensor( [[-0.0_4_8_2, 0.9_2_2_8, 0.4_9_5_1], [-0.2_5_4_7, 0.8_0_1_7, 0.8_5_2_7], [-0.0_0_6_9, 0.3_3_8_5, -0.0_0_8_9]] ).to(lowerCAmelCase__ ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) snake_case_ : Any = torch.tensor( [[-0.8_4_2_2, -0.8_4_3_4, -0.9_7_1_8], [-1.0_1_4_4, -0.5_5_6_5, -0.4_1_9_5], [-1.0_0_3_8, -0.4_4_8_4, -0.1_9_6_1]] ).to(lowerCAmelCase__ ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) snake_case_ : Optional[Any] = torch.tensor( [[0.2_8_5_2, -0.0_1_5_9, 0.9_7_3_5], [0.6_2_5_4, 0.1_8_5_8, 0.8_5_2_9], [-0.0_6_8_0, -0.4_1_1_6, 1.8_4_1_3]] ).to(lowerCAmelCase__ ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) def _A ( self :Optional[Any] ) -> List[Any]: '''simple docstring''' snake_case_ : List[Any] = ( MaskFormerForInstanceSegmentation.from_pretrained("facebook/maskformer-swin-small-coco" ) .to(lowerCAmelCase__ ) .eval() ) snake_case_ : List[str] = self.default_image_processor snake_case_ : str = prepare_img() snake_case_ : Any = image_processor(lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) snake_case_ : List[str] = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(lowerCAmelCase__ , (1, 3, 800, 1_088) ) with torch.no_grad(): snake_case_ : Tuple = model(**lowerCAmelCase__ ) # masks_queries_logits snake_case_ : Tuple = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) snake_case_ : Any = [ [-1.3_7_3_7_1_2_4, -1.7_7_2_4_9_3_7, -1.9_3_6_4_2_3_3], [-1.5_9_7_7_2_8_1, -1.9_8_6_7_9_3_9, -2.1_5_2_3_6_9_5], [-1.5_7_9_5_3_9_8, -1.9_2_6_9_8_3_2, -2.0_9_3_9_4_2], ] snake_case_ : Any = torch.tensor(lowerCAmelCase__ ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) # class_queries_logits snake_case_ : str = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) snake_case_ : List[str] = torch.tensor( [ [1.6_5_1_2E0_0, -5.2_5_7_2E0_0, -3.3_5_1_9E0_0], [3.6_1_6_9E-0_2, -5.9_0_2_5E0_0, -2.9_3_1_3E0_0], [1.0_7_6_6E-0_4, -7.7_6_3_0E0_0, -5.1_2_6_3E0_0], ] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) def _A ( self :Any ) -> Optional[Any]: '''simple docstring''' snake_case_ : Dict = ( MaskFormerForInstanceSegmentation.from_pretrained("facebook/maskformer-resnet101-coco-stuff" ) .to(lowerCAmelCase__ ) .eval() ) snake_case_ : int = self.default_image_processor snake_case_ : Optional[int] = prepare_img() snake_case_ : str = image_processor(lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) snake_case_ : List[Any] = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(lowerCAmelCase__ , (1, 3, 800, 1_088) ) with torch.no_grad(): snake_case_ : int = model(**lowerCAmelCase__ ) # masks_queries_logits snake_case_ : int = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) snake_case_ : str = [[-0.9_0_4_6, -2.6_3_6_6, -4.6_0_6_2], [-3.4_1_7_9, -5.7_8_9_0, -8.8_0_5_7], [-4.9_1_7_9, -7.6_5_6_0, -1_0.7_7_1_1]] snake_case_ : Optional[int] = torch.tensor(lowerCAmelCase__ ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) # class_queries_logits snake_case_ : Union[str, Any] = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) snake_case_ : List[str] = torch.tensor( [[4.7_1_8_8, -3.2_5_8_5, -2.8_8_5_7], [6.6_8_7_1, -2.9_1_8_1, -1.2_4_8_7], [7.2_4_4_9, -2.2_7_6_4, -2.1_8_7_4]] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) def _A ( self :Optional[int] ) -> str: '''simple docstring''' snake_case_ : List[str] = ( MaskFormerForInstanceSegmentation.from_pretrained("facebook/maskformer-swin-small-coco" ) .to(lowerCAmelCase__ ) .eval() ) snake_case_ : Any = self.default_image_processor snake_case_ : List[Any] = image_processor( [np.zeros((3, 800, 1_333) ), np.zeros((3, 800, 1_333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="pt" , ) snake_case_ : Dict = inputs['''pixel_values'''].to(lowerCAmelCase__ ) snake_case_ : int = [el.to(lowerCAmelCase__ ) for el in inputs['''mask_labels''']] snake_case_ : int = [el.to(lowerCAmelCase__ ) for el in inputs['''class_labels''']] with torch.no_grad(): snake_case_ : List[str] = model(**lowerCAmelCase__ ) self.assertTrue(outputs.loss is not None )
719
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Any ) -> str: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _A ( self :List[Any] ) -> List[str]: '''simple docstring''' snake_case_ : Any = 1 snake_case_ : Dict = 3 snake_case_ : Union[str, Any] = (32, 32) snake_case_ : Optional[int] = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) return image @property def _A ( self :Optional[int] ) -> Any: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : List[str] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) return model @property def _A ( self :Dict ) -> Any: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) return model @property def _A ( self :Dict ) -> Optional[int]: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : str = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5_006 , ) return RobertaSeriesModelWithTransformation(lowerCAmelCase__ ) @property def _A ( self :Any ) -> str: '''simple docstring''' def extract(*lowerCAmelCase__ :Any , **lowerCAmelCase__ :List[str] ): class A_ : """simple docstring""" def __init__( self :Optional[int] ) -> List[str]: '''simple docstring''' snake_case_ : str = torch.ones([0] ) def _A ( self :int , lowerCAmelCase__ :List[Any] ) -> Tuple: '''simple docstring''' self.pixel_values.to(lowerCAmelCase__ ) return self return Out() return extract def _A ( self :int ) -> Dict: '''simple docstring''' snake_case_ : str = "cpu" # ensure determinism for the device-dependent torch.Generator snake_case_ : str = self.dummy_cond_unet snake_case_ : Optional[int] = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) snake_case_ : Dict = self.dummy_vae snake_case_ : Dict = self.dummy_text_encoder snake_case_ : Optional[int] = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) snake_case_ : str = 77 snake_case_ : Any = self.dummy_image.to(lowerCAmelCase__ ) snake_case_ : Tuple = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk snake_case_ : Optional[Any] = AltDiffusionImgaImgPipeline( unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , vae=lowerCAmelCase__ , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , feature_extractor=self.dummy_extractor , ) snake_case_ : Union[str, Any] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowerCAmelCase__ ) snake_case_ : Optional[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) snake_case_ : Dict = "A painting of a squirrel eating a burger" snake_case_ : List[str] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) snake_case_ : Dict = alt_pipe( [prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , ) snake_case_ : Any = output.images snake_case_ : List[str] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) snake_case_ : Optional[Any] = alt_pipe( [prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , return_dict=lowerCAmelCase__ , )[0] snake_case_ : Tuple = image[0, -3:, -3:, -1] snake_case_ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) snake_case_ : int = np.array([0.4_4_2_7, 0.3_7_3_1, 0.4_2_4_9, 0.4_9_4_1, 0.4_5_4_6, 0.4_1_4_8, 0.4_1_9_3, 0.4_6_6_6, 0.4_4_9_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5E-3 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def _A ( self :int ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = self.dummy_cond_unet snake_case_ : Union[str, Any] = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) snake_case_ : int = self.dummy_vae snake_case_ : List[Any] = self.dummy_text_encoder snake_case_ : int = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) snake_case_ : int = 77 snake_case_ : Dict = self.dummy_image.to(lowerCAmelCase__ ) # put models in fp16 snake_case_ : Optional[Any] = unet.half() snake_case_ : Tuple = vae.half() snake_case_ : List[str] = bert.half() # make sure here that pndm scheduler skips prk snake_case_ : Optional[int] = AltDiffusionImgaImgPipeline( unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , vae=lowerCAmelCase__ , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , feature_extractor=self.dummy_extractor , ) snake_case_ : List[str] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowerCAmelCase__ ) snake_case_ : Optional[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) snake_case_ : List[Any] = "A painting of a squirrel eating a burger" snake_case_ : str = torch.manual_seed(0 ) snake_case_ : Any = alt_pipe( [prompt] , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def _A ( self :Optional[int] ) -> Any: '''simple docstring''' snake_case_ : Union[str, Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) # resize to resolution that is divisible by 8 but not 16 or 32 snake_case_ : str = init_image.resize((760, 504) ) snake_case_ : Optional[Any] = "BAAI/AltDiffusion" snake_case_ : int = AltDiffusionImgaImgPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() snake_case_ : Tuple = "A fantasy landscape, trending on artstation" snake_case_ : int = torch.manual_seed(0 ) snake_case_ : List[str] = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , strength=0.7_5 , guidance_scale=7.5 , generator=lowerCAmelCase__ , output_type="np" , ) snake_case_ : str = output.images[0] snake_case_ : List[Any] = image[255:258, 383:386, -1] assert image.shape == (504, 760, 3) snake_case_ : Tuple = np.array([0.9_3_5_8, 0.9_3_9_7, 0.9_5_9_9, 0.9_9_0_1, 1.0_0_0_0, 1.0_0_0_0, 0.9_8_8_2, 1.0_0_0_0, 1.0_0_0_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _A ( self :str ) -> Any: '''simple docstring''' snake_case_ : Optional[Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) snake_case_ : List[Any] = init_image.resize((768, 512) ) snake_case_ : Tuple = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy" ) snake_case_ : Any = "BAAI/AltDiffusion" snake_case_ : List[str] = AltDiffusionImgaImgPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() snake_case_ : Tuple = "A fantasy landscape, trending on artstation" snake_case_ : Tuple = torch.manual_seed(0 ) snake_case_ : List[Any] = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , strength=0.7_5 , guidance_scale=7.5 , generator=lowerCAmelCase__ , output_type="np" , ) snake_case_ : Optional[int] = output.images[0] assert image.shape == (512, 768, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1E-2
656
0
'''simple docstring''' import sys __lowerCamelCase : Union[str, Any] = ( """73167176531330624919225119674426574742355349194934""" """96983520312774506326239578318016984801869478851843""" """85861560789112949495459501737958331952853208805511""" """12540698747158523863050715693290963295227443043557""" """66896648950445244523161731856403098711121722383113""" """62229893423380308135336276614282806444486645238749""" """30358907296290491560440772390713810515859307960866""" """70172427121883998797908792274921901699720888093776""" """65727333001053367881220235421809751254540594752243""" """52584907711670556013604839586446706324415722155397""" """53697817977846174064955149290862569321978468622482""" """83972241375657056057490261407972968652414535100474""" """82166370484403199890008895243450658541227588666881""" """16427171479924442928230863465674813919123162824586""" """17866458359124566529476545682848912883142607690042""" """24219022671055626321111109370544217506941658960408""" """07198403850962455444362981230987879927244284909188""" """84580156166097919133875499200524063689912560717606""" """05886116467109405077541002256983155200055935729725""" """71636269561882670428252483600823257530420752963450""" ) def __UpperCAmelCase ( __magic_name__ = N )-> Optional[int]: """simple docstring""" snake_case_ : int = -sys.maxsize - 1 for i in range(len(UpperCamelCase__ ) - 12 ): snake_case_ : Tuple = 1 for j in range(13 ): product *= int(n[i + j] ) if product > largest_product: snake_case_ : str = product return largest_product if __name__ == "__main__": print(f'''{solution() = }''')
720
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. __lowerCamelCase : List[str] = {'''LayoutLMv2Config''', '''LayoutLMv3Config'''} @is_pipeline_test class A_ (unittest.TestCase ): """simple docstring""" a__ = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING a__ = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: a__ = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: a__ = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def _A ( self :Tuple , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :Dict ) -> Any: '''simple docstring''' snake_case_ : Optional[Any] = ZeroShotClassificationPipeline( model=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , candidate_labels=["polics", "health"] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def _A ( self :List[str] , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Any ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = classifier("Who are you voting for in 2020?" , candidate_labels="politics" ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) # No kwarg snake_case_ : List[Any] = classifier("Who are you voting for in 2020?" , ["politics"] ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) snake_case_ : Dict = classifier("Who are you voting for in 2020?" , candidate_labels=["politics"] ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) snake_case_ : int = classifier("Who are you voting for in 2020?" , candidate_labels="politics, public health" ) self.assertEqual( lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ) , 1.0 ) snake_case_ : Optional[int] = classifier("Who are you voting for in 2020?" , candidate_labels=["politics", "public health"] ) self.assertEqual( lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ) , 1.0 ) snake_case_ : str = classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template="This text is about {}" ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) # https://github.com/huggingface/transformers/issues/13846 snake_case_ : Dict = classifier(["I am happy"] , ["positive", "negative"] ) self.assertEqual( lowerCAmelCase__ , [ {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} for i in range(1 ) ] , ) snake_case_ : Tuple = classifier(["I am happy", "I am sad"] , ["positive", "negative"] ) self.assertEqual( lowerCAmelCase__ , [ {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} for i in range(2 ) ] , ) with self.assertRaises(lowerCAmelCase__ ): classifier("" , candidate_labels="politics" ) with self.assertRaises(lowerCAmelCase__ ): classifier(lowerCAmelCase__ , candidate_labels="politics" ) with self.assertRaises(lowerCAmelCase__ ): classifier("Who are you voting for in 2020?" , candidate_labels="" ) with self.assertRaises(lowerCAmelCase__ ): classifier("Who are you voting for in 2020?" , candidate_labels=lowerCAmelCase__ ) with self.assertRaises(lowerCAmelCase__ ): classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template="Not formatting template" , ) with self.assertRaises(lowerCAmelCase__ ): classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template=lowerCAmelCase__ , ) self.run_entailment_id(lowerCAmelCase__ ) def _A ( self :List[Any] , lowerCAmelCase__ :Pipeline ) -> Union[str, Any]: '''simple docstring''' snake_case_ : int = zero_shot_classifier.model.config snake_case_ : Optional[int] = config.labelaid snake_case_ : Tuple = zero_shot_classifier.entailment_id snake_case_ : Optional[Any] = {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2} self.assertEqual(zero_shot_classifier.entailment_id , -1 ) snake_case_ : Tuple = {"entailment": 0, "neutral": 1, "contradiction": 2} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) snake_case_ : str = {"ENTAIL": 0, "NON-ENTAIL": 1} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) snake_case_ : str = {"ENTAIL": 2, "NEUTRAL": 1, "CONTR": 0} self.assertEqual(zero_shot_classifier.entailment_id , 2 ) snake_case_ : List[str] = original_labelaid self.assertEqual(lowerCAmelCase__ , zero_shot_classifier.entailment_id ) @require_torch def _A ( self :Tuple ) -> Any: '''simple docstring''' snake_case_ : List[Any] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="pt" , ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( "Who are you voting for in 2020?" * 100 , candidate_labels=["politics", "public health", "science"] ) @require_torch def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="pt" , ) snake_case_ : int = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.3_3_3, 0.3_3_3, 0.3_3_3], } , ) @require_tf def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : List[str] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="tf" , ) snake_case_ : Optional[int] = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.3_3_3, 0.3_3_3, 0.3_3_3], } , ) @slow @require_torch def _A ( self :Union[str, Any] ) -> int: '''simple docstring''' snake_case_ : int = pipeline("zero-shot-classification" , model="roberta-large-mnli" , framework="pt" ) snake_case_ : str = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.9_7_6, 0.0_1_5, 0.0_0_9], } , ) snake_case_ : Optional[int] = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data." , candidate_labels=["machine learning", "statistics", "translation", "vision"] , multi_label=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], } , ) @slow @require_tf def _A ( self :List[str] ) -> str: '''simple docstring''' snake_case_ : int = pipeline("zero-shot-classification" , model="roberta-large-mnli" , framework="tf" ) snake_case_ : Optional[Any] = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.9_7_6, 0.0_1_5, 0.0_0_9], } , ) snake_case_ : Tuple = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data." , candidate_labels=["machine learning", "statistics", "translation", "vision"] , multi_label=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], } , )
656
0
'''simple docstring''' import os import tempfile import unittest import uuid from pathlib import Path from transformers.testing_utils import get_tests_dir, require_soundfile, require_torch, require_vision from transformers.tools.agent_types import AgentAudio, AgentImage, AgentText from transformers.utils import is_soundfile_availble, is_torch_available, is_vision_available if is_torch_available(): import torch if is_soundfile_availble(): import soundfile as sf if is_vision_available(): from PIL import Image def __UpperCAmelCase ( __magic_name__="" )-> str: """simple docstring""" snake_case_ : Optional[Any] = tempfile.mkdtemp() return os.path.join(lowerCamelCase__ ,str(uuid.uuida() ) + suffix ) @require_soundfile @require_torch class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Any ) -> List[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = torch.rand(12 , dtype=torch.floataa ) - 0.5 snake_case_ : Optional[int] = AgentAudio(lowerCAmelCase__ ) snake_case_ : int = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowerCAmelCase__ , agent_type.to_raw() , atol=1E-4 ) ) del agent_type # Ensure the path remains even after the object deletion self.assertTrue(os.path.exists(lowerCAmelCase__ ) ) # Ensure that the file contains the same value as the original tensor snake_case_, snake_case_ : str = sf.read(lowerCAmelCase__ ) self.assertTrue(torch.allclose(lowerCAmelCase__ , torch.tensor(lowerCAmelCase__ ) , atol=1E-4 ) ) def _A ( self :Optional[Any] ) -> Dict: '''simple docstring''' snake_case_ : Optional[Any] = torch.rand(12 , dtype=torch.floataa ) - 0.5 snake_case_ : int = get_new_path(suffix=".wav" ) sf.write(lowerCAmelCase__ , lowerCAmelCase__ , 16_000 ) snake_case_ : List[str] = AgentAudio(lowerCAmelCase__ ) self.assertTrue(torch.allclose(lowerCAmelCase__ , agent_type.to_raw() , atol=1E-4 ) ) self.assertEqual(agent_type.to_string() , lowerCAmelCase__ ) @require_vision @require_torch class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Any ) -> Any: '''simple docstring''' snake_case_ : str = torch.randint(0 , 256 , (64, 64, 3) ) snake_case_ : Optional[Any] = AgentImage(lowerCAmelCase__ ) snake_case_ : Optional[int] = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowerCAmelCase__ , agent_type._tensor , atol=1E-4 ) ) self.assertIsInstance(agent_type.to_raw() , Image.Image ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowerCAmelCase__ ) ) def _A ( self :Optional[int] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[int] = Path(get_tests_dir("fixtures/tests_samples/COCO" ) ) / "000000039769.png" snake_case_ : Dict = Image.open(lowerCAmelCase__ ) snake_case_ : Dict = AgentImage(lowerCAmelCase__ ) self.assertTrue(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowerCAmelCase__ ) ) def _A ( self :Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Tuple = Path(get_tests_dir("fixtures/tests_samples/COCO" ) ) / "000000039769.png" snake_case_ : Optional[Any] = Image.open(lowerCAmelCase__ ) snake_case_ : Optional[Any] = AgentImage(lowerCAmelCase__ ) self.assertFalse(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowerCAmelCase__ ) ) class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Optional[Any] ) -> Optional[Any]: '''simple docstring''' snake_case_ : int = "Hey!" snake_case_ : Optional[int] = AgentText(lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , agent_type.to_string() ) self.assertEqual(lowerCAmelCase__ , agent_type.to_raw() ) self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ )
721
'''simple docstring''' import argparse import pathlib import fairseq import torch from fairseq.models.roberta import RobertaModel as FairseqRobertaModel from fairseq.modules import TransformerSentenceEncoderLayer from packaging import version from transformers import XLMRobertaConfig, XLMRobertaXLForMaskedLM, XLMRobertaXLForSequenceClassification from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.models.roberta.modeling_roberta import RobertaAttention from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse('''1.0.0a'''): raise Exception('''requires fairseq >= 1.0.0a''') logging.set_verbosity_info() __lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) __lowerCamelCase : Union[str, Any] = '''Hello world! cécé herlolip''' def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : str = FairseqRobertaModel.from_pretrained(__magic_name__ ) roberta.eval() # disable dropout snake_case_ : Dict = roberta.model.encoder.sentence_encoder snake_case_ : List[str] = XLMRobertaConfig( vocab_size=roberta_sent_encoder.embed_tokens.num_embeddings ,hidden_size=roberta.cfg.model.encoder_embed_dim ,num_hidden_layers=roberta.cfg.model.encoder_layers ,num_attention_heads=roberta.cfg.model.encoder_attention_heads ,intermediate_size=roberta.cfg.model.encoder_ffn_embed_dim ,max_position_embeddings=514 ,type_vocab_size=1 ,layer_norm_eps=1E-5 ,) if classification_head: snake_case_ : List[str] = roberta.model.classification_heads["mnli"].out_proj.weight.shape[0] print("Our RoBERTa config:" ,__magic_name__ ) snake_case_ : List[str] = XLMRobertaXLForSequenceClassification(__magic_name__ ) if classification_head else XLMRobertaXLForMaskedLM(__magic_name__ ) model.eval() # Now let's copy all the weights. # Embeddings snake_case_ : List[Any] = roberta_sent_encoder.embed_tokens.weight snake_case_ : int = roberta_sent_encoder.embed_positions.weight snake_case_ : Union[str, Any] = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c RoBERTa doesn't use them. snake_case_ : Union[str, Any] = roberta_sent_encoder.layer_norm.weight snake_case_ : str = roberta_sent_encoder.layer_norm.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer snake_case_ : BertLayer = model.roberta.encoder.layer[i] snake_case_ : TransformerSentenceEncoderLayer = roberta_sent_encoder.layers[i] snake_case_ : RobertaAttention = layer.attention snake_case_ : Dict = roberta_layer.self_attn_layer_norm.weight snake_case_ : Dict = roberta_layer.self_attn_layer_norm.bias # self attention snake_case_ : BertSelfAttention = layer.attention.self assert ( roberta_layer.self_attn.k_proj.weight.data.shape == roberta_layer.self_attn.q_proj.weight.data.shape == roberta_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ) snake_case_ : Dict = roberta_layer.self_attn.q_proj.weight snake_case_ : Any = roberta_layer.self_attn.q_proj.bias snake_case_ : Optional[Any] = roberta_layer.self_attn.k_proj.weight snake_case_ : Optional[Any] = roberta_layer.self_attn.k_proj.bias snake_case_ : Optional[int] = roberta_layer.self_attn.v_proj.weight snake_case_ : Any = roberta_layer.self_attn.v_proj.bias # self-attention output snake_case_ : BertSelfOutput = layer.attention.output assert self_output.dense.weight.shape == roberta_layer.self_attn.out_proj.weight.shape snake_case_ : List[str] = roberta_layer.self_attn.out_proj.weight snake_case_ : Optional[int] = roberta_layer.self_attn.out_proj.bias # this one is final layer norm snake_case_ : int = roberta_layer.final_layer_norm.weight snake_case_ : Union[str, Any] = roberta_layer.final_layer_norm.bias # intermediate snake_case_ : BertIntermediate = layer.intermediate assert intermediate.dense.weight.shape == roberta_layer.fca.weight.shape snake_case_ : List[str] = roberta_layer.fca.weight snake_case_ : List[Any] = roberta_layer.fca.bias # output snake_case_ : BertOutput = layer.output assert bert_output.dense.weight.shape == roberta_layer.fca.weight.shape snake_case_ : Any = roberta_layer.fca.weight snake_case_ : Any = roberta_layer.fca.bias # end of layer if classification_head: snake_case_ : int = roberta.model.classification_heads["mnli"].dense.weight snake_case_ : Union[str, Any] = roberta.model.classification_heads["mnli"].dense.bias snake_case_ : Tuple = roberta.model.classification_heads["mnli"].out_proj.weight snake_case_ : str = roberta.model.classification_heads["mnli"].out_proj.bias else: # LM Head snake_case_ : Optional[Any] = roberta.model.encoder.lm_head.dense.weight snake_case_ : int = roberta.model.encoder.lm_head.dense.bias snake_case_ : Optional[Any] = roberta.model.encoder.lm_head.layer_norm.weight snake_case_ : Optional[int] = roberta.model.encoder.lm_head.layer_norm.bias snake_case_ : int = roberta.model.encoder.lm_head.weight snake_case_ : List[str] = roberta.model.encoder.lm_head.bias # Let's check that we get the same results. snake_case_ : torch.Tensor = roberta.encode(__magic_name__ ).unsqueeze(0 ) # batch of size 1 snake_case_ : Union[str, Any] = model(__magic_name__ )[0] if classification_head: snake_case_ : Optional[Any] = roberta.model.classification_heads["mnli"](roberta.extract_features(__magic_name__ ) ) else: snake_case_ : List[str] = roberta.model(__magic_name__ )[0] print(our_output.shape ,their_output.shape ) snake_case_ : str = torch.max(torch.abs(our_output - their_output ) ).item() print(F'''max_absolute_diff = {max_absolute_diff}''' ) # ~ 1e-7 snake_case_ : Any = torch.allclose(__magic_name__ ,__magic_name__ ,atol=1E-3 ) print("Do both models output the same tensors?" ,"🔥" if success else "💩" ) if not success: raise Exception("Something went wRoNg" ) pathlib.Path(__magic_name__ ).mkdir(parents=__magic_name__ ,exist_ok=__magic_name__ ) print(F'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(__magic_name__ ) if __name__ == "__main__": __lowerCamelCase : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--roberta_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--classification_head''', action='''store_true''', help='''Whether to convert a final classification head.''' ) __lowerCamelCase : Tuple = parser.parse_args() convert_xlm_roberta_xl_checkpoint_to_pytorch( args.roberta_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
656
0
'''simple docstring''' import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py __lowerCamelCase : str = '''.''' # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) __lowerCamelCase : Tuple = [ '''Assert''', '''AssignVariableOp''', '''EmptyTensorList''', '''MergeV2Checkpoints''', '''ReadVariableOp''', '''ResourceGather''', '''RestoreV2''', '''SaveV2''', '''ShardedFilename''', '''StatefulPartitionedCall''', '''StaticRegexFullMatch''', '''VarHandleOp''', ] def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Tuple = SavedModel() snake_case_ : Dict = [] with open(os.path.join(__magic_name__ ,"utils" ,"tf_ops" ,"onnx.json" ) ) as f: snake_case_ : Dict = json.load(__magic_name__ )["opsets"] for i in range(1 ,opset + 1 ): onnx_ops.extend(onnx_opsets[str(__magic_name__ )] ) with open(__magic_name__ ,"rb" ) as f: saved_model.ParseFromString(f.read() ) snake_case_ : Tuple = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want snake_case_ : str = sorted(__magic_name__ ) snake_case_ : Optional[int] = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(__magic_name__ ) if strict and len(__magic_name__ ) > 0: raise Exception(F'''Found the following incompatible ops for the opset {opset}:\n''' + incompatible_ops ) elif len(__magic_name__ ) > 0: print(F'''Found the following incompatible ops for the opset {opset}:''' ) print(*__magic_name__ ,sep="\n" ) else: print(F'''The saved model {saved_model_path} can properly be converted with ONNX.''' ) if __name__ == "__main__": __lowerCamelCase : Any = argparse.ArgumentParser() parser.add_argument('''--saved_model_path''', help='''Path of the saved model to check (the .pb file).''') parser.add_argument( '''--opset''', default=12, type=int, help='''The ONNX opset against which the model has to be tested.''' ) parser.add_argument( '''--framework''', choices=['''onnx'''], default='''onnx''', help='''Frameworks against which to test the saved model.''' ) parser.add_argument( '''--strict''', action='''store_true''', help='''Whether make the checking strict (raise errors) or not (raise warnings)''' ) __lowerCamelCase : Dict = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
700
'''simple docstring''' import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def __UpperCAmelCase ( __magic_name__ ,__magic_name__=() ,__magic_name__=None ,__magic_name__="no" ,__magic_name__="29500" )-> Optional[int]: """simple docstring""" snake_case_ : str = False snake_case_ : int = False if any(key.startswith("KAGGLE" ) for key in os.environ.keys() ): snake_case_ : Any = True elif "IPython" in sys.modules: snake_case_ : Union[str, Any] = "google.colab" in str(sys.modules["IPython"].get_ipython() ) try: snake_case_ : Any = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F'''Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.''' ) if (in_colab or in_kaggle) and (os.environ.get("TPU_NAME" ,__magic_name__ ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( "To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside " "your training function. Restart your notebook and make sure no cells initializes an " "`Accelerator`." ) if num_processes is None: snake_case_ : Tuple = 8 snake_case_ : Optional[int] = PrepareForLaunch(__magic_name__ ,distributed_type="TPU" ) print(F'''Launching a training on {num_processes} TPU cores.''' ) xmp.spawn(__magic_name__ ,args=__magic_name__ ,nprocs=__magic_name__ ,start_method="fork" ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print("Launching training on one GPU." ) else: print("Launching training on one CPU." ) function(*__magic_name__ ) else: if num_processes is None: raise ValueError( "You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call." ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( "To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized " "inside your training function. Restart your notebook and make sure no cells initializes an " "`Accelerator`." ) if torch.cuda.is_initialized(): raise ValueError( "To launch a multi-GPU training from your notebook, you need to avoid running any instruction " "using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA " "function." ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=__magic_name__ ,master_addr="127.0.01" ,master_port=__magic_name__ ,mixed_precision=__magic_name__ ): snake_case_ : Optional[int] = PrepareForLaunch(__magic_name__ ,distributed_type="MULTI_GPU" ) print(F'''Launching training on {num_processes} GPUs.''' ) try: start_processes(__magic_name__ ,args=__magic_name__ ,nprocs=__magic_name__ ,start_method="fork" ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( "CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. " "This likely stems from an outside import causing issues once the `notebook_launcher()` is called. " "Please review your imports and test them when running the `notebook_launcher()` to identify " "which one is problematic." ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): snake_case_ : Any = "1" print("Launching training on MPS." ) elif torch.cuda.is_available(): print("Launching training on one GPU." ) else: print("Launching training on CPU." ) function(*__magic_name__ ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__=() ,__magic_name__=2 )-> Dict: """simple docstring""" from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=__magic_name__ ,master_addr="127.0.01" ,master_port="29500" ,accelerate_mixed_precision="no" ,accelerate_debug_rdv_file=tmp_file.name ,accelerate_use_cpu="yes" ,): snake_case_ : Any = PrepareForLaunch(__magic_name__ ,debug=__magic_name__ ) start_processes(__magic_name__ ,args=__magic_name__ ,nprocs=__magic_name__ ,start_method="fork" )
656
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bert import BertTokenizer __lowerCamelCase : int = logging.get_logger(__name__) __lowerCamelCase : Optional[int] = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} __lowerCamelCase : int = { '''vocab_file''': { '''bert-base-uncased''': '''https://huggingface.co/bert-base-uncased/resolve/main/vocab.txt''', '''bert-large-uncased''': '''https://huggingface.co/bert-large-uncased/resolve/main/vocab.txt''', '''bert-base-cased''': '''https://huggingface.co/bert-base-cased/resolve/main/vocab.txt''', '''bert-large-cased''': '''https://huggingface.co/bert-large-cased/resolve/main/vocab.txt''', '''bert-base-multilingual-uncased''': ( '''https://huggingface.co/bert-base-multilingual-uncased/resolve/main/vocab.txt''' ), '''bert-base-multilingual-cased''': '''https://huggingface.co/bert-base-multilingual-cased/resolve/main/vocab.txt''', '''bert-base-chinese''': '''https://huggingface.co/bert-base-chinese/resolve/main/vocab.txt''', '''bert-base-german-cased''': '''https://huggingface.co/bert-base-german-cased/resolve/main/vocab.txt''', '''bert-large-uncased-whole-word-masking''': ( '''https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/vocab.txt''' ), '''bert-large-cased-whole-word-masking''': ( '''https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/vocab.txt''' ), '''bert-large-uncased-whole-word-masking-finetuned-squad''': ( '''https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt''' ), '''bert-large-cased-whole-word-masking-finetuned-squad''': ( '''https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt''' ), '''bert-base-cased-finetuned-mrpc''': ( '''https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/vocab.txt''' ), '''bert-base-german-dbmdz-cased''': '''https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/vocab.txt''', '''bert-base-german-dbmdz-uncased''': ( '''https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/vocab.txt''' ), '''TurkuNLP/bert-base-finnish-cased-v1''': ( '''https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/vocab.txt''' ), '''TurkuNLP/bert-base-finnish-uncased-v1''': ( '''https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/vocab.txt''' ), '''wietsedv/bert-base-dutch-cased''': ( '''https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''bert-base-uncased''': '''https://huggingface.co/bert-base-uncased/resolve/main/tokenizer.json''', '''bert-large-uncased''': '''https://huggingface.co/bert-large-uncased/resolve/main/tokenizer.json''', '''bert-base-cased''': '''https://huggingface.co/bert-base-cased/resolve/main/tokenizer.json''', '''bert-large-cased''': '''https://huggingface.co/bert-large-cased/resolve/main/tokenizer.json''', '''bert-base-multilingual-uncased''': ( '''https://huggingface.co/bert-base-multilingual-uncased/resolve/main/tokenizer.json''' ), '''bert-base-multilingual-cased''': ( '''https://huggingface.co/bert-base-multilingual-cased/resolve/main/tokenizer.json''' ), '''bert-base-chinese''': '''https://huggingface.co/bert-base-chinese/resolve/main/tokenizer.json''', '''bert-base-german-cased''': '''https://huggingface.co/bert-base-german-cased/resolve/main/tokenizer.json''', '''bert-large-uncased-whole-word-masking''': ( '''https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/tokenizer.json''' ), '''bert-large-cased-whole-word-masking''': ( '''https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/tokenizer.json''' ), '''bert-large-uncased-whole-word-masking-finetuned-squad''': ( '''https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json''' ), '''bert-large-cased-whole-word-masking-finetuned-squad''': ( '''https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json''' ), '''bert-base-cased-finetuned-mrpc''': ( '''https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/tokenizer.json''' ), '''bert-base-german-dbmdz-cased''': ( '''https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/tokenizer.json''' ), '''bert-base-german-dbmdz-uncased''': ( '''https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/tokenizer.json''' ), '''TurkuNLP/bert-base-finnish-cased-v1''': ( '''https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/tokenizer.json''' ), '''TurkuNLP/bert-base-finnish-uncased-v1''': ( '''https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/tokenizer.json''' ), '''wietsedv/bert-base-dutch-cased''': ( '''https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/tokenizer.json''' ), }, } __lowerCamelCase : List[str] = { '''bert-base-uncased''': 512, '''bert-large-uncased''': 512, '''bert-base-cased''': 512, '''bert-large-cased''': 512, '''bert-base-multilingual-uncased''': 512, '''bert-base-multilingual-cased''': 512, '''bert-base-chinese''': 512, '''bert-base-german-cased''': 512, '''bert-large-uncased-whole-word-masking''': 512, '''bert-large-cased-whole-word-masking''': 512, '''bert-large-uncased-whole-word-masking-finetuned-squad''': 512, '''bert-large-cased-whole-word-masking-finetuned-squad''': 512, '''bert-base-cased-finetuned-mrpc''': 512, '''bert-base-german-dbmdz-cased''': 512, '''bert-base-german-dbmdz-uncased''': 512, '''TurkuNLP/bert-base-finnish-cased-v1''': 512, '''TurkuNLP/bert-base-finnish-uncased-v1''': 512, '''wietsedv/bert-base-dutch-cased''': 512, } __lowerCamelCase : Optional[int] = { '''bert-base-uncased''': {'''do_lower_case''': True}, '''bert-large-uncased''': {'''do_lower_case''': True}, '''bert-base-cased''': {'''do_lower_case''': False}, '''bert-large-cased''': {'''do_lower_case''': False}, '''bert-base-multilingual-uncased''': {'''do_lower_case''': True}, '''bert-base-multilingual-cased''': {'''do_lower_case''': False}, '''bert-base-chinese''': {'''do_lower_case''': False}, '''bert-base-german-cased''': {'''do_lower_case''': False}, '''bert-large-uncased-whole-word-masking''': {'''do_lower_case''': True}, '''bert-large-cased-whole-word-masking''': {'''do_lower_case''': False}, '''bert-large-uncased-whole-word-masking-finetuned-squad''': {'''do_lower_case''': True}, '''bert-large-cased-whole-word-masking-finetuned-squad''': {'''do_lower_case''': False}, '''bert-base-cased-finetuned-mrpc''': {'''do_lower_case''': False}, '''bert-base-german-dbmdz-cased''': {'''do_lower_case''': False}, '''bert-base-german-dbmdz-uncased''': {'''do_lower_case''': True}, '''TurkuNLP/bert-base-finnish-cased-v1''': {'''do_lower_case''': False}, '''TurkuNLP/bert-base-finnish-uncased-v1''': {'''do_lower_case''': True}, '''wietsedv/bert-base-dutch-cased''': {'''do_lower_case''': False}, } class A_ (a_ ): """simple docstring""" a__ = VOCAB_FILES_NAMES a__ = PRETRAINED_VOCAB_FILES_MAP a__ = PRETRAINED_INIT_CONFIGURATION a__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ = BertTokenizer def __init__( self :int , lowerCAmelCase__ :List[str]=None , lowerCAmelCase__ :Dict=None , lowerCAmelCase__ :List[Any]=True , lowerCAmelCase__ :List[str]="[UNK]" , lowerCAmelCase__ :List[str]="[SEP]" , lowerCAmelCase__ :int="[PAD]" , lowerCAmelCase__ :Optional[int]="[CLS]" , lowerCAmelCase__ :List[str]="[MASK]" , lowerCAmelCase__ :Dict=True , lowerCAmelCase__ :Any=None , **lowerCAmelCase__ :List[Any] , ) -> List[Any]: '''simple docstring''' super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , tokenize_chinese_chars=lowerCAmelCase__ , strip_accents=lowerCAmelCase__ , **lowerCAmelCase__ , ) snake_case_ : Any = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCAmelCase__ ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase__ ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase__ ) != tokenize_chinese_chars ): snake_case_ : Dict = getattr(lowerCAmelCase__ , normalizer_state.pop("type" ) ) snake_case_ : int = do_lower_case snake_case_ : Tuple = strip_accents snake_case_ : int = tokenize_chinese_chars snake_case_ : Dict = normalizer_class(**lowerCAmelCase__ ) snake_case_ : Dict = do_lower_case def _A ( self :List[Any] , lowerCAmelCase__ :Any , lowerCAmelCase__ :List[str]=None ) -> List[Any]: '''simple docstring''' snake_case_ : Any = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _A ( self :List[Any] , lowerCAmelCase__ :List[int] , lowerCAmelCase__ :Optional[List[int]] = None ) -> List[int]: '''simple docstring''' snake_case_ : Optional[Any] = [self.sep_token_id] snake_case_ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _A ( self :Tuple , lowerCAmelCase__ :str , lowerCAmelCase__ :Optional[str] = None ) -> Tuple[str]: '''simple docstring''' snake_case_ : Optional[Any] = self._tokenizer.model.save(lowerCAmelCase__ , name=lowerCAmelCase__ ) return tuple(lowerCAmelCase__ )
701
'''simple docstring''' from collections import deque from math import floor from random import random from time import time class A_ : """simple docstring""" def __init__( self :Dict ) -> List[str]: '''simple docstring''' snake_case_ : int = {} def _A ( self :Any , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Optional[Any]=1 ) -> Any: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: snake_case_ : Optional[int] = [[w, v]] if not self.graph.get(lowerCAmelCase__ ): snake_case_ : Dict = [] def _A ( self :List[Any] ) -> Optional[int]: '''simple docstring''' return list(self.graph ) def _A ( self :str , lowerCAmelCase__ :Any , lowerCAmelCase__ :int ) -> List[Any]: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCAmelCase__ ) def _A ( self :List[str] , lowerCAmelCase__ :Optional[Any]=-2 , lowerCAmelCase__ :str=-1 ) -> str: '''simple docstring''' if s == d: return [] snake_case_ : str = [] snake_case_ : Optional[int] = [] if s == -2: snake_case_ : List[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Dict = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : str = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCAmelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCAmelCase__ ) != 0: snake_case_ : Union[str, Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return visited def _A ( self :Tuple , lowerCAmelCase__ :int=-1 ) -> int: '''simple docstring''' if c == -1: snake_case_ : Any = floor(random() * 10_000 ) + 10 for i in range(lowerCAmelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): snake_case_ : Optional[Any] = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCAmelCase__ , lowerCAmelCase__ , 1 ) def _A ( self :Tuple , lowerCAmelCase__ :Dict=-2 ) -> Dict: '''simple docstring''' snake_case_ : Union[str, Any] = deque() snake_case_ : Optional[Any] = [] if s == -2: snake_case_ : Tuple = list(self.graph )[0] d.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) while d: snake_case_ : Optional[int] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def _A ( self :List[str] , lowerCAmelCase__ :str ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = 0 for x in self.graph: for y in self.graph[x]: if y[1] == u: count += 1 return count def _A ( self :Any , lowerCAmelCase__ :int ) -> Optional[Any]: '''simple docstring''' return len(self.graph[u] ) def _A ( self :Tuple , lowerCAmelCase__ :List[str]=-2 ) -> Optional[Any]: '''simple docstring''' snake_case_ : str = [] snake_case_ : str = [] if s == -2: snake_case_ : Optional[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : int = s snake_case_ : Optional[int] = [] while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : List[str] = node[1] break # check if all the children are visited if s == ss: sorted_nodes.append(stack.pop() ) if len(lowerCAmelCase__ ) != 0: snake_case_ : int = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Union[str, Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return sorted_nodes def _A ( self :Dict ) -> Any: '''simple docstring''' snake_case_ : Dict = [] snake_case_ : Any = [] snake_case_ : str = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Optional[int] = -2 snake_case_ : Any = [] snake_case_ : List[Any] = s snake_case_ : int = False snake_case_ : Optional[int] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Any = len(lowerCAmelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Optional[int] = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[Any] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : str = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : List[str] = s snake_case_ : Optional[int] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return list(lowerCAmelCase__ ) def _A ( self :Tuple ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = [] snake_case_ : Tuple = [] snake_case_ : List[str] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : str = -2 snake_case_ : List[str] = [] snake_case_ : List[Any] = s snake_case_ : List[str] = False snake_case_ : Dict = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Any = len(lowerCAmelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Tuple = True if len(lowerCAmelCase__ ) != 0: snake_case_ : List[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : int = s snake_case_ : Union[str, Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return False def _A ( self :Optional[int] , lowerCAmelCase__ :Optional[int]=-2 , lowerCAmelCase__ :Tuple=-1 ) -> str: '''simple docstring''' snake_case_ : Optional[int] = time() self.dfs(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : Optional[Any] = time() return end - begin def _A ( self :Any , lowerCAmelCase__ :Tuple=-2 ) -> Optional[Any]: '''simple docstring''' snake_case_ : Any = time() self.bfs(lowerCAmelCase__ ) snake_case_ : Any = time() return end - begin class A_ : """simple docstring""" def __init__( self :Tuple ) -> List[str]: '''simple docstring''' snake_case_ : Optional[Any] = {} def _A ( self :str , lowerCAmelCase__ :Dict , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Union[str, Any]=1 ) -> str: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): # if there already is a edge if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: # if u does not exist snake_case_ : str = [[w, v]] # add the other way if self.graph.get(lowerCAmelCase__ ): # if there already is a edge if self.graph[v].count([w, u] ) == 0: self.graph[v].append([w, u] ) else: # if u does not exist snake_case_ : List[str] = [[w, u]] def _A ( self :Dict , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Optional[Any] ) -> Optional[Any]: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCAmelCase__ ) # the other way round if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[v]: if _[1] == u: self.graph[v].remove(lowerCAmelCase__ ) def _A ( self :Optional[Any] , lowerCAmelCase__ :Optional[Any]=-2 , lowerCAmelCase__ :Optional[int]=-1 ) -> int: '''simple docstring''' if s == d: return [] snake_case_ : Any = [] snake_case_ : Dict = [] if s == -2: snake_case_ : Optional[int] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[str] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCAmelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : str = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return visited def _A ( self :Optional[int] , lowerCAmelCase__ :str=-1 ) -> List[Any]: '''simple docstring''' if c == -1: snake_case_ : Optional[int] = floor(random() * 10_000 ) + 10 for i in range(lowerCAmelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): snake_case_ : str = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCAmelCase__ , lowerCAmelCase__ , 1 ) def _A ( self :Any , lowerCAmelCase__ :Optional[Any]=-2 ) -> List[Any]: '''simple docstring''' snake_case_ : List[str] = deque() snake_case_ : Optional[Any] = [] if s == -2: snake_case_ : List[Any] = list(self.graph )[0] d.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) while d: snake_case_ : Optional[int] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def _A ( self :str , lowerCAmelCase__ :Union[str, Any] ) -> Optional[Any]: '''simple docstring''' return len(self.graph[u] ) def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : Any = [] snake_case_ : Optional[Any] = [] snake_case_ : Optional[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = -2 snake_case_ : Optional[int] = [] snake_case_ : Tuple = s snake_case_ : Optional[Any] = False snake_case_ : Optional[int] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : Optional[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Optional[int] = len(lowerCAmelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Tuple = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[int] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : List[Any] = s snake_case_ : Dict = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return list(lowerCAmelCase__ ) def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Optional[Any] = [] snake_case_ : int = [] snake_case_ : List[str] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = -2 snake_case_ : int = [] snake_case_ : int = s snake_case_ : Optional[Any] = False snake_case_ : List[Any] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : Union[str, Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Tuple = len(lowerCAmelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Optional[Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[Any] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Tuple = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : Union[str, Any] = s snake_case_ : Tuple = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return False def _A ( self :Any ) -> Tuple: '''simple docstring''' return list(self.graph ) def _A ( self :Optional[Any] , lowerCAmelCase__ :Tuple=-2 , lowerCAmelCase__ :Optional[int]=-1 ) -> str: '''simple docstring''' snake_case_ : List[str] = time() self.dfs(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : List[Any] = time() return end - begin def _A ( self :Union[str, Any] , lowerCAmelCase__ :List[Any]=-2 ) -> int: '''simple docstring''' snake_case_ : List[str] = time() self.bfs(lowerCAmelCase__ ) snake_case_ : Tuple = time() return end - begin
656
0
'''simple docstring''' from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class A_ (a_ , a_ , a_ ): """simple docstring""" a__ = [r'''h\.\d+\.attn\.bias''', r'''h\.\d+\.attn\.masked_bias'''] @register_to_config def __init__( self :Optional[Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :Optional[int] = None , lowerCAmelCase__ :int = 50_257 , lowerCAmelCase__ :int = 1_024 , lowerCAmelCase__ :int = 768 , lowerCAmelCase__ :int = 12 , lowerCAmelCase__ :int = 12 , lowerCAmelCase__ :Optional[int] = None , lowerCAmelCase__ :str = "gelu_new" , lowerCAmelCase__ :float = 0.1 , lowerCAmelCase__ :float = 0.1 , lowerCAmelCase__ :float = 0.1 , lowerCAmelCase__ :float = 1E-5 , lowerCAmelCase__ :float = 0.0_2 , lowerCAmelCase__ :bool = True , lowerCAmelCase__ :bool = True , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , ) -> List[Any]: '''simple docstring''' super().__init__() snake_case_ : str = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' F''' `n_embd`: {n_embd} are not equal.''' ) snake_case_ : Any = prefix_inner_dim snake_case_ : List[Any] = prefix_hidden_dim snake_case_ : str = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) snake_case_ : Optional[int] = ( nn.Linear(self.prefix_hidden_dim , lowerCAmelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) snake_case_ : int = GPTaConfig( vocab_size=lowerCAmelCase__ , n_positions=lowerCAmelCase__ , n_embd=lowerCAmelCase__ , n_layer=lowerCAmelCase__ , n_head=lowerCAmelCase__ , n_inner=lowerCAmelCase__ , activation_function=lowerCAmelCase__ , resid_pdrop=lowerCAmelCase__ , embd_pdrop=lowerCAmelCase__ , attn_pdrop=lowerCAmelCase__ , layer_norm_epsilon=lowerCAmelCase__ , initializer_range=lowerCAmelCase__ , scale_attn_weights=lowerCAmelCase__ , use_cache=lowerCAmelCase__ , scale_attn_by_inverse_layer_idx=lowerCAmelCase__ , reorder_and_upcast_attn=lowerCAmelCase__ , ) snake_case_ : Dict = GPTaLMHeadModel(lowerCAmelCase__ ) def _A ( self :str , lowerCAmelCase__ :torch.Tensor , lowerCAmelCase__ :torch.Tensor , lowerCAmelCase__ :Optional[torch.Tensor] = None , lowerCAmelCase__ :Optional[torch.Tensor] = None , ) -> int: '''simple docstring''' snake_case_ : Dict = self.transformer.transformer.wte(lowerCAmelCase__ ) snake_case_ : List[Any] = self.encode_prefix(lowerCAmelCase__ ) snake_case_ : List[Any] = self.decode_prefix(lowerCAmelCase__ ) snake_case_ : List[str] = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: snake_case_ : int = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) snake_case_ : Optional[Any] = torch.cat((dummy_token, input_ids) , dim=1 ) snake_case_ : List[str] = self.transformer(inputs_embeds=lowerCAmelCase__ , labels=lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def _A ( self :Union[str, Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :torch.device ) -> torch.Tensor: '''simple docstring''' return torch.zeros(lowerCAmelCase__ , self.prefix_length , dtype=torch.intaa , device=lowerCAmelCase__ ) def _A ( self :List[Any] , lowerCAmelCase__ :Union[str, Any] ) -> Dict: '''simple docstring''' return self.encode_prefix(lowerCAmelCase__ ) @torch.no_grad() def _A ( self :Union[str, Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Tuple ) -> str: '''simple docstring''' snake_case_ : List[str] = torch.split(lowerCAmelCase__ , 1 , dim=0 ) snake_case_ : Any = [] snake_case_ : int = [] for feature in features: snake_case_ : Dict = self.decode_prefix(feature.to(lowerCAmelCase__ ) ) # back to the clip feature # Only support beam search for now snake_case_ : int = self.generate_beam( input_embeds=lowerCAmelCase__ , device=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) snake_case_ : List[Any] = torch.stack(lowerCAmelCase__ ) snake_case_ : str = torch.stack(lowerCAmelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def _A ( self :Any , lowerCAmelCase__ :Optional[Any]=None , lowerCAmelCase__ :int=None , lowerCAmelCase__ :List[Any]=None , lowerCAmelCase__ :int = 5 , lowerCAmelCase__ :int = 67 , lowerCAmelCase__ :float = 1.0 , lowerCAmelCase__ :Optional[int] = None , ) -> str: '''simple docstring''' snake_case_ : Union[str, Any] = eos_token_id snake_case_ : str = None snake_case_ : int = None snake_case_ : Tuple = torch.ones(lowerCAmelCase__ , device=lowerCAmelCase__ , dtype=torch.int ) snake_case_ : Dict = torch.zeros(lowerCAmelCase__ , device=lowerCAmelCase__ , dtype=torch.bool ) if input_embeds is not None: snake_case_ : List[str] = input_embeds else: snake_case_ : List[str] = self.transformer.transformer.wte(lowerCAmelCase__ ) for i in range(lowerCAmelCase__ ): snake_case_ : Any = self.transformer(inputs_embeds=lowerCAmelCase__ ) snake_case_ : Optional[Any] = outputs.logits snake_case_ : str = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) snake_case_ : List[str] = logits.softmax(-1 ).log() if scores is None: snake_case_ : Optional[Any] = logits.topk(lowerCAmelCase__ , -1 ) snake_case_ : List[str] = generated.expand(lowerCAmelCase__ , *generated.shape[1:] ) snake_case_ : int = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: snake_case_ : Union[str, Any] = next_tokens else: snake_case_ : Optional[Any] = tokens.expand(lowerCAmelCase__ , *tokens.shape[1:] ) snake_case_ : Any = torch.cat((tokens, next_tokens) , dim=1 ) else: snake_case_ : Tuple = -float(np.inf ) snake_case_ : List[Any] = 0 snake_case_ : Dict = scores[:, None] + logits seq_lengths[~is_stopped] += 1 snake_case_ : Any = scores_sum / seq_lengths[:, None] snake_case_ : Dict = scores_sum_average.view(-1 ).topk(lowerCAmelCase__ , -1 ) snake_case_ : Optional[Any] = next_tokens // scores_sum.shape[1] snake_case_ : Optional[int] = seq_lengths[next_tokens_source] snake_case_ : Optional[int] = next_tokens % scores_sum.shape[1] snake_case_ : List[str] = next_tokens.unsqueeze(1 ) snake_case_ : Dict = tokens[next_tokens_source] snake_case_ : Tuple = torch.cat((tokens, next_tokens) , dim=1 ) snake_case_ : Optional[Any] = generated[next_tokens_source] snake_case_ : List[str] = scores_sum_average * seq_lengths snake_case_ : Optional[int] = is_stopped[next_tokens_source] snake_case_ : List[str] = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) snake_case_ : int = torch.cat((generated, next_token_embed) , dim=1 ) snake_case_ : Any = is_stopped + next_tokens.eq(lowerCAmelCase__ ).squeeze() if is_stopped.all(): break snake_case_ : List[str] = scores / seq_lengths snake_case_ : Dict = scores.argsort(descending=lowerCAmelCase__ ) # tokens tensors are already padded to max_seq_length snake_case_ : int = [tokens[i] for i in order] snake_case_ : str = torch.stack(lowerCAmelCase__ , dim=0 ) snake_case_ : str = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
702
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __lowerCamelCase : List[str] = re.compile(R'''\s+''') def __UpperCAmelCase ( __magic_name__ )-> Union[str, Any]: """simple docstring""" return {"hash": hashlib.mda(re.sub(__magic_name__ ,"" ,example["content"] ).encode("utf-8" ) ).hexdigest()} def __UpperCAmelCase ( __magic_name__ )-> str: """simple docstring""" snake_case_ : Optional[Any] = [len(__magic_name__ ) for line in example["content"].splitlines()] return {"line_mean": np.mean(__magic_name__ ), "line_max": max(__magic_name__ )} def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" snake_case_ : Optional[int] = np.mean([c.isalnum() for c in example["content"]] ) return {"alpha_frac": alpha_frac} def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" if example["hash"] in uniques: uniques.remove(example["hash"] ) return True else: return False def __UpperCAmelCase ( __magic_name__ ,__magic_name__=5 )-> Tuple: """simple docstring""" snake_case_ : List[str] = ["auto-generated", "autogenerated", "automatically generated"] snake_case_ : Optional[Any] = example["content"].splitlines() for _, line in zip(range(__magic_name__ ) ,__magic_name__ ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def __UpperCAmelCase ( __magic_name__ ,__magic_name__=5 ,__magic_name__=0.05 )-> Optional[Any]: """simple docstring""" snake_case_ : str = ["unit tests", "test file", "configuration file"] snake_case_ : int = example["content"].splitlines() snake_case_ : Optional[Any] = 0 snake_case_ : Any = 0 # first test for _, line in zip(range(__magic_name__ ) ,__magic_name__ ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test snake_case_ : Tuple = example["content"].count("\n" ) snake_case_ : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count("config" ) count_test += line.lower().count("test" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def __UpperCAmelCase ( __magic_name__ )-> str: """simple docstring""" snake_case_ : List[Any] = ["def ", "class ", "for ", "while "] snake_case_ : Optional[Any] = example["content"].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def __UpperCAmelCase ( __magic_name__ ,__magic_name__=4 )-> Optional[int]: """simple docstring""" snake_case_ : Tuple = example["content"].splitlines() snake_case_ : Tuple = 0 for line in lines: counter += line.lower().count("=" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : Tuple = tokenizer(example["content"] ,truncation=__magic_name__ )["input_ids"] snake_case_ : int = len(example["content"] ) / len(__magic_name__ ) return {"ratio": ratio} def __UpperCAmelCase ( __magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : Union[str, Any] = {} results.update(get_hash(__magic_name__ ) ) results.update(line_stats(__magic_name__ ) ) results.update(alpha_stats(__magic_name__ ) ) results.update(char_token_ratio(__magic_name__ ) ) results.update(is_autogenerated(__magic_name__ ) ) results.update(is_config_or_test(__magic_name__ ) ) results.update(has_no_keywords(__magic_name__ ) ) results.update(has_few_assignments(__magic_name__ ) ) return results def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" if not check_uniques(__magic_name__ ,__magic_name__ ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def __UpperCAmelCase ( __magic_name__ )-> Dict: """simple docstring""" with open(__magic_name__ ,"rb" ) as f_in: with gzip.open(str(__magic_name__ ) + ".gz" ,"wb" ,compresslevel=6 ) as f_out: shutil.copyfileobj(__magic_name__ ,__magic_name__ ) os.unlink(__magic_name__ ) # Settings __lowerCamelCase : List[Any] = HfArgumentParser(PreprocessingArguments) __lowerCamelCase : str = parser.parse_args() if args.num_workers is None: __lowerCamelCase : List[Any] = multiprocessing.cpu_count() __lowerCamelCase : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __lowerCamelCase : Any = time.time() __lowerCamelCase : str = load_dataset(args.dataset_name, split='''train''') print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __lowerCamelCase : List[str] = time.time() __lowerCamelCase : Any = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __lowerCamelCase : Any = set(ds.unique('''hash''')) __lowerCamelCase : Optional[int] = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __lowerCamelCase : List[str] = time.time() __lowerCamelCase : Tuple = ds.filter(filter, fn_kwargs={'''uniques''': uniques, '''args''': args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __lowerCamelCase : List[str] = time.time() __lowerCamelCase , __lowerCamelCase : Tuple = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __lowerCamelCase : List[Any] = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / '''duplicate_clusters.json''', '''w''') as f: json.dump(duplicate_clusters, f) __lowerCamelCase : List[str] = output_dir / '''data''' data_dir.mkdir(exist_ok=True) __lowerCamelCase : int = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __lowerCamelCase : Union[str, Any] = str(data_dir / f'''file-{file_number+1:012}.json''') __lowerCamelCase : List[Any] = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
656
0
'''simple docstring''' import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, ) from transformers.models.efficientformer.modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_vision_available(): from PIL import Image from transformers import EfficientFormerImageProcessor class A_ : """simple docstring""" def __init__( self :str , lowerCAmelCase__ :Dict , lowerCAmelCase__ :int = 13 , lowerCAmelCase__ :int = 64 , lowerCAmelCase__ :int = 2 , lowerCAmelCase__ :int = 3 , lowerCAmelCase__ :int = 3 , lowerCAmelCase__ :bool = True , lowerCAmelCase__ :bool = True , lowerCAmelCase__ :int = 128 , lowerCAmelCase__ :List[str]=[16, 32, 64, 128] , lowerCAmelCase__ :int = 7 , lowerCAmelCase__ :int = 4 , lowerCAmelCase__ :int = 37 , lowerCAmelCase__ :str = "gelu" , lowerCAmelCase__ :float = 0.1 , lowerCAmelCase__ :float = 0.1 , lowerCAmelCase__ :int = 10 , lowerCAmelCase__ :float = 0.0_2 , lowerCAmelCase__ :int = 2 , lowerCAmelCase__ :int = 1 , lowerCAmelCase__ :int = 128 , lowerCAmelCase__ :List[int] = [2, 2, 2, 2] , lowerCAmelCase__ :int = 2 , lowerCAmelCase__ :int = 2 , ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = parent snake_case_ : Dict = batch_size snake_case_ : Dict = image_size snake_case_ : List[str] = patch_size snake_case_ : Tuple = num_channels snake_case_ : List[Any] = is_training snake_case_ : Union[str, Any] = use_labels snake_case_ : Tuple = hidden_size snake_case_ : Tuple = num_hidden_layers snake_case_ : Optional[int] = num_attention_heads snake_case_ : Tuple = intermediate_size snake_case_ : Tuple = hidden_act snake_case_ : Union[str, Any] = hidden_dropout_prob snake_case_ : int = attention_probs_dropout_prob snake_case_ : Union[str, Any] = type_sequence_label_size snake_case_ : Optional[int] = initializer_range snake_case_ : int = encoder_stride snake_case_ : Dict = num_attention_outputs snake_case_ : Union[str, Any] = embed_dim snake_case_ : str = embed_dim + 1 snake_case_ : Optional[int] = resolution snake_case_ : List[Any] = depths snake_case_ : Union[str, Any] = hidden_sizes snake_case_ : List[str] = dim snake_case_ : int = mlp_expansion_ratio def _A ( self :Optional[int] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ : Tuple = None if self.use_labels: snake_case_ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ : Tuple = self.get_config() return config, pixel_values, labels def _A ( self :int ) -> int: '''simple docstring''' return EfficientFormerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCAmelCase__ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , resolution=self.resolution , depths=self.depths , hidden_sizes=self.hidden_sizes , dim=self.dim , mlp_expansion_ratio=self.mlp_expansion_ratio , ) def _A ( self :Tuple , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :List[str] , lowerCAmelCase__ :int ) -> Any: '''simple docstring''' snake_case_ : List[Any] = TFEfficientFormerModel(config=lowerCAmelCase__ ) snake_case_ : Tuple = model(lowerCAmelCase__ , training=lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _A ( self :str , lowerCAmelCase__ :int , lowerCAmelCase__ :str , lowerCAmelCase__ :Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case_ : str = self.type_sequence_label_size snake_case_ : str = TFEfficientFormerForImageClassification(lowerCAmelCase__ ) snake_case_ : Tuple = model(lowerCAmelCase__ , labels=lowerCAmelCase__ , training=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images snake_case_ : str = 1 snake_case_ : Tuple = TFEfficientFormerForImageClassification(lowerCAmelCase__ ) snake_case_ : Union[str, Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case_ : Dict = model(lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _A ( self :Any ) -> Optional[int]: '''simple docstring''' snake_case_ : Any = self.prepare_config_and_inputs() snake_case_ : Union[str, Any] = config_and_inputs snake_case_ : Any = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class A_ (a_ , a_ , unittest.TestCase ): """simple docstring""" a__ = ( ( TFEfficientFormerModel, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerForImageClassification, ) if is_tf_available() else () ) a__ = ( { '''feature-extraction''': TFEfficientFormerModel, '''image-classification''': ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, ), } if is_tf_available() else {} ) a__ = False a__ = False a__ = False a__ = False a__ = False def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Dict = TFEfficientFormerModelTester(self ) snake_case_ : Union[str, Any] = ConfigTester( self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ , hidden_size=37 ) def _A ( self :List[Any] ) -> Dict: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="EfficientFormer does not use inputs_embeds" ) def _A ( self :int ) -> Optional[int]: '''simple docstring''' pass @unittest.skip(reason="EfficientFormer does not support input and output embeddings" ) def _A ( self :Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' pass def _A ( self :List[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ : Tuple = model_class(lowerCAmelCase__ ) snake_case_ : Optional[Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ : List[Any] = [*signature.parameters.keys()] snake_case_ : Optional[int] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) def _A ( self :Dict ) -> Any: '''simple docstring''' def check_hidden_states_output(lowerCAmelCase__ :Any , lowerCAmelCase__ :List[str] , lowerCAmelCase__ :List[str] ): snake_case_ : Dict = model_class(lowerCAmelCase__ ) snake_case_ : Union[str, Any] = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) , training=lowerCAmelCase__ ) snake_case_ : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states snake_case_ : List[str] = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(lowerCAmelCase__ ) , lowerCAmelCase__ ) if hasattr(self.model_tester , "encoder_seq_length" ): snake_case_ : Dict = self.model_tester.encoder_seq_length if hasattr(self.model_tester , "chunk_length" ) and self.model_tester.chunk_length > 1: snake_case_ : List[Any] = seq_length * self.model_tester.chunk_length else: snake_case_ : List[Any] = self.model_tester.seq_length self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) if config.is_encoder_decoder: snake_case_ : List[str] = outputs.decoder_hidden_states self.asseretIsInstance(lowerCAmelCase__ , (list, tuple) ) self.assertEqual(len(lowerCAmelCase__ ) , lowerCAmelCase__ ) snake_case_ : Any = getattr(self.model_tester , "seq_length" , lowerCAmelCase__ ) snake_case_ : Any = getattr(self.model_tester , "decoder_seq_length" , lowerCAmelCase__ ) self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [decoder_seq_length, self.model_tester.hidden_size] , ) snake_case_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ : Optional[Any] = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ : Tuple = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def _A ( self :Union[str, Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :Any , lowerCAmelCase__ :str=False ) -> Dict: '''simple docstring''' snake_case_ : int = super()._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__ ) if return_labels: if model_class.__name__ == "TFEfficientFormerForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def _A ( self :Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) @unittest.skip(reason="EfficientFormer does not implement masked image modeling yet" ) def _A ( self :Optional[int] ) -> List[str]: '''simple docstring''' snake_case_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCAmelCase__ ) def _A ( self :Dict ) -> int: '''simple docstring''' snake_case_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase__ ) @slow def _A ( self :Optional[int] ) -> List[str]: '''simple docstring''' for model_name in TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ : str = TFEfficientFormerModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> str: '''simple docstring''' snake_case_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ : Union[str, Any] = True snake_case_ : Tuple = getattr(self.model_tester , "seq_length" , lowerCAmelCase__ ) snake_case_ : Tuple = getattr(self.model_tester , "encoder_seq_length" , lowerCAmelCase__ ) snake_case_ : List[str] = getattr(self.model_tester , "key_length" , lowerCAmelCase__ ) snake_case_ : Optional[Any] = getattr(self.model_tester , "chunk_length" , lowerCAmelCase__ ) if chunk_length is not None and hasattr(self.model_tester , "num_hashes" ): snake_case_ : Optional[int] = encoder_seq_length * self.model_tester.num_hashes for model_class in self.all_model_classes: snake_case_ : Optional[int] = True snake_case_ : List[Any] = False snake_case_ : Tuple = True snake_case_ : Any = model_class(lowerCAmelCase__ ) snake_case_ : int = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) , training=lowerCAmelCase__ ) snake_case_ : Any = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(lowerCAmelCase__ ) , self.model_tester.num_attention_outputs ) # check that output_attentions also work using config del inputs_dict["output_attentions"] snake_case_ : Optional[Any] = True snake_case_ : Tuple = model_class(lowerCAmelCase__ ) snake_case_ : Optional[int] = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) , training=lowerCAmelCase__ ) snake_case_ : List[str] = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(lowerCAmelCase__ ) , self.model_tester.num_attention_outputs ) if chunk_length is not None: self.assertListEqual( list(attentions[0].shape[-4:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, chunk_length, encoder_key_length] , ) else: self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length] , ) def _A ( self :Tuple ) -> List[str]: '''simple docstring''' snake_case_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # Prepare our model snake_case_ : str = model_class(lowerCAmelCase__ ) # These are maximally general inputs for the model, with multiple None dimensions # Hopefully this will catch any conditionals that fail for flexible shapes snake_case_ : int = { key: tf.keras.Input(shape=val.shape[1:] , dtype=val.dtype , name=lowerCAmelCase__ ) for key, val in model.input_signature.items() if key in model.dummy_inputs } snake_case_ : int = model(lowerCAmelCase__ ) self.assertTrue(outputs_dict is not None ) def __UpperCAmelCase ( )-> Tuple: """simple docstring""" snake_case_ : List[str] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class A_ (unittest.TestCase ): """simple docstring""" @cached_property def _A ( self :List[Any] ) -> Optional[int]: '''simple docstring''' return ( EfficientFormerImageProcessor.from_pretrained("snap-research/efficientformer-l1-300" ) if is_vision_available() else None ) @slow def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Any = TFEfficientFormerForImageClassification.from_pretrained("snap-research/efficientformer-l1-300" ) snake_case_ : Dict = self.default_image_processor snake_case_ : Optional[Any] = prepare_img() snake_case_ : Any = image_processor(images=lowerCAmelCase__ , return_tensors="tf" ) # forward pass snake_case_ : int = model(**lowerCAmelCase__ , training=lowerCAmelCase__ ) # verify the logits snake_case_ : int = tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase__ ) snake_case_ : Union[str, Any] = tf.constant([-0.0_5_5_5, 0.4_8_2_5, -0.0_8_5_2] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1E-4 ) ) @slow def _A ( self :Tuple ) -> Dict: '''simple docstring''' snake_case_ : Optional[int] = TFEfficientFormerForImageClassificationWithTeacher.from_pretrained( "snap-research/efficientformer-l1-300" ) snake_case_ : Optional[int] = self.default_image_processor snake_case_ : str = prepare_img() snake_case_ : Union[str, Any] = image_processor(images=lowerCAmelCase__ , return_tensors="tf" ) # forward pass snake_case_ : List[Any] = model(**lowerCAmelCase__ , training=lowerCAmelCase__ ) # verify the logits snake_case_ : Optional[int] = tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase__ ) snake_case_ : Any = tf.constant([-0.1_3_1_2, 0.4_3_5_3, -1.0_4_9_9] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1E-4 ) )
703
'''simple docstring''' import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :str ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Union[str, Any] = torch.nn.Linear(10 , 10 ) snake_case_ : Dict = torch.optim.SGD(model.parameters() , 0.1 ) snake_case_ : Tuple = Accelerator() snake_case_ : Optional[Any] = accelerator.prepare(lowerCAmelCase__ ) try: pickle.loads(pickle.dumps(lowerCAmelCase__ ) ) except Exception as e: self.fail(F'''Accelerated optimizer pickling failed with {e}''' ) AcceleratorState._reset_state()
656
0
from __future__ import annotations __lowerCamelCase : Dict = [ [-1, 0], # left [0, -1], # down [1, 0], # right [0, 1], # up ] def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,)-> tuple[list[list[int]], list[list[int]]]: """simple docstring""" snake_case_ : List[Any] = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__magic_name__ ) ) ] # the reference grid snake_case_ : Optional[Any] = 1 snake_case_ : Union[str, Any] = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__magic_name__ ) ) ] # the action grid snake_case_ : List[Any] = init[0] snake_case_ : List[Any] = init[1] snake_case_ : Any = 0 snake_case_ : Tuple = g + heuristic[x][y] # cost from starting cell to destination cell snake_case_ : Tuple = [[f, g, x, y]] snake_case_ : List[Any] = False # flag that is set when search is complete snake_case_ : Tuple = False # flag set if we can't find expand while not found and not resign: if len(__magic_name__ ) == 0: raise ValueError("Algorithm is unable to find solution" ) else: # to choose the least costliest action so as to move closer to the goal cell.sort() cell.reverse() snake_case_ : Tuple = cell.pop() snake_case_ : List[str] = next_cell[2] snake_case_ : int = next_cell[3] snake_case_ : Optional[int] = next_cell[1] if x == goal[0] and y == goal[1]: snake_case_ : Tuple = True else: for i in range(len(__magic_name__ ) ): # to try out different valid actions snake_case_ : Optional[Any] = x + DIRECTIONS[i][0] snake_case_ : List[str] = y + DIRECTIONS[i][1] if xa >= 0 and xa < len(__magic_name__ ) and ya >= 0 and ya < len(grid[0] ): if closed[xa][ya] == 0 and grid[xa][ya] == 0: snake_case_ : Dict = g + cost snake_case_ : Any = ga + heuristic[xa][ya] cell.append([fa, ga, xa, ya] ) snake_case_ : Union[str, Any] = 1 snake_case_ : Optional[Any] = i snake_case_ : Optional[Any] = [] snake_case_ : Any = goal[0] snake_case_ : Tuple = goal[1] invpath.append([x, y] ) # we get the reverse path from here while x != init[0] or y != init[1]: snake_case_ : Any = x - DIRECTIONS[action[x][y]][0] snake_case_ : Tuple = y - DIRECTIONS[action[x][y]][1] snake_case_ : Tuple = xa snake_case_ : Tuple = ya invpath.append([x, y] ) snake_case_ : Any = [] for i in range(len(__magic_name__ ) ): path.append(invpath[len(__magic_name__ ) - 1 - i] ) return path, action if __name__ == "__main__": __lowerCamelCase : Optional[int] = [ [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0], [0, 0, 0, 0, 1, 0], ] __lowerCamelCase : Any = [0, 0] # all coordinates are given in format [y,x] __lowerCamelCase : Optional[Any] = [len(grid) - 1, len(grid[0]) - 1] __lowerCamelCase : Any = 1 # the cost map which pushes the path closer to the goal __lowerCamelCase : int = [[0 for row in range(len(grid[0]))] for col in range(len(grid))] for i in range(len(grid)): for j in range(len(grid[0])): __lowerCamelCase : Union[str, Any] = abs(i - goal[0]) + abs(j - goal[1]) if grid[i][j] == 1: # added extra penalty in the heuristic map __lowerCamelCase : Dict = 99 __lowerCamelCase : Dict = search(grid, init, goal, cost, heuristic) print('''ACTION MAP''') for i in range(len(action)): print(action[i]) for i in range(len(path)): print(path[i])
704
'''simple docstring''' import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __lowerCamelCase : Any = '''src/transformers''' # This is to make sure the transformers module imported is the one in the repo. __lowerCamelCase : List[str] = direct_transformers_import(PATH_TO_TRANSFORMERS) __lowerCamelCase : Optional[Any] = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __lowerCamelCase : Union[str, Any] = re.compile(R'''\[(.+?)\]\((https://huggingface\.co/.+?)\)''') __lowerCamelCase : Any = { '''DecisionTransformerConfig''', '''EncoderDecoderConfig''', '''MusicgenConfig''', '''RagConfig''', '''SpeechEncoderDecoderConfig''', '''TimmBackboneConfig''', '''VisionEncoderDecoderConfig''', '''VisionTextDualEncoderConfig''', '''LlamaConfig''', } def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : Tuple = None # source code of `config_class` snake_case_ : List[Any] = inspect.getsource(__magic_name__ ) snake_case_ : List[str] = _re_checkpoint.findall(__magic_name__ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith("/" ): snake_case_ : Optional[Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link snake_case_ : str = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: snake_case_ : Dict = ckpt_name break return checkpoint def __UpperCAmelCase ( )-> Dict: """simple docstring""" snake_case_ : Optional[int] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue snake_case_ : str = get_checkpoint_from_config_class(__magic_name__ ) snake_case_ : Union[str, Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(__magic_name__ ) if len(__magic_name__ ) > 0: snake_case_ : Tuple = "\n".join(sorted(__magic_name__ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
656
0
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from importlib import import_module from typing import Dict, List, Optional, Tuple import numpy as np from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch import nn from utils_ner import Split, TokenClassificationDataset, TokenClassificationTask import transformers from transformers import ( AutoConfig, AutoModelForTokenClassification, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process __lowerCamelCase : str = logging.getLogger(__name__) @dataclass class A_ : """simple docstring""" a__ = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) a__ = field( default=a_ , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) a__ = field( default='''NER''' , metadata={'''help''': '''Task type to fine tune in training (e.g. NER, POS, etc)'''} ) a__ = field( default=a_ , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''} ) a__ = field(default=a_ , metadata={'''help''': '''Set this flag to use fast tokenization.'''} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. a__ = field( default=a_ , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) @dataclass class A_ : """simple docstring""" a__ = field( metadata={'''help''': '''The input data dir. Should contain the .txt files for a CoNLL-2003-formatted task.'''} ) a__ = field( default=a_ , metadata={'''help''': '''Path to a file containing all labels. If not specified, CoNLL-2003 labels are used.'''} , ) a__ = field( default=128 , metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) a__ = field( default=a_ , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) def __UpperCAmelCase ( )-> List[str]: """simple docstring""" snake_case_ : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. snake_case_ : str = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: snake_case_ : List[Any] = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' " --overwrite_output_dir to overcome." ) snake_case_ : str = import_module("tasks" ) try: snake_case_ : Dict = getattr(__magic_name__ ,model_args.task_type ) snake_case_ : TokenClassificationTask = token_classification_task_clazz() except AttributeError: raise ValueError( F'''Task {model_args.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. ''' F'''Available tasks classes are: {TokenClassificationTask.__subclasses__()}''' ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" ,datefmt="%m/%d/%Y %H:%M:%S" ,level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN ,) logger.warning( "Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s" ,training_args.local_rank ,training_args.device ,training_args.n_gpu ,bool(training_args.local_rank != -1 ) ,training_args.fpaa ,) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("Training/evaluation parameters %s" ,__magic_name__ ) # Set seed set_seed(training_args.seed ) # Prepare CONLL-2003 task snake_case_ : Any = token_classification_task.get_labels(data_args.labels ) snake_case_ : Dict[int, str] = dict(enumerate(__magic_name__ ) ) snake_case_ : List[str] = len(__magic_name__ ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case_ : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path ,num_labels=__magic_name__ ,idalabel=__magic_name__ ,labelaid={label: i for i, label in enumerate(__magic_name__ )} ,cache_dir=model_args.cache_dir ,) snake_case_ : Optional[Any] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path ,cache_dir=model_args.cache_dir ,use_fast=model_args.use_fast ,) snake_case_ : Any = AutoModelForTokenClassification.from_pretrained( model_args.model_name_or_path ,from_tf=bool(".ckpt" in model_args.model_name_or_path ) ,config=__magic_name__ ,cache_dir=model_args.cache_dir ,) # Get datasets snake_case_ : str = ( TokenClassificationDataset( token_classification_task=__magic_name__ ,data_dir=data_args.data_dir ,tokenizer=__magic_name__ ,labels=__magic_name__ ,model_type=config.model_type ,max_seq_length=data_args.max_seq_length ,overwrite_cache=data_args.overwrite_cache ,mode=Split.train ,) if training_args.do_train else None ) snake_case_ : List[str] = ( TokenClassificationDataset( token_classification_task=__magic_name__ ,data_dir=data_args.data_dir ,tokenizer=__magic_name__ ,labels=__magic_name__ ,model_type=config.model_type ,max_seq_length=data_args.max_seq_length ,overwrite_cache=data_args.overwrite_cache ,mode=Split.dev ,) if training_args.do_eval else None ) def align_predictions(__magic_name__ ,__magic_name__ ) -> Tuple[List[int], List[int]]: snake_case_ : Dict = np.argmax(__magic_name__ ,axis=2 ) snake_case_ : Dict = preds.shape snake_case_ : Tuple = [[] for _ in range(__magic_name__ )] snake_case_ : Tuple = [[] for _ in range(__magic_name__ )] for i in range(__magic_name__ ): for j in range(__magic_name__ ): if label_ids[i, j] != nn.CrossEntropyLoss().ignore_index: out_label_list[i].append(label_map[label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) return preds_list, out_label_list def compute_metrics(__magic_name__ ) -> Dict: snake_case_ : int = align_predictions(p.predictions ,p.label_ids ) return { "accuracy_score": accuracy_score(__magic_name__ ,__magic_name__ ), "precision": precision_score(__magic_name__ ,__magic_name__ ), "recall": recall_score(__magic_name__ ,__magic_name__ ), "f1": fa_score(__magic_name__ ,__magic_name__ ), } # Data collator snake_case_ : Optional[int] = DataCollatorWithPadding(__magic_name__ ,pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer snake_case_ : Tuple = Trainer( model=__magic_name__ ,args=__magic_name__ ,train_dataset=__magic_name__ ,eval_dataset=__magic_name__ ,compute_metrics=__magic_name__ ,data_collator=__magic_name__ ,) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_process_zero(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation snake_case_ : Dict = {} if training_args.do_eval: logger.info("*** Evaluate ***" ) snake_case_ : Tuple = trainer.evaluate() snake_case_ : int = os.path.join(training_args.output_dir ,"eval_results.txt" ) if trainer.is_world_process_zero(): with open(__magic_name__ ,"w" ) as writer: logger.info("***** Eval results *****" ) for key, value in result.items(): logger.info(" %s = %s" ,__magic_name__ ,__magic_name__ ) writer.write("%s = %s\n" % (key, value) ) results.update(__magic_name__ ) # Predict if training_args.do_predict: snake_case_ : List[str] = TokenClassificationDataset( token_classification_task=__magic_name__ ,data_dir=data_args.data_dir ,tokenizer=__magic_name__ ,labels=__magic_name__ ,model_type=config.model_type ,max_seq_length=data_args.max_seq_length ,overwrite_cache=data_args.overwrite_cache ,mode=Split.test ,) snake_case_ : str = trainer.predict(__magic_name__ ) snake_case_ : List[Any] = align_predictions(__magic_name__ ,__magic_name__ ) snake_case_ : Optional[Any] = os.path.join(training_args.output_dir ,"test_results.txt" ) if trainer.is_world_process_zero(): with open(__magic_name__ ,"w" ) as writer: for key, value in metrics.items(): logger.info(" %s = %s" ,__magic_name__ ,__magic_name__ ) writer.write("%s = %s\n" % (key, value) ) # Save predictions snake_case_ : Optional[Any] = os.path.join(training_args.output_dir ,"test_predictions.txt" ) if trainer.is_world_process_zero(): with open(__magic_name__ ,"w" ) as writer: with open(os.path.join(data_args.data_dir ,"test.txt" ) ,"r" ) as f: token_classification_task.write_predictions_to_file(__magic_name__ ,__magic_name__ ,__magic_name__ ) return results def __UpperCAmelCase ( __magic_name__ )-> Dict: """simple docstring""" main() if __name__ == "__main__": main()
705
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCamelCase : List[str] = logging.get_logger(__name__) __lowerCamelCase : int = { '''microsoft/cvt-13''': '''https://huggingface.co/microsoft/cvt-13/resolve/main/config.json''', # See all Cvt models at https://huggingface.co/models?filter=cvt } class A_ (a_ ): """simple docstring""" a__ = '''cvt''' def __init__( self :List[Any] , lowerCAmelCase__ :Optional[int]=3 , lowerCAmelCase__ :Any=[7, 3, 3] , lowerCAmelCase__ :Dict=[4, 2, 2] , lowerCAmelCase__ :Union[str, Any]=[2, 1, 1] , lowerCAmelCase__ :Any=[64, 192, 384] , lowerCAmelCase__ :List[str]=[1, 3, 6] , lowerCAmelCase__ :str=[1, 2, 10] , lowerCAmelCase__ :Any=[4.0, 4.0, 4.0] , lowerCAmelCase__ :int=[0.0, 0.0, 0.0] , lowerCAmelCase__ :Optional[Any]=[0.0, 0.0, 0.0] , lowerCAmelCase__ :Dict=[0.0, 0.0, 0.1] , lowerCAmelCase__ :List[Any]=[True, True, True] , lowerCAmelCase__ :List[Any]=[False, False, True] , lowerCAmelCase__ :Dict=["dw_bn", "dw_bn", "dw_bn"] , lowerCAmelCase__ :Any=[3, 3, 3] , lowerCAmelCase__ :Tuple=[1, 1, 1] , lowerCAmelCase__ :Optional[int]=[2, 2, 2] , lowerCAmelCase__ :Union[str, Any]=[1, 1, 1] , lowerCAmelCase__ :Any=[1, 1, 1] , lowerCAmelCase__ :List[str]=0.0_2 , lowerCAmelCase__ :Dict=1E-1_2 , **lowerCAmelCase__ :Optional[Any] , ) -> str: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) snake_case_ : int = num_channels snake_case_ : int = patch_sizes snake_case_ : Optional[Any] = patch_stride snake_case_ : Dict = patch_padding snake_case_ : Tuple = embed_dim snake_case_ : Optional[int] = num_heads snake_case_ : Union[str, Any] = depth snake_case_ : Optional[int] = mlp_ratio snake_case_ : Tuple = attention_drop_rate snake_case_ : str = drop_rate snake_case_ : Tuple = drop_path_rate snake_case_ : Any = qkv_bias snake_case_ : Union[str, Any] = cls_token snake_case_ : int = qkv_projection_method snake_case_ : Any = kernel_qkv snake_case_ : Union[str, Any] = padding_kv snake_case_ : str = stride_kv snake_case_ : Dict = padding_q snake_case_ : Tuple = stride_q snake_case_ : Any = initializer_range snake_case_ : Any = layer_norm_eps
656
0
'''simple docstring''' import argparse import json import os import re import torch from transformers import BloomConfig, BloomModel from transformers.file_utils import CONFIG_NAME, WEIGHTS_NAME from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : Optional[int] = [ '''word_embeddings_layernorm.weight''', '''word_embeddings_layernorm.bias''', '''input_layernorm.weight''', '''input_layernorm.bias''', '''post_attention_layernorm.weight''', '''post_attention_layernorm.bias''', '''self_attention.dense.bias''', '''mlp.dense_4h_to_h.bias''', '''ln_f.weight''', '''ln_f.bias''', ] __lowerCamelCase : Dict = [ '''mlp.dense_4h_to_h.weight''', '''self_attention.dense.weight''', ] def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Optional[int]: """simple docstring""" snake_case_ : Optional[int] = { "word_embeddings.weight": "word_embeddings.weight", "word_embeddings.norm.weight": "word_embeddings_layernorm.weight", "word_embeddings.norm.bias": "word_embeddings_layernorm.bias", "weight": "ln_f.weight", "bias": "ln_f.bias", } if key in layer_rename_map: return layer_rename_map[key] # Handle transformer blocks snake_case_ : str = int(re.match(r".*layer_(\d*).*" ,__magic_name__ )[1] ) layer_number -= 3 return F'''h.{layer_number}.''' + key def __UpperCAmelCase ( __magic_name__ )-> Dict: """simple docstring""" if dtype == torch.bool: return 1 / 8 snake_case_ : str = re.search(r"[^\d](\d+)$" ,str(__magic_name__ ) ) if bit_search is None: raise ValueError(F'''`dtype` is not a valid dtype: {dtype}.''' ) snake_case_ : Tuple = int(bit_search.groups()[0] ) return bit_size // 8 def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" if bloom_config_file == "": snake_case_ : str = BloomConfig() else: snake_case_ : Tuple = BloomConfig.from_json_file(__magic_name__ ) if shard_model: snake_case_ : str = os.listdir(__magic_name__ ) snake_case_ : List[str] = sorted(filter(lambda __magic_name__ : s.startswith("layer" ) and "model_00" in s ,__magic_name__ ) ) snake_case_ : List[Any] = {"weight_map": {}, "metadata": {}} snake_case_ : Union[str, Any] = 0 snake_case_ : Tuple = None snake_case_ : Optional[Any] = BloomConfig() for j, file in enumerate(__magic_name__ ): print("Processing file: {}".format(__magic_name__ ) ) snake_case_ : List[Any] = None for i in range(__magic_name__ ): # load all TP files snake_case_ : List[Any] = file.replace("model_00" ,F'''model_0{i}''' ) snake_case_ : Optional[Any] = torch.load(os.path.join(__magic_name__ ,__magic_name__ ) ,map_location="cpu" ) # Rename keys in the transformers names snake_case_ : Union[str, Any] = list(temp.keys() ) for key in keys: snake_case_ : List[str] = temp.pop(__magic_name__ ) if tensors is None: snake_case_ : Union[str, Any] = temp else: for key in tensors.keys(): if any(key.endswith(__magic_name__ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel snake_case_ : Optional[Any] = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks snake_case_ : List[Any] = torch.cat([tensors[key], temp[key]] ,dim=__magic_name__ ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(__magic_name__ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): snake_case_ : int = tensors[key] / pretraining_tp torch.save( __magic_name__ ,os.path.join( __magic_name__ ,"pytorch_model_{}-of-{}.bin".format(str(j + 1 ).zfill(5 ) ,str(len(__magic_name__ ) ).zfill(5 ) ) ,) ,) for key in tensors.keys(): snake_case_ : Any = tensors[key] total_size += value.numel() * get_dtype_size(value.dtype ) if key not in index_dict["weight_map"]: snake_case_ : int = "pytorch_model_{}-of-{}.bin".format( str(j + 1 ).zfill(5 ) ,str(len(__magic_name__ ) ).zfill(5 ) ) snake_case_ : int = BloomConfig() snake_case_ : Dict = pytorch_dump_folder_path + "/" + CONFIG_NAME snake_case_ : Optional[Any] = total_size with open(__magic_name__ ,"w" ,encoding="utf-8" ) as f: f.write(config.to_json_string() ) with open(os.path.join(__magic_name__ ,WEIGHTS_NAME + ".index.json" ) ,"w" ,encoding="utf-8" ) as f: snake_case_ : Any = json.dumps(__magic_name__ ,indent=2 ,sort_keys=__magic_name__ ) + "\n" f.write(__magic_name__ ) else: snake_case_ : Optional[Any] = BloomModel(__magic_name__ ) snake_case_ : str = os.listdir(__magic_name__ ) snake_case_ : Optional[int] = sorted(filter(lambda __magic_name__ : s.startswith("layer" ) and "model_00" in s ,__magic_name__ ) ) snake_case_ : List[Any] = None for i, file in enumerate(__magic_name__ ): snake_case_ : Tuple = None for i in range(__magic_name__ ): # load all TP files snake_case_ : Tuple = file.replace("model_00" ,F'''model_0{i}''' ) snake_case_ : Dict = torch.load(os.path.join(__magic_name__ ,__magic_name__ ) ,map_location="cpu" ) # Rename keys in the transformers names snake_case_ : Tuple = list(temp.keys() ) for key in keys: snake_case_ : Tuple = temp.pop(__magic_name__ ) if tensors is None: snake_case_ : str = temp else: for key in tensors.keys(): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) if any(key.endswith(__magic_name__ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel snake_case_ : Tuple = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks snake_case_ : Union[str, Any] = torch.cat([tensors[key], temp[key]] ,dim=__magic_name__ ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(__magic_name__ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): snake_case_ : List[str] = tensors[key] / pretraining_tp snake_case_ : Optional[Any] = model.load_state_dict(__magic_name__ ,strict=__magic_name__ ) assert not other_keys.unexpected_keys, F'''The keys {other_keys.unexpected_keys} are unexpected''' if missing_keys is None: snake_case_ : Dict = set(other_keys.missing_keys ) else: snake_case_ : Optional[int] = missing_keys.intersection(set(other_keys.missing_keys ) ) assert not missing_keys, F'''The keys {missing_keys} are missing''' # Save pytorch-model os.makedirs(__magic_name__ ,exist_ok=__magic_name__ ) snake_case_ : Dict = pytorch_dump_folder_path + "/" + WEIGHTS_NAME snake_case_ : Optional[Any] = pytorch_dump_folder_path + "/" + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path} with dtype {config.torch_dtype}''' ) if config.torch_dtype is not None: snake_case_ : int = model.to(config.torch_dtype ) torch.save(model.state_dict() ,__magic_name__ ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(__magic_name__ ,"w" ,encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": __lowerCamelCase : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bloom_checkpoint_path''', default=None, type=str, required=True, help='''Path to the Megatron-LM checkpoint path.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--bloom_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--shard_model''', action='''store_true''', help='''An optional setting to shard the output model \nThis enables sharding the converted checkpoint''', ) parser.add_argument( '''--pretraining_tp''', default=4, type=int, help='''Pretraining TP rank that has been used when training the model in Megatron-LM \n''', ) __lowerCamelCase : Union[str, Any] = parser.parse_args() convert_bloom_checkpoint_to_pytorch( args.bloom_checkpoint_path, args.bloom_config_file, args.pytorch_dump_folder_path, args.shard_model, args.pretraining_tp, )
706
'''simple docstring''' import sacrebleu as scb from packaging import version from sacrebleu import TER import datasets __lowerCamelCase : str = '''\ @inproceedings{snover-etal-2006-study, title = "A Study of Translation Edit Rate with Targeted Human Annotation", author = "Snover, Matthew and Dorr, Bonnie and Schwartz, Rich and Micciulla, Linnea and Makhoul, John", booktitle = "Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers", month = aug # " 8-12", year = "2006", address = "Cambridge, Massachusetts, USA", publisher = "Association for Machine Translation in the Americas", url = "https://aclanthology.org/2006.amta-papers.25", pages = "223--231", } @inproceedings{post-2018-call, title = "A Call for Clarity in Reporting {BLEU} Scores", author = "Post, Matt", booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers", month = oct, year = "2018", address = "Belgium, Brussels", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W18-6319", pages = "186--191", } ''' __lowerCamelCase : Dict = '''\ TER (Translation Edit Rate, also called Translation Error Rate) is a metric to quantify the edit operations that a hypothesis requires to match a reference translation. We use the implementation that is already present in sacrebleu (https://github.com/mjpost/sacreBLEU#ter), which in turn is inspired by the TERCOM implementation, which can be found here: https://github.com/jhclark/tercom. The implementation here is slightly different from sacrebleu in terms of the required input format. The length of the references and hypotheses lists need to be the same, so you may need to transpose your references compared to sacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534 See the README.md file at https://github.com/mjpost/sacreBLEU#ter for more information. ''' __lowerCamelCase : int = ''' Produces TER scores alongside the number of edits and reference length. Args: predictions (list of str): The system stream (a sequence of segments). references (list of list of str): A list of one or more reference streams (each a sequence of segments). normalized (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. ignore_punct (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. support_zh_ja_chars (boolean): If `True`, tokenization/normalization supports processing of Chinese characters, as well as Japanese Kanji, Hiragana, Katakana, and Phonetic Extensions of Katakana. Only applies if `normalized = True`. Defaults to `False`. case_sensitive (boolean): If `False`, makes all predictions and references lowercase to ignore differences in case. Defaults to `False`. Returns: \'score\' (float): TER score (num_edits / sum_ref_lengths * 100) \'num_edits\' (int): The cumulative number of edits \'ref_length\' (float): The cumulative average reference length Examples: Example 1: >>> predictions = ["does this sentence match??", ... "what about this sentence?", ... "What did the TER metric user say to the developer?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"], ... ["Your jokes are...", "...TERrible"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {\'score\': 150.0, \'num_edits\': 15, \'ref_length\': 10.0} Example 2: >>> predictions = ["does this sentence match??", ... "what about this sentence?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {\'score\': 62.5, \'num_edits\': 5, \'ref_length\': 8.0} Example 3: >>> predictions = ["does this sentence match??", ... "what about this sentence?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... normalized=True, ... case_sensitive=True) >>> print(results) {\'score\': 57.14285714285714, \'num_edits\': 6, \'ref_length\': 10.5} Example 4: >>> predictions = ["does this sentence match??", ... "what about this sentence?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {\'score\': 0.0, \'num_edits\': 0, \'ref_length\': 8.0} Example 5: >>> predictions = ["does this sentence match??", ... "what about this sentence?", ... "What did the TER metric user say to the developer?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"], ... ["Your jokes are...", "...TERrible"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {\'score\': 100.0, \'num_edits\': 10, \'ref_length\': 10.0} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): """simple docstring""" def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' if version.parse(scb.__version__ ) < version.parse("1.4.12" ): raise ImportWarning( "To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n" "You can install it with `pip install \"sacrebleu>=1.4.12\"`." ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="http://www.cs.umd.edu/~snover/tercom/" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Sequence(datasets.Value("string" , id="sequence" ) , id="references" ), } ) , codebase_urls=["https://github.com/mjpost/sacreBLEU#ter"] , reference_urls=[ "https://github.com/jhclark/tercom", ] , ) def _A ( self :Any , lowerCAmelCase__ :Any , lowerCAmelCase__ :str , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , ) -> Optional[Any]: '''simple docstring''' snake_case_ : List[str] = len(references[0] ) if any(len(lowerCAmelCase__ ) != references_per_prediction for refs in references ): raise ValueError("Sacrebleu requires the same number of references for each prediction" ) snake_case_ : List[str] = [[refs[i] for refs in references] for i in range(lowerCAmelCase__ )] snake_case_ : List[str] = TER( normalized=lowerCAmelCase__ , no_punct=lowerCAmelCase__ , asian_support=lowerCAmelCase__ , case_sensitive=lowerCAmelCase__ , ) snake_case_ : Any = sb_ter.corpus_score(lowerCAmelCase__ , lowerCAmelCase__ ) return {"score": output.score, "num_edits": output.num_edits, "ref_length": output.ref_length}
656
0
'''simple docstring''' import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class A_ : """simple docstring""" def __init__( self :Optional[Any] , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Any=13 , lowerCAmelCase__ :Optional[Any]=7 , lowerCAmelCase__ :int=True , lowerCAmelCase__ :Tuple=True , lowerCAmelCase__ :List[Any]=99 , lowerCAmelCase__ :List[str]=32 , lowerCAmelCase__ :int=5 , lowerCAmelCase__ :Union[str, Any]=4 , lowerCAmelCase__ :Union[str, Any]=37 , lowerCAmelCase__ :Optional[int]="gelu" , lowerCAmelCase__ :List[str]=0.1 , lowerCAmelCase__ :int=0.1 , lowerCAmelCase__ :List[str]=50 , lowerCAmelCase__ :int=0.0_2 , lowerCAmelCase__ :Dict=True , lowerCAmelCase__ :List[str]=None , ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[Any] = parent snake_case_ : Any = batch_size snake_case_ : int = seq_length snake_case_ : Dict = is_training snake_case_ : str = use_input_mask snake_case_ : Any = vocab_size snake_case_ : List[str] = hidden_size snake_case_ : Tuple = num_hidden_layers snake_case_ : Optional[int] = num_attention_heads snake_case_ : List[Any] = intermediate_size snake_case_ : Optional[int] = hidden_act snake_case_ : Tuple = hidden_dropout_prob snake_case_ : Optional[Any] = attention_probs_dropout_prob snake_case_ : Union[str, Any] = max_position_embeddings snake_case_ : str = initializer_range snake_case_ : Dict = use_labels snake_case_ : List[str] = scope def _A ( self :List[Any] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ : List[Any] = None if self.use_input_mask: snake_case_ : str = random_attention_mask([self.batch_size, self.seq_length] ) if self.use_labels: snake_case_ : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ : Any = self.get_config() return config, input_ids, input_mask, token_labels def _A ( self :str ) -> List[str]: '''simple docstring''' return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=lowerCAmelCase__ , initializer_range=self.initializer_range , ) def _A ( self :Optional[Any] ) -> Union[str, Any]: '''simple docstring''' ( snake_case_ ) : List[str] = self.prepare_config_and_inputs() snake_case_ : Any = True snake_case_ : Optional[Any] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) snake_case_ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def _A ( self :Tuple , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :List[str] , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Tuple , **lowerCAmelCase__ :Dict , ) -> Optional[Any]: '''simple docstring''' snake_case_ : Optional[int] = BertGenerationEncoder(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case_ : int = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) snake_case_ : Any = model(lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _A ( self :str , lowerCAmelCase__ :Dict , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :str , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Optional[Any] , **lowerCAmelCase__ :Optional[Any] , ) -> int: '''simple docstring''' snake_case_ : str = True snake_case_ : Tuple = BertGenerationEncoder(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case_ : List[str] = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , encoder_attention_mask=lowerCAmelCase__ , ) snake_case_ : List[str] = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _A ( self :Union[str, Any] , lowerCAmelCase__ :str , lowerCAmelCase__ :int , lowerCAmelCase__ :Any , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :str , lowerCAmelCase__ :Optional[int] , **lowerCAmelCase__ :int , ) -> Tuple: '''simple docstring''' snake_case_ : Tuple = True snake_case_ : str = True snake_case_ : Union[str, Any] = BertGenerationDecoder(config=lowerCAmelCase__ ).to(lowerCAmelCase__ ).eval() # first forward pass snake_case_ : Dict = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , encoder_attention_mask=lowerCAmelCase__ , use_cache=lowerCAmelCase__ , ) snake_case_ : List[str] = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids snake_case_ : Union[str, Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case_ : List[str] = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and snake_case_ : List[str] = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case_ : str = torch.cat([input_mask, next_mask] , dim=-1 ) snake_case_ : str = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , encoder_attention_mask=lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ , )["hidden_states"][0] snake_case_ : List[Any] = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , encoder_attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__ , output_hidden_states=lowerCAmelCase__ , )["hidden_states"][0] # select random slice snake_case_ : List[str] = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case_ : Any = output_from_no_past[:, -3:, random_slice_idx].detach() snake_case_ : List[str] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) ) def _A ( self :int , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Dict , lowerCAmelCase__ :int , lowerCAmelCase__ :int , *lowerCAmelCase__ :Optional[Any] , ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = BertGenerationDecoder(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case_ : List[str] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _A ( self :Tuple ) -> List[Any]: '''simple docstring''' snake_case_ : int = self.prepare_config_and_inputs() snake_case_ : List[Any] = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class A_ (a_ , a_ , a_ , unittest.TestCase ): """simple docstring""" a__ = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () a__ = (BertGenerationDecoder,) if is_torch_available() else () a__ = ( {'''feature-extraction''': BertGenerationEncoder, '''text-generation''': BertGenerationDecoder} if is_torch_available() else {} ) def _A ( self :int ) -> str: '''simple docstring''' snake_case_ : List[str] = BertGenerationEncoderTester(self ) snake_case_ : str = ConfigTester(self , config_class=lowerCAmelCase__ , hidden_size=37 ) def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() def _A ( self :int ) -> List[Any]: '''simple docstring''' snake_case_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def _A ( self :Tuple ) -> Union[str, Any]: '''simple docstring''' snake_case_ : str = self.model_tester.prepare_config_and_inputs() snake_case_ : int = "bert" self.model_tester.create_and_check_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def _A ( self :List[str] ) -> Any: '''simple docstring''' snake_case_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*lowerCAmelCase__ ) def _A ( self :List[Any] ) -> Dict: '''simple docstring''' snake_case_ : int = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*lowerCAmelCase__ ) def _A ( self :Optional[Any] ) -> Any: '''simple docstring''' ( snake_case_ ) : List[Any] = self.model_tester.prepare_config_and_inputs_for_decoder() snake_case_ : Optional[Any] = None self.model_tester.create_and_check_model_as_decoder( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ) def _A ( self :Optional[Any] ) -> Dict: '''simple docstring''' snake_case_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*lowerCAmelCase__ ) @slow def _A ( self :Optional[Any] ) -> str: '''simple docstring''' snake_case_ : int = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) self.assertIsNotNone(lowerCAmelCase__ ) @require_torch class A_ (unittest.TestCase ): """simple docstring""" @slow def _A ( self :int ) -> int: '''simple docstring''' snake_case_ : Optional[int] = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) snake_case_ : str = torch.tensor([[101, 7_592, 1_010, 2_026, 3_899, 2_003, 10_140, 102]] ) with torch.no_grad(): snake_case_ : Optional[Any] = model(lowerCAmelCase__ )[0] snake_case_ : Optional[int] = torch.Size([1, 8, 1_024] ) self.assertEqual(output.shape , lowerCAmelCase__ ) snake_case_ : str = torch.tensor( [[[0.1_7_7_5, 0.0_0_8_3, -0.0_3_2_1], [1.6_0_0_2, 0.1_2_8_7, 0.3_9_1_2], [2.1_4_7_3, 0.5_7_9_1, 0.6_0_6_6]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowerCAmelCase__ , atol=1E-4 ) ) @require_torch class A_ (unittest.TestCase ): """simple docstring""" @slow def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : str = BertGenerationDecoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) snake_case_ : Optional[Any] = torch.tensor([[101, 7_592, 1_010, 2_026, 3_899, 2_003, 10_140, 102]] ) with torch.no_grad(): snake_case_ : List[Any] = model(lowerCAmelCase__ )[0] snake_case_ : Optional[Any] = torch.Size([1, 8, 50_358] ) self.assertEqual(output.shape , lowerCAmelCase__ ) snake_case_ : Optional[Any] = torch.tensor( [[[-0.5_7_8_8, -2.5_9_9_4, -3.7_0_5_4], [0.0_4_3_8, 4.7_9_9_7, 1.8_7_9_5], [1.5_8_6_2, 6.6_4_0_9, 4.4_6_3_8]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowerCAmelCase__ , atol=1E-4 ) )
707
'''simple docstring''' from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : Any = { "repo_name": ["test_repo1", "test_repo2", "test_repo3"], "path": ["test_1.py", "test_2.py", "unit_test.py"], "content": ["a " * 20, "a " * 30, "b " * 7], } snake_case_ : int = Dataset.from_dict(__magic_name__ ) return dataset class A_ (a_ ): """simple docstring""" def _A ( self :List[str] ) -> str: '''simple docstring''' snake_case_ : Union[str, Any] = get_dataset() snake_case_ : Optional[int] = make_duplicate_clusters(lowerCAmelCase__ , 0.8_5 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case_ : Optional[int] = get_dataset() snake_case_, snake_case_ : List[Any] = deduplicate_dataset(lowerCAmelCase__ ) self.assertEqual(len(lowerCAmelCase__ ) , 2 ) print(lowerCAmelCase__ ) self.assertEqual(duplicate_clusters[0][0]["copies"] , 2 ) self.assertEqual(duplicate_clusters[0][0]["is_extreme"] , lowerCAmelCase__ )
656
0
'''simple docstring''' import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class A_ (a_ , a_ , a_ ): """simple docstring""" @register_to_config def __init__( self :int , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :float , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :str , lowerCAmelCase__ :bool = False , ) -> int: '''simple docstring''' super().__init__() snake_case_ : Union[str, Any] = nn.Embedding(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : Dict = nn.Embedding(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : Any = False snake_case_ : Optional[int] = nn.Dropout(p=lowerCAmelCase__ ) snake_case_ : int = TaConfig( vocab_size=lowerCAmelCase__ , d_model=lowerCAmelCase__ , num_heads=lowerCAmelCase__ , d_kv=lowerCAmelCase__ , d_ff=lowerCAmelCase__ , dropout_rate=lowerCAmelCase__ , feed_forward_proj=lowerCAmelCase__ , is_decoder=lowerCAmelCase__ , is_encoder_decoder=lowerCAmelCase__ , ) snake_case_ : str = nn.ModuleList() for lyr_num in range(lowerCAmelCase__ ): snake_case_ : List[Any] = TaBlock(lowerCAmelCase__ ) self.encoders.append(lowerCAmelCase__ ) snake_case_ : Dict = TaLayerNorm(lowerCAmelCase__ ) snake_case_ : List[Any] = nn.Dropout(p=lowerCAmelCase__ ) def _A ( self :Optional[Any] , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Any ) -> Dict: '''simple docstring''' snake_case_ : Dict = self.token_embedder(lowerCAmelCase__ ) snake_case_ : Tuple = encoder_input_tokens.shape[1] snake_case_ : Union[str, Any] = torch.arange(lowerCAmelCase__ , device=encoder_input_tokens.device ) x += self.position_encoding(lowerCAmelCase__ ) snake_case_ : List[str] = self.dropout_pre(lowerCAmelCase__ ) # inverted the attention mask snake_case_ : int = encoder_input_tokens.size() snake_case_ : str = self.get_extended_attention_mask(lowerCAmelCase__ , lowerCAmelCase__ ) for lyr in self.encoders: snake_case_ : Optional[int] = lyr(lowerCAmelCase__ , lowerCAmelCase__ )[0] snake_case_ : Any = self.layer_norm(lowerCAmelCase__ ) return self.dropout_post(lowerCAmelCase__ ), encoder_inputs_mask
708
'''simple docstring''' from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable __lowerCamelCase : Dict = { '''configuration_gpt_neox_japanese''': ['''GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoXJapaneseConfig'''], '''tokenization_gpt_neox_japanese''': ['''GPTNeoXJapaneseTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ '''GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoXJapaneseForCausalLM''', '''GPTNeoXJapaneseLayer''', '''GPTNeoXJapaneseModel''', '''GPTNeoXJapanesePreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neox_japanese import GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXJapaneseConfig from .tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox_japanese import ( GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseLayer, GPTNeoXJapaneseModel, GPTNeoXJapanesePreTrainedModel, ) else: import sys __lowerCamelCase : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
656
0
'''simple docstring''' import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline __lowerCamelCase : List[str] = argparse.ArgumentParser('''Stable Diffusion script with intel optimization''', add_help=False) parser.add_argument('''--dpm''', action='''store_true''', help='''Enable DPMSolver or not''') parser.add_argument('''--steps''', default=None, type=int, help='''Num inference steps''') __lowerCamelCase : str = parser.parse_args() __lowerCamelCase : Any = '''cpu''' __lowerCamelCase : Union[str, Any] = '''a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings''' __lowerCamelCase : Union[str, Any] = '''path-to-your-trained-model''' __lowerCamelCase : int = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: __lowerCamelCase : List[str] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) __lowerCamelCase : List[str] = pipe.to(device) # to channels last __lowerCamelCase : List[Any] = pipe.unet.to(memory_format=torch.channels_last) __lowerCamelCase : List[Any] = pipe.vae.to(memory_format=torch.channels_last) __lowerCamelCase : Optional[Any] = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: __lowerCamelCase : Tuple = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex __lowerCamelCase : int = torch.randn(2, 4, 64, 64) __lowerCamelCase : List[str] = torch.rand(1) * 999 __lowerCamelCase : Tuple = torch.randn(2, 77, 768) __lowerCamelCase : List[str] = (sample, timestep, encoder_hidden_status) try: __lowerCamelCase : str = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: __lowerCamelCase : str = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) __lowerCamelCase : Union[str, Any] = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) __lowerCamelCase : List[Any] = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: __lowerCamelCase : Optional[Any] = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute __lowerCamelCase : List[str] = 666 __lowerCamelCase : Tuple = torch.Generator(device).manual_seed(seed) __lowerCamelCase : Any = {'''generator''': generator} if args.steps is not None: __lowerCamelCase : str = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): __lowerCamelCase : List[Any] = pipe(prompt, **generate_kwargs).images[0] # save image image.save('''generated.png''')
709
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ )-> list[int]: """simple docstring""" if length <= 0 or not isinstance(__magic_name__ ,__magic_name__ ): raise ValueError("Length must be a positive integer." ) return [n * (2 * n - 1) for n in range(__magic_name__ )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=10))
656
0
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> int: """simple docstring""" if len(__magic_name__ ) != len(__magic_name__ ): raise ValueError("The length of profit and weight must be same." ) if max_weight <= 0: raise ValueError("max_weight must greater than zero." ) if any(p < 0 for p in profit ): raise ValueError("Profit can not be negative." ) if any(w < 0 for w in weight ): raise ValueError("Weight can not be negative." ) # List created to store profit gained for the 1kg in case of each weight # respectively. Calculate and append profit/weight for each element. snake_case_ : Optional[Any] = [p / w for p, w in zip(__magic_name__ ,__magic_name__ )] # Creating a copy of the list and sorting profit/weight in ascending order snake_case_ : List[Any] = sorted(__magic_name__ ) # declaring useful variables snake_case_ : Optional[int] = len(__magic_name__ ) snake_case_ : Union[str, Any] = 0 snake_case_ : int = 0 snake_case_ : Union[str, Any] = 0 # loop till the total weight do not reach max limit e.g. 15 kg and till i<length while limit <= max_weight and i < length: # flag value for encountered greatest element in sorted_profit_by_weight snake_case_ : List[Any] = sorted_profit_by_weight[length - i - 1] snake_case_ : str = profit_by_weight.index(__magic_name__ ) snake_case_ : Union[str, Any] = -1 # check if the weight encountered is less than the total weight # encountered before. if max_weight - limit >= weight[index]: limit += weight[index] # Adding profit gained for the given weight 1 === # weight[index]/weight[index] gain += 1 * profit[index] else: # Since the weight encountered is greater than limit, therefore take the # required number of remaining kgs and calculate profit for it. # weight remaining / weight[index] gain += (max_weight - limit) / weight[index] * profit[index] break i += 1 return gain if __name__ == "__main__": print( '''Input profits, weights, and then max_weight (all positive ints) separated by ''' '''spaces.''' ) __lowerCamelCase : Any = [int(x) for x in input('''Input profits separated by spaces: ''').split()] __lowerCamelCase : str = [int(x) for x in input('''Input weights separated by spaces: ''').split()] __lowerCamelCase : Optional[Any] = int(input('''Max weight allowed: ''')) # Function Call calc_profit(profit, weight, max_weight)
710
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def __UpperCAmelCase ( __magic_name__=None )-> List[str]: """simple docstring""" if subparsers is not None: snake_case_ : List[str] = subparsers.add_parser("test" ) else: snake_case_ : List[Any] = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" ,default=__magic_name__ ,help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) ,) if subparsers is not None: parser.set_defaults(func=__magic_name__ ) return parser def __UpperCAmelCase ( __magic_name__ )-> Tuple: """simple docstring""" snake_case_ : Optional[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: snake_case_ : str = script_name else: snake_case_ : Any = F'''--config_file={args.config_file} {script_name}''' snake_case_ : Union[str, Any] = ["accelerate-launch"] + test_args.split() snake_case_ : Optional[int] = execute_subprocess_async(__magic_name__ ,env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : Dict = test_command_parser() snake_case_ : Dict = parser.parse_args() test_command(__magic_name__ ) if __name__ == "__main__": main()
656
0
'''simple docstring''' import inspect import tempfile import unittest from huggingface_hub import hf_hub_download from transformers import is_torch_available from transformers.testing_utils import is_flaky, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin __lowerCamelCase : str = 1E-4 if is_torch_available(): import torch from transformers import AutoformerConfig, AutoformerForPrediction, AutoformerModel from transformers.models.autoformer.modeling_autoformer import AutoformerDecoder, AutoformerEncoder @require_torch class A_ : """simple docstring""" def __init__( self :List[str] , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Optional[int]=16 , lowerCAmelCase__ :Tuple=13 , lowerCAmelCase__ :int=7 , lowerCAmelCase__ :Dict=14 , lowerCAmelCase__ :int=10 , lowerCAmelCase__ :str=19 , lowerCAmelCase__ :Optional[Any]=5 , lowerCAmelCase__ :int=4 , lowerCAmelCase__ :List[Any]=True , lowerCAmelCase__ :Any=16 , lowerCAmelCase__ :Optional[int]=2 , lowerCAmelCase__ :str=4 , lowerCAmelCase__ :Union[str, Any]=4 , lowerCAmelCase__ :int="gelu" , lowerCAmelCase__ :Dict=0.1 , lowerCAmelCase__ :int=0.1 , lowerCAmelCase__ :Optional[int]=[1, 2, 3, 4, 5] , lowerCAmelCase__ :Dict=25 , lowerCAmelCase__ :Any=5 , ) -> Tuple: '''simple docstring''' snake_case_ : Optional[Any] = d_model snake_case_ : Any = parent snake_case_ : Tuple = batch_size snake_case_ : Tuple = prediction_length snake_case_ : List[str] = context_length snake_case_ : List[str] = cardinality snake_case_ : Optional[Any] = num_time_features snake_case_ : Dict = lags_sequence snake_case_ : Union[str, Any] = embedding_dimension snake_case_ : List[str] = is_training snake_case_ : int = hidden_size snake_case_ : int = num_hidden_layers snake_case_ : Dict = num_attention_heads snake_case_ : Optional[Any] = intermediate_size snake_case_ : List[str] = hidden_act snake_case_ : int = hidden_dropout_prob snake_case_ : str = attention_probs_dropout_prob snake_case_ : Optional[int] = context_length snake_case_ : Dict = prediction_length + label_length snake_case_ : Dict = label_length snake_case_ : Union[str, Any] = moving_average snake_case_ : int = autocorrelation_factor def _A ( self :Union[str, Any] ) -> str: '''simple docstring''' return AutoformerConfig( d_model=self.d_model , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , prediction_length=self.prediction_length , context_length=self.context_length , label_length=self.label_length , lags_sequence=self.lags_sequence , num_time_features=self.num_time_features , num_static_categorical_features=1 , cardinality=[self.cardinality] , embedding_dimension=[self.embedding_dimension] , moving_average=self.moving_average , ) def _A ( self :Optional[int] , lowerCAmelCase__ :Union[str, Any] ) -> List[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = config.context_length + max(config.lags_sequence ) snake_case_ : Union[str, Any] = ids_tensor([self.batch_size, 1] , config.cardinality[0] ) snake_case_ : Union[str, Any] = floats_tensor([self.batch_size, _past_length, config.num_time_features] ) snake_case_ : Any = floats_tensor([self.batch_size, _past_length] ) snake_case_ : List[str] = floats_tensor([self.batch_size, _past_length] ) > 0.5 # decoder inputs snake_case_ : Tuple = floats_tensor([self.batch_size, config.prediction_length, config.num_time_features] ) snake_case_ : Optional[Any] = floats_tensor([self.batch_size, config.prediction_length] ) snake_case_ : Optional[Any] = { "past_values": past_values, "static_categorical_features": static_categorical_features, "past_time_features": past_time_features, "past_observed_mask": past_observed_mask, "future_time_features": future_time_features, "future_values": future_values, } return inputs_dict def _A ( self :Optional[int] ) -> Dict: '''simple docstring''' snake_case_ : Optional[int] = self.get_config() snake_case_ : Any = self.prepare_autoformer_inputs_dict(lowerCAmelCase__ ) return config, inputs_dict def _A ( self :Union[str, Any] ) -> Optional[Any]: '''simple docstring''' snake_case_ : Optional[int] = self.prepare_config_and_inputs() return config, inputs_dict def _A ( self :Optional[int] , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :Optional[Any] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : int = AutoformerModel(config=lowerCAmelCase__ ).to(lowerCAmelCase__ ).eval() snake_case_ : Union[str, Any] = model(**lowerCAmelCase__ ) snake_case_ : Dict = outputs.encoder_last_hidden_state snake_case_ : Optional[Any] = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: snake_case_ : str = model.get_encoder() encoder.save_pretrained(lowerCAmelCase__ ) snake_case_ : List[str] = AutoformerEncoder.from_pretrained(lowerCAmelCase__ ).to(lowerCAmelCase__ ) snake_case_ : str = model.create_network_inputs(**lowerCAmelCase__ ) snake_case_ : Any = model.decomposition_layer(transformer_inputs[:, : config.context_length, ...] ) snake_case_ : List[str] = torch.cat( (transformer_inputs[:, : config.context_length, ...], feature[:, : config.context_length, ...]) , dim=-1 , ) snake_case_ : str = encoder(inputs_embeds=lowerCAmelCase__ )[0] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1E-3 ) snake_case_ : Optional[int] = ( torch.mean(transformer_inputs[:, : config.context_length, ...] , dim=1 ) .unsqueeze(1 ) .repeat(1 , config.prediction_length , 1 ) ) snake_case_ : List[Any] = torch.zeros( [transformer_inputs.shape[0], config.prediction_length, transformer_inputs.shape[2]] , device=enc_input.device , ) snake_case_ : int = torch.cat( ( torch.cat((seasonal_input[:, -config.label_length :, ...], zeros) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) snake_case_ : Any = torch.cat( ( torch.cat((trend_input[:, -config.label_length :, ...], mean) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) with tempfile.TemporaryDirectory() as tmpdirname: snake_case_ : Optional[int] = model.get_decoder() decoder.save_pretrained(lowerCAmelCase__ ) snake_case_ : int = AutoformerDecoder.from_pretrained(lowerCAmelCase__ ).to(lowerCAmelCase__ ) snake_case_ : Union[str, Any] = decoder( trend=lowerCAmelCase__ , inputs_embeds=lowerCAmelCase__ , encoder_hidden_states=lowerCAmelCase__ , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1E-3 ) @require_torch class A_ (a_ , a_ , unittest.TestCase ): """simple docstring""" a__ = (AutoformerModel, AutoformerForPrediction) if is_torch_available() else () a__ = (AutoformerForPrediction,) if is_torch_available() else () a__ = {'''feature-extraction''': AutoformerModel} if is_torch_available() else {} a__ = False a__ = False a__ = False a__ = False a__ = False a__ = False def _A ( self :int ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[int] = AutoformerModelTester(self ) snake_case_ : Optional[Any] = ConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ ) def _A ( self :Tuple ) -> Dict: '''simple docstring''' self.config_tester.run_common_tests() def _A ( self :Optional[int] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[Any] = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: snake_case_ : Any = model_class(lowerCAmelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCAmelCase__ ) snake_case_ : int = model_class.from_pretrained(lowerCAmelCase__ , output_loading_info=lowerCAmelCase__ ) self.assertEqual(info["missing_keys"] , [] ) def _A ( self :Optional[int] ) -> str: '''simple docstring''' snake_case_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*lowerCAmelCase__ ) @unittest.skip(reason="Model has no tokens embeddings" ) def _A ( self :Optional[int] ) -> Optional[Any]: '''simple docstring''' pass def _A ( self :Dict ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[str] = inspect.signature(getattr(lowerCAmelCase__ , "forward" ) ) # The main input is the name of the argument after `self` snake_case_ : Any = list(model_signature.parameters.keys() )[1] self.assertEqual(AutoformerModel.main_input_name , lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ : Optional[Any] = model_class(lowerCAmelCase__ ) snake_case_ : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ : Tuple = [*signature.parameters.keys()] snake_case_ : Optional[int] = [ "past_values", "past_time_features", "past_observed_mask", "static_categorical_features", "static_real_features", "future_values", "future_time_features", ] if model.__class__.__name__ in ["AutoformerForPrediction"]: expected_arg_names.append("future_observed_mask" ) expected_arg_names.extend( [ "decoder_attention_mask", "head_mask", "decoder_head_mask", "cross_attn_head_mask", "encoder_outputs", "past_key_values", "output_hidden_states", "output_attentions", "use_cache", "return_dict", ] ) self.assertListEqual(arg_names[: len(lowerCAmelCase__ )] , lowerCAmelCase__ ) def _A ( self :Dict ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ : List[str] = True snake_case_ : Dict = getattr(self.model_tester , "seq_length" , lowerCAmelCase__ ) snake_case_ : str = getattr(self.model_tester , "decoder_seq_length" , lowerCAmelCase__ ) snake_case_ : int = getattr(self.model_tester , "encoder_seq_length" , lowerCAmelCase__ ) snake_case_ : Tuple = getattr(self.model_tester , "d_model" , lowerCAmelCase__ ) snake_case_ : int = getattr(self.model_tester , "num_attention_heads" , lowerCAmelCase__ ) snake_case_ : Optional[int] = d_model // num_attention_heads for model_class in self.all_model_classes: snake_case_ : int = True snake_case_ : Dict = False snake_case_ : Optional[Any] = True snake_case_ : Optional[Any] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() with torch.no_grad(): snake_case_ : List[str] = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) ) snake_case_ : Optional[int] = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(lowerCAmelCase__ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] snake_case_ : str = True snake_case_ : Optional[int] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() with torch.no_grad(): snake_case_ : Any = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) ) snake_case_ : List[Any] = outputs.encoder_attentions self.assertEqual(len(lowerCAmelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) snake_case_ : int = len(lowerCAmelCase__ ) snake_case_ : Dict = 7 if "last_hidden_state" in outputs: correct_outlen += 1 if "trend" in outputs: correct_outlen += 1 if "past_key_values" in outputs: correct_outlen += 1 # past_key_values have been returned if "loss" in outputs: correct_outlen += 1 if "params" in outputs: correct_outlen += 1 self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ ) # decoder attentions snake_case_ : Any = outputs.decoder_attentions self.assertIsInstance(lowerCAmelCase__ , (list, tuple) ) self.assertEqual(len(lowerCAmelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # cross attentions snake_case_ : List[Any] = outputs.cross_attentions self.assertIsInstance(lowerCAmelCase__ , (list, tuple) ) self.assertEqual(len(lowerCAmelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(cross_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # Check attention is always last and order is fine snake_case_ : List[Any] = True snake_case_ : Union[str, Any] = True snake_case_ : Dict = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() with torch.no_grad(): snake_case_ : List[str] = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) ) self.assertEqual(out_len + 2 , len(lowerCAmelCase__ ) ) snake_case_ : Optional[Any] = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(lowerCAmelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) @is_flaky() def _A ( self :List[Any] ) -> List[str]: '''simple docstring''' super().test_retain_grad_hidden_states_attentions() def __UpperCAmelCase ( __magic_name__="train-batch.pt" )-> int: """simple docstring""" snake_case_ : Any = hf_hub_download(repo_id="hf-internal-testing/tourism-monthly-batch" ,filename=__magic_name__ ,repo_type="dataset" ) snake_case_ : str = torch.load(__magic_name__ ,map_location=__magic_name__ ) return batch @require_torch @slow class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Dict ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = AutoformerModel.from_pretrained("huggingface/autoformer-tourism-monthly" ).to(lowerCAmelCase__ ) snake_case_ : int = prepare_batch() with torch.no_grad(): snake_case_ : Dict = model( past_values=batch["past_values"] , past_time_features=batch["past_time_features"] , past_observed_mask=batch["past_observed_mask"] , static_categorical_features=batch["static_categorical_features"] , future_values=batch["future_values"] , future_time_features=batch["future_time_features"] , )[0] snake_case_ : int = torch.Size( (64, model.config.prediction_length + model.config.label_length, model.config.feature_size) ) self.assertEqual(output.shape , lowerCAmelCase__ ) snake_case_ : Any = torch.tensor( [[0.3_5_9_3, -1.3_3_9_8, 0.6_3_3_0], [0.2_2_7_9, 1.5_3_9_6, -0.1_7_9_2], [0.0_4_5_0, 1.3_2_2_5, -0.2_3_3_5]] , device=lowerCAmelCase__ ) self.assertTrue(torch.allclose(output[0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) def _A ( self :Union[str, Any] ) -> str: '''simple docstring''' snake_case_ : str = AutoformerForPrediction.from_pretrained("huggingface/autoformer-tourism-monthly" ).to(lowerCAmelCase__ ) snake_case_ : Tuple = prepare_batch("val-batch.pt" ) with torch.no_grad(): snake_case_ : Any = model( past_values=batch["past_values"] , past_time_features=batch["past_time_features"] , past_observed_mask=batch["past_observed_mask"] , static_categorical_features=batch["static_categorical_features"] , ).encoder_last_hidden_state snake_case_ : Optional[int] = torch.Size((64, model.config.context_length, model.config.d_model) ) self.assertEqual(output.shape , lowerCAmelCase__ ) snake_case_ : List[Any] = torch.tensor( [[-0.0_7_3_4, -0.9_0_3_6, 0.8_3_5_8], [4.7_1_8_6, 2.4_1_1_3, 1.9_5_8_1], [1.7_9_5_3, 2.3_5_5_8, 1.2_9_7_0]] , device=lowerCAmelCase__ ) self.assertTrue(torch.allclose(output[0, :3, :3] , lowerCAmelCase__ , atol=lowerCAmelCase__ ) ) def _A ( self :List[str] ) -> List[Any]: '''simple docstring''' snake_case_ : str = AutoformerForPrediction.from_pretrained("huggingface/autoformer-tourism-monthly" ).to(lowerCAmelCase__ ) snake_case_ : int = prepare_batch("val-batch.pt" ) with torch.no_grad(): snake_case_ : Union[str, Any] = model.generate( static_categorical_features=batch["static_categorical_features"] , past_time_features=batch["past_time_features"] , past_values=batch["past_values"] , future_time_features=batch["future_time_features"] , past_observed_mask=batch["past_observed_mask"] , ) snake_case_ : Tuple = torch.Size((64, model.config.num_parallel_samples, model.config.prediction_length) ) self.assertEqual(outputs.sequences.shape , lowerCAmelCase__ ) snake_case_ : str = torch.tensor([3_130.6_763, 4_056.5_293, 7_053.0_786] , device=lowerCAmelCase__ ) snake_case_ : Union[str, Any] = outputs.sequences.mean(dim=1 ) self.assertTrue(torch.allclose(mean_prediction[0, -3:] , lowerCAmelCase__ , rtol=1E-1 ) )
711
'''simple docstring''' from scipy.stats import spearmanr import datasets __lowerCamelCase : str = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' __lowerCamelCase : int = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' __lowerCamelCase : List[str] = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): """simple docstring""" def _A ( self :str ) -> Union[str, Any]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } ) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"] , ) def _A ( self :Optional[int] , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :Optional[Any]=False ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[Any] = spearmanr(lowerCAmelCase__ , lowerCAmelCase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
656
0
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Optional[int]: """simple docstring""" if b == 0: return 1 if (b % 2) == 0: return actual_power(__magic_name__ ,int(b / 2 ) ) * actual_power(__magic_name__ ,int(b / 2 ) ) else: return a * actual_power(__magic_name__ ,int(b / 2 ) ) * actual_power(__magic_name__ ,int(b / 2 ) ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> float: """simple docstring""" if b < 0: return 1 / actual_power(__magic_name__ ,__magic_name__ ) return actual_power(__magic_name__ ,__magic_name__ ) if __name__ == "__main__": print(power(-2, -3))
712
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): __lowerCamelCase : Optional[Any] = get_tests_dir('''fixtures/test_sentencepiece.model''') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right __lowerCamelCase : str = 128022 __lowerCamelCase : List[Any] = 128028 @require_sentencepiece class A_ (a_ , unittest.TestCase ): """simple docstring""" a__ = MaMaaaTokenizer a__ = False a__ = False a__ = True def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' super().setUp() snake_case_ : int = ["</s>", "<unk>", "▁This", "▁is", "▁a", "▁t", "est", "\u0120", "<pad>"] snake_case_ : Any = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) snake_case_ : Optional[int] = Path(self.tmpdirname ) save_json(lowerCAmelCase__ , save_dir / VOCAB_FILES_NAMES["vocab_file"] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(lowerCAmelCase__ , save_dir / VOCAB_FILES_NAMES["spm_file"] ) snake_case_ : Union[str, Any] = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def _A ( self :List[Any] , **lowerCAmelCase__ :List[Any] ) -> str: '''simple docstring''' return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def _A ( self :Optional[int] , lowerCAmelCase__ :Any ) -> Optional[int]: '''simple docstring''' return ( "This is a test", "This is a test", ) def _A ( self :List[str] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : str = "</s>" snake_case_ : Union[str, Any] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase__ ) , lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = self.get_tokenizer() snake_case_ : Any = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "</s>" ) self.assertEqual(vocab_keys[1] , "<unk>" ) self.assertEqual(vocab_keys[-1] , "<s>" ) self.assertEqual(len(lowerCAmelCase__ ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip("Skip this test while all models are still to be uploaded." ) def _A ( self :List[Any] ) -> Union[str, Any]: '''simple docstring''' pass def _A ( self :Optional[int] ) -> int: '''simple docstring''' snake_case_ : int = self.get_tokenizer() snake_case_ : List[str] = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowerCAmelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [2, 3, 4, 5, 6] , ) snake_case_ : Any = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(lowerCAmelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) snake_case_ : Any = tokenizer.convert_tokens_to_string(lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , "This is a test" ) @slow def _A ( self :Any ) -> List[Any]: '''simple docstring''' snake_case_ : int = {"input_ids": [[128_022, 110_108, 397, 11, 38_272, 2_247, 124_811, 285, 18_105, 1_586, 207, 7, 39_534, 4_428, 397, 1_019, 18_105, 1_586, 207, 7, 41_337, 16_786, 241, 7, 20_214, 17, 125_690, 10_398, 7, 44_378, 58_069, 68_342, 7_798, 7_343, 11, 299, 33_310, 4, 158, 37_350, 94_077, 4_569, 299, 33_310, 90, 4, 52_840, 290, 4, 31_270, 112, 299, 682, 4, 52_840, 39_953, 14_079, 193, 52_519, 90_894, 17_894, 120_697, 11, 40_445, 551, 17, 1_019, 52_519, 90_894, 17_756, 963, 11, 40_445, 480, 17, 9_792, 1_120, 5_173, 1_393, 6_240, 16_786, 241, 120_996, 28, 1_245, 1_393, 118_240, 11_123, 1_019, 93_612, 2_691, 10_618, 98_058, 120_409, 1_928, 279, 4, 40_683, 367, 178, 207, 1_019, 103, 103_121, 506, 65_296, 5, 2], [128_022, 21_217, 367, 117, 125_450, 128, 719, 7, 7_308, 40, 93_612, 12_669, 1_116, 16_704, 71, 17_785, 3_699, 15_592, 35, 144, 9_584, 241, 11_943, 713, 950, 799, 2_247, 88_427, 150, 149, 118_813, 120_706, 1_019, 106_906, 81_518, 28, 1_224, 22_799, 397, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [128_022, 1_658, 123_311, 5_155, 5_578, 4_722, 279, 14_947, 2_366, 1_120, 1_197, 14, 1_348, 9_232, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase__ , model_name="facebook/m2m100_418M" , revision="c168bae485c864188cf9aa0e4108b0b6934dc91e" , ) @require_torch @require_sentencepiece @require_tokenizers class A_ (unittest.TestCase ): """simple docstring""" a__ = '''facebook/m2m100_418M''' a__ = [ '''In my opinion, there are two levels of response from the French government.''', '''NSA Affair Emphasizes Complete Lack of Debate on Intelligence''', ] a__ = [ '''Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.''', '''L\'affaire NSA souligne l\'absence totale de débat sur le renseignement''', ] # fmt: off a__ = [EN_CODE, 593, 1949, 115781, 4, 71586, 4234, 60633, 126233, 432, 123808, 15592, 1197, 117132, 120618, 5, 2] @classmethod def _A ( cls :str ) -> int: '''simple docstring''' snake_case_ : MaMaaaTokenizer = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en" , tgt_lang="fr" ) snake_case_ : List[str] = 1 return cls def _A ( self :Tuple ) -> Union[str, Any]: '''simple docstring''' self.assertEqual(self.tokenizer.get_lang_id("ar" ) , 128_006 ) self.assertEqual(self.tokenizer.get_lang_id("en" ) , 128_022 ) self.assertEqual(self.tokenizer.get_lang_id("ro" ) , 128_076 ) self.assertEqual(self.tokenizer.get_lang_id("mr" ) , 128_063 ) def _A ( self :Optional[int] ) -> List[str]: '''simple docstring''' snake_case_ : Dict = self.tokenizer.get_vocab() self.assertEqual(len(lowerCAmelCase__ ) , self.tokenizer.vocab_size ) self.assertEqual(vocab["<unk>"] , 3 ) self.assertIn(self.tokenizer.get_lang_token("en" ) , lowerCAmelCase__ ) def _A ( self :Any ) -> Dict: '''simple docstring''' snake_case_ : List[str] = "en" snake_case_ : Dict = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' self.assertIn(lowerCAmelCase__ , self.tokenizer.all_special_ids ) # fmt: off snake_case_ : Dict = [FR_CODE, 5_364, 82, 8_642, 4, 294, 47, 8, 14_028, 136, 3_286, 9_706, 6, 90_797, 6, 144_012, 162, 88_128, 30_061, 5, 2] # fmt: on snake_case_ : List[str] = self.tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) snake_case_ : str = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertNotIn(self.tokenizer.eos_token , lowerCAmelCase__ ) def _A ( self :Tuple ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = tempfile.mkdtemp() snake_case_ : int = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(lowerCAmelCase__ ) snake_case_ : List[str] = MaMaaaTokenizer.from_pretrained(lowerCAmelCase__ ) self.assertDictEqual(new_tok.lang_token_to_id , lowerCAmelCase__ ) @require_torch def _A ( self :Optional[Any] ) -> str: '''simple docstring''' snake_case_ : Union[str, Any] = "en" snake_case_ : Tuple = "fr" snake_case_ : Optional[int] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=lowerCAmelCase__ , return_tensors="pt" ) snake_case_ : Dict = shift_tokens_right( batch["labels"] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: snake_case_ : str = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : List[str] = "mr" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) snake_case_ : int = "zh" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def _A ( self :str ) -> int: '''simple docstring''' snake_case_ : Dict = "mr" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) snake_case_ : Tuple = "zh" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Optional[int] = self.tokenizer._build_translation_inputs("A test" , return_tensors="pt" , src_lang="en" , tgt_lang="ar" ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { # en_XX, A, test, EOS "input_ids": [[128_022, 58, 4_183, 2]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 128_006, } , )
656
0
'''simple docstring''' import json import os from pathlib import Path import pytest from datasets.download.download_config import DownloadConfig from datasets.download.download_manager import DownloadManager from datasets.utils.file_utils import hash_url_to_filename __lowerCamelCase : List[Any] = '''http://www.mocksite.com/file1.txt''' __lowerCamelCase : Dict = '''"text": ["foo", "foo"]''' __lowerCamelCase : str = '''6d8ce9aa78a471c7477201efbeabd3bb01ac2e7d100a6dc024ba1608361f90a8''' class A_ : """simple docstring""" a__ = 200 a__ = {'''Content-Length''': '''100'''} a__ = {} def _A ( self :Dict , **lowerCAmelCase__ :Optional[int] ) -> List[Any]: '''simple docstring''' return [bytes(lowerCAmelCase__ , "utf-8" )] def __UpperCAmelCase ( *__magic_name__ ,**__magic_name__ )-> str: """simple docstring""" return MockResponse() @pytest.mark.parametrize("urls_type" ,[str, list, dict] ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Dict: """simple docstring""" import requests monkeypatch.setattr(__magic_name__ ,"request" ,__magic_name__ ) snake_case_ : Any = URL if issubclass(__magic_name__ ,__magic_name__ ): snake_case_ : Dict = url elif issubclass(__magic_name__ ,__magic_name__ ): snake_case_ : Optional[Any] = [url] elif issubclass(__magic_name__ ,__magic_name__ ): snake_case_ : Tuple = {"train": url} snake_case_ : Optional[Any] = "dummy" snake_case_ : Dict = "downloads" snake_case_ : int = tmp_path snake_case_ : List[str] = DownloadConfig( cache_dir=os.path.join(__magic_name__ ,__magic_name__ ) ,use_etag=__magic_name__ ,) snake_case_ : List[Any] = DownloadManager(dataset_name=__magic_name__ ,download_config=__magic_name__ ) snake_case_ : Any = dl_manager.download(__magic_name__ ) snake_case_ : int = urls for downloaded_paths in [downloaded_paths]: if isinstance(__magic_name__ ,__magic_name__ ): snake_case_ : int = [downloaded_paths] snake_case_ : int = [urls] elif isinstance(__magic_name__ ,__magic_name__ ): assert "train" in downloaded_paths.keys() snake_case_ : Tuple = downloaded_paths.values() snake_case_ : List[str] = urls.values() assert downloaded_paths for downloaded_path, input_url in zip(__magic_name__ ,__magic_name__ ): assert downloaded_path == dl_manager.downloaded_paths[input_url] snake_case_ : Union[str, Any] = Path(__magic_name__ ) snake_case_ : Tuple = downloaded_path.parts assert parts[-1] == HASH assert parts[-2] == cache_subdir assert downloaded_path.exists() snake_case_ : List[Any] = downloaded_path.read_text() assert content == CONTENT snake_case_ : Tuple = downloaded_path.with_suffix(".json" ) assert metadata_downloaded_path.exists() snake_case_ : str = json.loads(metadata_downloaded_path.read_text() ) assert metadata_content == {"url": URL, "etag": None} @pytest.mark.parametrize("paths_type" ,[str, list, dict] ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Dict: """simple docstring""" snake_case_ : Tuple = str(__magic_name__ ) if issubclass(__magic_name__ ,__magic_name__ ): snake_case_ : Optional[Any] = filename elif issubclass(__magic_name__ ,__magic_name__ ): snake_case_ : str = [filename] elif issubclass(__magic_name__ ,__magic_name__ ): snake_case_ : List[Any] = {"train": filename} snake_case_ : str = "dummy" snake_case_ : int = xz_file.parent snake_case_ : str = "extracted" snake_case_ : Optional[Any] = DownloadConfig( cache_dir=__magic_name__ ,use_etag=__magic_name__ ,) snake_case_ : str = DownloadManager(dataset_name=__magic_name__ ,download_config=__magic_name__ ) snake_case_ : int = dl_manager.extract(__magic_name__ ) snake_case_ : int = paths for extracted_paths in [extracted_paths]: if isinstance(__magic_name__ ,__magic_name__ ): snake_case_ : List[str] = [extracted_paths] snake_case_ : int = [paths] elif isinstance(__magic_name__ ,__magic_name__ ): assert "train" in extracted_paths.keys() snake_case_ : List[Any] = extracted_paths.values() snake_case_ : Optional[Any] = paths.values() assert extracted_paths for extracted_path, input_path in zip(__magic_name__ ,__magic_name__ ): assert extracted_path == dl_manager.extracted_paths[input_path] snake_case_ : str = Path(__magic_name__ ) snake_case_ : int = extracted_path.parts assert parts[-1] == hash_url_to_filename(__magic_name__ ,etag=__magic_name__ ) assert parts[-2] == extracted_subdir assert extracted_path.exists() snake_case_ : Union[str, Any] = extracted_path.read_text() snake_case_ : Dict = text_file.read_text() assert extracted_file_content == expected_file_content def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" assert path.endswith(".jsonl" ) for num_items, line in enumerate(__magic_name__ ,start=1 ): snake_case_ : Union[str, Any] = json.loads(line.decode("utf-8" ) ) assert item.keys() == {"col_1", "col_2", "col_3"} assert num_items == 4 @pytest.mark.parametrize("archive_jsonl" ,["tar_jsonl_path", "zip_jsonl_path"] ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : List[str] = request.getfixturevalue(__magic_name__ ) snake_case_ : Dict = DownloadManager() for num_jsonl, (path, file) in enumerate(dl_manager.iter_archive(__magic_name__ ) ,start=1 ): _test_jsonl(__magic_name__ ,__magic_name__ ) assert num_jsonl == 2 @pytest.mark.parametrize("archive_nested_jsonl" ,["tar_nested_jsonl_path", "zip_nested_jsonl_path"] ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Any: """simple docstring""" snake_case_ : Optional[Any] = request.getfixturevalue(__magic_name__ ) snake_case_ : List[str] = DownloadManager() for num_tar, (path, file) in enumerate(dl_manager.iter_archive(__magic_name__ ) ,start=1 ): for num_jsonl, (subpath, subfile) in enumerate(dl_manager.iter_archive(__magic_name__ ) ,start=1 ): _test_jsonl(__magic_name__ ,__magic_name__ ) assert num_tar == 1 assert num_jsonl == 2 def __UpperCAmelCase ( __magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Union[str, Any] = DownloadManager() for num_file, file in enumerate(dl_manager.iter_files(__magic_name__ ) ,start=1 ): assert os.path.basename(__magic_name__ ) == ("test.txt" if num_file == 1 else "train.txt") assert num_file == 2
713
'''simple docstring''' import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py __lowerCamelCase : str = '''.''' # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) __lowerCamelCase : Tuple = [ '''Assert''', '''AssignVariableOp''', '''EmptyTensorList''', '''MergeV2Checkpoints''', '''ReadVariableOp''', '''ResourceGather''', '''RestoreV2''', '''SaveV2''', '''ShardedFilename''', '''StatefulPartitionedCall''', '''StaticRegexFullMatch''', '''VarHandleOp''', ] def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Tuple = SavedModel() snake_case_ : Dict = [] with open(os.path.join(__magic_name__ ,"utils" ,"tf_ops" ,"onnx.json" ) ) as f: snake_case_ : Dict = json.load(__magic_name__ )["opsets"] for i in range(1 ,opset + 1 ): onnx_ops.extend(onnx_opsets[str(__magic_name__ )] ) with open(__magic_name__ ,"rb" ) as f: saved_model.ParseFromString(f.read() ) snake_case_ : Tuple = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want snake_case_ : str = sorted(__magic_name__ ) snake_case_ : Optional[int] = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(__magic_name__ ) if strict and len(__magic_name__ ) > 0: raise Exception(F'''Found the following incompatible ops for the opset {opset}:\n''' + incompatible_ops ) elif len(__magic_name__ ) > 0: print(F'''Found the following incompatible ops for the opset {opset}:''' ) print(*__magic_name__ ,sep="\n" ) else: print(F'''The saved model {saved_model_path} can properly be converted with ONNX.''' ) if __name__ == "__main__": __lowerCamelCase : Any = argparse.ArgumentParser() parser.add_argument('''--saved_model_path''', help='''Path of the saved model to check (the .pb file).''') parser.add_argument( '''--opset''', default=12, type=int, help='''The ONNX opset against which the model has to be tested.''' ) parser.add_argument( '''--framework''', choices=['''onnx'''], default='''onnx''', help='''Frameworks against which to test the saved model.''' ) parser.add_argument( '''--strict''', action='''store_true''', help='''Whether make the checking strict (raise errors) or not (raise warnings)''' ) __lowerCamelCase : Dict = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
656
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCamelCase : List[str] = { '''configuration_jukebox''': [ '''JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''JukeboxConfig''', '''JukeboxPriorConfig''', '''JukeboxVQVAEConfig''', ], '''tokenization_jukebox''': ['''JukeboxTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Dict = [ '''JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST''', '''JukeboxModel''', '''JukeboxPreTrainedModel''', '''JukeboxVQVAE''', '''JukeboxPrior''', ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys __lowerCamelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
714
'''simple docstring''' import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal __lowerCamelCase : Optional[Any] = datasets.utils.logging.get_logger(__name__) __lowerCamelCase : List[str] = ['''names''', '''prefix'''] __lowerCamelCase : int = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] __lowerCamelCase : str = ['''encoding_errors''', '''on_bad_lines'''] __lowerCamelCase : Optional[Any] = ['''date_format'''] @dataclass class A_ (datasets.BuilderConfig ): """simple docstring""" a__ = "," a__ = None a__ = "infer" a__ = None a__ = None a__ = None a__ = None a__ = None a__ = True a__ = None a__ = None a__ = None a__ = None a__ = False a__ = None a__ = None a__ = None a__ = True a__ = True a__ = False a__ = True a__ = None a__ = "." a__ = None a__ = '"' a__ = 0 a__ = None a__ = None a__ = None a__ = None a__ = True a__ = True a__ = 0 a__ = True a__ = False a__ = None a__ = 10000 a__ = None a__ = "strict" a__ = "error" a__ = None def _A ( self :List[str] ) -> Any: '''simple docstring''' if self.delimiter is not None: snake_case_ : Tuple = self.delimiter if self.column_names is not None: snake_case_ : List[Any] = self.column_names @property def _A ( self :Optional[Any] ) -> int: '''simple docstring''' snake_case_ : Optional[int] = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class A_ (datasets.ArrowBasedBuilder ): """simple docstring""" a__ = CsvConfig def _A ( self :Optional[Any] ) -> Optional[Any]: '''simple docstring''' return datasets.DatasetInfo(features=self.config.features ) def _A ( self :Tuple , lowerCAmelCase__ :Dict ) -> List[Any]: '''simple docstring''' if not self.config.data_files: raise ValueError(F'''At least one data file must be specified, but got data_files={self.config.data_files}''' ) snake_case_ : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): snake_case_ : int = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ : List[str] = [files] snake_case_ : Tuple = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] snake_case_ : str = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ : str = [files] snake_case_ : Any = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def _A ( self :List[Any] , lowerCAmelCase__ :pa.Table ) -> pa.Table: '''simple docstring''' if self.config.features is not None: snake_case_ : int = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast snake_case_ : Optional[Any] = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example snake_case_ : Dict = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def _A ( self :Dict , lowerCAmelCase__ :Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str snake_case_ : str = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): snake_case_ : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): snake_case_ : Optional[int] = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(F'''Failed to read file \'{file}\' with error {type(lowerCAmelCase__ )}: {e}''' ) raise
656
0
import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def __UpperCAmelCase ( __magic_name__ ,__magic_name__=False )-> Dict: """simple docstring""" snake_case_ : Any = OmegaConf.load(__magic_name__ ) if display: print(yaml.dump(OmegaConf.to_container(__magic_name__ ) ) ) return config def __UpperCAmelCase ( __magic_name__ ,__magic_name__=None ,__magic_name__=None )-> List[str]: """simple docstring""" if conf_path is None: snake_case_ : str = "./model_checkpoints/vqgan_only.yaml" snake_case_ : List[Any] = load_config(__magic_name__ ,display=__magic_name__ ) snake_case_ : Dict = VQModel(**config.model.params ) if ckpt_path is None: snake_case_ : Union[str, Any] = "./model_checkpoints/vqgan_only.pt" snake_case_ : Optional[Any] = torch.load(__magic_name__ ,map_location=__magic_name__ ) if ".ckpt" in ckpt_path: snake_case_ : Optional[int] = sd["state_dict"] model.load_state_dict(__magic_name__ ,strict=__magic_name__ ) model.to(__magic_name__ ) del sd return model def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : Tuple = model.encode(__magic_name__ ) print(F'''VQGAN --- {model.__class__.__name__}: latent shape: {z.shape[2:]}''' ) snake_case_ : List[Any] = model.decode(__magic_name__ ) return xrec def __UpperCAmelCase ( __magic_name__ ,__magic_name__=False )-> List[Any]: """simple docstring""" snake_case_ : List[str] = string.rsplit("." ,1 ) if reload: snake_case_ : Dict = importlib.import_module(__magic_name__ ) importlib.reload(__magic_name__ ) return getattr(importlib.import_module(__magic_name__ ,package=__magic_name__ ) ,cls ) def __UpperCAmelCase ( __magic_name__ )-> Tuple: """simple docstring""" if "target" not in config: raise KeyError("Expected key `target` to instantiate." ) return get_obj_from_str(config["target"] )(**config.get("params" ,{} ) ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__=True ,__magic_name__=True )-> Optional[int]: """simple docstring""" snake_case_ : Any = instantiate_from_config(__magic_name__ ) if sd is not None: model.load_state_dict(__magic_name__ ) if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ )-> str: """simple docstring""" if ckpt: snake_case_ : Any = torch.load(__magic_name__ ,map_location="cpu" ) snake_case_ : Any = pl_sd["global_step"] print(F'''loaded model from global step {global_step}.''' ) else: snake_case_ : List[str] = {"state_dict": None} snake_case_ : List[Any] = None snake_case_ : str = load_model_from_config(config.model ,pl_sd["state_dict"] ,gpu=__magic_name__ ,eval_mode=__magic_name__ )["model"] return model, global_step
715
'''simple docstring''' import json import os import unittest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A_ (a_ , unittest.TestCase ): """simple docstring""" a__ = MgpstrTokenizer a__ = False a__ = {} a__ = False def _A ( self :List[str] ) -> List[str]: '''simple docstring''' super().setUp() # fmt: off snake_case_ : Dict = ["[GO]", "[s]", "0", "1", "2", "3", "4", "5", "6", "7", "8", "9", "a", "b", "c", "d", "e", "f", "g", "h", "i", "j", "k", "l", "m", "n", "o", "p", "q", "r", "s", "t", "u", "v", "w", "x", "y", "z"] # fmt: on snake_case_ : List[str] = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) snake_case_ : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase__ ) + "\n" ) def _A ( self :Optional[Any] , **lowerCAmelCase__ :Optional[Any] ) -> Dict: '''simple docstring''' return MgpstrTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def _A ( self :Dict , lowerCAmelCase__ :Any ) -> str: '''simple docstring''' snake_case_ : Dict = "tester" snake_case_ : Tuple = "tester" return input_text, output_text @unittest.skip("MGP-STR always lower cases letters." ) def _A ( self :Dict ) -> str: '''simple docstring''' pass def _A ( self :Tuple ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[str] = self.get_tokenizers(do_lower_case=lowerCAmelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): snake_case_ : Tuple = "[SPECIAL_TOKEN]" tokenizer.add_special_tokens({"cls_token": special_token} ) snake_case_ : str = tokenizer.encode([special_token] , add_special_tokens=lowerCAmelCase__ ) self.assertEqual(len(lowerCAmelCase__ ) , 1 ) snake_case_ : Tuple = tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) self.assertTrue(special_token not in decoded ) def _A ( self :int ) -> List[str]: '''simple docstring''' snake_case_ : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): snake_case_, snake_case_ : str = self.get_input_output_texts(lowerCAmelCase__ ) snake_case_ : Union[str, Any] = tokenizer.tokenize(lowerCAmelCase__ ) snake_case_ : List[Any] = tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) snake_case_ : Dict = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : List[str] = tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) self.assertNotEqual(len(lowerCAmelCase__ ) , 0 ) snake_case_ : List[str] = tokenizer.decode(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertEqual(text_a.replace(" " , "" ) , lowerCAmelCase__ ) @unittest.skip("MGP-STR tokenizer only handles one sequence." ) def _A ( self :Union[str, Any] ) -> Any: '''simple docstring''' pass @unittest.skip("inputs cannot be pretokenized in MgpstrTokenizer" ) def _A ( self :int ) -> Dict: '''simple docstring''' pass
656
0
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> str: """simple docstring""" if a < 0 or b < 0: raise ValueError("the value of both inputs must be positive" ) snake_case_ : Union[str, Any] = str(bin(__magic_name__ ) )[2:] # remove the leading "0b" snake_case_ : Optional[Any] = str(bin(__magic_name__ ) )[2:] # remove the leading "0b" snake_case_ : Any = max(len(__magic_name__ ) ,len(__magic_name__ ) ) return "0b" + "".join( str(int(char_a == "1" and char_b == "1" ) ) for char_a, char_b in zip(a_binary.zfill(__magic_name__ ) ,b_binary.zfill(__magic_name__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
716
'''simple docstring''' from __future__ import annotations import math import numpy as np from numpy.linalg import norm def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> float: """simple docstring""" return math.sqrt(sum(pow(a - b ,2 ) for a, b in zip(__magic_name__ ,__magic_name__ ) ) ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> list[list[list[float] | float]]: """simple docstring""" if dataset.ndim != value_array.ndim: snake_case_ : int = ( "Wrong input data's dimensions... " F'''dataset : {dataset.ndim}, value_array : {value_array.ndim}''' ) raise ValueError(__magic_name__ ) try: if dataset.shape[1] != value_array.shape[1]: snake_case_ : Dict = ( "Wrong input data's shape... " F'''dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}''' ) raise ValueError(__magic_name__ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError("Wrong shape" ) if dataset.dtype != value_array.dtype: snake_case_ : Dict = ( "Input data have different datatype... " F'''dataset : {dataset.dtype}, value_array : {value_array.dtype}''' ) raise TypeError(__magic_name__ ) snake_case_ : Optional[int] = [] for value in value_array: snake_case_ : List[str] = euclidean(__magic_name__ ,dataset[0] ) snake_case_ : int = dataset[0].tolist() for dataset_value in dataset[1:]: snake_case_ : Optional[Any] = euclidean(__magic_name__ ,__magic_name__ ) if dist > temp_dist: snake_case_ : Tuple = temp_dist snake_case_ : Optional[int] = dataset_value.tolist() answer.append([vector, dist] ) return answer def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> float: """simple docstring""" return np.dot(__magic_name__ ,__magic_name__ ) / (norm(__magic_name__ ) * norm(__magic_name__ )) if __name__ == "__main__": import doctest doctest.testmod()
656
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCamelCase : Any = logging.get_logger(__name__) __lowerCamelCase : List[Any] = { '''facebook/xlm-roberta-xl''': '''https://huggingface.co/facebook/xlm-roberta-xl/resolve/main/config.json''', '''facebook/xlm-roberta-xxl''': '''https://huggingface.co/facebook/xlm-roberta-xxl/resolve/main/config.json''', # See all XLM-RoBERTa-XL models at https://huggingface.co/models?filter=xlm-roberta-xl } class A_ (a_ ): """simple docstring""" a__ = '''xlm-roberta-xl''' def __init__( self :int , lowerCAmelCase__ :str=250_880 , lowerCAmelCase__ :Any=2_560 , lowerCAmelCase__ :Any=36 , lowerCAmelCase__ :Optional[Any]=32 , lowerCAmelCase__ :Optional[int]=10_240 , lowerCAmelCase__ :int="gelu" , lowerCAmelCase__ :List[Any]=0.1 , lowerCAmelCase__ :int=0.1 , lowerCAmelCase__ :Optional[Any]=514 , lowerCAmelCase__ :Tuple=1 , lowerCAmelCase__ :Union[str, Any]=0.0_2 , lowerCAmelCase__ :Any=1E-0_5 , lowerCAmelCase__ :Union[str, Any]=1 , lowerCAmelCase__ :List[Any]=0 , lowerCAmelCase__ :Tuple=2 , lowerCAmelCase__ :Union[str, Any]="absolute" , lowerCAmelCase__ :List[Any]=True , lowerCAmelCase__ :Dict=None , **lowerCAmelCase__ :Union[str, Any] , ) -> Tuple: '''simple docstring''' super().__init__(pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) snake_case_ : Tuple = vocab_size snake_case_ : Any = hidden_size snake_case_ : List[str] = num_hidden_layers snake_case_ : Dict = num_attention_heads snake_case_ : Tuple = hidden_act snake_case_ : int = intermediate_size snake_case_ : Optional[Any] = hidden_dropout_prob snake_case_ : Optional[Any] = attention_probs_dropout_prob snake_case_ : Union[str, Any] = max_position_embeddings snake_case_ : Dict = type_vocab_size snake_case_ : Any = initializer_range snake_case_ : str = layer_norm_eps snake_case_ : Optional[Any] = position_embedding_type snake_case_ : Optional[Any] = use_cache snake_case_ : List[str] = classifier_dropout class A_ (a_ ): """simple docstring""" @property def _A ( self :int ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": snake_case_ : Optional[Any] = {0: "batch", 1: "choice", 2: "sequence"} else: snake_case_ : Optional[Any] = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
717
'''simple docstring''' import fire from utils import calculate_rouge, save_json def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__=None ,**__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : int = [x.strip() for x in open(__magic_name__ ).readlines()] snake_case_ : Optional[int] = [x.strip() for x in open(__magic_name__ ).readlines()][: len(__magic_name__ )] snake_case_ : List[Any] = calculate_rouge(__magic_name__ ,__magic_name__ ,**__magic_name__ ) if save_path is not None: save_json(__magic_name__ ,__magic_name__ ,indent=__magic_name__ ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
656
0
'''simple docstring''' import requests from bsa import BeautifulSoup def __UpperCAmelCase ( __magic_name__ = "AAPL" )-> str: """simple docstring""" snake_case_ : int = F'''https://in.finance.yahoo.com/quote/{symbol}?s={symbol}''' snake_case_ : Dict = BeautifulSoup(requests.get(__magic_name__ ).text ,"html.parser" ) snake_case_ : Dict = "My(6px) Pos(r) smartphone_Mt(6px)" return soup.find("div" ,class_=class_ ).find("span" ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(f'''Current {symbol:<4} stock price is {stock_price(symbol):>8}''')
718
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) __lowerCamelCase : Optional[Any] = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', f'''encoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', f'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.weight''', f'''encoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.bias''', f'''encoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.weight''', f'''encoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.bias''', f'''encoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.encoder.layers.{i}.norm1.weight''', f'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.norm1.bias''', f'''encoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.weight''', f'''encoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.bias''', f'''encoder.layers.{i}.final_layer_norm.bias''')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', f'''decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', f'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.cross_attn.out_proj.weight''', f'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.cross_attn.out_proj.bias''', f'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.weight''', f'''decoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.bias''', f'''decoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.weight''', f'''decoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.bias''', f'''decoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm1.weight''', f'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm1.bias''', f'''decoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.weight''', f'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.bias''', f'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.weight''', f'''decoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.bias''', f'''decoder.layers.{i}.final_layer_norm.bias''')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qcontent_proj.weight''', f'''decoder.layers.{i}.sa_qcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kcontent_proj.weight''', f'''decoder.layers.{i}.sa_kcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qpos_proj.weight''', f'''decoder.layers.{i}.sa_qpos_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kpos_proj.weight''', f'''decoder.layers.{i}.sa_kpos_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_v_proj.weight''', f'''decoder.layers.{i}.sa_v_proj.weight''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qcontent_proj.weight''', f'''decoder.layers.{i}.ca_qcontent_proj.weight''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kcontent_proj.weight''', f'''decoder.layers.{i}.ca_kcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kpos_proj.weight''', f'''decoder.layers.{i}.ca_kpos_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_v_proj.weight''', f'''decoder.layers.{i}.ca_v_proj.weight''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight''', f'''decoder.layers.{i}.ca_qpos_sine_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qcontent_proj.bias''', f'''decoder.layers.{i}.sa_qcontent_proj.bias''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kcontent_proj.bias''', f'''decoder.layers.{i}.sa_kcontent_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_qpos_proj.bias''', f'''decoder.layers.{i}.sa_qpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_kpos_proj.bias''', f'''decoder.layers.{i}.sa_kpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_v_proj.bias''', f'''decoder.layers.{i}.sa_v_proj.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qcontent_proj.bias''', f'''decoder.layers.{i}.ca_qcontent_proj.bias''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kcontent_proj.bias''', f'''decoder.layers.{i}.ca_kcontent_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_kpos_proj.bias''', f'''decoder.layers.{i}.ca_kpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_v_proj.bias''', f'''decoder.layers.{i}.ca_v_proj.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias''', f'''decoder.layers.{i}.ca_qpos_sine_proj.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> int: """simple docstring""" snake_case_ : Optional[Any] = state_dict.pop(__magic_name__ ) snake_case_ : Any = val def __UpperCAmelCase ( __magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : Any = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: snake_case_ : Optional[Any] = key.replace("backbone.0.body" ,"backbone.conv_encoder.model" ) snake_case_ : int = value else: snake_case_ : int = value return new_state_dict def __UpperCAmelCase ( __magic_name__ ,__magic_name__=False )-> Optional[int]: """simple docstring""" snake_case_ : str = "" if is_panoptic: snake_case_ : Dict = "conditional_detr." # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) snake_case_ : Any = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) snake_case_ : Optional[int] = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case_ : Tuple = in_proj_weight[:256, :] snake_case_ : List[Any] = in_proj_bias[:256] snake_case_ : Optional[Any] = in_proj_weight[256:512, :] snake_case_ : Optional[int] = in_proj_bias[256:512] snake_case_ : Optional[int] = in_proj_weight[-256:, :] snake_case_ : str = in_proj_bias[-256:] def __UpperCAmelCase ( )-> Optional[Any]: """simple docstring""" snake_case_ : Optional[int] = "http://images.cocodataset.org/val2017/000000039769.jpg" snake_case_ : Optional[Any] = Image.open(requests.get(__magic_name__ ,stream=__magic_name__ ).raw ) return im @torch.no_grad() def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Optional[Any] = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: snake_case_ : Optional[Any] = "resnet101" if "dc5" in model_name: snake_case_ : List[str] = True snake_case_ : Tuple = "panoptic" in model_name if is_panoptic: snake_case_ : List[Any] = 250 else: snake_case_ : Optional[Any] = 91 snake_case_ : Optional[int] = "huggingface/label-files" snake_case_ : Dict = "coco-detection-id2label.json" snake_case_ : List[Any] = json.load(open(hf_hub_download(__magic_name__ ,__magic_name__ ,repo_type="dataset" ) ,"r" ) ) snake_case_ : Optional[int] = {int(__magic_name__ ): v for k, v in idalabel.items()} snake_case_ : int = idalabel snake_case_ : Dict = {v: k for k, v in idalabel.items()} # load image processor snake_case_ : Optional[int] = "coco_panoptic" if is_panoptic else "coco_detection" snake_case_ : str = ConditionalDetrImageProcessor(format=__magic_name__ ) # prepare image snake_case_ : str = prepare_img() snake_case_ : int = image_processor(images=__magic_name__ ,return_tensors="pt" ) snake_case_ : Union[str, Any] = encoding["pixel_values"] logger.info(F'''Converting model {model_name}...''' ) # load original model from torch hub snake_case_ : Union[str, Any] = torch.hub.load("DeppMeng/ConditionalDETR" ,__magic_name__ ,pretrained=__magic_name__ ).eval() snake_case_ : Any = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: snake_case_ : Any = "conditional_detr." + src rename_key(__magic_name__ ,__magic_name__ ,__magic_name__ ) snake_case_ : Tuple = rename_backbone_keys(__magic_name__ ) # query, key and value matrices need special treatment read_in_q_k_v(__magic_name__ ,is_panoptic=__magic_name__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them snake_case_ : int = "conditional_detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("conditional_detr" ) and not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ) ): snake_case_ : Any = state_dict.pop(__magic_name__ ) snake_case_ : Optional[int] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: snake_case_ : Tuple = state_dict.pop(__magic_name__ ) snake_case_ : Any = val elif key.startswith("bbox_attention" ) or key.startswith("mask_head" ): continue else: snake_case_ : Union[str, Any] = state_dict.pop(__magic_name__ ) snake_case_ : List[Any] = val else: if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): snake_case_ : Any = state_dict.pop(__magic_name__ ) snake_case_ : List[Any] = val # finally, create HuggingFace model and load state dict snake_case_ : Optional[int] = ConditionalDetrForSegmentation(__magic_name__ ) if is_panoptic else ConditionalDetrForObjectDetection(__magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() model.push_to_hub(repo_id=__magic_name__ ,organization="DepuMeng" ,commit_message="Add model" ) # verify our conversion snake_case_ : Dict = conditional_detr(__magic_name__ ) snake_case_ : Union[str, Any] = model(__magic_name__ ) assert torch.allclose(outputs.logits ,original_outputs["pred_logits"] ,atol=1E-4 ) assert torch.allclose(outputs.pred_boxes ,original_outputs["pred_boxes"] ,atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks ,original_outputs["pred_masks"] ,atol=1E-4 ) # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) image_processor.save_pretrained(__magic_name__ ) if __name__ == "__main__": __lowerCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) __lowerCamelCase : int = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
656
0
import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal __lowerCamelCase : Optional[Any] = datasets.utils.logging.get_logger(__name__) __lowerCamelCase : List[str] = ['''names''', '''prefix'''] __lowerCamelCase : int = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] __lowerCamelCase : str = ['''encoding_errors''', '''on_bad_lines'''] __lowerCamelCase : Optional[Any] = ['''date_format'''] @dataclass class A_ (datasets.BuilderConfig ): """simple docstring""" a__ = ''',''' a__ = None a__ = '''infer''' a__ = None a__ = None a__ = None a__ = None a__ = None a__ = True a__ = None a__ = None a__ = None a__ = None a__ = False a__ = None a__ = None a__ = None a__ = True a__ = True a__ = False a__ = True a__ = None a__ = '''.''' a__ = None a__ = '''"''' a__ = 0 a__ = None a__ = None a__ = None a__ = None a__ = True a__ = True a__ = 0 a__ = True a__ = False a__ = None a__ = 10000 a__ = None a__ = '''strict''' a__ = '''error''' a__ = None def _A ( self :List[str] ) -> Any: '''simple docstring''' if self.delimiter is not None: snake_case_ : Tuple = self.delimiter if self.column_names is not None: snake_case_ : List[Any] = self.column_names @property def _A ( self :Optional[Any] ) -> int: '''simple docstring''' snake_case_ : Optional[int] = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class A_ (datasets.ArrowBasedBuilder ): """simple docstring""" a__ = CsvConfig def _A ( self :Optional[Any] ) -> Optional[Any]: '''simple docstring''' return datasets.DatasetInfo(features=self.config.features ) def _A ( self :Tuple , lowerCAmelCase__ :Dict ) -> List[Any]: '''simple docstring''' if not self.config.data_files: raise ValueError(F'''At least one data file must be specified, but got data_files={self.config.data_files}''' ) snake_case_ : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): snake_case_ : int = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ : List[str] = [files] snake_case_ : Tuple = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] snake_case_ : str = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ : str = [files] snake_case_ : Any = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def _A ( self :List[Any] , lowerCAmelCase__ :pa.Table ) -> pa.Table: '''simple docstring''' if self.config.features is not None: snake_case_ : int = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast snake_case_ : Optional[Any] = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example snake_case_ : Dict = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def _A ( self :Dict , lowerCAmelCase__ :Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str snake_case_ : str = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): snake_case_ : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): snake_case_ : Optional[int] = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(F'''Failed to read file \'{file}\' with error {type(lowerCAmelCase__ )}: {e}''' ) raise
719
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Any ) -> str: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _A ( self :List[Any] ) -> List[str]: '''simple docstring''' snake_case_ : Any = 1 snake_case_ : Dict = 3 snake_case_ : Union[str, Any] = (32, 32) snake_case_ : Optional[int] = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) return image @property def _A ( self :Optional[int] ) -> Any: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : List[str] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) return model @property def _A ( self :Dict ) -> Any: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) return model @property def _A ( self :Dict ) -> Optional[int]: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : str = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5_006 , ) return RobertaSeriesModelWithTransformation(lowerCAmelCase__ ) @property def _A ( self :Any ) -> str: '''simple docstring''' def extract(*lowerCAmelCase__ :Any , **lowerCAmelCase__ :List[str] ): class A_ : """simple docstring""" def __init__( self :Optional[int] ) -> List[str]: '''simple docstring''' snake_case_ : str = torch.ones([0] ) def _A ( self :int , lowerCAmelCase__ :List[Any] ) -> Tuple: '''simple docstring''' self.pixel_values.to(lowerCAmelCase__ ) return self return Out() return extract def _A ( self :int ) -> Dict: '''simple docstring''' snake_case_ : str = "cpu" # ensure determinism for the device-dependent torch.Generator snake_case_ : str = self.dummy_cond_unet snake_case_ : Optional[int] = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) snake_case_ : Dict = self.dummy_vae snake_case_ : Dict = self.dummy_text_encoder snake_case_ : Optional[int] = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) snake_case_ : str = 77 snake_case_ : Any = self.dummy_image.to(lowerCAmelCase__ ) snake_case_ : Tuple = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk snake_case_ : Optional[Any] = AltDiffusionImgaImgPipeline( unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , vae=lowerCAmelCase__ , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , feature_extractor=self.dummy_extractor , ) snake_case_ : Union[str, Any] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowerCAmelCase__ ) snake_case_ : Optional[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) snake_case_ : Dict = "A painting of a squirrel eating a burger" snake_case_ : List[str] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) snake_case_ : Dict = alt_pipe( [prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , ) snake_case_ : Any = output.images snake_case_ : List[str] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) snake_case_ : Optional[Any] = alt_pipe( [prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , return_dict=lowerCAmelCase__ , )[0] snake_case_ : Tuple = image[0, -3:, -3:, -1] snake_case_ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) snake_case_ : int = np.array([0.4_4_2_7, 0.3_7_3_1, 0.4_2_4_9, 0.4_9_4_1, 0.4_5_4_6, 0.4_1_4_8, 0.4_1_9_3, 0.4_6_6_6, 0.4_4_9_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5E-3 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def _A ( self :int ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = self.dummy_cond_unet snake_case_ : Union[str, Any] = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) snake_case_ : int = self.dummy_vae snake_case_ : List[Any] = self.dummy_text_encoder snake_case_ : int = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) snake_case_ : int = 77 snake_case_ : Dict = self.dummy_image.to(lowerCAmelCase__ ) # put models in fp16 snake_case_ : Optional[Any] = unet.half() snake_case_ : Tuple = vae.half() snake_case_ : List[str] = bert.half() # make sure here that pndm scheduler skips prk snake_case_ : Optional[int] = AltDiffusionImgaImgPipeline( unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , vae=lowerCAmelCase__ , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , feature_extractor=self.dummy_extractor , ) snake_case_ : List[str] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowerCAmelCase__ ) snake_case_ : Optional[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) snake_case_ : List[Any] = "A painting of a squirrel eating a burger" snake_case_ : str = torch.manual_seed(0 ) snake_case_ : Any = alt_pipe( [prompt] , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def _A ( self :Optional[int] ) -> Any: '''simple docstring''' snake_case_ : Union[str, Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) # resize to resolution that is divisible by 8 but not 16 or 32 snake_case_ : str = init_image.resize((760, 504) ) snake_case_ : Optional[Any] = "BAAI/AltDiffusion" snake_case_ : int = AltDiffusionImgaImgPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() snake_case_ : Tuple = "A fantasy landscape, trending on artstation" snake_case_ : int = torch.manual_seed(0 ) snake_case_ : List[str] = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , strength=0.7_5 , guidance_scale=7.5 , generator=lowerCAmelCase__ , output_type="np" , ) snake_case_ : str = output.images[0] snake_case_ : List[Any] = image[255:258, 383:386, -1] assert image.shape == (504, 760, 3) snake_case_ : Tuple = np.array([0.9_3_5_8, 0.9_3_9_7, 0.9_5_9_9, 0.9_9_0_1, 1.0_0_0_0, 1.0_0_0_0, 0.9_8_8_2, 1.0_0_0_0, 1.0_0_0_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _A ( self :str ) -> Any: '''simple docstring''' snake_case_ : Optional[Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) snake_case_ : List[Any] = init_image.resize((768, 512) ) snake_case_ : Tuple = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy" ) snake_case_ : Any = "BAAI/AltDiffusion" snake_case_ : List[str] = AltDiffusionImgaImgPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() snake_case_ : Tuple = "A fantasy landscape, trending on artstation" snake_case_ : Tuple = torch.manual_seed(0 ) snake_case_ : List[Any] = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , strength=0.7_5 , guidance_scale=7.5 , generator=lowerCAmelCase__ , output_type="np" , ) snake_case_ : Optional[int] = output.images[0] assert image.shape == (512, 768, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1E-2
656
0
'''simple docstring''' import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : Optional[Any] = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) __lowerCamelCase : Optional[Any] = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', f'''encoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', f'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.weight''', f'''encoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.bias''', f'''encoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.weight''', f'''encoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.bias''', f'''encoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.encoder.layers.{i}.norm1.weight''', f'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.norm1.bias''', f'''encoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.weight''', f'''encoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.bias''', f'''encoder.layers.{i}.final_layer_norm.bias''')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', f'''decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', f'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.multihead_attn.out_proj.weight''', f'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.multihead_attn.out_proj.bias''', f'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.weight''', f'''decoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.bias''', f'''decoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.weight''', f'''decoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.bias''', f'''decoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm1.weight''', f'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm1.bias''', f'''decoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.weight''', f'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.bias''', f'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.weight''', f'''decoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.bias''', f'''decoder.layers.{i}.final_layer_norm.bias''')) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.encoder.norm.weight''', '''encoder.layernorm.weight'''), ('''transformer.encoder.norm.bias''', '''encoder.layernorm.bias'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ] ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Dict = state_dict.pop(__magic_name__ ) snake_case_ : List[Any] = val def __UpperCAmelCase ( __magic_name__ )-> List[str]: """simple docstring""" snake_case_ : str = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: snake_case_ : Any = key.replace("backbone.0.body" ,"backbone.conv_encoder.model" ) snake_case_ : List[Any] = value else: snake_case_ : Dict = value return new_state_dict def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : str = "" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) snake_case_ : Optional[int] = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) snake_case_ : List[Any] = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case_ : Tuple = in_proj_weight[:256, :] snake_case_ : str = in_proj_bias[:256] snake_case_ : Dict = in_proj_weight[256:512, :] snake_case_ : Optional[Any] = in_proj_bias[256:512] snake_case_ : Optional[int] = in_proj_weight[-256:, :] snake_case_ : Tuple = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention snake_case_ : Tuple = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) snake_case_ : Any = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case_ : int = in_proj_weight[:256, :] snake_case_ : List[Any] = in_proj_bias[:256] snake_case_ : List[Any] = in_proj_weight[256:512, :] snake_case_ : List[str] = in_proj_bias[256:512] snake_case_ : Dict = in_proj_weight[-256:, :] snake_case_ : int = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention snake_case_ : Optional[int] = state_dict.pop( F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight''' ) snake_case_ : Union[str, Any] = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) of cross-attention to the state dict snake_case_ : int = in_proj_weight_cross_attn[:256, :] snake_case_ : Optional[Any] = in_proj_bias_cross_attn[:256] snake_case_ : Any = in_proj_weight_cross_attn[256:512, :] snake_case_ : Optional[int] = in_proj_bias_cross_attn[256:512] snake_case_ : Optional[Any] = in_proj_weight_cross_attn[-256:, :] snake_case_ : Any = in_proj_bias_cross_attn[-256:] def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" snake_case_ : Optional[int] = image.size snake_case_ : Dict = max(__magic_name__ ,__magic_name__ ) snake_case_ : Optional[int] = 800 if "detection" in checkpoint_url else 1000 snake_case_ : Optional[int] = target_max_size / current_max_size snake_case_ : str = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def __UpperCAmelCase ( __magic_name__ )-> Tuple: """simple docstring""" snake_case_ : Union[str, Any] = F.to_tensor(__magic_name__ ) snake_case_ : List[Any] = F.normalize(__magic_name__ ,mean=[0.485, 0.456, 0.406] ,std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" logger.info("Converting model..." ) # load original state dict snake_case_ : str = torch.hub.load_state_dict_from_url(__magic_name__ ,map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(__magic_name__ ,__magic_name__ ,__magic_name__ ) snake_case_ : Dict = rename_backbone_keys(__magic_name__ ) # query, key and value matrices need special treatment read_in_q_k_v(__magic_name__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them snake_case_ : Tuple = "model." for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): snake_case_ : str = state_dict.pop(__magic_name__ ) snake_case_ : int = val # create HuggingFace model and load state dict snake_case_ : List[str] = TableTransformerConfig( backbone="resnet18" ,mask_loss_coefficient=1 ,dice_loss_coefficient=1 ,ce_loss_coefficient=1 ,bbox_loss_coefficient=5 ,giou_loss_coefficient=2 ,eos_coefficient=0.4 ,class_cost=1 ,bbox_cost=5 ,giou_cost=2 ,) if "detection" in checkpoint_url: snake_case_ : str = 15 snake_case_ : Tuple = 2 snake_case_ : Any = {0: "table", 1: "table rotated"} snake_case_ : List[str] = idalabel snake_case_ : Optional[int] = {v: k for k, v in idalabel.items()} else: snake_case_ : int = 125 snake_case_ : List[Any] = 6 snake_case_ : str = { 0: "table", 1: "table column", 2: "table row", 3: "table column header", 4: "table projected row header", 5: "table spanning cell", } snake_case_ : Any = idalabel snake_case_ : int = {v: k for k, v in idalabel.items()} snake_case_ : List[str] = DetrImageProcessor( format="coco_detection" ,max_size=800 if "detection" in checkpoint_url else 1000 ) snake_case_ : Optional[Any] = TableTransformerForObjectDetection(__magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() # verify our conversion snake_case_ : Dict = "example_pdf.png" if "detection" in checkpoint_url else "example_table.png" snake_case_ : List[str] = hf_hub_download(repo_id="nielsr/example-pdf" ,repo_type="dataset" ,filename=__magic_name__ ) snake_case_ : Any = Image.open(__magic_name__ ).convert("RGB" ) snake_case_ : Optional[int] = normalize(resize(__magic_name__ ,__magic_name__ ) ).unsqueeze(0 ) snake_case_ : Optional[Any] = model(__magic_name__ ) if "detection" in checkpoint_url: snake_case_ : Optional[Any] = (1, 15, 3) snake_case_ : Union[str, Any] = torch.tensor( [[-6.7_897, -16.9_985, 6.7_937], [-8.0_186, -22.2_192, 6.9_677], [-7.3_117, -21.0_708, 7.4_055]] ) snake_case_ : int = torch.tensor([[0.4_867, 0.1_767, 0.6_732], [0.6_718, 0.4_479, 0.3_830], [0.4_716, 0.1_760, 0.6_364]] ) else: snake_case_ : Any = (1, 125, 7) snake_case_ : Dict = torch.tensor( [[-18.1_430, -8.3_214, 4.8_274], [-18.4_685, -7.1_361, -4.2_667], [-26.3_693, -9.3_429, -4.9_962]] ) snake_case_ : Optional[Any] = torch.tensor([[0.4_983, 0.5_595, 0.9_440], [0.4_916, 0.6_315, 0.5_954], [0.6_108, 0.8_637, 0.1_135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] ,__magic_name__ ,atol=1E-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] ,__magic_name__ ,atol=1E-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) image_processor.save_pretrained(__magic_name__ ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) snake_case_ : str = ( "microsoft/table-transformer-detection" if "detection" in checkpoint_url else "microsoft/table-transformer-structure-recognition" ) model.push_to_hub(__magic_name__ ) image_processor.push_to_hub(__magic_name__ ) if __name__ == "__main__": __lowerCamelCase : List[str] = argparse.ArgumentParser() parser.add_argument( '''--checkpoint_url''', default='''https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth''', type=str, choices=[ '''https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth''', '''https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth''', ], help='''URL of the Table Transformer checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) __lowerCamelCase : Dict = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
720
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. __lowerCamelCase : List[str] = {'''LayoutLMv2Config''', '''LayoutLMv3Config'''} @is_pipeline_test class A_ (unittest.TestCase ): """simple docstring""" a__ = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING a__ = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: a__ = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: a__ = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def _A ( self :Tuple , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :Dict ) -> Any: '''simple docstring''' snake_case_ : Optional[Any] = ZeroShotClassificationPipeline( model=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , candidate_labels=["polics", "health"] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def _A ( self :List[str] , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Any ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = classifier("Who are you voting for in 2020?" , candidate_labels="politics" ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) # No kwarg snake_case_ : List[Any] = classifier("Who are you voting for in 2020?" , ["politics"] ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) snake_case_ : Dict = classifier("Who are you voting for in 2020?" , candidate_labels=["politics"] ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) snake_case_ : int = classifier("Who are you voting for in 2020?" , candidate_labels="politics, public health" ) self.assertEqual( lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ) , 1.0 ) snake_case_ : Optional[int] = classifier("Who are you voting for in 2020?" , candidate_labels=["politics", "public health"] ) self.assertEqual( lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ) , 1.0 ) snake_case_ : str = classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template="This text is about {}" ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) # https://github.com/huggingface/transformers/issues/13846 snake_case_ : Dict = classifier(["I am happy"] , ["positive", "negative"] ) self.assertEqual( lowerCAmelCase__ , [ {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} for i in range(1 ) ] , ) snake_case_ : Tuple = classifier(["I am happy", "I am sad"] , ["positive", "negative"] ) self.assertEqual( lowerCAmelCase__ , [ {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} for i in range(2 ) ] , ) with self.assertRaises(lowerCAmelCase__ ): classifier("" , candidate_labels="politics" ) with self.assertRaises(lowerCAmelCase__ ): classifier(lowerCAmelCase__ , candidate_labels="politics" ) with self.assertRaises(lowerCAmelCase__ ): classifier("Who are you voting for in 2020?" , candidate_labels="" ) with self.assertRaises(lowerCAmelCase__ ): classifier("Who are you voting for in 2020?" , candidate_labels=lowerCAmelCase__ ) with self.assertRaises(lowerCAmelCase__ ): classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template="Not formatting template" , ) with self.assertRaises(lowerCAmelCase__ ): classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template=lowerCAmelCase__ , ) self.run_entailment_id(lowerCAmelCase__ ) def _A ( self :List[Any] , lowerCAmelCase__ :Pipeline ) -> Union[str, Any]: '''simple docstring''' snake_case_ : int = zero_shot_classifier.model.config snake_case_ : Optional[int] = config.labelaid snake_case_ : Tuple = zero_shot_classifier.entailment_id snake_case_ : Optional[Any] = {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2} self.assertEqual(zero_shot_classifier.entailment_id , -1 ) snake_case_ : Tuple = {"entailment": 0, "neutral": 1, "contradiction": 2} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) snake_case_ : str = {"ENTAIL": 0, "NON-ENTAIL": 1} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) snake_case_ : str = {"ENTAIL": 2, "NEUTRAL": 1, "CONTR": 0} self.assertEqual(zero_shot_classifier.entailment_id , 2 ) snake_case_ : List[str] = original_labelaid self.assertEqual(lowerCAmelCase__ , zero_shot_classifier.entailment_id ) @require_torch def _A ( self :Tuple ) -> Any: '''simple docstring''' snake_case_ : List[Any] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="pt" , ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( "Who are you voting for in 2020?" * 100 , candidate_labels=["politics", "public health", "science"] ) @require_torch def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="pt" , ) snake_case_ : int = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.3_3_3, 0.3_3_3, 0.3_3_3], } , ) @require_tf def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : List[str] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="tf" , ) snake_case_ : Optional[int] = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.3_3_3, 0.3_3_3, 0.3_3_3], } , ) @slow @require_torch def _A ( self :Union[str, Any] ) -> int: '''simple docstring''' snake_case_ : int = pipeline("zero-shot-classification" , model="roberta-large-mnli" , framework="pt" ) snake_case_ : str = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.9_7_6, 0.0_1_5, 0.0_0_9], } , ) snake_case_ : Optional[int] = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data." , candidate_labels=["machine learning", "statistics", "translation", "vision"] , multi_label=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], } , ) @slow @require_tf def _A ( self :List[str] ) -> str: '''simple docstring''' snake_case_ : int = pipeline("zero-shot-classification" , model="roberta-large-mnli" , framework="tf" ) snake_case_ : Optional[Any] = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.9_7_6, 0.0_1_5, 0.0_0_9], } , ) snake_case_ : Tuple = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data." , candidate_labels=["machine learning", "statistics", "translation", "vision"] , multi_label=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], } , )
656
0
'''simple docstring''' import argparse import torch from transformers import ( SpeechTaConfig, SpeechTaFeatureExtractor, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaProcessor, SpeechTaTokenizer, logging, ) from transformers.tokenization_utils import AddedToken logging.set_verbosity_info() __lowerCamelCase : List[Any] = logging.get_logger('''transformers.models.speecht5''') __lowerCamelCase : Any = { '''speech_encoder_prenet.layer_norm''': '''speecht5.encoder.prenet.feature_projection.layer_norm''', '''speech_encoder_prenet.post_extract_proj''': '''speecht5.encoder.prenet.feature_projection.projection''', '''speech_encoder_prenet.pos_conv.0''': '''speecht5.encoder.prenet.pos_conv_embed.conv''', '''speech_encoder_prenet.mask_emb''': '''speecht5.encoder.prenet.masked_spec_embed''', } __lowerCamelCase : Any = { '''text_encoder_prenet.encoder_prenet.0''': '''speecht5.encoder.prenet.embed_tokens''', '''text_encoder_prenet.encoder_prenet.1.alpha''': '''speecht5.encoder.prenet.encode_positions.alpha''', } __lowerCamelCase : List[str] = { '''speech_decoder_prenet.decoder_prenet.0.0.prenet.0.0''': '''speecht5.decoder.prenet.layers.0''', '''speech_decoder_prenet.decoder_prenet.0.0.prenet.1.0''': '''speecht5.decoder.prenet.layers.1''', '''speech_decoder_prenet.decoder_prenet.0.1''': '''speecht5.decoder.prenet.final_layer''', '''speech_decoder_prenet.decoder_prenet.1.alpha''': '''speecht5.decoder.prenet.encode_positions.alpha''', '''speech_decoder_prenet.spkembs_layer.0''': '''speecht5.decoder.prenet.speaker_embeds_layer''', } __lowerCamelCase : str = { '''speech_decoder_postnet.feat_out''': '''speech_decoder_postnet.feat_out''', '''speech_decoder_postnet.prob_out''': '''speech_decoder_postnet.prob_out''', '''speech_decoder_postnet.postnet.postnet.0.0''': '''speech_decoder_postnet.layers.0.conv''', '''speech_decoder_postnet.postnet.postnet.0.1''': '''speech_decoder_postnet.layers.0.batch_norm''', '''speech_decoder_postnet.postnet.postnet.1.0''': '''speech_decoder_postnet.layers.1.conv''', '''speech_decoder_postnet.postnet.postnet.1.1''': '''speech_decoder_postnet.layers.1.batch_norm''', '''speech_decoder_postnet.postnet.postnet.2.0''': '''speech_decoder_postnet.layers.2.conv''', '''speech_decoder_postnet.postnet.postnet.2.1''': '''speech_decoder_postnet.layers.2.batch_norm''', '''speech_decoder_postnet.postnet.postnet.3.0''': '''speech_decoder_postnet.layers.3.conv''', '''speech_decoder_postnet.postnet.postnet.3.1''': '''speech_decoder_postnet.layers.3.batch_norm''', '''speech_decoder_postnet.postnet.postnet.4.0''': '''speech_decoder_postnet.layers.4.conv''', '''speech_decoder_postnet.postnet.postnet.4.1''': '''speech_decoder_postnet.layers.4.batch_norm''', } __lowerCamelCase : Dict = { '''text_decoder_prenet.embed_tokens''': '''speecht5.decoder.prenet.embed_tokens''', } __lowerCamelCase : Union[str, Any] = { '''text_decoder_postnet.output_projection''': '''text_decoder_postnet.lm_head''', } __lowerCamelCase : Tuple = { '''encoder.layers.*.self_attn.k_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.k_proj''', '''encoder.layers.*.self_attn.v_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.v_proj''', '''encoder.layers.*.self_attn.q_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.q_proj''', '''encoder.layers.*.self_attn.out_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.out_proj''', '''encoder.layers.*.self_attn_layer_norm''': '''speecht5.encoder.wrapped_encoder.layers.*.layer_norm''', '''encoder.layers.*.fc1''': '''speecht5.encoder.wrapped_encoder.layers.*.feed_forward.intermediate_dense''', '''encoder.layers.*.fc2''': '''speecht5.encoder.wrapped_encoder.layers.*.feed_forward.output_dense''', '''encoder.layers.*.final_layer_norm''': '''speecht5.encoder.wrapped_encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''speecht5.encoder.wrapped_encoder.layer_norm''', '''encoder.pos_emb.pe_k''': '''speecht5.encoder.wrapped_encoder.embed_positions.pe_k''', } __lowerCamelCase : List[str] = { '''decoder.layers.*.self_attn.k_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.k_proj''', '''decoder.layers.*.self_attn.v_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.v_proj''', '''decoder.layers.*.self_attn.q_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.q_proj''', '''decoder.layers.*.self_attn.out_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.out_proj''', '''decoder.layers.*.self_attn_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn_layer_norm''', '''decoder.layers.*.encoder_attn.k_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.k_proj''', '''decoder.layers.*.encoder_attn.v_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.v_proj''', '''decoder.layers.*.encoder_attn.q_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.q_proj''', '''decoder.layers.*.encoder_attn.out_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.out_proj''', '''decoder.layers.*.encoder_attn_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn_layer_norm''', '''decoder.layers.*.fc1''': '''speecht5.decoder.wrapped_decoder.layers.*.feed_forward.intermediate_dense''', '''decoder.layers.*.fc2''': '''speecht5.decoder.wrapped_decoder.layers.*.feed_forward.output_dense''', '''decoder.layers.*.final_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.final_layer_norm''', } __lowerCamelCase : List[Any] = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_TEXT_DECODER_PRENET, **MAPPING_TEXT_DECODER_POSTNET, } __lowerCamelCase : Union[str, Any] = { **MAPPING_TEXT_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __lowerCamelCase : int = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __lowerCamelCase : Union[str, Any] = [] __lowerCamelCase : Tuple = [ '''encoder.version''', '''encoder.layers.*.norm_k.weight''', '''encoder.layers.*.norm_k.bias''', '''decoder.version''', '''decoder.layers.*.norm_k.weight''', '''decoder.layers.*.norm_k.bias''', '''decoder.pos_emb.pe_k''', '''speech_encoder_prenet.embed_positions._float_tensor''', '''text_decoder_prenet.embed_positions._float_tensor''', ] __lowerCamelCase : Any = IGNORE_KEYS + [ '''encoder.proj''', '''text_encoder_prenet.*''', '''speech_decoder_prenet.*''', '''speech_decoder_postnet.*''', ] __lowerCamelCase : Any = IGNORE_KEYS + [ '''encoder.proj''', '''speech_encoder_prenet.*''', '''text_decoder_prenet.*''', '''text_decoder_postnet.*''', ] __lowerCamelCase : Dict = IGNORE_KEYS + [ '''encoder.proj''', '''text_encoder_prenet.*''', '''text_decoder_prenet.*''', '''text_decoder_postnet.*''', ] def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" for attribute in key.split("." ): snake_case_ : Optional[int] = getattr(__magic_name__ ,__magic_name__ ) if weight_type is not None: snake_case_ : Union[str, Any] = getattr(__magic_name__ ,__magic_name__ ).shape else: snake_case_ : Tuple = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'''Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": snake_case_ : Any = value elif weight_type == "weight_g": snake_case_ : List[Any] = value elif weight_type == "weight_v": snake_case_ : Tuple = value elif weight_type == "bias": snake_case_ : List[Any] = value elif weight_type == "running_mean": snake_case_ : int = value elif weight_type == "running_var": snake_case_ : int = value elif weight_type == "num_batches_tracked": snake_case_ : Union[str, Any] = value else: snake_case_ : Dict = value logger.info(F'''{key + ('.' + weight_type if weight_type is not None else '')} was initialized from {full_name}.''' ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Optional[Any]: """simple docstring""" for key in ignore_keys: if key.endswith(".*" ): if name.startswith(key[:-1] ): return True elif ".*." in key: snake_case_ : Any = key.split(".*." ) if prefix in name and suffix in name: return True elif key in name: return True return False def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Optional[int]: """simple docstring""" snake_case_ : Tuple = [] if task == "s2t": snake_case_ : Tuple = hf_model.speechta.encoder.prenet.feature_encoder snake_case_ : Optional[Any] = MAPPING_S2T snake_case_ : str = IGNORE_KEYS_S2T elif task == "t2s": snake_case_ : List[str] = None snake_case_ : int = MAPPING_T2S snake_case_ : Any = IGNORE_KEYS_T2S elif task == "s2s": snake_case_ : List[str] = hf_model.speechta.encoder.prenet.feature_encoder snake_case_ : List[Any] = MAPPING_S2S snake_case_ : List[str] = IGNORE_KEYS_S2S else: raise ValueError(F'''Unsupported task: {task}''' ) for name, value in fairseq_dict.items(): if should_ignore(__magic_name__ ,__magic_name__ ): logger.info(F'''{name} was ignored''' ) continue snake_case_ : List[str] = False if "conv_layers" in name: load_conv_layer( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,hf_model.config.feat_extract_norm == "group" ,) snake_case_ : Optional[int] = True else: for key, mapped_key in MAPPING.items(): # mapped_key = "speecht5." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if "*" in key: snake_case_ : Dict = key.split(".*." ) if prefix in name and suffix in name: snake_case_ : List[Any] = suffix # if key in name or key.split("w2v_model.")[-1] == name.split(".")[0]: if key in name: snake_case_ : Dict = True if "*" in mapped_key: snake_case_ : Tuple = name.split(__magic_name__ )[0].split("." )[-2] snake_case_ : Union[str, Any] = mapped_key.replace("*" ,__magic_name__ ) if "weight_g" in name: snake_case_ : Union[str, Any] = "weight_g" elif "weight_v" in name: snake_case_ : Union[str, Any] = "weight_v" elif "bias" in name: snake_case_ : List[Any] = "bias" elif "weight" in name: snake_case_ : List[str] = "weight" elif "running_mean" in name: snake_case_ : List[Any] = "running_mean" elif "running_var" in name: snake_case_ : str = "running_var" elif "num_batches_tracked" in name: snake_case_ : Optional[int] = "num_batches_tracked" else: snake_case_ : Any = None set_recursively(__magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ) continue if not is_used: unused_weights.append(__magic_name__ ) logger.warning(F'''Unused weights: {unused_weights}''' ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ )-> Any: """simple docstring""" snake_case_ : Dict = full_name.split("conv_layers." )[-1] snake_case_ : Dict = name.split("." ) snake_case_ : List[Any] = int(items[0] ) snake_case_ : List[str] = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) snake_case_ : Optional[int] = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) snake_case_ : Dict = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.''' ) snake_case_ : List[Any] = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.''' ) snake_case_ : Any = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(__magic_name__ ) @torch.no_grad() def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__=None ,__magic_name__=None ,__magic_name__=None ,)-> Optional[Any]: """simple docstring""" if config_path is not None: snake_case_ : Optional[int] = SpeechTaConfig.from_pretrained(__magic_name__ ) else: snake_case_ : int = SpeechTaConfig() if task == "s2t": snake_case_ : Optional[int] = config.max_text_positions snake_case_ : List[Any] = SpeechTaForSpeechToText(__magic_name__ ) elif task == "t2s": snake_case_ : str = 1876 snake_case_ : Any = 600 snake_case_ : Any = config.max_speech_positions snake_case_ : Optional[Any] = SpeechTaForTextToSpeech(__magic_name__ ) elif task == "s2s": snake_case_ : Tuple = 1876 snake_case_ : str = config.max_speech_positions snake_case_ : Dict = SpeechTaForSpeechToSpeech(__magic_name__ ) else: raise ValueError(F'''Unknown task name: {task}''' ) if vocab_path: snake_case_ : Dict = SpeechTaTokenizer(__magic_name__ ,model_max_length=config.max_text_positions ) # Mask token behaves like a normal word, i.e. include the space before it snake_case_ : Tuple = AddedToken("<mask>" ,lstrip=__magic_name__ ,rstrip=__magic_name__ ) snake_case_ : Union[str, Any] = mask_token tokenizer.add_special_tokens({"mask_token": mask_token} ) tokenizer.add_tokens(["<ctc_blank>"] ) snake_case_ : int = SpeechTaFeatureExtractor() snake_case_ : Dict = SpeechTaProcessor(tokenizer=__magic_name__ ,feature_extractor=__magic_name__ ) processor.save_pretrained(__magic_name__ ) snake_case_ : Optional[Any] = torch.load(__magic_name__ ) recursively_load_weights(fairseq_checkpoint["model"] ,__magic_name__ ,__magic_name__ ) model.save_pretrained(__magic_name__ ) if repo_id: print("Pushing to the hub..." ) processor.push_to_hub(__magic_name__ ) model.push_to_hub(__magic_name__ ) if __name__ == "__main__": __lowerCamelCase : List[Any] = argparse.ArgumentParser() parser.add_argument( '''--task''', default='''s2t''', type=str, help='''Type of the SpeechT5 model you\'d like to convert. Should be one of \'s2t\', \'t2s\', \'s2s\'.''', ) parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--vocab_path''', default=None, type=str, help='''Path to SentencePiece model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) __lowerCamelCase : List[str] = parser.parse_args() convert_speechta_checkpoint( args.task, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.vocab_path, args.push_to_hub, )
721
'''simple docstring''' import argparse import pathlib import fairseq import torch from fairseq.models.roberta import RobertaModel as FairseqRobertaModel from fairseq.modules import TransformerSentenceEncoderLayer from packaging import version from transformers import XLMRobertaConfig, XLMRobertaXLForMaskedLM, XLMRobertaXLForSequenceClassification from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.models.roberta.modeling_roberta import RobertaAttention from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse('''1.0.0a'''): raise Exception('''requires fairseq >= 1.0.0a''') logging.set_verbosity_info() __lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) __lowerCamelCase : Union[str, Any] = '''Hello world! cécé herlolip''' def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : str = FairseqRobertaModel.from_pretrained(__magic_name__ ) roberta.eval() # disable dropout snake_case_ : Dict = roberta.model.encoder.sentence_encoder snake_case_ : List[str] = XLMRobertaConfig( vocab_size=roberta_sent_encoder.embed_tokens.num_embeddings ,hidden_size=roberta.cfg.model.encoder_embed_dim ,num_hidden_layers=roberta.cfg.model.encoder_layers ,num_attention_heads=roberta.cfg.model.encoder_attention_heads ,intermediate_size=roberta.cfg.model.encoder_ffn_embed_dim ,max_position_embeddings=514 ,type_vocab_size=1 ,layer_norm_eps=1E-5 ,) if classification_head: snake_case_ : List[str] = roberta.model.classification_heads["mnli"].out_proj.weight.shape[0] print("Our RoBERTa config:" ,__magic_name__ ) snake_case_ : List[str] = XLMRobertaXLForSequenceClassification(__magic_name__ ) if classification_head else XLMRobertaXLForMaskedLM(__magic_name__ ) model.eval() # Now let's copy all the weights. # Embeddings snake_case_ : List[Any] = roberta_sent_encoder.embed_tokens.weight snake_case_ : int = roberta_sent_encoder.embed_positions.weight snake_case_ : Union[str, Any] = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c RoBERTa doesn't use them. snake_case_ : Union[str, Any] = roberta_sent_encoder.layer_norm.weight snake_case_ : str = roberta_sent_encoder.layer_norm.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer snake_case_ : BertLayer = model.roberta.encoder.layer[i] snake_case_ : TransformerSentenceEncoderLayer = roberta_sent_encoder.layers[i] snake_case_ : RobertaAttention = layer.attention snake_case_ : Dict = roberta_layer.self_attn_layer_norm.weight snake_case_ : Dict = roberta_layer.self_attn_layer_norm.bias # self attention snake_case_ : BertSelfAttention = layer.attention.self assert ( roberta_layer.self_attn.k_proj.weight.data.shape == roberta_layer.self_attn.q_proj.weight.data.shape == roberta_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ) snake_case_ : Dict = roberta_layer.self_attn.q_proj.weight snake_case_ : Any = roberta_layer.self_attn.q_proj.bias snake_case_ : Optional[Any] = roberta_layer.self_attn.k_proj.weight snake_case_ : Optional[Any] = roberta_layer.self_attn.k_proj.bias snake_case_ : Optional[int] = roberta_layer.self_attn.v_proj.weight snake_case_ : Any = roberta_layer.self_attn.v_proj.bias # self-attention output snake_case_ : BertSelfOutput = layer.attention.output assert self_output.dense.weight.shape == roberta_layer.self_attn.out_proj.weight.shape snake_case_ : List[str] = roberta_layer.self_attn.out_proj.weight snake_case_ : Optional[int] = roberta_layer.self_attn.out_proj.bias # this one is final layer norm snake_case_ : int = roberta_layer.final_layer_norm.weight snake_case_ : Union[str, Any] = roberta_layer.final_layer_norm.bias # intermediate snake_case_ : BertIntermediate = layer.intermediate assert intermediate.dense.weight.shape == roberta_layer.fca.weight.shape snake_case_ : List[str] = roberta_layer.fca.weight snake_case_ : List[Any] = roberta_layer.fca.bias # output snake_case_ : BertOutput = layer.output assert bert_output.dense.weight.shape == roberta_layer.fca.weight.shape snake_case_ : Any = roberta_layer.fca.weight snake_case_ : Any = roberta_layer.fca.bias # end of layer if classification_head: snake_case_ : int = roberta.model.classification_heads["mnli"].dense.weight snake_case_ : Union[str, Any] = roberta.model.classification_heads["mnli"].dense.bias snake_case_ : Tuple = roberta.model.classification_heads["mnli"].out_proj.weight snake_case_ : str = roberta.model.classification_heads["mnli"].out_proj.bias else: # LM Head snake_case_ : Optional[Any] = roberta.model.encoder.lm_head.dense.weight snake_case_ : int = roberta.model.encoder.lm_head.dense.bias snake_case_ : Optional[Any] = roberta.model.encoder.lm_head.layer_norm.weight snake_case_ : Optional[int] = roberta.model.encoder.lm_head.layer_norm.bias snake_case_ : int = roberta.model.encoder.lm_head.weight snake_case_ : List[str] = roberta.model.encoder.lm_head.bias # Let's check that we get the same results. snake_case_ : torch.Tensor = roberta.encode(__magic_name__ ).unsqueeze(0 ) # batch of size 1 snake_case_ : Union[str, Any] = model(__magic_name__ )[0] if classification_head: snake_case_ : Optional[Any] = roberta.model.classification_heads["mnli"](roberta.extract_features(__magic_name__ ) ) else: snake_case_ : List[str] = roberta.model(__magic_name__ )[0] print(our_output.shape ,their_output.shape ) snake_case_ : str = torch.max(torch.abs(our_output - their_output ) ).item() print(F'''max_absolute_diff = {max_absolute_diff}''' ) # ~ 1e-7 snake_case_ : Any = torch.allclose(__magic_name__ ,__magic_name__ ,atol=1E-3 ) print("Do both models output the same tensors?" ,"🔥" if success else "💩" ) if not success: raise Exception("Something went wRoNg" ) pathlib.Path(__magic_name__ ).mkdir(parents=__magic_name__ ,exist_ok=__magic_name__ ) print(F'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(__magic_name__ ) if __name__ == "__main__": __lowerCamelCase : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--roberta_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--classification_head''', action='''store_true''', help='''Whether to convert a final classification head.''' ) __lowerCamelCase : Tuple = parser.parse_args() convert_xlm_roberta_xl_checkpoint_to_pytorch( args.roberta_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
656
0
'''simple docstring''' import collections import inspect import unittest from transformers import FocalNetConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, ) from transformers.models.focalnet.modeling_focalnet import FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A_ : """simple docstring""" def __init__( self :str , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :List[str]=13 , lowerCAmelCase__ :Any=32 , lowerCAmelCase__ :Tuple=2 , lowerCAmelCase__ :Tuple=3 , lowerCAmelCase__ :Dict=16 , lowerCAmelCase__ :Optional[int]=[32, 64, 128] , lowerCAmelCase__ :int=[1, 2, 1] , lowerCAmelCase__ :Union[str, Any]=[2, 2, 4] , lowerCAmelCase__ :Optional[int]=2 , lowerCAmelCase__ :Tuple=2.0 , lowerCAmelCase__ :Optional[Any]=True , lowerCAmelCase__ :List[str]=0.0 , lowerCAmelCase__ :List[Any]=0.0 , lowerCAmelCase__ :int=0.1 , lowerCAmelCase__ :Any="gelu" , lowerCAmelCase__ :int=False , lowerCAmelCase__ :str=True , lowerCAmelCase__ :Optional[int]=0.0_2 , lowerCAmelCase__ :str=1E-5 , lowerCAmelCase__ :Optional[Any]=True , lowerCAmelCase__ :Dict=None , lowerCAmelCase__ :Any=True , lowerCAmelCase__ :Union[str, Any]=10 , lowerCAmelCase__ :Dict=8 , lowerCAmelCase__ :Optional[int]=["stage1", "stage2"] , lowerCAmelCase__ :Union[str, Any]=[1, 2] , ) -> Optional[Any]: '''simple docstring''' snake_case_ : Optional[Any] = parent snake_case_ : Dict = batch_size snake_case_ : str = image_size snake_case_ : Dict = patch_size snake_case_ : Union[str, Any] = num_channels snake_case_ : Optional[int] = embed_dim snake_case_ : List[Any] = hidden_sizes snake_case_ : Optional[int] = depths snake_case_ : Tuple = num_heads snake_case_ : str = window_size snake_case_ : List[Any] = mlp_ratio snake_case_ : List[Any] = qkv_bias snake_case_ : List[str] = hidden_dropout_prob snake_case_ : Optional[Any] = attention_probs_dropout_prob snake_case_ : Tuple = drop_path_rate snake_case_ : Union[str, Any] = hidden_act snake_case_ : Optional[int] = use_absolute_embeddings snake_case_ : str = patch_norm snake_case_ : Optional[int] = layer_norm_eps snake_case_ : Union[str, Any] = initializer_range snake_case_ : str = is_training snake_case_ : Any = scope snake_case_ : Optional[int] = use_labels snake_case_ : Dict = type_sequence_label_size snake_case_ : Optional[Any] = encoder_stride snake_case_ : str = out_features snake_case_ : List[str] = out_indices def _A ( self :int ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ : List[str] = None if self.use_labels: snake_case_ : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ : Dict = self.get_config() return config, pixel_values, labels def _A ( self :str ) -> Optional[int]: '''simple docstring''' return FocalNetConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , hidden_sizes=self.hidden_sizes , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def _A ( self :List[str] , lowerCAmelCase__ :Dict , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Optional[Any] ) -> Optional[Any]: '''simple docstring''' snake_case_ : int = FocalNetModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case_ : int = model(lowerCAmelCase__ ) snake_case_ : Dict = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) snake_case_ : Optional[Any] = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _A ( self :str , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :Any ) -> int: '''simple docstring''' snake_case_ : int = FocalNetBackbone(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case_ : Optional[int] = model(lowerCAmelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size, 8, 8] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[:-1] ) # verify backbone works with out_features=None snake_case_ : str = None snake_case_ : Optional[int] = FocalNetBackbone(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case_ : Any = model(lowerCAmelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size * 2, 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _A ( self :List[str] , lowerCAmelCase__ :int , lowerCAmelCase__ :List[str] , lowerCAmelCase__ :int ) -> Tuple: '''simple docstring''' snake_case_ : Any = FocalNetForMaskedImageModeling(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case_ : str = model(lowerCAmelCase__ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images snake_case_ : Optional[Any] = 1 snake_case_ : int = FocalNetForMaskedImageModeling(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case_ : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case_ : Tuple = model(lowerCAmelCase__ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _A ( self :Optional[int] , lowerCAmelCase__ :Dict , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :List[Any] ) -> List[str]: '''simple docstring''' snake_case_ : Dict = self.type_sequence_label_size snake_case_ : Any = FocalNetForImageClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case_ : Optional[int] = model(lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images snake_case_ : Tuple = 1 snake_case_ : List[str] = FocalNetForImageClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() snake_case_ : List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case_ : int = model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _A ( self :List[Any] ) -> int: '''simple docstring''' snake_case_ : Union[str, Any] = self.prepare_config_and_inputs() snake_case_ : List[str] = config_and_inputs snake_case_ : Union[str, Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class A_ (a_ , a_ , unittest.TestCase ): """simple docstring""" a__ = ( ( FocalNetModel, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetBackbone, ) if is_torch_available() else () ) a__ = ( {'''feature-extraction''': FocalNetModel, '''image-classification''': FocalNetForImageClassification} if is_torch_available() else {} ) a__ = False a__ = False a__ = False a__ = False a__ = False def _A ( self :Any ) -> Any: '''simple docstring''' snake_case_ : Union[str, Any] = FocalNetModelTester(self ) snake_case_ : Dict = ConfigTester(self , config_class=lowerCAmelCase__ , embed_dim=37 , has_text_modality=lowerCAmelCase__ ) def _A ( self :Dict ) -> Dict: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _A ( self :List[Any] ) -> Any: '''simple docstring''' return def _A ( self :Optional[int] ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def _A ( self :List[Any] ) -> Tuple: '''simple docstring''' snake_case_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowerCAmelCase__ ) def _A ( self :str ) -> Tuple: '''simple docstring''' snake_case_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCAmelCase__ ) def _A ( self :Any ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase__ ) @unittest.skip(reason="FocalNet does not use inputs_embeds" ) def _A ( self :Tuple ) -> Any: '''simple docstring''' pass @unittest.skip(reason="FocalNet does not use feedforward chunking" ) def _A ( self :Any ) -> List[str]: '''simple docstring''' pass def _A ( self :Tuple ) -> List[Any]: '''simple docstring''' snake_case_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: snake_case_ : str = model_class(lowerCAmelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case_ : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase__ , nn.Linear ) ) def _A ( self :List[Any] ) -> Optional[Any]: '''simple docstring''' snake_case_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: snake_case_ : int = model_class(lowerCAmelCase__ ) snake_case_ : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ : Optional[int] = [*signature.parameters.keys()] snake_case_ : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) def _A ( self :List[Any] , lowerCAmelCase__ :Dict , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :str , lowerCAmelCase__ :str ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[Any] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() with torch.no_grad(): snake_case_ : int = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) ) snake_case_ : int = outputs.hidden_states snake_case_ : Dict = getattr( self.model_tester , "expected_num_hidden_layers" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowerCAmelCase__ ) , lowerCAmelCase__ ) # FocalNet has a different seq_length snake_case_ : Tuple = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ : List[Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) snake_case_ : List[str] = outputs.reshaped_hidden_states self.assertEqual(len(lowerCAmelCase__ ) , lowerCAmelCase__ ) snake_case_ : Union[str, Any] = reshaped_hidden_states[0].shape snake_case_ : Union[str, Any] = ( reshaped_hidden_states[0].view(lowerCAmelCase__ , lowerCAmelCase__ , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _A ( self :Optional[int] ) -> Dict: '''simple docstring''' snake_case_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ : Optional[Any] = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes[:-1]: snake_case_ : Tuple = True self.check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ : Tuple = True self.check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def _A ( self :Optional[Any] ) -> Dict: '''simple docstring''' snake_case_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ : Optional[Any] = 3 snake_case_ : Tuple = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) snake_case_ : Optional[Any] = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ : Union[str, Any] = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) snake_case_ : List[Any] = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes[:-1]: snake_case_ : Union[str, Any] = True self.check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ : Union[str, Any] = True self.check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , (padded_height, padded_width) ) @slow def _A ( self :Optional[Any] ) -> str: '''simple docstring''' for model_name in FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ : Any = FocalNetModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> Tuple: '''simple docstring''' snake_case_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ : Tuple = _config_zero_init(lowerCAmelCase__ ) for model_class in self.all_model_classes: snake_case_ : str = model_class(config=lowerCAmelCase__ ) for name, param in model.named_parameters(): if "embeddings" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class A_ (unittest.TestCase ): """simple docstring""" @cached_property def _A ( self :Tuple ) -> int: '''simple docstring''' return AutoImageProcessor.from_pretrained("microsoft/focalnet-tiny" ) if is_vision_available() else None @slow def _A ( self :str ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[str] = FocalNetForImageClassification.from_pretrained("microsoft/focalnet-tiny" ).to(lowerCAmelCase__ ) snake_case_ : str = self.default_image_processor snake_case_ : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) snake_case_ : Any = image_processor(images=lowerCAmelCase__ , return_tensors="pt" ).to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): snake_case_ : str = model(**lowerCAmelCase__ ) # verify the logits snake_case_ : List[str] = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase__ ) snake_case_ : Tuple = torch.tensor([0.2_1_6_6, -0.4_3_6_8, 0.2_1_9_1] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1E-4 ) ) self.assertTrue(outputs.logits.argmax(dim=-1 ).item() , 281 ) @require_torch class A_ (a_ , unittest.TestCase ): """simple docstring""" a__ = (FocalNetBackbone,) if is_torch_available() else () a__ = FocalNetConfig a__ = False def _A ( self :Any ) -> List[Any]: '''simple docstring''' snake_case_ : Any = FocalNetModelTester(self )
700
'''simple docstring''' import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def __UpperCAmelCase ( __magic_name__ ,__magic_name__=() ,__magic_name__=None ,__magic_name__="no" ,__magic_name__="29500" )-> Optional[int]: """simple docstring""" snake_case_ : str = False snake_case_ : int = False if any(key.startswith("KAGGLE" ) for key in os.environ.keys() ): snake_case_ : Any = True elif "IPython" in sys.modules: snake_case_ : Union[str, Any] = "google.colab" in str(sys.modules["IPython"].get_ipython() ) try: snake_case_ : Any = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F'''Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.''' ) if (in_colab or in_kaggle) and (os.environ.get("TPU_NAME" ,__magic_name__ ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( "To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside " "your training function. Restart your notebook and make sure no cells initializes an " "`Accelerator`." ) if num_processes is None: snake_case_ : Tuple = 8 snake_case_ : Optional[int] = PrepareForLaunch(__magic_name__ ,distributed_type="TPU" ) print(F'''Launching a training on {num_processes} TPU cores.''' ) xmp.spawn(__magic_name__ ,args=__magic_name__ ,nprocs=__magic_name__ ,start_method="fork" ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print("Launching training on one GPU." ) else: print("Launching training on one CPU." ) function(*__magic_name__ ) else: if num_processes is None: raise ValueError( "You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call." ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( "To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized " "inside your training function. Restart your notebook and make sure no cells initializes an " "`Accelerator`." ) if torch.cuda.is_initialized(): raise ValueError( "To launch a multi-GPU training from your notebook, you need to avoid running any instruction " "using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA " "function." ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=__magic_name__ ,master_addr="127.0.01" ,master_port=__magic_name__ ,mixed_precision=__magic_name__ ): snake_case_ : Optional[int] = PrepareForLaunch(__magic_name__ ,distributed_type="MULTI_GPU" ) print(F'''Launching training on {num_processes} GPUs.''' ) try: start_processes(__magic_name__ ,args=__magic_name__ ,nprocs=__magic_name__ ,start_method="fork" ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( "CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. " "This likely stems from an outside import causing issues once the `notebook_launcher()` is called. " "Please review your imports and test them when running the `notebook_launcher()` to identify " "which one is problematic." ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): snake_case_ : Any = "1" print("Launching training on MPS." ) elif torch.cuda.is_available(): print("Launching training on one GPU." ) else: print("Launching training on CPU." ) function(*__magic_name__ ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__=() ,__magic_name__=2 )-> Dict: """simple docstring""" from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=__magic_name__ ,master_addr="127.0.01" ,master_port="29500" ,accelerate_mixed_precision="no" ,accelerate_debug_rdv_file=tmp_file.name ,accelerate_use_cpu="yes" ,): snake_case_ : Any = PrepareForLaunch(__magic_name__ ,debug=__magic_name__ ) start_processes(__magic_name__ ,args=__magic_name__ ,nprocs=__magic_name__ ,start_method="fork" )
656
0
'''simple docstring''' import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch __lowerCamelCase : Union[str, Any] = '''sshleifer/bart-tiny-random''' __lowerCamelCase : Optional[int] = '''patrickvonplaten/t5-tiny-random''' @require_torch class A_ (unittest.TestCase ): """simple docstring""" @cached_property def _A ( self :Tuple ) -> Optional[Any]: '''simple docstring''' return AutoConfig.from_pretrained(lowerCAmelCase__ ) def _A ( self :List[Any] ) -> Any: '''simple docstring''' snake_case_ : int = create_student_by_copying_alternating_layers(lowerCAmelCase__ , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.num_hidden_layers , 1 ) def _A ( self :int ) -> Dict: '''simple docstring''' snake_case_ : int = create_student_by_copying_alternating_layers(lowerCAmelCase__ , tempfile.mkdtemp() , e=1 , d=lowerCAmelCase__ ) def _A ( self :Tuple ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = create_student_by_copying_alternating_layers(lowerCAmelCase__ , tempfile.mkdtemp() , e=1 , d=lowerCAmelCase__ ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , self.teacher_config.encoder_layers ) def _A ( self :Dict ) -> str: '''simple docstring''' snake_case_ : Dict = create_student_by_copying_alternating_layers(lowerCAmelCase__ , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , 1 ) def _A ( self :Any ) -> Tuple: '''simple docstring''' with self.assertRaises(lowerCAmelCase__ ): create_student_by_copying_alternating_layers(lowerCAmelCase__ , tempfile.mkdtemp() , e=lowerCAmelCase__ , d=lowerCAmelCase__ )
701
'''simple docstring''' from collections import deque from math import floor from random import random from time import time class A_ : """simple docstring""" def __init__( self :Dict ) -> List[str]: '''simple docstring''' snake_case_ : int = {} def _A ( self :Any , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Optional[Any]=1 ) -> Any: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: snake_case_ : Optional[int] = [[w, v]] if not self.graph.get(lowerCAmelCase__ ): snake_case_ : Dict = [] def _A ( self :List[Any] ) -> Optional[int]: '''simple docstring''' return list(self.graph ) def _A ( self :str , lowerCAmelCase__ :Any , lowerCAmelCase__ :int ) -> List[Any]: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCAmelCase__ ) def _A ( self :List[str] , lowerCAmelCase__ :Optional[Any]=-2 , lowerCAmelCase__ :str=-1 ) -> str: '''simple docstring''' if s == d: return [] snake_case_ : str = [] snake_case_ : Optional[int] = [] if s == -2: snake_case_ : List[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Dict = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : str = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCAmelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCAmelCase__ ) != 0: snake_case_ : Union[str, Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return visited def _A ( self :Tuple , lowerCAmelCase__ :int=-1 ) -> int: '''simple docstring''' if c == -1: snake_case_ : Any = floor(random() * 10_000 ) + 10 for i in range(lowerCAmelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): snake_case_ : Optional[Any] = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCAmelCase__ , lowerCAmelCase__ , 1 ) def _A ( self :Tuple , lowerCAmelCase__ :Dict=-2 ) -> Dict: '''simple docstring''' snake_case_ : Union[str, Any] = deque() snake_case_ : Optional[Any] = [] if s == -2: snake_case_ : Tuple = list(self.graph )[0] d.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) while d: snake_case_ : Optional[int] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def _A ( self :List[str] , lowerCAmelCase__ :str ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = 0 for x in self.graph: for y in self.graph[x]: if y[1] == u: count += 1 return count def _A ( self :Any , lowerCAmelCase__ :int ) -> Optional[Any]: '''simple docstring''' return len(self.graph[u] ) def _A ( self :Tuple , lowerCAmelCase__ :List[str]=-2 ) -> Optional[Any]: '''simple docstring''' snake_case_ : str = [] snake_case_ : str = [] if s == -2: snake_case_ : Optional[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : int = s snake_case_ : Optional[int] = [] while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : List[str] = node[1] break # check if all the children are visited if s == ss: sorted_nodes.append(stack.pop() ) if len(lowerCAmelCase__ ) != 0: snake_case_ : int = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Union[str, Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return sorted_nodes def _A ( self :Dict ) -> Any: '''simple docstring''' snake_case_ : Dict = [] snake_case_ : Any = [] snake_case_ : str = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Optional[int] = -2 snake_case_ : Any = [] snake_case_ : List[Any] = s snake_case_ : int = False snake_case_ : Optional[int] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Any = len(lowerCAmelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Optional[int] = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[Any] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : str = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : List[str] = s snake_case_ : Optional[int] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return list(lowerCAmelCase__ ) def _A ( self :Tuple ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = [] snake_case_ : Tuple = [] snake_case_ : List[str] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : str = -2 snake_case_ : List[str] = [] snake_case_ : List[Any] = s snake_case_ : List[str] = False snake_case_ : Dict = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Any = len(lowerCAmelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Tuple = True if len(lowerCAmelCase__ ) != 0: snake_case_ : List[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : int = s snake_case_ : Union[str, Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return False def _A ( self :Optional[int] , lowerCAmelCase__ :Optional[int]=-2 , lowerCAmelCase__ :Tuple=-1 ) -> str: '''simple docstring''' snake_case_ : Optional[int] = time() self.dfs(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : Optional[Any] = time() return end - begin def _A ( self :Any , lowerCAmelCase__ :Tuple=-2 ) -> Optional[Any]: '''simple docstring''' snake_case_ : Any = time() self.bfs(lowerCAmelCase__ ) snake_case_ : Any = time() return end - begin class A_ : """simple docstring""" def __init__( self :Tuple ) -> List[str]: '''simple docstring''' snake_case_ : Optional[Any] = {} def _A ( self :str , lowerCAmelCase__ :Dict , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Union[str, Any]=1 ) -> str: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): # if there already is a edge if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: # if u does not exist snake_case_ : str = [[w, v]] # add the other way if self.graph.get(lowerCAmelCase__ ): # if there already is a edge if self.graph[v].count([w, u] ) == 0: self.graph[v].append([w, u] ) else: # if u does not exist snake_case_ : List[str] = [[w, u]] def _A ( self :Dict , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Optional[Any] ) -> Optional[Any]: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCAmelCase__ ) # the other way round if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[v]: if _[1] == u: self.graph[v].remove(lowerCAmelCase__ ) def _A ( self :Optional[Any] , lowerCAmelCase__ :Optional[Any]=-2 , lowerCAmelCase__ :Optional[int]=-1 ) -> int: '''simple docstring''' if s == d: return [] snake_case_ : Any = [] snake_case_ : Dict = [] if s == -2: snake_case_ : Optional[int] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[str] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCAmelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : str = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return visited def _A ( self :Optional[int] , lowerCAmelCase__ :str=-1 ) -> List[Any]: '''simple docstring''' if c == -1: snake_case_ : Optional[int] = floor(random() * 10_000 ) + 10 for i in range(lowerCAmelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): snake_case_ : str = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCAmelCase__ , lowerCAmelCase__ , 1 ) def _A ( self :Any , lowerCAmelCase__ :Optional[Any]=-2 ) -> List[Any]: '''simple docstring''' snake_case_ : List[str] = deque() snake_case_ : Optional[Any] = [] if s == -2: snake_case_ : List[Any] = list(self.graph )[0] d.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) while d: snake_case_ : Optional[int] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def _A ( self :str , lowerCAmelCase__ :Union[str, Any] ) -> Optional[Any]: '''simple docstring''' return len(self.graph[u] ) def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : Any = [] snake_case_ : Optional[Any] = [] snake_case_ : Optional[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = -2 snake_case_ : Optional[int] = [] snake_case_ : Tuple = s snake_case_ : Optional[Any] = False snake_case_ : Optional[int] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : Optional[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Optional[int] = len(lowerCAmelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Tuple = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[int] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : List[Any] = s snake_case_ : Dict = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return list(lowerCAmelCase__ ) def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Optional[Any] = [] snake_case_ : int = [] snake_case_ : List[str] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = -2 snake_case_ : int = [] snake_case_ : int = s snake_case_ : Optional[Any] = False snake_case_ : List[Any] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : Union[str, Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Tuple = len(lowerCAmelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Optional[Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[Any] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Tuple = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : Union[str, Any] = s snake_case_ : Tuple = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return False def _A ( self :Any ) -> Tuple: '''simple docstring''' return list(self.graph ) def _A ( self :Optional[Any] , lowerCAmelCase__ :Tuple=-2 , lowerCAmelCase__ :Optional[int]=-1 ) -> str: '''simple docstring''' snake_case_ : List[str] = time() self.dfs(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : List[Any] = time() return end - begin def _A ( self :Union[str, Any] , lowerCAmelCase__ :List[Any]=-2 ) -> int: '''simple docstring''' snake_case_ : List[str] = time() self.bfs(lowerCAmelCase__ ) snake_case_ : Tuple = time() return end - begin
656
0
'''simple docstring''' from abc import ABC, abstractmethod from argparse import ArgumentParser class A_ (a_ ): """simple docstring""" @staticmethod @abstractmethod def _A ( lowerCAmelCase__ :ArgumentParser ) -> Optional[int]: '''simple docstring''' raise NotImplementedError() @abstractmethod def _A ( self :Any ) -> List[str]: '''simple docstring''' raise NotImplementedError()
702
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __lowerCamelCase : List[str] = re.compile(R'''\s+''') def __UpperCAmelCase ( __magic_name__ )-> Union[str, Any]: """simple docstring""" return {"hash": hashlib.mda(re.sub(__magic_name__ ,"" ,example["content"] ).encode("utf-8" ) ).hexdigest()} def __UpperCAmelCase ( __magic_name__ )-> str: """simple docstring""" snake_case_ : Optional[Any] = [len(__magic_name__ ) for line in example["content"].splitlines()] return {"line_mean": np.mean(__magic_name__ ), "line_max": max(__magic_name__ )} def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" snake_case_ : Optional[int] = np.mean([c.isalnum() for c in example["content"]] ) return {"alpha_frac": alpha_frac} def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" if example["hash"] in uniques: uniques.remove(example["hash"] ) return True else: return False def __UpperCAmelCase ( __magic_name__ ,__magic_name__=5 )-> Tuple: """simple docstring""" snake_case_ : List[str] = ["auto-generated", "autogenerated", "automatically generated"] snake_case_ : Optional[Any] = example["content"].splitlines() for _, line in zip(range(__magic_name__ ) ,__magic_name__ ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def __UpperCAmelCase ( __magic_name__ ,__magic_name__=5 ,__magic_name__=0.05 )-> Optional[Any]: """simple docstring""" snake_case_ : str = ["unit tests", "test file", "configuration file"] snake_case_ : int = example["content"].splitlines() snake_case_ : Optional[Any] = 0 snake_case_ : Any = 0 # first test for _, line in zip(range(__magic_name__ ) ,__magic_name__ ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test snake_case_ : Tuple = example["content"].count("\n" ) snake_case_ : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count("config" ) count_test += line.lower().count("test" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def __UpperCAmelCase ( __magic_name__ )-> str: """simple docstring""" snake_case_ : List[Any] = ["def ", "class ", "for ", "while "] snake_case_ : Optional[Any] = example["content"].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def __UpperCAmelCase ( __magic_name__ ,__magic_name__=4 )-> Optional[int]: """simple docstring""" snake_case_ : Tuple = example["content"].splitlines() snake_case_ : Tuple = 0 for line in lines: counter += line.lower().count("=" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : Tuple = tokenizer(example["content"] ,truncation=__magic_name__ )["input_ids"] snake_case_ : int = len(example["content"] ) / len(__magic_name__ ) return {"ratio": ratio} def __UpperCAmelCase ( __magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : Union[str, Any] = {} results.update(get_hash(__magic_name__ ) ) results.update(line_stats(__magic_name__ ) ) results.update(alpha_stats(__magic_name__ ) ) results.update(char_token_ratio(__magic_name__ ) ) results.update(is_autogenerated(__magic_name__ ) ) results.update(is_config_or_test(__magic_name__ ) ) results.update(has_no_keywords(__magic_name__ ) ) results.update(has_few_assignments(__magic_name__ ) ) return results def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> Tuple: """simple docstring""" if not check_uniques(__magic_name__ ,__magic_name__ ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def __UpperCAmelCase ( __magic_name__ )-> Dict: """simple docstring""" with open(__magic_name__ ,"rb" ) as f_in: with gzip.open(str(__magic_name__ ) + ".gz" ,"wb" ,compresslevel=6 ) as f_out: shutil.copyfileobj(__magic_name__ ,__magic_name__ ) os.unlink(__magic_name__ ) # Settings __lowerCamelCase : List[Any] = HfArgumentParser(PreprocessingArguments) __lowerCamelCase : str = parser.parse_args() if args.num_workers is None: __lowerCamelCase : List[Any] = multiprocessing.cpu_count() __lowerCamelCase : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __lowerCamelCase : Any = time.time() __lowerCamelCase : str = load_dataset(args.dataset_name, split='''train''') print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __lowerCamelCase : List[str] = time.time() __lowerCamelCase : Any = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __lowerCamelCase : Any = set(ds.unique('''hash''')) __lowerCamelCase : Optional[int] = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __lowerCamelCase : List[str] = time.time() __lowerCamelCase : Tuple = ds.filter(filter, fn_kwargs={'''uniques''': uniques, '''args''': args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __lowerCamelCase : List[str] = time.time() __lowerCamelCase , __lowerCamelCase : Tuple = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __lowerCamelCase : List[Any] = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / '''duplicate_clusters.json''', '''w''') as f: json.dump(duplicate_clusters, f) __lowerCamelCase : List[str] = output_dir / '''data''' data_dir.mkdir(exist_ok=True) __lowerCamelCase : int = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __lowerCamelCase : Union[str, Any] = str(data_dir / f'''file-{file_number+1:012}.json''') __lowerCamelCase : List[Any] = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
656
0
'''simple docstring''' from collections import deque from math import floor from random import random from time import time class A_ : """simple docstring""" def __init__( self :Dict ) -> List[str]: '''simple docstring''' snake_case_ : int = {} def _A ( self :Any , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Optional[Any]=1 ) -> Any: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: snake_case_ : Optional[int] = [[w, v]] if not self.graph.get(lowerCAmelCase__ ): snake_case_ : Dict = [] def _A ( self :List[Any] ) -> Optional[int]: '''simple docstring''' return list(self.graph ) def _A ( self :str , lowerCAmelCase__ :Any , lowerCAmelCase__ :int ) -> List[Any]: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCAmelCase__ ) def _A ( self :List[str] , lowerCAmelCase__ :Optional[Any]=-2 , lowerCAmelCase__ :str=-1 ) -> str: '''simple docstring''' if s == d: return [] snake_case_ : str = [] snake_case_ : Optional[int] = [] if s == -2: snake_case_ : List[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Dict = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : str = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCAmelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCAmelCase__ ) != 0: snake_case_ : Union[str, Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return visited def _A ( self :Tuple , lowerCAmelCase__ :int=-1 ) -> int: '''simple docstring''' if c == -1: snake_case_ : Any = floor(random() * 10_000 ) + 10 for i in range(lowerCAmelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): snake_case_ : Optional[Any] = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCAmelCase__ , lowerCAmelCase__ , 1 ) def _A ( self :Tuple , lowerCAmelCase__ :Dict=-2 ) -> Dict: '''simple docstring''' snake_case_ : Union[str, Any] = deque() snake_case_ : Optional[Any] = [] if s == -2: snake_case_ : Tuple = list(self.graph )[0] d.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) while d: snake_case_ : Optional[int] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def _A ( self :List[str] , lowerCAmelCase__ :str ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = 0 for x in self.graph: for y in self.graph[x]: if y[1] == u: count += 1 return count def _A ( self :Any , lowerCAmelCase__ :int ) -> Optional[Any]: '''simple docstring''' return len(self.graph[u] ) def _A ( self :Tuple , lowerCAmelCase__ :List[str]=-2 ) -> Optional[Any]: '''simple docstring''' snake_case_ : str = [] snake_case_ : str = [] if s == -2: snake_case_ : Optional[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : int = s snake_case_ : Optional[int] = [] while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : List[str] = node[1] break # check if all the children are visited if s == ss: sorted_nodes.append(stack.pop() ) if len(lowerCAmelCase__ ) != 0: snake_case_ : int = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Union[str, Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return sorted_nodes def _A ( self :Dict ) -> Any: '''simple docstring''' snake_case_ : Dict = [] snake_case_ : Any = [] snake_case_ : str = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Optional[int] = -2 snake_case_ : Any = [] snake_case_ : List[Any] = s snake_case_ : int = False snake_case_ : Optional[int] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Any = len(lowerCAmelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Optional[int] = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[Any] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : str = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : List[str] = s snake_case_ : Optional[int] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return list(lowerCAmelCase__ ) def _A ( self :Tuple ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = [] snake_case_ : Tuple = [] snake_case_ : List[str] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : str = -2 snake_case_ : List[str] = [] snake_case_ : List[Any] = s snake_case_ : List[str] = False snake_case_ : Dict = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Any = len(lowerCAmelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Tuple = True if len(lowerCAmelCase__ ) != 0: snake_case_ : List[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : int = s snake_case_ : Union[str, Any] = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return False def _A ( self :Optional[int] , lowerCAmelCase__ :Optional[int]=-2 , lowerCAmelCase__ :Tuple=-1 ) -> str: '''simple docstring''' snake_case_ : Optional[int] = time() self.dfs(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : Optional[Any] = time() return end - begin def _A ( self :Any , lowerCAmelCase__ :Tuple=-2 ) -> Optional[Any]: '''simple docstring''' snake_case_ : Any = time() self.bfs(lowerCAmelCase__ ) snake_case_ : Any = time() return end - begin class A_ : """simple docstring""" def __init__( self :Tuple ) -> List[str]: '''simple docstring''' snake_case_ : Optional[Any] = {} def _A ( self :str , lowerCAmelCase__ :Dict , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Union[str, Any]=1 ) -> str: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): # if there already is a edge if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: # if u does not exist snake_case_ : str = [[w, v]] # add the other way if self.graph.get(lowerCAmelCase__ ): # if there already is a edge if self.graph[v].count([w, u] ) == 0: self.graph[v].append([w, u] ) else: # if u does not exist snake_case_ : List[str] = [[w, u]] def _A ( self :Dict , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Optional[Any] ) -> Optional[Any]: '''simple docstring''' if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(lowerCAmelCase__ ) # the other way round if self.graph.get(lowerCAmelCase__ ): for _ in self.graph[v]: if _[1] == u: self.graph[v].remove(lowerCAmelCase__ ) def _A ( self :Optional[Any] , lowerCAmelCase__ :Optional[Any]=-2 , lowerCAmelCase__ :Optional[int]=-1 ) -> int: '''simple docstring''' if s == d: return [] snake_case_ : Any = [] snake_case_ : Dict = [] if s == -2: snake_case_ : Optional[int] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : List[str] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(lowerCAmelCase__ ) return visited else: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : str = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : str = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return visited def _A ( self :Optional[int] , lowerCAmelCase__ :str=-1 ) -> List[Any]: '''simple docstring''' if c == -1: snake_case_ : Optional[int] = floor(random() * 10_000 ) + 10 for i in range(lowerCAmelCase__ ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): snake_case_ : str = floor(random() * c ) + 1 if n != i: self.add_pair(lowerCAmelCase__ , lowerCAmelCase__ , 1 ) def _A ( self :Any , lowerCAmelCase__ :Optional[Any]=-2 ) -> List[Any]: '''simple docstring''' snake_case_ : List[str] = deque() snake_case_ : Optional[Any] = [] if s == -2: snake_case_ : List[Any] = list(self.graph )[0] d.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) while d: snake_case_ : Optional[int] = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def _A ( self :str , lowerCAmelCase__ :Union[str, Any] ) -> Optional[Any]: '''simple docstring''' return len(self.graph[u] ) def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : Any = [] snake_case_ : Optional[Any] = [] snake_case_ : Optional[Any] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = -2 snake_case_ : Optional[int] = [] snake_case_ : Tuple = s snake_case_ : Optional[Any] = False snake_case_ : Optional[int] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : Optional[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Optional[int] = len(lowerCAmelCase__ ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Tuple = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[int] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Optional[Any] = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : List[Any] = s snake_case_ : Dict = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return list(lowerCAmelCase__ ) def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Optional[Any] = [] snake_case_ : int = [] snake_case_ : List[str] = list(self.graph )[0] stack.append(lowerCAmelCase__ ) visited.append(lowerCAmelCase__ ) snake_case_ : Tuple = -2 snake_case_ : int = [] snake_case_ : int = s snake_case_ : Optional[Any] = False snake_case_ : List[Any] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: snake_case_ : Union[str, Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): snake_case_ : Tuple = len(lowerCAmelCase__ ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) snake_case_ : Optional[Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() snake_case_ : Optional[Any] = True if len(lowerCAmelCase__ ) != 0: snake_case_ : Tuple = stack[len(lowerCAmelCase__ ) - 1] else: snake_case_ : Optional[int] = False indirect_parents.append(lowerCAmelCase__ ) snake_case_ : Union[str, Any] = s snake_case_ : Tuple = ss # check if se have reached the starting point if len(lowerCAmelCase__ ) == 0: return False def _A ( self :Any ) -> Tuple: '''simple docstring''' return list(self.graph ) def _A ( self :Optional[Any] , lowerCAmelCase__ :Tuple=-2 , lowerCAmelCase__ :Optional[int]=-1 ) -> str: '''simple docstring''' snake_case_ : List[str] = time() self.dfs(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : List[Any] = time() return end - begin def _A ( self :Union[str, Any] , lowerCAmelCase__ :List[Any]=-2 ) -> int: '''simple docstring''' snake_case_ : List[str] = time() self.bfs(lowerCAmelCase__ ) snake_case_ : Tuple = time() return end - begin
703
'''simple docstring''' import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :str ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Union[str, Any] = torch.nn.Linear(10 , 10 ) snake_case_ : Dict = torch.optim.SGD(model.parameters() , 0.1 ) snake_case_ : Tuple = Accelerator() snake_case_ : Optional[Any] = accelerator.prepare(lowerCAmelCase__ ) try: pickle.loads(pickle.dumps(lowerCAmelCase__ ) ) except Exception as e: self.fail(F'''Accelerated optimizer pickling failed with {e}''' ) AcceleratorState._reset_state()
656
0
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class A_ (a_ , a_ , unittest.TestCase ): """simple docstring""" a__ = IFImgaImgSuperResolutionPipeline a__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''width''', '''height'''} a__ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'''original_image'''} ) a__ = PipelineTesterMixin.required_optional_params - {'''latents'''} def _A ( self :Tuple ) -> List[Any]: '''simple docstring''' return self._get_superresolution_dummy_components() def _A ( self :Union[str, Any] , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :List[Any]=0 ) -> str: '''simple docstring''' if str(lowerCAmelCase__ ).startswith("mps" ): snake_case_ : Union[str, Any] = torch.manual_seed(lowerCAmelCase__ ) else: snake_case_ : Optional[Any] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) snake_case_ : Tuple = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) snake_case_ : List[Any] = floats_tensor((1, 3, 16, 16) , rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) snake_case_ : str = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def _A ( self :Tuple ) -> List[Any]: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def _A ( self :str ) -> Tuple: '''simple docstring''' self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def _A ( self :Optional[Any] ) -> Optional[Any]: '''simple docstring''' super().test_save_load_floataa(expected_max_diff=1E-1 ) def _A ( self :Dict ) -> Union[str, Any]: '''simple docstring''' self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def _A ( self :Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' self._test_save_load_local() def _A ( self :Dict ) -> Tuple: '''simple docstring''' self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
704
'''simple docstring''' import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __lowerCamelCase : Any = '''src/transformers''' # This is to make sure the transformers module imported is the one in the repo. __lowerCamelCase : List[str] = direct_transformers_import(PATH_TO_TRANSFORMERS) __lowerCamelCase : Optional[Any] = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __lowerCamelCase : Union[str, Any] = re.compile(R'''\[(.+?)\]\((https://huggingface\.co/.+?)\)''') __lowerCamelCase : Any = { '''DecisionTransformerConfig''', '''EncoderDecoderConfig''', '''MusicgenConfig''', '''RagConfig''', '''SpeechEncoderDecoderConfig''', '''TimmBackboneConfig''', '''VisionEncoderDecoderConfig''', '''VisionTextDualEncoderConfig''', '''LlamaConfig''', } def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : Tuple = None # source code of `config_class` snake_case_ : List[Any] = inspect.getsource(__magic_name__ ) snake_case_ : List[str] = _re_checkpoint.findall(__magic_name__ ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith("/" ): snake_case_ : Optional[Any] = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link snake_case_ : str = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: snake_case_ : Dict = ckpt_name break return checkpoint def __UpperCAmelCase ( )-> Dict: """simple docstring""" snake_case_ : Optional[int] = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue snake_case_ : str = get_checkpoint_from_config_class(__magic_name__ ) snake_case_ : Union[str, Any] = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(__magic_name__ ) if len(__magic_name__ ) > 0: snake_case_ : Tuple = "\n".join(sorted(__magic_name__ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
656
0
'''simple docstring''' __lowerCamelCase : dict[str, float] = { "joule": 1.0, "kilojoule": 1000, "megajoule": 1000000, "gigajoule": 1000000000, "wattsecond": 1.0, "watthour": 3600, "kilowatthour": 3600000, "newtonmeter": 1.0, "calorie_nutr": 4_186.8, "kilocalorie_nutr": 4_186_800.00, "electronvolt": 1.6_0217_6634E-19, "britishthermalunit_it": 1_055.05_585, "footpound": 1.355_818, } def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> float: """simple docstring""" if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: snake_case_ : int = ( F'''Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n''' F'''Valid values are: {', '.join(__magic_name__ )}''' ) raise ValueError(__magic_name__ ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
705
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCamelCase : List[str] = logging.get_logger(__name__) __lowerCamelCase : int = { '''microsoft/cvt-13''': '''https://huggingface.co/microsoft/cvt-13/resolve/main/config.json''', # See all Cvt models at https://huggingface.co/models?filter=cvt } class A_ (a_ ): """simple docstring""" a__ = '''cvt''' def __init__( self :List[Any] , lowerCAmelCase__ :Optional[int]=3 , lowerCAmelCase__ :Any=[7, 3, 3] , lowerCAmelCase__ :Dict=[4, 2, 2] , lowerCAmelCase__ :Union[str, Any]=[2, 1, 1] , lowerCAmelCase__ :Any=[64, 192, 384] , lowerCAmelCase__ :List[str]=[1, 3, 6] , lowerCAmelCase__ :str=[1, 2, 10] , lowerCAmelCase__ :Any=[4.0, 4.0, 4.0] , lowerCAmelCase__ :int=[0.0, 0.0, 0.0] , lowerCAmelCase__ :Optional[Any]=[0.0, 0.0, 0.0] , lowerCAmelCase__ :Dict=[0.0, 0.0, 0.1] , lowerCAmelCase__ :List[Any]=[True, True, True] , lowerCAmelCase__ :List[Any]=[False, False, True] , lowerCAmelCase__ :Dict=["dw_bn", "dw_bn", "dw_bn"] , lowerCAmelCase__ :Any=[3, 3, 3] , lowerCAmelCase__ :Tuple=[1, 1, 1] , lowerCAmelCase__ :Optional[int]=[2, 2, 2] , lowerCAmelCase__ :Union[str, Any]=[1, 1, 1] , lowerCAmelCase__ :Any=[1, 1, 1] , lowerCAmelCase__ :List[str]=0.0_2 , lowerCAmelCase__ :Dict=1E-1_2 , **lowerCAmelCase__ :Optional[Any] , ) -> str: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) snake_case_ : int = num_channels snake_case_ : int = patch_sizes snake_case_ : Optional[Any] = patch_stride snake_case_ : Dict = patch_padding snake_case_ : Tuple = embed_dim snake_case_ : Optional[int] = num_heads snake_case_ : Union[str, Any] = depth snake_case_ : Optional[int] = mlp_ratio snake_case_ : Tuple = attention_drop_rate snake_case_ : str = drop_rate snake_case_ : Tuple = drop_path_rate snake_case_ : Any = qkv_bias snake_case_ : Union[str, Any] = cls_token snake_case_ : int = qkv_projection_method snake_case_ : Any = kernel_qkv snake_case_ : Union[str, Any] = padding_kv snake_case_ : str = stride_kv snake_case_ : Dict = padding_q snake_case_ : Tuple = stride_q snake_case_ : Any = initializer_range snake_case_ : Any = layer_norm_eps
656
0
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. __lowerCamelCase : List[str] = {'''LayoutLMv2Config''', '''LayoutLMv3Config'''} @is_pipeline_test class A_ (unittest.TestCase ): """simple docstring""" a__ = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING a__ = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: a__ = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: a__ = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def _A ( self :Tuple , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :Dict ) -> Any: '''simple docstring''' snake_case_ : Optional[Any] = ZeroShotClassificationPipeline( model=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , candidate_labels=["polics", "health"] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def _A ( self :List[str] , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Any ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = classifier("Who are you voting for in 2020?" , candidate_labels="politics" ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) # No kwarg snake_case_ : List[Any] = classifier("Who are you voting for in 2020?" , ["politics"] ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) snake_case_ : Dict = classifier("Who are you voting for in 2020?" , candidate_labels=["politics"] ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) snake_case_ : int = classifier("Who are you voting for in 2020?" , candidate_labels="politics, public health" ) self.assertEqual( lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ) , 1.0 ) snake_case_ : Optional[int] = classifier("Who are you voting for in 2020?" , candidate_labels=["politics", "public health"] ) self.assertEqual( lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["scores"] ) ) , 1.0 ) snake_case_ : str = classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template="This text is about {}" ) self.assertEqual(lowerCAmelCase__ , {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ )]} ) # https://github.com/huggingface/transformers/issues/13846 snake_case_ : Dict = classifier(["I am happy"] , ["positive", "negative"] ) self.assertEqual( lowerCAmelCase__ , [ {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} for i in range(1 ) ] , ) snake_case_ : Tuple = classifier(["I am happy", "I am sad"] , ["positive", "negative"] ) self.assertEqual( lowerCAmelCase__ , [ {"sequence": ANY(lowerCAmelCase__ ), "labels": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )], "scores": [ANY(lowerCAmelCase__ ), ANY(lowerCAmelCase__ )]} for i in range(2 ) ] , ) with self.assertRaises(lowerCAmelCase__ ): classifier("" , candidate_labels="politics" ) with self.assertRaises(lowerCAmelCase__ ): classifier(lowerCAmelCase__ , candidate_labels="politics" ) with self.assertRaises(lowerCAmelCase__ ): classifier("Who are you voting for in 2020?" , candidate_labels="" ) with self.assertRaises(lowerCAmelCase__ ): classifier("Who are you voting for in 2020?" , candidate_labels=lowerCAmelCase__ ) with self.assertRaises(lowerCAmelCase__ ): classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template="Not formatting template" , ) with self.assertRaises(lowerCAmelCase__ ): classifier( "Who are you voting for in 2020?" , candidate_labels="politics" , hypothesis_template=lowerCAmelCase__ , ) self.run_entailment_id(lowerCAmelCase__ ) def _A ( self :List[Any] , lowerCAmelCase__ :Pipeline ) -> Union[str, Any]: '''simple docstring''' snake_case_ : int = zero_shot_classifier.model.config snake_case_ : Optional[int] = config.labelaid snake_case_ : Tuple = zero_shot_classifier.entailment_id snake_case_ : Optional[Any] = {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2} self.assertEqual(zero_shot_classifier.entailment_id , -1 ) snake_case_ : Tuple = {"entailment": 0, "neutral": 1, "contradiction": 2} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) snake_case_ : str = {"ENTAIL": 0, "NON-ENTAIL": 1} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) snake_case_ : str = {"ENTAIL": 2, "NEUTRAL": 1, "CONTR": 0} self.assertEqual(zero_shot_classifier.entailment_id , 2 ) snake_case_ : List[str] = original_labelaid self.assertEqual(lowerCAmelCase__ , zero_shot_classifier.entailment_id ) @require_torch def _A ( self :Tuple ) -> Any: '''simple docstring''' snake_case_ : List[Any] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="pt" , ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( "Who are you voting for in 2020?" * 100 , candidate_labels=["politics", "public health", "science"] ) @require_torch def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="pt" , ) snake_case_ : int = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.3_3_3, 0.3_3_3, 0.3_3_3], } , ) @require_tf def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' snake_case_ : List[str] = pipeline( "zero-shot-classification" , model="sshleifer/tiny-distilbert-base-cased-distilled-squad" , framework="tf" , ) snake_case_ : Optional[int] = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["science", "public health", "politics"], "scores": [0.3_3_3, 0.3_3_3, 0.3_3_3], } , ) @slow @require_torch def _A ( self :Union[str, Any] ) -> int: '''simple docstring''' snake_case_ : int = pipeline("zero-shot-classification" , model="roberta-large-mnli" , framework="pt" ) snake_case_ : str = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.9_7_6, 0.0_1_5, 0.0_0_9], } , ) snake_case_ : Optional[int] = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data." , candidate_labels=["machine learning", "statistics", "translation", "vision"] , multi_label=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], } , ) @slow @require_tf def _A ( self :List[str] ) -> str: '''simple docstring''' snake_case_ : int = pipeline("zero-shot-classification" , model="roberta-large-mnli" , framework="tf" ) snake_case_ : Optional[Any] = zero_shot_classifier( "Who are you voting for in 2020?" , candidate_labels=["politics", "public health", "science"] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": "Who are you voting for in 2020?", "labels": ["politics", "public health", "science"], "scores": [0.9_7_6, 0.0_1_5, 0.0_0_9], } , ) snake_case_ : Tuple = zero_shot_classifier( "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks" " in an encoder-decoder configuration. The best performing models also connect the encoder and decoder" " through an attention mechanism. We propose a new simple network architecture, the Transformer, based" " solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two" " machine translation tasks show these models to be superior in quality while being more parallelizable" " and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014" " English-to-German translation task, improving over the existing best results, including ensembles by" " over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new" " single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small" " fraction of the training costs of the best models from the literature. We show that the Transformer" " generalizes well to other tasks by applying it successfully to English constituency parsing both with" " large and limited training data." , candidate_labels=["machine learning", "statistics", "translation", "vision"] , multi_label=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { "sequence": ( "The dominant sequence transduction models are based on complex recurrent or convolutional neural" " networks in an encoder-decoder configuration. The best performing models also connect the" " encoder and decoder through an attention mechanism. We propose a new simple network" " architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence" " and convolutions entirely. Experiments on two machine translation tasks show these models to be" " superior in quality while being more parallelizable and requiring significantly less time to" " train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task," " improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014" " English-to-French translation task, our model establishes a new single-model state-of-the-art" " BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training" " costs of the best models from the literature. We show that the Transformer generalizes well to" " other tasks by applying it successfully to English constituency parsing both with large and" " limited training data." ), "labels": ["translation", "machine learning", "vision", "statistics"], "scores": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], } , )
706
'''simple docstring''' import sacrebleu as scb from packaging import version from sacrebleu import TER import datasets __lowerCamelCase : str = '''\ @inproceedings{snover-etal-2006-study, title = "A Study of Translation Edit Rate with Targeted Human Annotation", author = "Snover, Matthew and Dorr, Bonnie and Schwartz, Rich and Micciulla, Linnea and Makhoul, John", booktitle = "Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers", month = aug # " 8-12", year = "2006", address = "Cambridge, Massachusetts, USA", publisher = "Association for Machine Translation in the Americas", url = "https://aclanthology.org/2006.amta-papers.25", pages = "223--231", } @inproceedings{post-2018-call, title = "A Call for Clarity in Reporting {BLEU} Scores", author = "Post, Matt", booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers", month = oct, year = "2018", address = "Belgium, Brussels", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W18-6319", pages = "186--191", } ''' __lowerCamelCase : Dict = '''\ TER (Translation Edit Rate, also called Translation Error Rate) is a metric to quantify the edit operations that a hypothesis requires to match a reference translation. We use the implementation that is already present in sacrebleu (https://github.com/mjpost/sacreBLEU#ter), which in turn is inspired by the TERCOM implementation, which can be found here: https://github.com/jhclark/tercom. The implementation here is slightly different from sacrebleu in terms of the required input format. The length of the references and hypotheses lists need to be the same, so you may need to transpose your references compared to sacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534 See the README.md file at https://github.com/mjpost/sacreBLEU#ter for more information. ''' __lowerCamelCase : int = ''' Produces TER scores alongside the number of edits and reference length. Args: predictions (list of str): The system stream (a sequence of segments). references (list of list of str): A list of one or more reference streams (each a sequence of segments). normalized (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. ignore_punct (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. support_zh_ja_chars (boolean): If `True`, tokenization/normalization supports processing of Chinese characters, as well as Japanese Kanji, Hiragana, Katakana, and Phonetic Extensions of Katakana. Only applies if `normalized = True`. Defaults to `False`. case_sensitive (boolean): If `False`, makes all predictions and references lowercase to ignore differences in case. Defaults to `False`. Returns: \'score\' (float): TER score (num_edits / sum_ref_lengths * 100) \'num_edits\' (int): The cumulative number of edits \'ref_length\' (float): The cumulative average reference length Examples: Example 1: >>> predictions = ["does this sentence match??", ... "what about this sentence?", ... "What did the TER metric user say to the developer?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"], ... ["Your jokes are...", "...TERrible"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {\'score\': 150.0, \'num_edits\': 15, \'ref_length\': 10.0} Example 2: >>> predictions = ["does this sentence match??", ... "what about this sentence?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {\'score\': 62.5, \'num_edits\': 5, \'ref_length\': 8.0} Example 3: >>> predictions = ["does this sentence match??", ... "what about this sentence?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... normalized=True, ... case_sensitive=True) >>> print(results) {\'score\': 57.14285714285714, \'num_edits\': 6, \'ref_length\': 10.5} Example 4: >>> predictions = ["does this sentence match??", ... "what about this sentence?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {\'score\': 0.0, \'num_edits\': 0, \'ref_length\': 8.0} Example 5: >>> predictions = ["does this sentence match??", ... "what about this sentence?", ... "What did the TER metric user say to the developer?"] >>> references = [["does this sentence match", "does this sentence match!?!"], ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"], ... ["Your jokes are...", "...TERrible"]] >>> ter = datasets.load_metric("ter") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {\'score\': 100.0, \'num_edits\': 10, \'ref_length\': 10.0} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): """simple docstring""" def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' if version.parse(scb.__version__ ) < version.parse("1.4.12" ): raise ImportWarning( "To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n" "You can install it with `pip install \"sacrebleu>=1.4.12\"`." ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="http://www.cs.umd.edu/~snover/tercom/" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Sequence(datasets.Value("string" , id="sequence" ) , id="references" ), } ) , codebase_urls=["https://github.com/mjpost/sacreBLEU#ter"] , reference_urls=[ "https://github.com/jhclark/tercom", ] , ) def _A ( self :Any , lowerCAmelCase__ :Any , lowerCAmelCase__ :str , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :bool = False , ) -> Optional[Any]: '''simple docstring''' snake_case_ : List[str] = len(references[0] ) if any(len(lowerCAmelCase__ ) != references_per_prediction for refs in references ): raise ValueError("Sacrebleu requires the same number of references for each prediction" ) snake_case_ : List[str] = [[refs[i] for refs in references] for i in range(lowerCAmelCase__ )] snake_case_ : List[str] = TER( normalized=lowerCAmelCase__ , no_punct=lowerCAmelCase__ , asian_support=lowerCAmelCase__ , case_sensitive=lowerCAmelCase__ , ) snake_case_ : Any = sb_ter.corpus_score(lowerCAmelCase__ , lowerCAmelCase__ ) return {"score": output.score, "num_edits": output.num_edits, "ref_length": output.ref_length}
656
0
'''simple docstring''' from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCamelCase : Optional[Any] = { '''configuration_cpmant''': ['''CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''CpmAntConfig'''], '''tokenization_cpmant''': ['''CpmAntTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : List[str] = [ '''CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''CpmAntForCausalLM''', '''CpmAntModel''', '''CpmAntPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_cpmant import CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP, CpmAntConfig from .tokenization_cpmant import CpmAntTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_cpmant import ( CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST, CpmAntForCausalLM, CpmAntModel, CpmAntPreTrainedModel, ) else: import sys __lowerCamelCase : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
707
'''simple docstring''' from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : Any = { "repo_name": ["test_repo1", "test_repo2", "test_repo3"], "path": ["test_1.py", "test_2.py", "unit_test.py"], "content": ["a " * 20, "a " * 30, "b " * 7], } snake_case_ : int = Dataset.from_dict(__magic_name__ ) return dataset class A_ (a_ ): """simple docstring""" def _A ( self :List[str] ) -> str: '''simple docstring''' snake_case_ : Union[str, Any] = get_dataset() snake_case_ : Optional[int] = make_duplicate_clusters(lowerCAmelCase__ , 0.8_5 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case_ : Optional[int] = get_dataset() snake_case_, snake_case_ : List[Any] = deduplicate_dataset(lowerCAmelCase__ ) self.assertEqual(len(lowerCAmelCase__ ) , 2 ) print(lowerCAmelCase__ ) self.assertEqual(duplicate_clusters[0][0]["copies"] , 2 ) self.assertEqual(duplicate_clusters[0][0]["is_extreme"] , lowerCAmelCase__ )
656
0
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> List[Any]: """simple docstring""" snake_case_ : Optional[int] = [0 for i in range(r + 1 )] # nc0 = 1 snake_case_ : Tuple = 1 for i in range(1 ,n + 1 ): # to compute current row from previous row. snake_case_ : Union[str, Any] = min(__magic_name__ ,__magic_name__ ) while j > 0: c[j] += c[j - 1] j -= 1 return c[r] print(binomial_coefficient(n=10, r=5))
708
'''simple docstring''' from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable __lowerCamelCase : Dict = { '''configuration_gpt_neox_japanese''': ['''GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoXJapaneseConfig'''], '''tokenization_gpt_neox_japanese''': ['''GPTNeoXJapaneseTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ '''GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoXJapaneseForCausalLM''', '''GPTNeoXJapaneseLayer''', '''GPTNeoXJapaneseModel''', '''GPTNeoXJapanesePreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neox_japanese import GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXJapaneseConfig from .tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox_japanese import ( GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseLayer, GPTNeoXJapaneseModel, GPTNeoXJapanesePreTrainedModel, ) else: import sys __lowerCamelCase : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
656
0
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ )-> bool: """simple docstring""" if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> bool: """simple docstring""" if curr_ind == len(__magic_name__ ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 ,len(__magic_name__ ) ): if valid_connection(__magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ ): # Insert current vertex into path as next transition snake_case_ : int = next_ver # Validate created path if util_hamilton_cycle(__magic_name__ ,__magic_name__ ,curr_ind + 1 ): return True # Backtrack snake_case_ : Union[str, Any] = -1 return False def __UpperCAmelCase ( __magic_name__ ,__magic_name__ = 0 )-> list[int]: """simple docstring""" snake_case_ : List[str] = [-1] * (len(__magic_name__ ) + 1) # initialize start and end of path with starting index snake_case_ : str = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(__magic_name__ ,__magic_name__ ,1 ) else []
709
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ )-> list[int]: """simple docstring""" if length <= 0 or not isinstance(__magic_name__ ,__magic_name__ ): raise ValueError("Length must be a positive integer." ) return [n * (2 * n - 1) for n in range(__magic_name__ )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=10))
656
0
'''simple docstring''' def __UpperCAmelCase ( __magic_name__ = 10**12 )-> int: """simple docstring""" snake_case_ : Dict = 1 snake_case_ : str = 0 snake_case_ : Dict = 1 snake_case_ : Union[str, Any] = 1 while numerator <= 2 * min_total - 1: prev_numerator += 2 * numerator numerator += 2 * prev_numerator prev_denominator += 2 * denominator denominator += 2 * prev_denominator return (denominator + 1) // 2 if __name__ == "__main__": print(f'''{solution() = }''')
710
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def __UpperCAmelCase ( __magic_name__=None )-> List[str]: """simple docstring""" if subparsers is not None: snake_case_ : List[str] = subparsers.add_parser("test" ) else: snake_case_ : List[Any] = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" ,default=__magic_name__ ,help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) ,) if subparsers is not None: parser.set_defaults(func=__magic_name__ ) return parser def __UpperCAmelCase ( __magic_name__ )-> Tuple: """simple docstring""" snake_case_ : Optional[Any] = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: snake_case_ : str = script_name else: snake_case_ : Any = F'''--config_file={args.config_file} {script_name}''' snake_case_ : Union[str, Any] = ["accelerate-launch"] + test_args.split() snake_case_ : Optional[int] = execute_subprocess_async(__magic_name__ ,env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : Dict = test_command_parser() snake_case_ : Dict = parser.parse_args() test_command(__magic_name__ ) if __name__ == "__main__": main()
656
0
'''simple docstring''' import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, temp_seed, temporary_assignment, zip_dict, ) from .utils import require_tf, require_torch def __UpperCAmelCase ( __magic_name__ )-> Dict: # picklable for multiprocessing """simple docstring""" return x.sum() def __UpperCAmelCase ( __magic_name__ )-> List[Any]: # picklable for multiprocessing """simple docstring""" return i + 1 @dataclass class A_ : """simple docstring""" a__ = 42 a__ = 42 class A_ (a_ ): """simple docstring""" def _A ( self :Dict ) -> Any: '''simple docstring''' snake_case_ : List[str] = {} snake_case_ : Optional[Any] = [] snake_case_ : str = 1 snake_case_ : Optional[int] = [1, 2] snake_case_ : Optional[int] = {"a": 1, "b": 2} snake_case_ : List[str] = {"a": [1, 2], "b": [3, 4]} snake_case_ : Dict = {"a": {"1": 1}, "b": 2} snake_case_ : Any = {"a": 1, "b": 2, "c": 3, "d": 4} snake_case_ : Optional[Any] = {} snake_case_ : Any = [] snake_case_ : List[Any] = 2 snake_case_ : Tuple = [2, 3] snake_case_ : Any = {"a": 2, "b": 3} snake_case_ : str = {"a": [2, 3], "b": [4, 5]} snake_case_ : Optional[int] = {"a": {"1": 2}, "b": 3} snake_case_ : int = {"a": 2, "b": 3, "c": 4, "d": 5} self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) snake_case_ : Optional[int] = 2 self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ , num_proc=lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ , num_proc=lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ , num_proc=lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ , num_proc=lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ , num_proc=lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ , num_proc=lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ , num_proc=lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ , num_proc=lowerCAmelCase__ ) , lowerCAmelCase__ ) snake_case_ : List[Any] = {"a": np.eye(2 ), "b": np.zeros(3 ), "c": np.ones(2 )} snake_case_ : Union[str, Any] = {"a": 2, "b": 0, "c": 2} snake_case_ : Dict = { "a": np.eye(2 ).astype(lowerCAmelCase__ ), "b": np.zeros(3 ).astype(lowerCAmelCase__ ), "c": np.ones(2 ).astype(lowerCAmelCase__ ), } self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ , map_numpy=lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(lowerCAmelCase__ , lowerCAmelCase__ , map_numpy=lowerCAmelCase__ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) self.assertEqual(map_nested(lowerCAmelCase__ , lowerCAmelCase__ , map_numpy=lowerCAmelCase__ , num_proc=lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(lowerCAmelCase__ , lowerCAmelCase__ , map_numpy=lowerCAmelCase__ , num_proc=lowerCAmelCase__ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) with self.assertRaises(lowerCAmelCase__ ): # can't pickle a local lambda map_nested(lambda lowerCAmelCase__ : x + 1 , lowerCAmelCase__ , num_proc=lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = {"a": 1, "b": 2} snake_case_ : List[Any] = {"a": 3, "b": 4} snake_case_ : Optional[int] = {"a": 5, "b": 6} snake_case_ : Optional[Any] = sorted([("a", (1, 3, 5)), ("b", (2, 4, 6))] ) self.assertEqual(sorted(zip_dict(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) ) , lowerCAmelCase__ ) def _A ( self :Tuple ) -> str: '''simple docstring''' class A_ : """simple docstring""" a__ = '''bar''' snake_case_ : int = Foo() self.assertEqual(foo.my_attr , "bar" ) with temporary_assignment(lowerCAmelCase__ , "my_attr" , "BAR" ): self.assertEqual(foo.my_attr , "BAR" ) self.assertEqual(foo.my_attr , "bar" ) @pytest.mark.parametrize( "iterable_length, num_proc, expected_num_proc" ,[ (1, None, 1), (1, 1, 1), (2, None, 1), (2, 1, 1), (2, 2, 1), (2, 3, 1), (3, 2, 1), (16, 16, 16), (16, 17, 16), (17, 16, 16), ] ,) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> int: """simple docstring""" with patch("datasets.utils.py_utils._single_map_nested" ) as mock_single_map_nested, patch( "datasets.parallel.parallel.Pool" ) as mock_multiprocessing_pool: snake_case_ : Any = {F'''{i}''': i for i in range(__magic_name__ )} snake_case_ : Any = map_nested(lambda __magic_name__ : x + 10 ,__magic_name__ ,num_proc=__magic_name__ ,parallel_min_length=16 ) if expected_num_proc == 1: assert mock_single_map_nested.called assert not mock_multiprocessing_pool.called else: assert not mock_single_map_nested.called assert mock_multiprocessing_pool.called assert mock_multiprocessing_pool.call_args[0][0] == expected_num_proc class A_ (a_ ): """simple docstring""" @require_tf def _A ( self :Any ) -> Any: '''simple docstring''' import tensorflow as tf from tensorflow.keras import layers snake_case_ : List[str] = layers.Dense(2 ) def gen_random_output(): snake_case_ : List[Any] = tf.random.uniform((1, 3) ) return model(lowerCAmelCase__ ).numpy() with temp_seed(42 , set_tensorflow=lowerCAmelCase__ ): snake_case_ : Union[str, Any] = gen_random_output() with temp_seed(42 , set_tensorflow=lowerCAmelCase__ ): snake_case_ : str = gen_random_output() snake_case_ : str = gen_random_output() np.testing.assert_equal(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @require_torch def _A ( self :Tuple ) -> Any: '''simple docstring''' import torch def gen_random_output(): snake_case_ : Dict = torch.nn.Linear(3 , 2 ) snake_case_ : int = torch.rand(1 , 3 ) return model(lowerCAmelCase__ ).detach().numpy() with temp_seed(42 , set_pytorch=lowerCAmelCase__ ): snake_case_ : Any = gen_random_output() with temp_seed(42 , set_pytorch=lowerCAmelCase__ ): snake_case_ : Any = gen_random_output() snake_case_ : Tuple = gen_random_output() np.testing.assert_equal(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) def _A ( self :Any ) -> Tuple: '''simple docstring''' def gen_random_output(): return np.random.rand(1 , 3 ) with temp_seed(42 ): snake_case_ : List[str] = gen_random_output() with temp_seed(42 ): snake_case_ : int = gen_random_output() snake_case_ : Any = gen_random_output() np.testing.assert_equal(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @pytest.mark.parametrize("input_data" ,[{}] ) def __UpperCAmelCase ( __magic_name__ )-> Dict: """simple docstring""" snake_case_ : str = NestedDataStructure(__magic_name__ ).data assert output_data == input_data @pytest.mark.parametrize( "data, expected_output" ,[ ({}, []), ([], []), ("foo", ["foo"]), (["foo", "bar"], ["foo", "bar"]), ([["foo", "bar"]], ["foo", "bar"]), ([[["foo"], ["bar"]]], ["foo", "bar"]), ([[["foo"], "bar"]], ["foo", "bar"]), ({"a": 1, "b": 2}, [1, 2]), ({"a": [1, 2], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[1, 2]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[[3], [4]]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [[3, 4]]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, 4]}, [1, 2, 3, 4]), ({"a": [[[1], [2]]], "b": [3, [4]]}, [1, 2, 3, 4]), ({"a": {"1": 1}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": 2}, [1, 2]), ({"a": {"1": [1]}, "b": [2]}, [1, 2]), ] ,) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : Any = NestedDataStructure(__magic_name__ ).flatten() assert output == expected_output def __UpperCAmelCase ( )-> Union[str, Any]: """simple docstring""" snake_case_ : Tuple = A(x=1 ,y="foobar" ) snake_case_ : Tuple = {"x": 1, "y": "foobar"} assert asdict(__magic_name__ ) == expected_output snake_case_ : int = {"a": {"b": A(x=10 ,y="foo" )}, "c": [A(x=20 ,y="bar" )]} snake_case_ : str = {"a": {"b": {"x": 10, "y": "foo"}}, "c": [{"x": 20, "y": "bar"}]} assert asdict(__magic_name__ ) == expected_output with pytest.raises(__magic_name__ ): asdict([1, A(x=10 ,y="foo" )] ) def __UpperCAmelCase ( __magic_name__ )-> Tuple: """simple docstring""" return text.split() def __UpperCAmelCase ( __magic_name__ )-> Optional[Any]: """simple docstring""" yield (time.time(), content) time.sleep(2 ) yield (time.time(), content) def __UpperCAmelCase ( )-> Tuple: """simple docstring""" with Pool(2 ) as pool: snake_case_ : List[Any] = list(iflatmap_unordered(__magic_name__ ,_split_text ,kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(__magic_name__ ) == 20 # check multiprocess from pathos (uses dill for pickling) with multiprocess.Pool(2 ) as pool: snake_case_ : List[str] = list(iflatmap_unordered(__magic_name__ ,_split_text ,kwargs_iterable=[{"text": "hello there"}] * 10 ) ) assert out.count("hello" ) == 10 assert out.count("there" ) == 10 assert len(__magic_name__ ) == 20 # check that we get items as fast as possible with Pool(2 ) as pool: snake_case_ : Any = [] for yield_time, content in iflatmap_unordered( __magic_name__ ,_aseconds_generator_of_aitems_with_timing ,kwargs_iterable=[{"content": "a"}, {"content": "b"}] ): assert yield_time < time.time() + 0.1, "we should each item directly after it was yielded" out.append(__magic_name__ ) assert out.count("a" ) == 2 assert out.count("b" ) == 2 assert len(__magic_name__ ) == 4
711
'''simple docstring''' from scipy.stats import spearmanr import datasets __lowerCamelCase : str = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' __lowerCamelCase : int = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' __lowerCamelCase : List[str] = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): """simple docstring""" def _A ( self :str ) -> Union[str, Any]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } ) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"] , ) def _A ( self :Optional[int] , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :Optional[Any]=False ) -> List[Any]: '''simple docstring''' snake_case_ : Optional[Any] = spearmanr(lowerCAmelCase__ , lowerCAmelCase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
656
0
'''simple docstring''' import gc import importlib.metadata import tempfile import unittest from packaging import version from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoTokenizer, BitsAndBytesConfig, pipeline, ) from transformers.testing_utils import ( is_torch_available, require_accelerate, require_bitsandbytes, require_torch, require_torch_gpu, require_torch_multi_gpu, slow, ) def __UpperCAmelCase ( __magic_name__ )-> str: """simple docstring""" if model.config.model_type == "gpt2": return model.transformer.h[0].mlp.c_fc return model.transformer.h[0].mlp.dense_ah_to_h if is_torch_available(): import torch import torch.nn as nn class A_ (nn.Module ): """simple docstring""" def __init__( self :Optional[int] , lowerCAmelCase__ :nn.Module , lowerCAmelCase__ :int ) -> List[Any]: '''simple docstring''' super().__init__() snake_case_ : int = module snake_case_ : Dict = nn.Sequential( nn.Linear(module.in_features , lowerCAmelCase__ , bias=lowerCAmelCase__ ) , nn.Linear(lowerCAmelCase__ , module.out_features , bias=lowerCAmelCase__ ) , ) snake_case_ : List[str] = (2.0 / (5 * min(module.in_features , module.out_features ))) ** 0.5 nn.init.normal_(self.adapter[0].weight , std=lowerCAmelCase__ ) nn.init.zeros_(self.adapter[1].weight ) self.adapter.to(module.weight.device ) def _A ( self :Optional[int] , lowerCAmelCase__ :str , *lowerCAmelCase__ :Optional[Any] , **lowerCAmelCase__ :Tuple ) -> List[str]: '''simple docstring''' return self.module(lowerCAmelCase__ , *lowerCAmelCase__ , **lowerCAmelCase__ ) + self.adapter(lowerCAmelCase__ ) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class A_ (unittest.TestCase ): """simple docstring""" a__ = '''bigscience/bloom-1b7''' # Constant values a__ = 2.109659552692574 a__ = '''Hello my name is''' a__ = set() EXPECTED_OUTPUTS.add('''Hello my name is John and I am a professional photographer. I''' ) EXPECTED_OUTPUTS.add('''Hello my name is John.\nI am a friend of your father.\n''' ) EXPECTED_OUTPUTS.add('''Hello my name is John Doe, I am a student at the University''' ) a__ = 10 def _A ( self :List[Any] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Union[str, Any] = AutoTokenizer.from_pretrained(self.model_name ) class A_ (a_ ): """simple docstring""" def _A ( self :Optional[int] ) -> Union[str, Any]: '''simple docstring''' super().setUp() # Models and tokenizer snake_case_ : List[Any] = AutoModelForCausalLM.from_pretrained( self.model_name , torch_dtype=torch.floataa , device_map="auto" ) snake_case_ : str = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=lowerCAmelCase__ , device_map="auto" ) def _A ( self :Any ) -> Any: '''simple docstring''' del self.model_fpaa del self.model_abit gc.collect() torch.cuda.empty_cache() def _A ( self :Dict ) -> Dict: '''simple docstring''' snake_case_ : List[Any] = self.model_abit.config self.assertTrue(hasattr(lowerCAmelCase__ , "quantization_config" ) ) snake_case_ : List[Any] = config.to_dict() snake_case_ : Optional[int] = config.to_diff_dict() snake_case_ : int = config.to_json_string() def _A ( self :List[str] ) -> int: '''simple docstring''' from bitsandbytes.nn import Paramsabit snake_case_ : Union[str, Any] = self.model_fpaa.get_memory_footprint() snake_case_ : Any = self.model_abit.get_memory_footprint() self.assertAlmostEqual(mem_fpaa / mem_abit , self.EXPECTED_RELATIVE_DIFFERENCE ) snake_case_ : str = get_some_linear_layer(self.model_abit ) self.assertTrue(linear.weight.__class__ == Paramsabit ) def _A ( self :Union[str, Any] ) -> List[Any]: '''simple docstring''' from transformers import TaPreTrainedModel self.model_fpaa.get_memory_footprint() self.model_abit.get_memory_footprint() for name, module in self.model_abit.named_modules(): if isinstance(lowerCAmelCase__ , torch.nn.Linear ): if name not in ["lm_head"] + TaPreTrainedModel._keep_in_fpaa_modules: # 4-bit parameters are packed in uint8 variables self.assertTrue(module.weight.dtype == torch.uinta ) def _A ( self :Dict ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = self.tokenizer(self.input_text , return_tensors="pt" ) snake_case_ : Union[str, Any] = self.model_abit.generate(input_ids=encoded_input["input_ids"].to(0 ) , max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=lowerCAmelCase__ ) , self.EXPECTED_OUTPUTS ) def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case_ : Optional[int] = BitsAndBytesConfig() snake_case_ : int = True snake_case_ : Tuple = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=lowerCAmelCase__ , device_map="auto" ) snake_case_ : List[str] = self.tokenizer(self.input_text , return_tensors="pt" ) snake_case_ : Any = model_abit_from_config.generate( input_ids=encoded_input["input_ids"].to(0 ) , max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=lowerCAmelCase__ ) , self.EXPECTED_OUTPUTS ) def _A ( self :Optional[int] ) -> Optional[Any]: '''simple docstring''' with self.assertRaises(lowerCAmelCase__ ), tempfile.TemporaryDirectory() as tmpdirname: self.model_abit.save_pretrained(lowerCAmelCase__ ) def _A ( self :Optional[int] ) -> str: '''simple docstring''' snake_case_ : Tuple = BitsAndBytesConfig() with self.assertRaises(lowerCAmelCase__ ): snake_case_ : Union[str, Any] = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=lowerCAmelCase__ , load_in_abit=lowerCAmelCase__ , device_map="auto" , bnb_abit_quant_type="nf4" , ) def _A ( self :Any ) -> int: '''simple docstring''' with self.assertRaises(lowerCAmelCase__ ): # Tries with `str` self.model_abit.to("cpu" ) with self.assertRaises(lowerCAmelCase__ ): # Tries with a `dtype`` self.model_abit.to(torch.floataa ) with self.assertRaises(lowerCAmelCase__ ): # Tries with a `device` self.model_abit.to(torch.device("cuda:0" ) ) with self.assertRaises(lowerCAmelCase__ ): # Tries with a `device` self.model_abit.float() with self.assertRaises(lowerCAmelCase__ ): # Tries with a `device` self.model_abit.half() # Test if we did not break anything snake_case_ : Any = self.tokenizer(self.input_text , return_tensors="pt" ) snake_case_ : Optional[Any] = self.model_fpaa.to(torch.floataa ) snake_case_ : Optional[int] = self.model_fpaa.generate(input_ids=encoded_input["input_ids"].to(0 ) , max_new_tokens=10 ) # Check this does not throw an error snake_case_ : Union[str, Any] = self.model_fpaa.to("cpu" ) # Check this does not throw an error snake_case_ : Optional[Any] = self.model_fpaa.half() # Check this does not throw an error snake_case_ : int = self.model_fpaa.float() def _A ( self :Optional[Any] ) -> Dict: '''simple docstring''' snake_case_ : Any = AutoModelForSeqaSeqLM.from_pretrained("t5-small" , load_in_abit=lowerCAmelCase__ , device_map="auto" ) self.assertTrue(model.decoder.block[0].layer[2].DenseReluDense.wo.weight.dtype == torch.floataa ) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class A_ (unittest.TestCase ): """simple docstring""" @classmethod def _A ( cls :List[Any] ) -> Dict: '''simple docstring''' snake_case_ : Any = "t5-small" snake_case_ : Union[str, Any] = "google/flan-t5-small" # flan-t5 uses dense-act instead of dense-relu-dense snake_case_ : int = AutoTokenizer.from_pretrained(cls.model_name ) snake_case_ : str = "Translate in German: Hello, my dog is cute" def _A ( self :int ) -> Optional[Any]: '''simple docstring''' gc.collect() torch.cuda.empty_cache() def _A ( self :Dict ) -> Any: '''simple docstring''' from transformers import TaForConditionalGeneration snake_case_ : List[Any] = TaForConditionalGeneration._keep_in_fpaa_modules snake_case_ : Optional[Any] = None # test with `t5-small` snake_case_ : Any = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=lowerCAmelCase__ , device_map="auto" ) snake_case_ : Union[str, Any] = self.tokenizer(self.input_text , return_tensors="pt" ).to(0 ) snake_case_ : int = model.generate(**lowerCAmelCase__ ) # test with `flan-t5-small` snake_case_ : Tuple = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=lowerCAmelCase__ , device_map="auto" ) snake_case_ : List[str] = self.tokenizer(self.input_text , return_tensors="pt" ).to(0 ) snake_case_ : Optional[Any] = model.generate(**lowerCAmelCase__ ) snake_case_ : List[str] = modules def _A ( self :Tuple ) -> Any: '''simple docstring''' import bitsandbytes as bnb from transformers import TaForConditionalGeneration # test with `t5-small` snake_case_ : Tuple = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=lowerCAmelCase__ , device_map="auto" ) # there was a bug with decoders - this test checks that it is fixed self.assertTrue(isinstance(model.decoder.block[0].layer[0].SelfAttention.q , bnb.nn.Linearabit ) ) snake_case_ : Any = self.tokenizer(self.input_text , return_tensors="pt" ).to(0 ) snake_case_ : List[str] = model.generate(**lowerCAmelCase__ ) # test with `flan-t5-small` snake_case_ : Tuple = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=lowerCAmelCase__ , device_map="auto" ) snake_case_ : Dict = self.tokenizer(self.input_text , return_tensors="pt" ).to(0 ) snake_case_ : Optional[Any] = model.generate(**lowerCAmelCase__ ) class A_ (a_ ): """simple docstring""" def _A ( self :List[str] ) -> List[str]: '''simple docstring''' super().setUp() # model_name snake_case_ : Union[str, Any] = "bigscience/bloom-560m" snake_case_ : Optional[Any] = "t5-small" # Different types of model snake_case_ : Optional[int] = AutoModel.from_pretrained(self.model_name , load_in_abit=lowerCAmelCase__ , device_map="auto" ) # Sequence classification model snake_case_ : str = AutoModelForSequenceClassification.from_pretrained( self.model_name , load_in_abit=lowerCAmelCase__ , device_map="auto" ) # CausalLM model snake_case_ : Union[str, Any] = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=lowerCAmelCase__ , device_map="auto" ) # Seq2seq model snake_case_ : Optional[int] = AutoModelForSeqaSeqLM.from_pretrained( self.seq_to_seq_name , load_in_abit=lowerCAmelCase__ , device_map="auto" ) def _A ( self :Dict ) -> Optional[int]: '''simple docstring''' del self.base_model del self.sequence_model del self.model_abit del self.seq_to_seq_model gc.collect() torch.cuda.empty_cache() def _A ( self :Any ) -> Dict: '''simple docstring''' from bitsandbytes.nn import Paramsabit self.assertTrue(self.base_model.h[-1].mlp.dense_ah_to_h.weight.__class__ == Paramsabit ) # Other heads should be nn.Parameter self.assertTrue(self.model_abit.lm_head.weight.__class__ == torch.nn.Parameter ) self.assertTrue(self.sequence_model.score.weight.__class__ == torch.nn.Parameter ) self.assertTrue(self.seq_to_seq_model.lm_head.weight.__class__ == torch.nn.Parameter ) class A_ (a_ ): """simple docstring""" def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' super().setUp() def _A ( self :Optional[Any] ) -> List[Any]: '''simple docstring''' del self.pipe gc.collect() torch.cuda.empty_cache() def _A ( self :Any ) -> Dict: '''simple docstring''' snake_case_ : Optional[int] = pipeline( "text-generation" , model=self.model_name , model_kwargs={"device_map": "auto", "load_in_4bit": True, "torch_dtype": torch.floataa} , max_new_tokens=self.MAX_NEW_TOKENS , ) # Real second forward pass snake_case_ : Dict = self.pipe(self.input_text ) self.assertIn(pipeline_output[0]["generated_text"] , self.EXPECTED_OUTPUTS ) @require_torch_multi_gpu class A_ (a_ ): """simple docstring""" def _A ( self :Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' super().setUp() def _A ( self :Optional[Any] ) -> List[Any]: '''simple docstring''' snake_case_ : Any = AutoModelForCausalLM.from_pretrained( self.model_name , load_in_abit=lowerCAmelCase__ , device_map="balanced" ) # Check correct device map self.assertEqual(set(model_parallel.hf_device_map.values() ) , {0, 1} ) # Check that inference pass works on the model snake_case_ : Any = self.tokenizer(self.input_text , return_tensors="pt" ) # Second real batch snake_case_ : Tuple = model_parallel.generate(input_ids=encoded_input["input_ids"].to(0 ) , max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_parallel[0] , skip_special_tokens=lowerCAmelCase__ ) , self.EXPECTED_OUTPUTS ) class A_ (a_ ): """simple docstring""" def _A ( self :List[Any] ) -> str: '''simple docstring''' snake_case_ : Dict = "facebook/opt-350m" super().setUp() def _A ( self :Any ) -> Tuple: '''simple docstring''' if version.parse(importlib.metadata.version("bitsandbytes" ) ) < version.parse("0.37.0" ): return # Step 1: freeze all parameters snake_case_ : str = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=lowerCAmelCase__ ) self.assertEqual(set(model.hf_device_map.values() ) , {torch.cuda.current_device()} ) for param in model.parameters(): snake_case_ : Optional[Any] = False # freeze the model - train adapters later if param.ndim == 1: # cast the small parameters (e.g. layernorm) to fp32 for stability snake_case_ : Optional[Any] = param.data.to(torch.floataa ) # Step 2: add adapters for _, module in model.named_modules(): if "OPTAttention" in repr(type(lowerCAmelCase__ ) ): snake_case_ : Union[str, Any] = LoRALayer(module.q_proj , rank=16 ) snake_case_ : Dict = LoRALayer(module.k_proj , rank=16 ) snake_case_ : Optional[Any] = LoRALayer(module.v_proj , rank=16 ) # Step 3: dummy batch snake_case_ : Optional[int] = self.tokenizer("Test batch " , return_tensors="pt" ).to(0 ) # Step 4: Check if the gradient is not None with torch.cuda.amp.autocast(): snake_case_ : Optional[int] = model.forward(**lowerCAmelCase__ ) out.logits.norm().backward() for module in model.modules(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertTrue(module.adapter[1].weight.grad is not None ) self.assertTrue(module.adapter[1].weight.grad.norm().item() > 0 ) elif isinstance(lowerCAmelCase__ , nn.Embedding ): self.assertTrue(module.weight.grad is None ) class A_ (a_ ): """simple docstring""" a__ = '''gpt2-xl''' a__ = 3.3191854854152187
712
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): __lowerCamelCase : Optional[Any] = get_tests_dir('''fixtures/test_sentencepiece.model''') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right __lowerCamelCase : str = 128022 __lowerCamelCase : List[Any] = 128028 @require_sentencepiece class A_ (a_ , unittest.TestCase ): """simple docstring""" a__ = MaMaaaTokenizer a__ = False a__ = False a__ = True def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' super().setUp() snake_case_ : int = ["</s>", "<unk>", "▁This", "▁is", "▁a", "▁t", "est", "\u0120", "<pad>"] snake_case_ : Any = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) snake_case_ : Optional[int] = Path(self.tmpdirname ) save_json(lowerCAmelCase__ , save_dir / VOCAB_FILES_NAMES["vocab_file"] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(lowerCAmelCase__ , save_dir / VOCAB_FILES_NAMES["spm_file"] ) snake_case_ : Union[str, Any] = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def _A ( self :List[Any] , **lowerCAmelCase__ :List[Any] ) -> str: '''simple docstring''' return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def _A ( self :Optional[int] , lowerCAmelCase__ :Any ) -> Optional[int]: '''simple docstring''' return ( "This is a test", "This is a test", ) def _A ( self :List[str] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : str = "</s>" snake_case_ : Union[str, Any] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase__ ) , lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = self.get_tokenizer() snake_case_ : Any = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "</s>" ) self.assertEqual(vocab_keys[1] , "<unk>" ) self.assertEqual(vocab_keys[-1] , "<s>" ) self.assertEqual(len(lowerCAmelCase__ ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip("Skip this test while all models are still to be uploaded." ) def _A ( self :List[Any] ) -> Union[str, Any]: '''simple docstring''' pass def _A ( self :Optional[int] ) -> int: '''simple docstring''' snake_case_ : int = self.get_tokenizer() snake_case_ : List[str] = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowerCAmelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [2, 3, 4, 5, 6] , ) snake_case_ : Any = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(lowerCAmelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) snake_case_ : Any = tokenizer.convert_tokens_to_string(lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , "This is a test" ) @slow def _A ( self :Any ) -> List[Any]: '''simple docstring''' snake_case_ : int = {"input_ids": [[128_022, 110_108, 397, 11, 38_272, 2_247, 124_811, 285, 18_105, 1_586, 207, 7, 39_534, 4_428, 397, 1_019, 18_105, 1_586, 207, 7, 41_337, 16_786, 241, 7, 20_214, 17, 125_690, 10_398, 7, 44_378, 58_069, 68_342, 7_798, 7_343, 11, 299, 33_310, 4, 158, 37_350, 94_077, 4_569, 299, 33_310, 90, 4, 52_840, 290, 4, 31_270, 112, 299, 682, 4, 52_840, 39_953, 14_079, 193, 52_519, 90_894, 17_894, 120_697, 11, 40_445, 551, 17, 1_019, 52_519, 90_894, 17_756, 963, 11, 40_445, 480, 17, 9_792, 1_120, 5_173, 1_393, 6_240, 16_786, 241, 120_996, 28, 1_245, 1_393, 118_240, 11_123, 1_019, 93_612, 2_691, 10_618, 98_058, 120_409, 1_928, 279, 4, 40_683, 367, 178, 207, 1_019, 103, 103_121, 506, 65_296, 5, 2], [128_022, 21_217, 367, 117, 125_450, 128, 719, 7, 7_308, 40, 93_612, 12_669, 1_116, 16_704, 71, 17_785, 3_699, 15_592, 35, 144, 9_584, 241, 11_943, 713, 950, 799, 2_247, 88_427, 150, 149, 118_813, 120_706, 1_019, 106_906, 81_518, 28, 1_224, 22_799, 397, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [128_022, 1_658, 123_311, 5_155, 5_578, 4_722, 279, 14_947, 2_366, 1_120, 1_197, 14, 1_348, 9_232, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase__ , model_name="facebook/m2m100_418M" , revision="c168bae485c864188cf9aa0e4108b0b6934dc91e" , ) @require_torch @require_sentencepiece @require_tokenizers class A_ (unittest.TestCase ): """simple docstring""" a__ = '''facebook/m2m100_418M''' a__ = [ '''In my opinion, there are two levels of response from the French government.''', '''NSA Affair Emphasizes Complete Lack of Debate on Intelligence''', ] a__ = [ '''Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.''', '''L\'affaire NSA souligne l\'absence totale de débat sur le renseignement''', ] # fmt: off a__ = [EN_CODE, 593, 1949, 115781, 4, 71586, 4234, 60633, 126233, 432, 123808, 15592, 1197, 117132, 120618, 5, 2] @classmethod def _A ( cls :str ) -> int: '''simple docstring''' snake_case_ : MaMaaaTokenizer = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en" , tgt_lang="fr" ) snake_case_ : List[str] = 1 return cls def _A ( self :Tuple ) -> Union[str, Any]: '''simple docstring''' self.assertEqual(self.tokenizer.get_lang_id("ar" ) , 128_006 ) self.assertEqual(self.tokenizer.get_lang_id("en" ) , 128_022 ) self.assertEqual(self.tokenizer.get_lang_id("ro" ) , 128_076 ) self.assertEqual(self.tokenizer.get_lang_id("mr" ) , 128_063 ) def _A ( self :Optional[int] ) -> List[str]: '''simple docstring''' snake_case_ : Dict = self.tokenizer.get_vocab() self.assertEqual(len(lowerCAmelCase__ ) , self.tokenizer.vocab_size ) self.assertEqual(vocab["<unk>"] , 3 ) self.assertIn(self.tokenizer.get_lang_token("en" ) , lowerCAmelCase__ ) def _A ( self :Any ) -> Dict: '''simple docstring''' snake_case_ : List[str] = "en" snake_case_ : Dict = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , lowerCAmelCase__ ) def _A ( self :Union[str, Any] ) -> Dict: '''simple docstring''' self.assertIn(lowerCAmelCase__ , self.tokenizer.all_special_ids ) # fmt: off snake_case_ : Dict = [FR_CODE, 5_364, 82, 8_642, 4, 294, 47, 8, 14_028, 136, 3_286, 9_706, 6, 90_797, 6, 144_012, 162, 88_128, 30_061, 5, 2] # fmt: on snake_case_ : List[str] = self.tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) snake_case_ : str = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=lowerCAmelCase__ ) self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertNotIn(self.tokenizer.eos_token , lowerCAmelCase__ ) def _A ( self :Tuple ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = tempfile.mkdtemp() snake_case_ : int = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(lowerCAmelCase__ ) snake_case_ : List[str] = MaMaaaTokenizer.from_pretrained(lowerCAmelCase__ ) self.assertDictEqual(new_tok.lang_token_to_id , lowerCAmelCase__ ) @require_torch def _A ( self :Optional[Any] ) -> str: '''simple docstring''' snake_case_ : Union[str, Any] = "en" snake_case_ : Tuple = "fr" snake_case_ : Optional[int] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=lowerCAmelCase__ , return_tensors="pt" ) snake_case_ : Dict = shift_tokens_right( batch["labels"] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: snake_case_ : str = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def _A ( self :Optional[Any] ) -> Tuple: '''simple docstring''' snake_case_ : List[str] = "mr" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) snake_case_ : int = "zh" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def _A ( self :str ) -> int: '''simple docstring''' snake_case_ : Dict = "mr" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) snake_case_ : Tuple = "zh" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Optional[int] = self.tokenizer._build_translation_inputs("A test" , return_tensors="pt" , src_lang="en" , tgt_lang="ar" ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , { # en_XX, A, test, EOS "input_ids": [[128_022, 58, 4_183, 2]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 128_006, } , )
656
0
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional from packaging import version if TYPE_CHECKING: from ... import PreTrainedTokenizer, TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import is_torch_available, logging __lowerCamelCase : List[str] = logging.get_logger(__name__) __lowerCamelCase : Tuple = { '''bigscience/bloom''': '''https://huggingface.co/bigscience/bloom/resolve/main/config.json''', '''bigscience/bloom-560m''': '''https://huggingface.co/bigscience/bloom-560m/blob/main/config.json''', '''bigscience/bloom-1b1''': '''https://huggingface.co/bigscience/bloom-1b1/blob/main/config.json''', '''bigscience/bloom-1b7''': '''https://huggingface.co/bigscience/bloom-1b7/blob/main/config.json''', '''bigscience/bloom-3b''': '''https://huggingface.co/bigscience/bloom-3b/blob/main/config.json''', '''bigscience/bloom-7b1''': '''https://huggingface.co/bigscience/bloom-7b1/blob/main/config.json''', } class A_ (a_ ): """simple docstring""" a__ = '''bloom''' a__ = ['''past_key_values'''] a__ = { '''num_hidden_layers''': '''n_layer''', '''num_attention_heads''': '''n_head''', } def __init__( self :Union[str, Any] , lowerCAmelCase__ :List[Any]=250_880 , lowerCAmelCase__ :List[str]=64 , lowerCAmelCase__ :Optional[Any]=2 , lowerCAmelCase__ :Union[str, Any]=8 , lowerCAmelCase__ :Optional[Any]=1E-5 , lowerCAmelCase__ :List[str]=0.0_2 , lowerCAmelCase__ :Union[str, Any]=True , lowerCAmelCase__ :Optional[Any]=1 , lowerCAmelCase__ :str=2 , lowerCAmelCase__ :Optional[int]=False , lowerCAmelCase__ :str=0.0 , lowerCAmelCase__ :Tuple=0.0 , lowerCAmelCase__ :Optional[Any]=1 , lowerCAmelCase__ :Optional[int]=False , **lowerCAmelCase__ :str , ) -> Dict: '''simple docstring''' snake_case_ : str = vocab_size # Backward compatibility with n_embed kwarg snake_case_ : Any = kwargs.pop("n_embed" , lowerCAmelCase__ ) snake_case_ : List[Any] = hidden_size if n_embed is None else n_embed snake_case_ : str = n_layer snake_case_ : Optional[Any] = n_head snake_case_ : int = layer_norm_epsilon snake_case_ : Dict = initializer_range snake_case_ : Tuple = use_cache snake_case_ : Any = pretraining_tp snake_case_ : Any = apply_residual_connection_post_layernorm snake_case_ : List[Any] = hidden_dropout snake_case_ : Dict = attention_dropout snake_case_ : Union[str, Any] = bos_token_id snake_case_ : List[str] = eos_token_id snake_case_ : Any = slow_but_exact super().__init__(bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) class A_ (a_ ): """simple docstring""" a__ = version.parse('''1.12''' ) def __init__( self :Optional[int] , lowerCAmelCase__ :PretrainedConfig , lowerCAmelCase__ :str = "default" , lowerCAmelCase__ :List[PatchingSpec] = None , lowerCAmelCase__ :bool = False , ) -> Dict: '''simple docstring''' super().__init__(lowerCAmelCase__ , task=lowerCAmelCase__ , patching_specs=lowerCAmelCase__ , use_past=lowerCAmelCase__ ) if not getattr(self._config , "pad_token_id" , lowerCAmelCase__ ): # TODO: how to do that better? snake_case_ : List[Any] = 0 @property def _A ( self :Optional[int] ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' snake_case_ : List[str] = OrderedDict({"input_ids": {0: "batch", 1: "sequence"}} ) if self.use_past: # BLOOM stores values on dynamic axis 2. For more details see: https://github.com/huggingface/transformers/pull/18344 self.fill_with_past_key_values_(lowerCAmelCase__ , direction="inputs" , inverted_values_shape=lowerCAmelCase__ ) snake_case_ : int = {0: "batch", 1: "past_sequence + sequence"} else: snake_case_ : int = {0: "batch", 1: "sequence"} return common_inputs @property def _A ( self :str ) -> int: '''simple docstring''' return self._config.n_layer @property def _A ( self :int ) -> int: '''simple docstring''' return self._config.n_head @property def _A ( self :Tuple ) -> float: '''simple docstring''' return 1E-3 def _A ( self :Optional[int] , lowerCAmelCase__ :"PreTrainedTokenizer" , lowerCAmelCase__ :int = -1 , lowerCAmelCase__ :int = -1 , lowerCAmelCase__ :bool = False , lowerCAmelCase__ :Optional["TensorType"] = None , ) -> Mapping[str, Any]: '''simple docstring''' snake_case_ : int = super(lowerCAmelCase__ , self ).generate_dummy_inputs( lowerCAmelCase__ , batch_size=lowerCAmelCase__ , seq_length=lowerCAmelCase__ , is_pair=lowerCAmelCase__ , framework=lowerCAmelCase__ ) # We need to order the input in the way they appears in the forward() snake_case_ : Any = OrderedDict({"input_ids": common_inputs["input_ids"]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch snake_case_ : Optional[Any] = common_inputs["input_ids"].shape # Not using the same length for past_key_values snake_case_ : Any = seqlen + 2 snake_case_ : Union[str, Any] = self._config.hidden_size // self.num_attention_heads snake_case_ : str = ( batch * self.num_attention_heads, head_dim, past_key_values_length, ) snake_case_ : str = ( batch * self.num_attention_heads, past_key_values_length, head_dim, ) snake_case_ : Dict = [ (torch.zeros(lowerCAmelCase__ ), torch.zeros(lowerCAmelCase__ )) for _ in range(self.num_layers ) ] snake_case_ : str = common_inputs["attention_mask"] if self.use_past: snake_case_ : Any = ordered_inputs["attention_mask"].dtype snake_case_ : Union[str, Any] = torch.cat( [ordered_inputs["attention_mask"], torch.ones(lowerCAmelCase__ , lowerCAmelCase__ , dtype=lowerCAmelCase__ )] , dim=1 ) return ordered_inputs @property def _A ( self :Optional[int] ) -> int: '''simple docstring''' return 13
713
'''simple docstring''' import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py __lowerCamelCase : str = '''.''' # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) __lowerCamelCase : Tuple = [ '''Assert''', '''AssignVariableOp''', '''EmptyTensorList''', '''MergeV2Checkpoints''', '''ReadVariableOp''', '''ResourceGather''', '''RestoreV2''', '''SaveV2''', '''ShardedFilename''', '''StatefulPartitionedCall''', '''StaticRegexFullMatch''', '''VarHandleOp''', ] def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Tuple = SavedModel() snake_case_ : Dict = [] with open(os.path.join(__magic_name__ ,"utils" ,"tf_ops" ,"onnx.json" ) ) as f: snake_case_ : Dict = json.load(__magic_name__ )["opsets"] for i in range(1 ,opset + 1 ): onnx_ops.extend(onnx_opsets[str(__magic_name__ )] ) with open(__magic_name__ ,"rb" ) as f: saved_model.ParseFromString(f.read() ) snake_case_ : Tuple = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want snake_case_ : str = sorted(__magic_name__ ) snake_case_ : Optional[int] = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(__magic_name__ ) if strict and len(__magic_name__ ) > 0: raise Exception(F'''Found the following incompatible ops for the opset {opset}:\n''' + incompatible_ops ) elif len(__magic_name__ ) > 0: print(F'''Found the following incompatible ops for the opset {opset}:''' ) print(*__magic_name__ ,sep="\n" ) else: print(F'''The saved model {saved_model_path} can properly be converted with ONNX.''' ) if __name__ == "__main__": __lowerCamelCase : Any = argparse.ArgumentParser() parser.add_argument('''--saved_model_path''', help='''Path of the saved model to check (the .pb file).''') parser.add_argument( '''--opset''', default=12, type=int, help='''The ONNX opset against which the model has to be tested.''' ) parser.add_argument( '''--framework''', choices=['''onnx'''], default='''onnx''', help='''Frameworks against which to test the saved model.''' ) parser.add_argument( '''--strict''', action='''store_true''', help='''Whether make the checking strict (raise errors) or not (raise warnings)''' ) __lowerCamelCase : Dict = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
656
0
def __UpperCAmelCase ( __magic_name__ = 100_0000 )-> int: """simple docstring""" snake_case_ : Any = set(range(3 ,__magic_name__ ,2 ) ) primes.add(2 ) for p in range(3 ,__magic_name__ ,2 ): if p not in primes: continue primes.difference_update(set(range(p * p ,__magic_name__ ,__magic_name__ ) ) ) snake_case_ : Optional[Any] = [float(__magic_name__ ) for n in range(limit + 1 )] for p in primes: for n in range(__magic_name__ ,limit + 1 ,__magic_name__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'''{solution() = }''')
714
'''simple docstring''' import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal __lowerCamelCase : Optional[Any] = datasets.utils.logging.get_logger(__name__) __lowerCamelCase : List[str] = ['''names''', '''prefix'''] __lowerCamelCase : int = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] __lowerCamelCase : str = ['''encoding_errors''', '''on_bad_lines'''] __lowerCamelCase : Optional[Any] = ['''date_format'''] @dataclass class A_ (datasets.BuilderConfig ): """simple docstring""" a__ = "," a__ = None a__ = "infer" a__ = None a__ = None a__ = None a__ = None a__ = None a__ = True a__ = None a__ = None a__ = None a__ = None a__ = False a__ = None a__ = None a__ = None a__ = True a__ = True a__ = False a__ = True a__ = None a__ = "." a__ = None a__ = '"' a__ = 0 a__ = None a__ = None a__ = None a__ = None a__ = True a__ = True a__ = 0 a__ = True a__ = False a__ = None a__ = 10000 a__ = None a__ = "strict" a__ = "error" a__ = None def _A ( self :List[str] ) -> Any: '''simple docstring''' if self.delimiter is not None: snake_case_ : Tuple = self.delimiter if self.column_names is not None: snake_case_ : List[Any] = self.column_names @property def _A ( self :Optional[Any] ) -> int: '''simple docstring''' snake_case_ : Optional[int] = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__ ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class A_ (datasets.ArrowBasedBuilder ): """simple docstring""" a__ = CsvConfig def _A ( self :Optional[Any] ) -> Optional[Any]: '''simple docstring''' return datasets.DatasetInfo(features=self.config.features ) def _A ( self :Tuple , lowerCAmelCase__ :Dict ) -> List[Any]: '''simple docstring''' if not self.config.data_files: raise ValueError(F'''At least one data file must be specified, but got data_files={self.config.data_files}''' ) snake_case_ : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowerCAmelCase__ , (str, list, tuple) ): snake_case_ : int = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ : List[str] = [files] snake_case_ : Tuple = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] snake_case_ : str = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ : str = [files] snake_case_ : Any = [dl_manager.iter_files(lowerCAmelCase__ ) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={"files": files} ) ) return splits def _A ( self :List[Any] , lowerCAmelCase__ :pa.Table ) -> pa.Table: '''simple docstring''' if self.config.features is not None: snake_case_ : int = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__ ) for feature in self.config.features.values() ): # cheaper cast snake_case_ : Optional[Any] = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__ ) else: # more expensive cast; allows str <-> int/float or str to Audio for example snake_case_ : Dict = table_cast(lowerCAmelCase__ , lowerCAmelCase__ ) return pa_table def _A ( self :Dict , lowerCAmelCase__ :Union[str, Any] ) -> Optional[int]: '''simple docstring''' snake_case_ : Tuple = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str snake_case_ : str = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__ ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__ ) ): snake_case_ : Tuple = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(lowerCAmelCase__ ): snake_case_ : Optional[int] = pa.Table.from_pandas(lowerCAmelCase__ ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__ ) except ValueError as e: logger.error(F'''Failed to read file \'{file}\' with error {type(lowerCAmelCase__ )}: {e}''' ) raise
656
0
import doctest import glob import importlib import inspect import os import re from contextlib import contextmanager from functools import wraps from unittest.mock import patch import numpy as np import pytest from absl.testing import parameterized import datasets from datasets import load_metric from .utils import for_all_test_methods, local, slow # mark all tests as integration __lowerCamelCase : Dict = pytest.mark.integration __lowerCamelCase : Any = {'''comet'''} __lowerCamelCase : Optional[int] = importlib.util.find_spec('''fairseq''') is not None __lowerCamelCase : int = {'''code_eval'''} __lowerCamelCase : Dict = os.name == '''nt''' __lowerCamelCase : str = {'''bertscore''', '''frugalscore''', '''perplexity'''} __lowerCamelCase : List[str] = importlib.util.find_spec('''transformers''') is not None def __UpperCAmelCase ( __magic_name__ )-> Union[str, Any]: """simple docstring""" @wraps(__magic_name__ ) def wrapper(self ,__magic_name__ ): if not _has_fairseq and metric_name in REQUIRE_FAIRSEQ: self.skipTest("\"test requires Fairseq\"" ) else: test_case(self ,__magic_name__ ) return wrapper def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" @wraps(__magic_name__ ) def wrapper(self ,__magic_name__ ): if not _has_transformers and metric_name in REQUIRE_TRANSFORMERS: self.skipTest("\"test requires transformers\"" ) else: test_case(self ,__magic_name__ ) return wrapper def __UpperCAmelCase ( __magic_name__ )-> Optional[int]: """simple docstring""" @wraps(__magic_name__ ) def wrapper(self ,__magic_name__ ): if _on_windows and metric_name in UNSUPPORTED_ON_WINDOWS: self.skipTest("\"test not supported on Windows\"" ) else: test_case(self ,__magic_name__ ) return wrapper def __UpperCAmelCase ( )-> Optional[Any]: """simple docstring""" snake_case_ : Optional[Any] = [metric_dir.split(os.sep )[-2] for metric_dir in glob.glob("./metrics/*/" )] return [{"testcase_name": x, "metric_name": x} for x in metrics if x != "gleu"] # gleu is unfinished @parameterized.named_parameters(get_local_metric_names() ) @for_all_test_methods( a_ , a_ , a_ ) @local class A_ (parameterized.TestCase ): """simple docstring""" a__ = {} a__ = None @pytest.mark.filterwarnings("ignore:metric_module_factory is deprecated:FutureWarning" ) @pytest.mark.filterwarnings("ignore:load_metric is deprecated:FutureWarning" ) def _A ( self :List[str] , lowerCAmelCase__ :List[Any] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Dict = "[...]" snake_case_ : Union[str, Any] = importlib.import_module( datasets.load.metric_module_factory(os.path.join("metrics" , lowerCAmelCase__ ) ).module_path ) snake_case_ : Optional[Any] = datasets.load.import_main_class(metric_module.__name__ , dataset=lowerCAmelCase__ ) # check parameters snake_case_ : int = inspect.signature(metric._compute ).parameters self.assertTrue(all(p.kind != p.VAR_KEYWORD for p in parameters.values() ) ) # no **kwargs # run doctest with self.patch_intensive_calls(lowerCAmelCase__ , metric_module.__name__ ): with self.use_local_metrics(): try: snake_case_ : Optional[Any] = doctest.testmod(lowerCAmelCase__ , verbose=lowerCAmelCase__ , raise_on_error=lowerCAmelCase__ ) except doctest.UnexpectedException as e: raise e.exc_info[1] # raise the exception that doctest caught self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @slow def _A ( self :Union[str, Any] , lowerCAmelCase__ :Optional[int] ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[Any] = "[...]" snake_case_ : List[Any] = importlib.import_module( datasets.load.metric_module_factory(os.path.join("metrics" , lowerCAmelCase__ ) ).module_path ) # run doctest with self.use_local_metrics(): snake_case_ : Any = doctest.testmod(lowerCAmelCase__ , verbose=lowerCAmelCase__ , raise_on_error=lowerCAmelCase__ ) self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @contextmanager def _A ( self :List[str] , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :List[str] ) -> Optional[Any]: '''simple docstring''' if metric_name in self.INTENSIVE_CALLS_PATCHER: with self.INTENSIVE_CALLS_PATCHER[metric_name](lowerCAmelCase__ ): yield else: yield @contextmanager def _A ( self :Optional[int] ) -> Optional[int]: '''simple docstring''' def load_local_metric(lowerCAmelCase__ :Optional[Any] , *lowerCAmelCase__ :Any , **lowerCAmelCase__ :List[str] ): return load_metric(os.path.join("metrics" , lowerCAmelCase__ ) , *lowerCAmelCase__ , **lowerCAmelCase__ ) with patch("datasets.load_metric" ) as mock_load_metric: snake_case_ : List[str] = load_local_metric yield @classmethod def _A ( cls :Optional[Any] , lowerCAmelCase__ :str ) -> str: '''simple docstring''' def wrapper(lowerCAmelCase__ :Any ): snake_case_ : Optional[int] = contextmanager(lowerCAmelCase__ ) snake_case_ : List[Any] = patcher return patcher return wrapper @LocalMetricTest.register_intensive_calls_patcher("bleurt" ) def __UpperCAmelCase ( __magic_name__ )-> List[Any]: """simple docstring""" import tensorflow.compat.va as tf from bleurt.score import Predictor tf.flags.DEFINE_string("sv" ,"" ,"" ) # handle pytest cli flags class A_ (a_ ): """simple docstring""" def _A ( self :Any , lowerCAmelCase__ :Optional[Any] ) -> str: '''simple docstring''' assert len(input_dict["input_ids"] ) == 2 return np.array([1.0_3, 1.0_4] ) # mock predict_fn which is supposed to do a forward pass with a bleurt model with patch("bleurt.score._create_predictor" ) as mock_create_predictor: snake_case_ : Dict = MockedPredictor() yield @LocalMetricTest.register_intensive_calls_patcher("bertscore" ) def __UpperCAmelCase ( __magic_name__ )-> Tuple: """simple docstring""" import torch def bert_cos_score_idf(__magic_name__ ,__magic_name__ ,*__magic_name__ ,**__magic_name__ ): return torch.tensor([[1.0, 1.0, 1.0]] * len(__magic_name__ ) ) # mock get_model which is supposed to do download a bert model # mock bert_cos_score_idf which is supposed to do a forward pass with a bert model with patch("bert_score.scorer.get_model" ), patch( "bert_score.scorer.bert_cos_score_idf" ) as mock_bert_cos_score_idf: snake_case_ : List[Any] = bert_cos_score_idf yield @LocalMetricTest.register_intensive_calls_patcher("comet" ) def __UpperCAmelCase ( __magic_name__ )-> Optional[int]: """simple docstring""" def load_from_checkpoint(__magic_name__ ): class A_ : """simple docstring""" def _A ( self :Tuple , lowerCAmelCase__ :Tuple , *lowerCAmelCase__ :Optional[Any] , **lowerCAmelCase__ :str ) -> List[Any]: '''simple docstring''' assert len(lowerCAmelCase__ ) == 2 snake_case_ : Tuple = [0.1_9, 0.9_2] return scores, sum(lowerCAmelCase__ ) / len(lowerCAmelCase__ ) return Model() # mock load_from_checkpoint which is supposed to do download a bert model # mock load_from_checkpoint which is supposed to do download a bert model with patch("comet.download_model" ) as mock_download_model: snake_case_ : Union[str, Any] = None with patch("comet.load_from_checkpoint" ) as mock_load_from_checkpoint: snake_case_ : int = load_from_checkpoint yield def __UpperCAmelCase ( )-> List[Any]: """simple docstring""" snake_case_ : Optional[int] = load_metric(os.path.join("metrics" ,"seqeval" ) ) snake_case_ : int = "ERROR" snake_case_ : int = F'''Scheme should be one of [IOB1, IOB2, IOE1, IOE2, IOBES, BILOU], got {wrong_scheme}''' with pytest.raises(__magic_name__ ,match=re.escape(__magic_name__ ) ): metric.compute(predictions=[] ,references=[] ,scheme=__magic_name__ )
715
'''simple docstring''' import json import os import unittest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A_ (a_ , unittest.TestCase ): """simple docstring""" a__ = MgpstrTokenizer a__ = False a__ = {} a__ = False def _A ( self :List[str] ) -> List[str]: '''simple docstring''' super().setUp() # fmt: off snake_case_ : Dict = ["[GO]", "[s]", "0", "1", "2", "3", "4", "5", "6", "7", "8", "9", "a", "b", "c", "d", "e", "f", "g", "h", "i", "j", "k", "l", "m", "n", "o", "p", "q", "r", "s", "t", "u", "v", "w", "x", "y", "z"] # fmt: on snake_case_ : List[str] = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) snake_case_ : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase__ ) + "\n" ) def _A ( self :Optional[Any] , **lowerCAmelCase__ :Optional[Any] ) -> Dict: '''simple docstring''' return MgpstrTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def _A ( self :Dict , lowerCAmelCase__ :Any ) -> str: '''simple docstring''' snake_case_ : Dict = "tester" snake_case_ : Tuple = "tester" return input_text, output_text @unittest.skip("MGP-STR always lower cases letters." ) def _A ( self :Dict ) -> str: '''simple docstring''' pass def _A ( self :Tuple ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[str] = self.get_tokenizers(do_lower_case=lowerCAmelCase__ ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): snake_case_ : Tuple = "[SPECIAL_TOKEN]" tokenizer.add_special_tokens({"cls_token": special_token} ) snake_case_ : str = tokenizer.encode([special_token] , add_special_tokens=lowerCAmelCase__ ) self.assertEqual(len(lowerCAmelCase__ ) , 1 ) snake_case_ : Tuple = tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) self.assertTrue(special_token not in decoded ) def _A ( self :int ) -> List[str]: '''simple docstring''' snake_case_ : Dict = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): snake_case_, snake_case_ : str = self.get_input_output_texts(lowerCAmelCase__ ) snake_case_ : Union[str, Any] = tokenizer.tokenize(lowerCAmelCase__ ) snake_case_ : List[Any] = tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) snake_case_ : Dict = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) snake_case_ : List[str] = tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) self.assertNotEqual(len(lowerCAmelCase__ ) , 0 ) snake_case_ : List[str] = tokenizer.decode(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertEqual(text_a.replace(" " , "" ) , lowerCAmelCase__ ) @unittest.skip("MGP-STR tokenizer only handles one sequence." ) def _A ( self :Union[str, Any] ) -> Any: '''simple docstring''' pass @unittest.skip("inputs cannot be pretokenized in MgpstrTokenizer" ) def _A ( self :int ) -> Dict: '''simple docstring''' pass
656
0
'''simple docstring''' import os def __UpperCAmelCase ( )-> int: """simple docstring""" snake_case_ : Any = os.path.dirname(os.path.realpath(__magic_name__ ) ) snake_case_ : List[Any] = os.path.join(__magic_name__ ,"triangle.txt" ) with open(__magic_name__ ) as f: snake_case_ : List[Any] = f.readlines() snake_case_ : Tuple = [] for line in triangle: snake_case_ : Any = [] for number in line.strip().split(" " ): numbers_from_line.append(int(__magic_name__ ) ) a.append(__magic_name__ ) for i in range(1 ,len(__magic_name__ ) ): for j in range(len(a[i] ) ): snake_case_ : Any = a[i - 1][j] if j != len(a[i - 1] ) else 0 snake_case_ : Optional[Any] = a[i - 1][j - 1] if j > 0 else 0 a[i][j] += max(__magic_name__ ,__magic_name__ ) return max(a[-1] ) if __name__ == "__main__": print(solution())
716
'''simple docstring''' from __future__ import annotations import math import numpy as np from numpy.linalg import norm def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> float: """simple docstring""" return math.sqrt(sum(pow(a - b ,2 ) for a, b in zip(__magic_name__ ,__magic_name__ ) ) ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> list[list[list[float] | float]]: """simple docstring""" if dataset.ndim != value_array.ndim: snake_case_ : int = ( "Wrong input data's dimensions... " F'''dataset : {dataset.ndim}, value_array : {value_array.ndim}''' ) raise ValueError(__magic_name__ ) try: if dataset.shape[1] != value_array.shape[1]: snake_case_ : Dict = ( "Wrong input data's shape... " F'''dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}''' ) raise ValueError(__magic_name__ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError("Wrong shape" ) if dataset.dtype != value_array.dtype: snake_case_ : Dict = ( "Input data have different datatype... " F'''dataset : {dataset.dtype}, value_array : {value_array.dtype}''' ) raise TypeError(__magic_name__ ) snake_case_ : Optional[int] = [] for value in value_array: snake_case_ : List[str] = euclidean(__magic_name__ ,dataset[0] ) snake_case_ : int = dataset[0].tolist() for dataset_value in dataset[1:]: snake_case_ : Optional[Any] = euclidean(__magic_name__ ,__magic_name__ ) if dist > temp_dist: snake_case_ : Tuple = temp_dist snake_case_ : Optional[int] = dataset_value.tolist() answer.append([vector, dist] ) return answer def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> float: """simple docstring""" return np.dot(__magic_name__ ,__magic_name__ ) / (norm(__magic_name__ ) * norm(__magic_name__ )) if __name__ == "__main__": import doctest doctest.testmod()
656
0
'''simple docstring''' import numpy as np # Importing the Keras libraries and packages import tensorflow as tf from tensorflow.keras import layers, models if __name__ == "__main__": # Initialising the CNN # (Sequential- Building the model layer by layer) __lowerCamelCase : int = models.Sequential() # Step 1 - Convolution # Here 64,64 is the length & breadth of dataset images and 3 is for the RGB channel # (3,3) is the kernel size (filter matrix) classifier.add( layers.ConvaD(32, (3, 3), input_shape=(64, 64, 3), activation='''relu''') ) # Step 2 - Pooling classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Adding a second convolutional layer classifier.add(layers.ConvaD(32, (3, 3), activation='''relu''')) classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Step 3 - Flattening classifier.add(layers.Flatten()) # Step 4 - Full connection classifier.add(layers.Dense(units=128, activation='''relu''')) classifier.add(layers.Dense(units=1, activation='''sigmoid''')) # Compiling the CNN classifier.compile( optimizer='''adam''', loss='''binary_crossentropy''', metrics=['''accuracy'''] ) # Part 2 - Fitting the CNN to the images # Load Trained model weights # from keras.models import load_model # regressor=load_model('cnn.h5') __lowerCamelCase : List[str] = tf.keras.preprocessing.image.ImageDataGenerator( rescale=1.0 / 255, shear_range=0.2, zoom_range=0.2, horizontal_flip=True ) __lowerCamelCase : Optional[int] = tf.keras.preprocessing.image.ImageDataGenerator(rescale=1.0 / 255) __lowerCamelCase : List[Any] = train_datagen.flow_from_directory( '''dataset/training_set''', target_size=(64, 64), batch_size=32, class_mode='''binary''' ) __lowerCamelCase : Tuple = test_datagen.flow_from_directory( '''dataset/test_set''', target_size=(64, 64), batch_size=32, class_mode='''binary''' ) classifier.fit_generator( training_set, steps_per_epoch=5, epochs=30, validation_data=test_set ) classifier.save('''cnn.h5''') # Part 3 - Making new predictions __lowerCamelCase : Optional[Any] = tf.keras.preprocessing.image.load_img( '''dataset/single_prediction/image.png''', target_size=(64, 64) ) __lowerCamelCase : Union[str, Any] = tf.keras.preprocessing.image.img_to_array(test_image) __lowerCamelCase : Optional[int] = np.expand_dims(test_image, axis=0) __lowerCamelCase : Optional[int] = classifier.predict(test_image) # training_set.class_indices if result[0][0] == 0: __lowerCamelCase : Union[str, Any] = '''Normal''' if result[0][0] == 1: __lowerCamelCase : Optional[int] = '''Abnormality detected'''
717
'''simple docstring''' import fire from utils import calculate_rouge, save_json def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__=None ,**__magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : int = [x.strip() for x in open(__magic_name__ ).readlines()] snake_case_ : Optional[int] = [x.strip() for x in open(__magic_name__ ).readlines()][: len(__magic_name__ )] snake_case_ : List[Any] = calculate_rouge(__magic_name__ ,__magic_name__ ,**__magic_name__ ) if save_path is not None: save_json(__magic_name__ ,__magic_name__ ,indent=__magic_name__ ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
656
0
'''simple docstring''' def __UpperCAmelCase ( )-> int: """simple docstring""" return 1 def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" return 0 if x < 0 else five_pence(x - 5 ) + two_pence(__magic_name__ ) def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" return 0 if x < 0 else ten_pence(x - 10 ) + five_pence(__magic_name__ ) def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" return 0 if x < 0 else twenty_pence(x - 20 ) + ten_pence(__magic_name__ ) def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" return 0 if x < 0 else fifty_pence(x - 50 ) + twenty_pence(__magic_name__ ) def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" return 0 if x < 0 else one_pound(x - 100 ) + fifty_pence(__magic_name__ ) def __UpperCAmelCase ( __magic_name__ )-> int: """simple docstring""" return 0 if x < 0 else two_pound(x - 200 ) + one_pound(__magic_name__ ) def __UpperCAmelCase ( __magic_name__ = 200 )-> int: """simple docstring""" return two_pound(__magic_name__ ) if __name__ == "__main__": print(solution(int(input().strip())))
718
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) __lowerCamelCase : Optional[Any] = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', f'''encoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', f'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.weight''', f'''encoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.bias''', f'''encoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.weight''', f'''encoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.bias''', f'''encoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.encoder.layers.{i}.norm1.weight''', f'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.norm1.bias''', f'''encoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.weight''', f'''encoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.bias''', f'''encoder.layers.{i}.final_layer_norm.bias''')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', f'''decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', f'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.cross_attn.out_proj.weight''', f'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.cross_attn.out_proj.bias''', f'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.weight''', f'''decoder.layers.{i}.fc1.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.bias''', f'''decoder.layers.{i}.fc1.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.weight''', f'''decoder.layers.{i}.fc2.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.bias''', f'''decoder.layers.{i}.fc2.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm1.weight''', f'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm1.bias''', f'''decoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.weight''', f'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.bias''', f'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.weight''', f'''decoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.bias''', f'''decoder.layers.{i}.final_layer_norm.bias''')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qcontent_proj.weight''', f'''decoder.layers.{i}.sa_qcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kcontent_proj.weight''', f'''decoder.layers.{i}.sa_kcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qpos_proj.weight''', f'''decoder.layers.{i}.sa_qpos_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kpos_proj.weight''', f'''decoder.layers.{i}.sa_kpos_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_v_proj.weight''', f'''decoder.layers.{i}.sa_v_proj.weight''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qcontent_proj.weight''', f'''decoder.layers.{i}.ca_qcontent_proj.weight''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kcontent_proj.weight''', f'''decoder.layers.{i}.ca_kcontent_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kpos_proj.weight''', f'''decoder.layers.{i}.ca_kpos_proj.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_v_proj.weight''', f'''decoder.layers.{i}.ca_v_proj.weight''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight''', f'''decoder.layers.{i}.ca_qpos_sine_proj.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_qcontent_proj.bias''', f'''decoder.layers.{i}.sa_qcontent_proj.bias''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.sa_kcontent_proj.bias''', f'''decoder.layers.{i}.sa_kcontent_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_qpos_proj.bias''', f'''decoder.layers.{i}.sa_qpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_kpos_proj.bias''', f'''decoder.layers.{i}.sa_kpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.sa_v_proj.bias''', f'''decoder.layers.{i}.sa_v_proj.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qcontent_proj.bias''', f'''decoder.layers.{i}.ca_qcontent_proj.bias''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_kcontent_proj.bias''', f'''decoder.layers.{i}.ca_kcontent_proj.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_kpos_proj.bias''', f'''decoder.layers.{i}.ca_kpos_proj.bias''')) rename_keys.append((f'''transformer.decoder.layers.{i}.ca_v_proj.bias''', f'''decoder.layers.{i}.ca_v_proj.bias''')) rename_keys.append( (f'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias''', f'''decoder.layers.{i}.ca_qpos_sine_proj.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ ,__magic_name__ )-> int: """simple docstring""" snake_case_ : Optional[Any] = state_dict.pop(__magic_name__ ) snake_case_ : Any = val def __UpperCAmelCase ( __magic_name__ )-> Optional[Any]: """simple docstring""" snake_case_ : Any = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: snake_case_ : Optional[Any] = key.replace("backbone.0.body" ,"backbone.conv_encoder.model" ) snake_case_ : int = value else: snake_case_ : int = value return new_state_dict def __UpperCAmelCase ( __magic_name__ ,__magic_name__=False )-> Optional[int]: """simple docstring""" snake_case_ : str = "" if is_panoptic: snake_case_ : Dict = "conditional_detr." # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) snake_case_ : Any = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) snake_case_ : Optional[int] = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case_ : Tuple = in_proj_weight[:256, :] snake_case_ : List[Any] = in_proj_bias[:256] snake_case_ : Optional[Any] = in_proj_weight[256:512, :] snake_case_ : Optional[int] = in_proj_bias[256:512] snake_case_ : Optional[int] = in_proj_weight[-256:, :] snake_case_ : str = in_proj_bias[-256:] def __UpperCAmelCase ( )-> Optional[Any]: """simple docstring""" snake_case_ : Optional[int] = "http://images.cocodataset.org/val2017/000000039769.jpg" snake_case_ : Optional[Any] = Image.open(requests.get(__magic_name__ ,stream=__magic_name__ ).raw ) return im @torch.no_grad() def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> List[str]: """simple docstring""" snake_case_ : Optional[Any] = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: snake_case_ : Optional[Any] = "resnet101" if "dc5" in model_name: snake_case_ : List[str] = True snake_case_ : Tuple = "panoptic" in model_name if is_panoptic: snake_case_ : List[Any] = 250 else: snake_case_ : Optional[Any] = 91 snake_case_ : Optional[int] = "huggingface/label-files" snake_case_ : Dict = "coco-detection-id2label.json" snake_case_ : List[Any] = json.load(open(hf_hub_download(__magic_name__ ,__magic_name__ ,repo_type="dataset" ) ,"r" ) ) snake_case_ : Optional[int] = {int(__magic_name__ ): v for k, v in idalabel.items()} snake_case_ : int = idalabel snake_case_ : Dict = {v: k for k, v in idalabel.items()} # load image processor snake_case_ : Optional[int] = "coco_panoptic" if is_panoptic else "coco_detection" snake_case_ : str = ConditionalDetrImageProcessor(format=__magic_name__ ) # prepare image snake_case_ : str = prepare_img() snake_case_ : int = image_processor(images=__magic_name__ ,return_tensors="pt" ) snake_case_ : Union[str, Any] = encoding["pixel_values"] logger.info(F'''Converting model {model_name}...''' ) # load original model from torch hub snake_case_ : Union[str, Any] = torch.hub.load("DeppMeng/ConditionalDETR" ,__magic_name__ ,pretrained=__magic_name__ ).eval() snake_case_ : Any = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: snake_case_ : Any = "conditional_detr." + src rename_key(__magic_name__ ,__magic_name__ ,__magic_name__ ) snake_case_ : Tuple = rename_backbone_keys(__magic_name__ ) # query, key and value matrices need special treatment read_in_q_k_v(__magic_name__ ,is_panoptic=__magic_name__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them snake_case_ : int = "conditional_detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("conditional_detr" ) and not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ) ): snake_case_ : Any = state_dict.pop(__magic_name__ ) snake_case_ : Optional[int] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: snake_case_ : Tuple = state_dict.pop(__magic_name__ ) snake_case_ : Any = val elif key.startswith("bbox_attention" ) or key.startswith("mask_head" ): continue else: snake_case_ : Union[str, Any] = state_dict.pop(__magic_name__ ) snake_case_ : List[Any] = val else: if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): snake_case_ : Any = state_dict.pop(__magic_name__ ) snake_case_ : List[Any] = val # finally, create HuggingFace model and load state dict snake_case_ : Optional[int] = ConditionalDetrForSegmentation(__magic_name__ ) if is_panoptic else ConditionalDetrForObjectDetection(__magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() model.push_to_hub(repo_id=__magic_name__ ,organization="DepuMeng" ,commit_message="Add model" ) # verify our conversion snake_case_ : Dict = conditional_detr(__magic_name__ ) snake_case_ : Union[str, Any] = model(__magic_name__ ) assert torch.allclose(outputs.logits ,original_outputs["pred_logits"] ,atol=1E-4 ) assert torch.allclose(outputs.pred_boxes ,original_outputs["pred_boxes"] ,atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks ,original_outputs["pred_masks"] ,atol=1E-4 ) # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) image_processor.save_pretrained(__magic_name__ ) if __name__ == "__main__": __lowerCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) __lowerCamelCase : int = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
656
0
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> float: """simple docstring""" return math.sqrt(sum(pow(a - b ,2 ) for a, b in zip(__magic_name__ ,__magic_name__ ) ) ) def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> list[list[list[float] | float]]: """simple docstring""" if dataset.ndim != value_array.ndim: snake_case_ : int = ( "Wrong input data's dimensions... " F'''dataset : {dataset.ndim}, value_array : {value_array.ndim}''' ) raise ValueError(__magic_name__ ) try: if dataset.shape[1] != value_array.shape[1]: snake_case_ : Dict = ( "Wrong input data's shape... " F'''dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}''' ) raise ValueError(__magic_name__ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError("Wrong shape" ) if dataset.dtype != value_array.dtype: snake_case_ : Dict = ( "Input data have different datatype... " F'''dataset : {dataset.dtype}, value_array : {value_array.dtype}''' ) raise TypeError(__magic_name__ ) snake_case_ : Optional[int] = [] for value in value_array: snake_case_ : List[str] = euclidean(__magic_name__ ,dataset[0] ) snake_case_ : int = dataset[0].tolist() for dataset_value in dataset[1:]: snake_case_ : Optional[Any] = euclidean(__magic_name__ ,__magic_name__ ) if dist > temp_dist: snake_case_ : Tuple = temp_dist snake_case_ : Optional[int] = dataset_value.tolist() answer.append([vector, dist] ) return answer def __UpperCAmelCase ( __magic_name__ ,__magic_name__ )-> float: """simple docstring""" return np.dot(__magic_name__ ,__magic_name__ ) / (norm(__magic_name__ ) * norm(__magic_name__ )) if __name__ == "__main__": import doctest doctest.testmod()
719
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Any ) -> str: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _A ( self :List[Any] ) -> List[str]: '''simple docstring''' snake_case_ : Any = 1 snake_case_ : Dict = 3 snake_case_ : Union[str, Any] = (32, 32) snake_case_ : Optional[int] = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) return image @property def _A ( self :Optional[int] ) -> Any: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : List[str] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) return model @property def _A ( self :Dict ) -> Any: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) return model @property def _A ( self :Dict ) -> Optional[int]: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : str = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5_006 , ) return RobertaSeriesModelWithTransformation(lowerCAmelCase__ ) @property def _A ( self :Any ) -> str: '''simple docstring''' def extract(*lowerCAmelCase__ :Any , **lowerCAmelCase__ :List[str] ): class A_ : """simple docstring""" def __init__( self :Optional[int] ) -> List[str]: '''simple docstring''' snake_case_ : str = torch.ones([0] ) def _A ( self :int , lowerCAmelCase__ :List[Any] ) -> Tuple: '''simple docstring''' self.pixel_values.to(lowerCAmelCase__ ) return self return Out() return extract def _A ( self :int ) -> Dict: '''simple docstring''' snake_case_ : str = "cpu" # ensure determinism for the device-dependent torch.Generator snake_case_ : str = self.dummy_cond_unet snake_case_ : Optional[int] = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) snake_case_ : Dict = self.dummy_vae snake_case_ : Dict = self.dummy_text_encoder snake_case_ : Optional[int] = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) snake_case_ : str = 77 snake_case_ : Any = self.dummy_image.to(lowerCAmelCase__ ) snake_case_ : Tuple = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk snake_case_ : Optional[Any] = AltDiffusionImgaImgPipeline( unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , vae=lowerCAmelCase__ , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , feature_extractor=self.dummy_extractor , ) snake_case_ : Union[str, Any] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowerCAmelCase__ ) snake_case_ : Optional[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) snake_case_ : Dict = "A painting of a squirrel eating a burger" snake_case_ : List[str] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) snake_case_ : Dict = alt_pipe( [prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , ) snake_case_ : Any = output.images snake_case_ : List[str] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) snake_case_ : Optional[Any] = alt_pipe( [prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , return_dict=lowerCAmelCase__ , )[0] snake_case_ : Tuple = image[0, -3:, -3:, -1] snake_case_ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) snake_case_ : int = np.array([0.4_4_2_7, 0.3_7_3_1, 0.4_2_4_9, 0.4_9_4_1, 0.4_5_4_6, 0.4_1_4_8, 0.4_1_9_3, 0.4_6_6_6, 0.4_4_9_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5E-3 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def _A ( self :int ) -> List[str]: '''simple docstring''' snake_case_ : Union[str, Any] = self.dummy_cond_unet snake_case_ : Union[str, Any] = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) snake_case_ : int = self.dummy_vae snake_case_ : List[Any] = self.dummy_text_encoder snake_case_ : int = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) snake_case_ : int = 77 snake_case_ : Dict = self.dummy_image.to(lowerCAmelCase__ ) # put models in fp16 snake_case_ : Optional[Any] = unet.half() snake_case_ : Tuple = vae.half() snake_case_ : List[str] = bert.half() # make sure here that pndm scheduler skips prk snake_case_ : Optional[int] = AltDiffusionImgaImgPipeline( unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , vae=lowerCAmelCase__ , text_encoder=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , feature_extractor=self.dummy_extractor , ) snake_case_ : List[str] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowerCAmelCase__ ) snake_case_ : Optional[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) snake_case_ : List[Any] = "A painting of a squirrel eating a burger" snake_case_ : str = torch.manual_seed(0 ) snake_case_ : Any = alt_pipe( [prompt] , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , image=lowerCAmelCase__ , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def _A ( self :Optional[int] ) -> Any: '''simple docstring''' snake_case_ : Union[str, Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) # resize to resolution that is divisible by 8 but not 16 or 32 snake_case_ : str = init_image.resize((760, 504) ) snake_case_ : Optional[Any] = "BAAI/AltDiffusion" snake_case_ : int = AltDiffusionImgaImgPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() snake_case_ : Tuple = "A fantasy landscape, trending on artstation" snake_case_ : int = torch.manual_seed(0 ) snake_case_ : List[str] = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , strength=0.7_5 , guidance_scale=7.5 , generator=lowerCAmelCase__ , output_type="np" , ) snake_case_ : str = output.images[0] snake_case_ : List[Any] = image[255:258, 383:386, -1] assert image.shape == (504, 760, 3) snake_case_ : Tuple = np.array([0.9_3_5_8, 0.9_3_9_7, 0.9_5_9_9, 0.9_9_0_1, 1.0_0_0_0, 1.0_0_0_0, 0.9_8_8_2, 1.0_0_0_0, 1.0_0_0_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class A_ (unittest.TestCase ): """simple docstring""" def _A ( self :Optional[Any] ) -> Optional[int]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _A ( self :str ) -> Any: '''simple docstring''' snake_case_ : Optional[Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) snake_case_ : List[Any] = init_image.resize((768, 512) ) snake_case_ : Tuple = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy" ) snake_case_ : Any = "BAAI/AltDiffusion" snake_case_ : List[str] = AltDiffusionImgaImgPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) pipe.enable_attention_slicing() snake_case_ : Tuple = "A fantasy landscape, trending on artstation" snake_case_ : Tuple = torch.manual_seed(0 ) snake_case_ : List[Any] = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , strength=0.7_5 , guidance_scale=7.5 , generator=lowerCAmelCase__ , output_type="np" , ) snake_case_ : Optional[int] = output.images[0] assert image.shape == (512, 768, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1E-2
656
0