code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase=1_3 , __UpperCAmelCase=3_2 , __UpperCAmelCase=3 , __UpperCAmelCase=4 , __UpperCAmelCase=[1_0, 2_0, 3_0, 4_0] , __UpperCAmelCase=[2, 2, 3, 2] , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=3_7 , __UpperCAmelCase="gelu" , __UpperCAmelCase=1_0 , __UpperCAmelCase=0.02 , __UpperCAmelCase=["stage2", "stage3", "stage4"] , __UpperCAmelCase=[2, 3, 4] , __UpperCAmelCase=None , ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = parent lowerCAmelCase__ :Any = batch_size lowerCAmelCase__ :List[Any] = image_size lowerCAmelCase__ :Optional[Any] = num_channels lowerCAmelCase__ :List[Any] = num_stages lowerCAmelCase__ :int = hidden_sizes lowerCAmelCase__ :Any = depths lowerCAmelCase__ :Any = is_training lowerCAmelCase__ :List[Any] = use_labels lowerCAmelCase__ :Optional[int] = intermediate_size lowerCAmelCase__ :Tuple = hidden_act lowerCAmelCase__ :Dict = num_labels lowerCAmelCase__ :Dict = initializer_range lowerCAmelCase__ :List[Any] = out_features lowerCAmelCase__ :Union[str, Any] = out_indices lowerCAmelCase__ :Dict = scope def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ :List[Any] = None if self.use_labels: lowerCAmelCase__ :Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) lowerCAmelCase__ :str = self.get_config() return config, pixel_values, labels def snake_case ( self ): '''simple docstring''' return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=_SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = ConvNextModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCAmelCase__ :Dict = model(_SCREAMING_SNAKE_CASE ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Any = ConvNextForImageClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCAmelCase__ :Tuple = model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :int = ConvNextBackbone(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCAmelCase__ :Optional[int] = model(_SCREAMING_SNAKE_CASE ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None lowerCAmelCase__ :Union[str, Any] = None lowerCAmelCase__ :List[Any] = ConvNextBackbone(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowerCAmelCase__ :Union[str, Any] = model(_SCREAMING_SNAKE_CASE ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[Any] = self.prepare_config_and_inputs() lowerCAmelCase__ :Tuple = config_and_inputs lowerCAmelCase__ :Union[str, Any] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( a , a , unittest.TestCase ): """simple docstring""" __magic_name__ :Tuple = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) __magic_name__ :Optional[int] = ( {"""feature-extraction""": ConvNextModel, """image-classification""": ConvNextForImageClassification} if is_torch_available() else {} ) __magic_name__ :str = True __magic_name__ :Union[str, Any] = False __magic_name__ :Optional[int] = False __magic_name__ :Dict = False __magic_name__ :List[str] = False def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[Any] = ConvNextModelTester(self ) lowerCAmelCase__ :List[str] = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , has_text_modality=_SCREAMING_SNAKE_CASE , hidden_size=3_7 ) def snake_case ( self ): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case ( self ): '''simple docstring''' return @unittest.skip(reason='ConvNext does not use inputs_embeds' ) def snake_case ( self ): '''simple docstring''' pass @unittest.skip(reason='ConvNext does not support input and output embeddings' ) def snake_case ( self ): '''simple docstring''' pass @unittest.skip(reason='ConvNext does not use feedforward chunking' ) def snake_case ( self ): '''simple docstring''' pass def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ :Optional[int] = model_class(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ :List[Any] = [*signature.parameters.keys()] lowerCAmelCase__ :Any = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_SCREAMING_SNAKE_CASE ) def snake_case ( self ): '''simple docstring''' def check_hidden_states_output(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): lowerCAmelCase__ :List[Any] = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): lowerCAmelCase__ :Tuple = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) lowerCAmelCase__ :int = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowerCAmelCase__ :List[Any] = self.model_tester.num_stages self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowerCAmelCase__ :int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ :Union[str, Any] = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ :Dict = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_SCREAMING_SNAKE_CASE ) @slow def snake_case ( self ): '''simple docstring''' for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ :List[Any] = ConvNextModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def __A () ->Union[str, Any]: """simple docstring""" lowerCAmelCase__ :Union[str, Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case ( self ): '''simple docstring''' return AutoImageProcessor.from_pretrained('facebook/convnext-tiny-224' ) if is_vision_available() else None @slow def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = ConvNextForImageClassification.from_pretrained('facebook/convnext-tiny-224' ).to(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Union[str, Any] = self.default_image_processor lowerCAmelCase__ :Union[str, Any] = prepare_img() lowerCAmelCase__ :Any = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='pt' ).to(_SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): lowerCAmelCase__ :int = model(**_SCREAMING_SNAKE_CASE ) # verify the logits lowerCAmelCase__ :Union[str, Any] = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :int = torch.tensor([-0.02_60, -0.47_39, 0.19_11] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1E-4 ) ) @require_torch class _lowerCAmelCase ( unittest.TestCase , a ): """simple docstring""" __magic_name__ :Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () __magic_name__ :Optional[int] = ConvNextConfig __magic_name__ :Optional[int] = False def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[Any] = ConvNextModelTester(self )
293
import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class _lowerCamelCase ( UpperCamelCase , unittest.TestCase ): """simple docstring""" snake_case = DDIMPipeline snake_case = UNCONDITIONAL_IMAGE_GENERATION_PARAMS snake_case = PipelineTesterMixin.required_optional_params - { "num_images_per_prompt", "latents", "callback", "callback_steps", } snake_case = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS snake_case = False def _snake_case ( self )->List[str]: '''simple docstring''' torch.manual_seed(0 ) A_ : List[str] = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) A_ : Optional[Any] = DDIMScheduler() A_ : str = {'''unet''': unet, '''scheduler''': scheduler} return components def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=0 )->Optional[Any]: '''simple docstring''' if str(_SCREAMING_SNAKE_CASE ).startswith('''mps''' ): A_ : Any = torch.manual_seed(_SCREAMING_SNAKE_CASE ) else: A_ : Optional[int] = torch.Generator(device=_SCREAMING_SNAKE_CASE ).manual_seed(_SCREAMING_SNAKE_CASE ) A_ : Any = { '''batch_size''': 1, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def _snake_case ( self )->List[Any]: '''simple docstring''' A_ : Optional[int] = '''cpu''' A_ : Dict = self.get_dummy_components() A_ : str = self.pipeline_class(**_SCREAMING_SNAKE_CASE ) pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) A_ : str = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE ) A_ : Any = pipe(**_SCREAMING_SNAKE_CASE ).images A_ : int = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 32, 32, 3) ) A_ : List[Any] = np.array( [1.000e00, 5.717e-01, 4.717e-01, 1.000e00, 0.000e00, 1.000e00, 3.000e-04, 0.000e00, 9.000e-04] ) A_ : str = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1e-3 ) def _snake_case ( self )->Union[str, Any]: '''simple docstring''' super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def _snake_case ( self )->Optional[int]: '''simple docstring''' super().test_save_load_local(expected_max_difference=3e-3 ) def _snake_case ( self )->Optional[int]: '''simple docstring''' super().test_save_load_optional_components(expected_max_difference=3e-3 ) def _snake_case ( self )->Any: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class _lowerCamelCase ( unittest.TestCase ): """simple docstring""" def _snake_case ( self )->Union[str, Any]: '''simple docstring''' A_ : int = '''google/ddpm-cifar10-32''' A_ : Tuple = UNetaDModel.from_pretrained(_SCREAMING_SNAKE_CASE ) A_ : str = DDIMScheduler() A_ : str = DDIMPipeline(unet=_SCREAMING_SNAKE_CASE , scheduler=_SCREAMING_SNAKE_CASE ) ddim.to(_SCREAMING_SNAKE_CASE ) ddim.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) A_ : Optional[int] = torch.manual_seed(0 ) A_ : Any = ddim(generator=_SCREAMING_SNAKE_CASE , eta=0.0 , output_type='''numpy''' ).images A_ : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) A_ : Any = np.array([0.1_7_2_3, 0.1_6_1_7, 0.1_6_0_0, 0.1_6_2_6, 0.1_4_9_7, 0.1_5_1_3, 0.1_5_0_5, 0.1_4_4_2, 0.1_4_5_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _snake_case ( self )->List[str]: '''simple docstring''' A_ : Tuple = '''google/ddpm-ema-bedroom-256''' A_ : int = UNetaDModel.from_pretrained(_SCREAMING_SNAKE_CASE ) A_ : Any = DDIMScheduler.from_pretrained(_SCREAMING_SNAKE_CASE ) A_ : Optional[Any] = DDIMPipeline(unet=_SCREAMING_SNAKE_CASE , scheduler=_SCREAMING_SNAKE_CASE ) ddpm.to(_SCREAMING_SNAKE_CASE ) ddpm.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) A_ : Dict = torch.manual_seed(0 ) A_ : List[str] = ddpm(generator=_SCREAMING_SNAKE_CASE , output_type='''numpy''' ).images A_ : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) A_ : Tuple = np.array([0.0_0_6_0, 0.0_2_0_1, 0.0_3_4_4, 0.0_0_2_4, 0.0_0_1_8, 0.0_0_0_2, 0.0_0_2_2, 0.0_0_0_0, 0.0_0_6_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
186
0
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowercase__ ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): A__ : List[Any] =StableDiffusionXLImgaImgPipeline A__ : Optional[Any] =TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} A__ : Optional[int] =PipelineTesterMixin.required_optional_params - {"""latents"""} A__ : List[Any] =TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS A__ : List[str] =IMAGE_TO_IMAGE_IMAGE_PARAMS A__ : str =IMAGE_TO_IMAGE_IMAGE_PARAMS def A_ ( self : int ): torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , attention_head_dim=(2, 4) , use_linear_projection=UpperCAmelCase_ , addition_embed_type='text_time' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) SCREAMING_SNAKE_CASE__ = EulerDiscreteScheduler( beta_start=0.00_085 , beta_end=0.012 , steps_offset=1 , beta_schedule='scaled_linear' , timestep_spacing='leading' , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=32 , ) SCREAMING_SNAKE_CASE__ = CLIPTextModel(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' , local_files_only=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = CLIPTextModelWithProjection(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' , local_files_only=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'text_encoder_2': text_encoder_a, 'tokenizer_2': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def A_ ( self : int , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : List[str]=0 ): SCREAMING_SNAKE_CASE__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCAmelCase_ ) ).to(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = image / 2 + 0.5 if str(UpperCAmelCase_ ).startswith('mps' ): SCREAMING_SNAKE_CASE__ = torch.manual_seed(UpperCAmelCase_ ) else: SCREAMING_SNAKE_CASE__ = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 5.0, 'output_type': 'numpy', 'strength': 0.75, } return inputs def A_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE__ = 'cpu' # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = StableDiffusionXLImgaImgPipeline(**UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = sd_pipe.to(UpperCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = sd_pipe(**UpperCAmelCase_ ).images SCREAMING_SNAKE_CASE__ = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) SCREAMING_SNAKE_CASE__ = np.array([0.4_656, 0.4_840, 0.4_439, 0.6_698, 0.5_574, 0.4_524, 0.5_799, 0.5_943, 0.5_165] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def A_ ( self : str ): super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def A_ ( self : List[Any] ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def A_ ( self : int ): pass def A_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = StableDiffusionXLImgaImgPipeline(**UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = sd_pipe.to(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = sd_pipe.to(UpperCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) # forward without prompt embeds SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = 3 * ['this is a negative prompt'] SCREAMING_SNAKE_CASE__ = negative_prompt SCREAMING_SNAKE_CASE__ = 3 * [inputs['prompt']] SCREAMING_SNAKE_CASE__ = sd_pipe(**UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = output.images[0, -3:, -3:, -1] # forward with prompt embeds SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = 3 * ['this is a negative prompt'] SCREAMING_SNAKE_CASE__ = 3 * [inputs.pop('prompt' )] ( ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ) = sd_pipe.encode_prompt(UpperCAmelCase_ , negative_prompt=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = sd_pipe( **UpperCAmelCase_ , prompt_embeds=UpperCAmelCase_ , negative_prompt_embeds=UpperCAmelCase_ , pooled_prompt_embeds=UpperCAmelCase_ , negative_pooled_prompt_embeds=UpperCAmelCase_ , ) SCREAMING_SNAKE_CASE__ = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @slow @require_torch_gpu class lowercase__ ( unittest.TestCase ): def A_ ( self : Any ): super().tearDown() gc.collect() torch.cuda.empty_cache() def A_ ( self : Union[str, Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Union[str, Any]="cpu" , UpperCAmelCase_ : Any=torch.floataa , UpperCAmelCase_ : Dict=0 ): SCREAMING_SNAKE_CASE__ = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = np.random.RandomState(UpperCAmelCase_ ).standard_normal((1, 4, 64, 64) ) SCREAMING_SNAKE_CASE__ = torch.from_numpy(UpperCAmelCase_ ).to(device=UpperCAmelCase_ , dtype=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = { 'prompt': 'a photograph of an astronaut riding a horse', 'latents': latents, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def A_ ( self : List[str] ): SCREAMING_SNAKE_CASE__ = DiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-base' ) pipe.to(UpperCAmelCase_ ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = self.get_inputs(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = pipe(**UpperCAmelCase_ ).images SCREAMING_SNAKE_CASE__ = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) SCREAMING_SNAKE_CASE__ = np.array([0.49_493, 0.47_896, 0.40_798, 0.54_214, 0.53_212, 0.48_202, 0.47_656, 0.46_329, 0.48_506] ) assert np.abs(image_slice - expected_slice ).max() < 7e-3
169
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor __snake_case = logging.get_logger(__name__) class lowercase__ ( _UpperCAmelCase ): def __init__( self : Dict , *UpperCAmelCase_ : int , **UpperCAmelCase_ : List[Any] ): warnings.warn( 'The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use LayoutLMv2ImageProcessor instead.' , UpperCAmelCase_ , ) super().__init__(*UpperCAmelCase_ , **UpperCAmelCase_ )
169
1
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets A : str = '\\n@inproceedings{pillutla-etal:mauve:neurips2021,\n title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers},\n author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid},\n booktitle = {NeurIPS},\n year = {2021}\n}\n\n' A : Tuple = '\\nMAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure.\n\nMAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences.\n\nFor details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021).\n\nThis metrics is a wrapper around the official implementation of MAUVE:\nhttps://github.com/krishnap25/mauve\n' A : List[Any] = '\nCalculates MAUVE scores between two lists of generated text and reference text.\nArgs:\n predictions: list of generated text to score. Each predictions\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\nOptional Args:\n num_buckets: the size of the histogram to quantize P and Q. Options: \'auto\' (default) or an integer\n pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1\n kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9\n kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5\n kmeans_max_iter: maximum number of k-means iterations. Default 500\n featurize_model_name: name of the model from which features are obtained. Default \'gpt2-large\' Use one of [\'gpt2\', \'gpt2-medium\', \'gpt2-large\', \'gpt2-xl\'].\n device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU\n max_text_length: maximum number of tokens to consider. Default 1024\n divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25\n mauve_scaling_factor: "c" from the paper. Default 5.\n verbose: If True (default), print running time updates\n seed: random seed to initialize k-means cluster assignments.\nReturns:\n mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer,\n frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer,\n divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve,\n p_hist: a discrete distribution, which is a quantized version of the text distribution p_text,\n q_hist: same as above, but with q_text.\nExamples:\n\n >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest\n >>> import datasets\n >>> mauve = datasets.load_metric(\'mauve\')\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP\n >>> print(out.mauve) # doctest: +SKIP\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A ( datasets.Metric ): '''simple docstring''' def lowerCamelCase__ (self : Union[str, Any] ) -> Any: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""https://github.com/krishnap25/mauve""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/krishnap25/mauve"""] , reference_urls=[ """https://arxiv.org/abs/2102.01454""", """https://github.com/krishnap25/mauve""", ] , ) def lowerCamelCase__ (self : Tuple , _UpperCAmelCase : List[str] , _UpperCAmelCase : Dict , _UpperCAmelCase : Any=None , _UpperCAmelCase : Optional[Any]=None , _UpperCAmelCase : Optional[int]=None , _UpperCAmelCase : List[str]=None , _UpperCAmelCase : Any="auto" , _UpperCAmelCase : Optional[int]=-1 , _UpperCAmelCase : Tuple=0.9 , _UpperCAmelCase : List[str]=5 , _UpperCAmelCase : Any=500 , _UpperCAmelCase : int="gpt2-large" , _UpperCAmelCase : Optional[Any]=-1 , _UpperCAmelCase : Any=1024 , _UpperCAmelCase : int=25 , _UpperCAmelCase : Dict=5 , _UpperCAmelCase : Optional[Any]=True , _UpperCAmelCase : Dict=25 , ) -> Union[str, Any]: """simple docstring""" lowercase__ = compute_mauve( p_text=_UpperCAmelCase , q_text=_UpperCAmelCase , p_features=_UpperCAmelCase , q_features=_UpperCAmelCase , p_tokens=_UpperCAmelCase , q_tokens=_UpperCAmelCase , num_buckets=_UpperCAmelCase , pca_max_data=_UpperCAmelCase , kmeans_explained_var=_UpperCAmelCase , kmeans_num_redo=_UpperCAmelCase , kmeans_max_iter=_UpperCAmelCase , featurize_model_name=_UpperCAmelCase , device_id=_UpperCAmelCase , max_text_length=_UpperCAmelCase , divergence_curve_discretization_size=_UpperCAmelCase , mauve_scaling_factor=_UpperCAmelCase , verbose=_UpperCAmelCase , seed=_UpperCAmelCase , ) return out
305
class A : '''simple docstring''' def __init__(self : List[str] ) -> Tuple: """simple docstring""" lowercase__ = 0 lowercase__ = 0 lowercase__ = {} def lowerCamelCase__ (self : Dict , _UpperCAmelCase : Tuple ) -> Optional[int]: """simple docstring""" if vertex not in self.adjacency: lowercase__ = {} self.num_vertices += 1 def lowerCamelCase__ (self : List[Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : int , _UpperCAmelCase : List[str] ) -> Tuple: """simple docstring""" self.add_vertex(_UpperCAmelCase ) self.add_vertex(_UpperCAmelCase ) if head == tail: return lowercase__ = weight lowercase__ = weight def lowerCamelCase__ (self : List[str] ) -> Optional[int]: """simple docstring""" lowercase__ = self.get_edges() for edge in edges: lowercase__ , lowercase__ , lowercase__ = edge edges.remove((tail, head, weight) ) for i in range(len(_UpperCAmelCase ) ): lowercase__ = list(edges[i] ) edges.sort(key=lambda _UpperCAmelCase : e[2] ) for i in range(len(_UpperCAmelCase ) - 1 ): if edges[i][2] >= edges[i + 1][2]: lowercase__ = edges[i][2] + 1 for edge in edges: lowercase__ , lowercase__ , lowercase__ = edge lowercase__ = weight lowercase__ = weight def __str__(self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" lowercase__ = """""" for tail in self.adjacency: for head in self.adjacency[tail]: lowercase__ = self.adjacency[head][tail] string += f'''{head} -> {tail} == {weight}\n''' return string.rstrip("""\n""" ) def lowerCamelCase__ (self : Any ) -> str: """simple docstring""" lowercase__ = [] for tail in self.adjacency: for head in self.adjacency[tail]: output.append((tail, head, self.adjacency[head][tail]) ) return output def lowerCamelCase__ (self : Optional[int] ) -> Optional[int]: """simple docstring""" return self.adjacency.keys() @staticmethod def lowerCamelCase__ (_UpperCAmelCase : List[str]=None , _UpperCAmelCase : Any=None ) -> Union[str, Any]: """simple docstring""" lowercase__ = Graph() if vertices is None: lowercase__ = [] if edges is None: lowercase__ = [] for vertex in vertices: g.add_vertex(_UpperCAmelCase ) for edge in edges: g.add_edge(*_UpperCAmelCase ) return g class A : '''simple docstring''' def __init__(self : Optional[Any] ) -> str: """simple docstring""" lowercase__ = {} lowercase__ = {} def __len__(self : Optional[Any] ) -> Dict: """simple docstring""" return len(self.parent ) def lowerCamelCase__ (self : str , _UpperCAmelCase : Dict ) -> Any: """simple docstring""" if item in self.parent: return self.find(_UpperCAmelCase ) lowercase__ = item lowercase__ = 0 return item def lowerCamelCase__ (self : List[str] , _UpperCAmelCase : Dict ) -> Any: """simple docstring""" if item not in self.parent: return self.make_set(_UpperCAmelCase ) if item != self.parent[item]: lowercase__ = self.find(self.parent[item] ) return self.parent[item] def lowerCamelCase__ (self : List[Any] , _UpperCAmelCase : Any , _UpperCAmelCase : List[Any] ) -> Optional[int]: """simple docstring""" lowercase__ = self.find(_UpperCAmelCase ) lowercase__ = self.find(_UpperCAmelCase ) if roota == roota: return roota if self.rank[roota] > self.rank[roota]: lowercase__ = roota return roota if self.rank[roota] < self.rank[roota]: lowercase__ = roota return roota if self.rank[roota] == self.rank[roota]: self.rank[roota] += 1 lowercase__ = roota return roota return None @staticmethod def lowerCamelCase__ (_UpperCAmelCase : str ) -> Optional[int]: """simple docstring""" lowercase__ = graph.num_vertices lowercase__ = Graph.UnionFind() lowercase__ = [] while num_components > 1: lowercase__ = {} for vertex in graph.get_vertices(): lowercase__ = -1 lowercase__ = graph.get_edges() for edge in edges: lowercase__ , lowercase__ , lowercase__ = edge edges.remove((tail, head, weight) ) for edge in edges: lowercase__ , lowercase__ , lowercase__ = edge lowercase__ = union_find.find(_UpperCAmelCase ) lowercase__ = union_find.find(_UpperCAmelCase ) if seta != seta: if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: lowercase__ = [head, tail, weight] if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: lowercase__ = [head, tail, weight] for vertex in cheap_edge: if cheap_edge[vertex] != -1: lowercase__ , lowercase__ , lowercase__ = cheap_edge[vertex] if union_find.find(_UpperCAmelCase ) != union_find.find(_UpperCAmelCase ): union_find.union(_UpperCAmelCase , _UpperCAmelCase ) mst_edges.append(cheap_edge[vertex] ) lowercase__ = num_components - 1 lowercase__ = Graph.build(edges=_UpperCAmelCase ) return mst
305
1
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def __a ( _SCREAMING_SNAKE_CASE ) ->int: a__: Optional[int] = 384 if "tiny" in model_name: a__: Any = [3, 3, 9, 3] a__: int = [96, 192, 384, 768] if "small" in model_name: a__: List[Any] = [3, 3, 27, 3] a__: List[Any] = [96, 192, 384, 768] if "base" in model_name: a__: Optional[int] = [3, 3, 27, 3] a__: List[str] = [128, 256, 512, 1024] a__: str = 512 if "large" in model_name: a__: Dict = [3, 3, 27, 3] a__: int = [192, 384, 768, 1536] a__: Optional[int] = 768 if "xlarge" in model_name: a__: Dict = [3, 3, 27, 3] a__: Dict = [256, 512, 1024, 2048] a__: Union[str, Any] = 1024 # set label information a__: Dict = 150 a__: Any = 'huggingface/label-files' a__: Dict = 'ade20k-id2label.json' a__: Tuple = json.load(open(hf_hub_download(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , repo_type='dataset' ) , 'r' ) ) a__: Tuple = {int(_SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} a__: Optional[int] = {v: k for k, v in idalabel.items()} a__: Tuple = ConvNextConfig( depths=_SCREAMING_SNAKE_CASE , hidden_sizes=_SCREAMING_SNAKE_CASE , out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) a__: Union[str, Any] = UperNetConfig( backbone_config=_SCREAMING_SNAKE_CASE , auxiliary_in_channels=_SCREAMING_SNAKE_CASE , num_labels=_SCREAMING_SNAKE_CASE , idalabel=_SCREAMING_SNAKE_CASE , labelaid=_SCREAMING_SNAKE_CASE , ) return config def __a ( _SCREAMING_SNAKE_CASE ) ->int: a__: Tuple = [] # fmt: off # stem rename_keys.append(('backbone.downsample_layers.0.0.weight', 'backbone.embeddings.patch_embeddings.weight') ) rename_keys.append(('backbone.downsample_layers.0.0.bias', 'backbone.embeddings.patch_embeddings.bias') ) rename_keys.append(('backbone.downsample_layers.0.1.weight', 'backbone.embeddings.layernorm.weight') ) rename_keys.append(('backbone.downsample_layers.0.1.bias', 'backbone.embeddings.layernorm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F'backbone.stages.{i}.{j}.gamma', F'backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter') ) rename_keys.append((F'backbone.stages.{i}.{j}.depthwise_conv.weight', F'backbone.encoder.stages.{i}.layers.{j}.dwconv.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.depthwise_conv.bias', F'backbone.encoder.stages.{i}.layers.{j}.dwconv.bias') ) rename_keys.append((F'backbone.stages.{i}.{j}.norm.weight', F'backbone.encoder.stages.{i}.layers.{j}.layernorm.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.norm.bias', F'backbone.encoder.stages.{i}.layers.{j}.layernorm.bias') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv1.weight', F'backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv1.bias', F'backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv2.weight', F'backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv2.bias', F'backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias') ) if i > 0: rename_keys.append((F'backbone.downsample_layers.{i}.0.weight', F'backbone.encoder.stages.{i}.downsampling_layer.0.weight') ) rename_keys.append((F'backbone.downsample_layers.{i}.0.bias', F'backbone.encoder.stages.{i}.downsampling_layer.0.bias') ) rename_keys.append((F'backbone.downsample_layers.{i}.1.weight', F'backbone.encoder.stages.{i}.downsampling_layer.1.weight') ) rename_keys.append((F'backbone.downsample_layers.{i}.1.bias', F'backbone.encoder.stages.{i}.downsampling_layer.1.bias') ) rename_keys.append((F'backbone.norm{i}.weight', F'backbone.hidden_states_norms.stage{i+1}.weight') ) rename_keys.append((F'backbone.norm{i}.bias', F'backbone.hidden_states_norms.stage{i+1}.bias') ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Dict: a__: Union[str, Any] = dct.pop(_SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = val def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: a__: Optional[Any] = { 'upernet-convnext-tiny': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth', 'upernet-convnext-small': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth', 'upernet-convnext-base': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth', 'upernet-convnext-large': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth', 'upernet-convnext-xlarge': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth', } a__: Union[str, Any] = model_name_to_url[model_name] a__: Tuple = torch.hub.load_state_dict_from_url(_SCREAMING_SNAKE_CASE , map_location='cpu' )['state_dict'] a__: List[Any] = get_upernet_config(_SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = UperNetForSemanticSegmentation(_SCREAMING_SNAKE_CASE ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): a__: Optional[Any] = state_dict.pop(_SCREAMING_SNAKE_CASE ) if "bn" in key: a__: List[Any] = key.replace('bn' , 'batch_norm' ) a__: List[Any] = val # rename keys a__: List[Any] = create_rename_keys(_SCREAMING_SNAKE_CASE ) for src, dest in rename_keys: rename_key(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) model.load_state_dict(_SCREAMING_SNAKE_CASE ) # verify on image a__: Optional[int] = 'https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg' a__: str = Image.open(requests.get(_SCREAMING_SNAKE_CASE , stream=_SCREAMING_SNAKE_CASE ).raw ).convert('RGB' ) a__: Tuple = SegformerImageProcessor() a__: List[str] = processor(_SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values with torch.no_grad(): a__: Dict = model(_SCREAMING_SNAKE_CASE ) if model_name == "upernet-convnext-tiny": a__: Optional[int] = torch.tensor( [[-8.8_110, -8.8_110, -8.6_521], [-8.8_110, -8.8_110, -8.6_521], [-8.7_746, -8.7_746, -8.6_130]] ) elif model_name == "upernet-convnext-small": a__: str = torch.tensor( [[-8.8_236, -8.8_236, -8.6_771], [-8.8_236, -8.8_236, -8.6_771], [-8.7_638, -8.7_638, -8.6_240]] ) elif model_name == "upernet-convnext-base": a__: str = torch.tensor( [[-8.8_558, -8.8_558, -8.6_905], [-8.8_558, -8.8_558, -8.6_905], [-8.7_669, -8.7_669, -8.6_021]] ) elif model_name == "upernet-convnext-large": a__: Any = torch.tensor( [[-8.6_660, -8.6_660, -8.6_210], [-8.6_660, -8.6_660, -8.6_210], [-8.6_310, -8.6_310, -8.5_964]] ) elif model_name == "upernet-convnext-xlarge": a__: Union[str, Any] = torch.tensor( [[-8.4_980, -8.4_980, -8.3_977], [-8.4_980, -8.4_980, -8.3_977], [-8.4_379, -8.4_379, -8.3_412]] ) print('Logits:' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) print(F'Saving processor to {pytorch_dump_folder_path}' ) processor.save_pretrained(_SCREAMING_SNAKE_CASE ) if push_to_hub: print(F'Pushing model and processor for {model_name} to hub' ) model.push_to_hub(F'openmmlab/{model_name}' ) processor.push_to_hub(F'openmmlab/{model_name}' ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='upernet-convnext-tiny', type=str, choices=[f"upernet-convnext-{size}" for size in ['tiny', 'small', 'base', 'large', 'xlarge']], help='Name of the ConvNext UperNet model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) lowercase__ = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
203
"""simple docstring""" import unittest from knapsack import knapsack as k class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: List[Any] = 0 a__: Dict = [0] a__: int = [0] a__: Optional[Any] = len(lowercase) self.assertEqual(k.knapsack(lowercase , lowercase , lowercase , lowercase) , 0) a__: str = [60] a__: Dict = [10] a__: List[str] = len(lowercase) self.assertEqual(k.knapsack(lowercase , lowercase , lowercase , lowercase) , 0) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: int = 3 a__: str = [1, 2, 3] a__: Dict = [3, 2, 1] a__: Optional[int] = len(lowercase) self.assertEqual(k.knapsack(lowercase , lowercase , lowercase , lowercase) , 5) def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Any = 50 a__: Optional[int] = [60, 1_00, 1_20] a__: str = [10, 20, 30] a__: int = len(lowercase) self.assertEqual(k.knapsack(lowercase , lowercase , lowercase , lowercase) , 2_20) if __name__ == "__main__": unittest.main()
203
1
import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class UpperCAmelCase_ ( _A ): '''simple docstring''' def __init__( self : Optional[int] , UpperCamelCase__ : Dict , UpperCamelCase__ : List[Any]=None , UpperCamelCase__ : Union[str, Any]=True , UpperCamelCase__ : str=None , **UpperCamelCase__ : List[Any] ) -> int: """simple docstring""" __magic_name__ = parent __magic_name__ = config_class __magic_name__ = has_text_modality __magic_name__ = kwargs __magic_name__ = common_properties def _lowercase ( self : List[Any] ) -> str: """simple docstring""" __magic_name__ = self.config_class(**self.inputs_dict ) __magic_name__ = ( ["""hidden_size""", """num_attention_heads""", """num_hidden_layers"""] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(["""vocab_size"""] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(UpperCamelCase__ , UpperCamelCase__ ) , msg=F'''`{prop}` does not exist''' ) # Test that config has the common properties as setter for idx, name in enumerate(UpperCamelCase__ ): try: setattr(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) self.parent.assertEqual( getattr(UpperCamelCase__ , UpperCamelCase__ ) , UpperCamelCase__ , msg=F'''`{name} value {idx} expected, but was {getattr(UpperCamelCase__ , UpperCamelCase__ )}''' ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(UpperCamelCase__ ): try: __magic_name__ = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(UpperCamelCase__ , UpperCamelCase__ ) , UpperCamelCase__ , msg=F'''`{name} value {idx} expected, but was {getattr(UpperCamelCase__ , UpperCamelCase__ )}''' ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def _lowercase ( self : List[str] ) -> Any: """simple docstring""" __magic_name__ = self.config_class(**self.inputs_dict ) __magic_name__ = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] , UpperCamelCase__ ) def _lowercase ( self : Tuple ) -> Optional[Any]: """simple docstring""" __magic_name__ = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __magic_name__ = os.path.join(UpperCamelCase__ , """config.json""" ) config_first.to_json_file(UpperCamelCase__ ) __magic_name__ = self.config_class.from_json_file(UpperCamelCase__ ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def _lowercase ( self : int ) -> List[str]: """simple docstring""" __magic_name__ = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(UpperCamelCase__ ) __magic_name__ = self.config_class.from_pretrained(UpperCamelCase__ ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def _lowercase ( self : str ) -> Tuple: """simple docstring""" __magic_name__ = self.config_class(**self.inputs_dict ) __magic_name__ = """test""" with tempfile.TemporaryDirectory() as tmpdirname: __magic_name__ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) config_first.save_pretrained(UpperCamelCase__ ) __magic_name__ = self.config_class.from_pretrained(UpperCamelCase__ , subfolder=UpperCamelCase__ ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def _lowercase ( self : List[Any] ) -> List[Any]: """simple docstring""" __magic_name__ = self.config_class(**self.inputs_dict , num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) , 5 ) self.parent.assertEqual(len(config.labelaid ) , 5 ) __magic_name__ = 3 self.parent.assertEqual(len(config.idalabel ) , 3 ) self.parent.assertEqual(len(config.labelaid ) , 3 ) def _lowercase ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" if self.config_class.is_composition: return __magic_name__ = self.config_class() self.parent.assertIsNotNone(UpperCamelCase__ ) def _lowercase ( self : Optional[int] ) -> List[Any]: """simple docstring""" __magic_name__ = copy.deepcopy(UpperCamelCase__ ) __magic_name__ = self.config_class(**UpperCamelCase__ ) __magic_name__ = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(("""torch_dtype""", config.torch_dtype, torch.floataa) ) elif getattr(UpperCamelCase__ , UpperCamelCase__ ) != value: wrong_values.append((key, getattr(UpperCamelCase__ , UpperCamelCase__ ), value) ) if len(UpperCamelCase__ ) > 0: __magic_name__ = """\n""".join([F'''- {v[0]}: got {v[1]} instead of {v[2]}''' for v in wrong_values] ) raise ValueError(F'''The following keys were not properly set in the config:\n{errors}''' ) def _lowercase ( self : Optional[int] ) -> List[Any]: """simple docstring""" self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
88
"""simple docstring""" from math import ceil def SCREAMING_SNAKE_CASE ( _lowerCamelCase : Union[str, Any] ,_lowerCamelCase : Union[str, Any] ) -> int: _lowerCAmelCase : Dict = list(range(0 ,_lowerCamelCase ) ) _lowerCAmelCase : Tuple = [item for sublist in list(device_map.values() ) for item in sublist] # Duplicate check _lowerCAmelCase : Union[str, Any] = [] for i in device_map_blocks: if device_map_blocks.count(_lowerCamelCase ) > 1 and i not in duplicate_blocks: duplicate_blocks.append(_lowerCamelCase ) # Missing blocks _lowerCAmelCase : int = [i for i in blocks if i not in device_map_blocks] _lowerCAmelCase : List[Any] = [i for i in device_map_blocks if i not in blocks] if len(_lowerCamelCase ) != 0: raise ValueError( """Duplicate attention blocks specified in device_map. Attention blocks must be specified to one device.""" """ These attention blocks were specified more than once: """ + str(_lowerCamelCase ) ) if len(_lowerCamelCase ) != 0: raise ValueError( """There are attention blocks for this model that are not specified in the device_map. Add these attention """ """blocks to a device on the device_map: """ + str(_lowerCamelCase ) ) if len(_lowerCamelCase ) != 0: raise ValueError( """The device_map contains more attention blocks than this model has. Remove these from the device_map:""" + str(_lowerCamelCase ) ) def SCREAMING_SNAKE_CASE ( _lowerCamelCase : int ,_lowerCamelCase : Tuple ) -> str: _lowerCAmelCase : Optional[Any] = list(range(_lowerCamelCase ) ) _lowerCAmelCase : Optional[Any] = int(ceil(n_layers / len(_lowerCamelCase ) ) ) _lowerCAmelCase : Optional[int] = [layers[i : i + n_blocks] for i in range(0 ,_lowerCamelCase ,_lowerCamelCase )] return dict(zip(_lowerCamelCase ,_lowerCamelCase ) )
44
0
'''simple docstring''' import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class SCREAMING_SNAKE_CASE__ ( snake_case_): @require_torch def UpperCAmelCase_ ( self )-> List[str]: '''simple docstring''' UpperCamelCase = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' UpperCamelCase = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' UpperCamelCase = '\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache UpperCamelCase = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(A_ ) BertModel.from_pretrained(A_ ) BertTokenizer.from_pretrained(A_ ) pipeline(task='fill-mask' , model=A_ ) # baseline - just load from_pretrained with normal network UpperCamelCase = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed UpperCamelCase = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files UpperCamelCase = '1' UpperCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def UpperCAmelCase_ ( self )-> Tuple: '''simple docstring''' UpperCamelCase = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' UpperCamelCase = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' UpperCamelCase = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache UpperCamelCase = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(A_ ) BertModel.from_pretrained(A_ ) BertTokenizer.from_pretrained(A_ ) pipeline(task='fill-mask' , model=A_ ) # baseline - just load from_pretrained with normal network UpperCamelCase = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed UpperCamelCase = self.get_env() UpperCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def UpperCAmelCase_ ( self )-> str: '''simple docstring''' UpperCamelCase = '\nfrom transformers import BertConfig, BertModel, BertTokenizer\n ' UpperCamelCase = '\nmname = "hf-internal-testing/tiny-random-bert-sharded"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint("success")\n ' UpperCamelCase = '\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled")\nsocket.socket = offline_socket\n ' # baseline - just load from_pretrained with normal network UpperCamelCase = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed UpperCamelCase = self.get_env() UpperCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # next emulate no network UpperCamelCase = [sys.executable, '-c', '\n'.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files UpperCamelCase = '1' UpperCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def UpperCAmelCase_ ( self )-> int: '''simple docstring''' UpperCamelCase = '\nfrom transformers import pipeline\n ' UpperCamelCase = '\nmname = "hf-internal-testing/tiny-random-bert"\npipe = pipeline(model=mname)\n ' UpperCamelCase = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled")\nsocket.socket = offline_socket\n ' UpperCamelCase = self.get_env() UpperCamelCase = '1' UpperCamelCase = [sys.executable, '-c', '\n'.join([load, mock, run] )] UpperCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( 'You cannot infer task automatically within `pipeline` when using offline mode' , result.stderr.decode().replace('\n' , '' ) , ) @require_torch def UpperCAmelCase_ ( self )-> Dict: '''simple docstring''' UpperCamelCase = '\nfrom transformers import AutoModel\n ' UpperCamelCase = '\nmname = "hf-internal-testing/test_dynamic_model"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint("success")\n ' # baseline - just load from_pretrained with normal network UpperCamelCase = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed UpperCamelCase = self.get_env() UpperCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files UpperCamelCase = '1' UpperCamelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() )
251
'''simple docstring''' from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig lowerCAmelCase : Any = logging.get_logger(__name__) # General docstring lowerCAmelCase : Tuple = 'MobileNetV1Config' # Base docstring lowerCAmelCase : Dict = 'google/mobilenet_v1_1.0_224' lowerCAmelCase : Any = [1, 10_24, 7, 7] # Image classification docstring lowerCAmelCase : Optional[Any] = 'google/mobilenet_v1_1.0_224' lowerCAmelCase : List[str] = 'tabby, tabby cat' lowerCAmelCase : str = [ 'google/mobilenet_v1_1.0_224', 'google/mobilenet_v1_0.75_192', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def A_( A : Union[str, Any] , A : Optional[Any] , A : Optional[Any]=None): UpperCamelCase = {} if isinstance(A , A): UpperCamelCase = model.mobilenet_va else: UpperCamelCase = model UpperCamelCase = 'MobilenetV1/Conv2d_0/' UpperCamelCase = backbone.conv_stem.convolution.weight UpperCamelCase = backbone.conv_stem.normalization.bias UpperCamelCase = backbone.conv_stem.normalization.weight UpperCamelCase = backbone.conv_stem.normalization.running_mean UpperCamelCase = backbone.conv_stem.normalization.running_var for i in range(13): UpperCamelCase = i + 1 UpperCamelCase = i * 2 UpperCamelCase = backbone.layer[pt_index] UpperCamelCase = f'''MobilenetV1/Conv2d_{tf_index}_depthwise/''' UpperCamelCase = pointer.convolution.weight UpperCamelCase = pointer.normalization.bias UpperCamelCase = pointer.normalization.weight UpperCamelCase = pointer.normalization.running_mean UpperCamelCase = pointer.normalization.running_var UpperCamelCase = backbone.layer[pt_index + 1] UpperCamelCase = f'''MobilenetV1/Conv2d_{tf_index}_pointwise/''' UpperCamelCase = pointer.convolution.weight UpperCamelCase = pointer.normalization.bias UpperCamelCase = pointer.normalization.weight UpperCamelCase = pointer.normalization.running_mean UpperCamelCase = pointer.normalization.running_var if isinstance(A , A): UpperCamelCase = 'MobilenetV1/Logits/Conv2d_1c_1x1/' UpperCamelCase = model.classifier.weight UpperCamelCase = model.classifier.bias return tf_to_pt_map def A_( A : int , A : str , A : Optional[int]): try: import numpy as np import tensorflow as tf except ImportError: logger.error( 'Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see ' 'https://www.tensorflow.org/install/ for installation instructions.') raise # Load weights from TF model UpperCamelCase = tf.train.list_variables(A) UpperCamelCase = {} for name, shape in init_vars: logger.info(f'''Loading TF weight {name} with shape {shape}''') UpperCamelCase = tf.train.load_variable(A , A) UpperCamelCase = array # Build TF to PyTorch weights loading map UpperCamelCase = _build_tf_to_pytorch_map(A , A , A) for name, pointer in tf_to_pt_map.items(): logger.info(f'''Importing {name}''') if name not in tf_weights: logger.info(f'''{name} not in tf pre-trained weights, skipping''') continue UpperCamelCase = tf_weights[name] if "depthwise_weights" in name: logger.info('Transposing depthwise') UpperCamelCase = np.transpose(A , (2, 3, 0, 1)) elif "weights" in name: logger.info('Transposing') if len(pointer.shape) == 2: # copying into linear layer UpperCamelCase = array.squeeze().transpose() else: UpperCamelCase = np.transpose(A , (3, 2, 0, 1)) if pointer.shape != array.shape: raise ValueError(f'''Pointer shape {pointer.shape} and array shape {array.shape} mismatched''') logger.info(f'''Initialize PyTorch weight {name} {array.shape}''') UpperCamelCase = torch.from_numpy(A) tf_weights.pop(A , A) tf_weights.pop(name + '/RMSProp' , A) tf_weights.pop(name + '/RMSProp_1' , A) tf_weights.pop(name + '/ExponentialMovingAverage' , A) logger.info(f'''Weights not copied to PyTorch model: {", ".join(tf_weights.keys())}''') return model def A_( A : torch.Tensor , A : nn.Convad): UpperCamelCase , UpperCamelCase = features.shape[-2:] UpperCamelCase , UpperCamelCase = conv_layer.stride UpperCamelCase , UpperCamelCase = conv_layer.kernel_size if in_height % stride_height == 0: UpperCamelCase = max(kernel_height - stride_height , 0) else: UpperCamelCase = max(kernel_height - (in_height % stride_height) , 0) if in_width % stride_width == 0: UpperCamelCase = max(kernel_width - stride_width , 0) else: UpperCamelCase = max(kernel_width - (in_width % stride_width) , 0) UpperCamelCase = pad_along_width // 2 UpperCamelCase = pad_along_width - pad_left UpperCamelCase = pad_along_height // 2 UpperCamelCase = pad_along_height - pad_top UpperCamelCase = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(A , A , 'constant' , 0.0) class SCREAMING_SNAKE_CASE__ ( nn.Module): def __init__( self , A_ , A_ , A_ , A_ , A_ = 1 , A_ = 1 , A_ = False , A_ = True , A_ = True , )-> None: '''simple docstring''' super().__init__() UpperCamelCase = config if in_channels % groups != 0: raise ValueError(F'''Input channels ({in_channels}) are not divisible by {groups} groups.''' ) if out_channels % groups != 0: raise ValueError(F'''Output channels ({out_channels}) are not divisible by {groups} groups.''' ) UpperCamelCase = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) UpperCamelCase = nn.Convad( in_channels=A_ , out_channels=A_ , kernel_size=A_ , stride=A_ , padding=A_ , groups=A_ , bias=A_ , padding_mode='zeros' , ) if use_normalization: UpperCamelCase = nn.BatchNormad( num_features=A_ , eps=config.layer_norm_eps , momentum=0.9_997 , affine=A_ , track_running_stats=A_ , ) else: UpperCamelCase = None if use_activation: if isinstance(A_ , A_ ): UpperCamelCase = ACTaFN[use_activation] elif isinstance(config.hidden_act , A_ ): UpperCamelCase = ACTaFN[config.hidden_act] else: UpperCamelCase = config.hidden_act else: UpperCamelCase = None def UpperCAmelCase_ ( self , A_ )-> torch.Tensor: '''simple docstring''' if self.config.tf_padding: UpperCamelCase = apply_tf_padding(A_ , self.convolution ) UpperCamelCase = self.convolution(A_ ) if self.normalization is not None: UpperCamelCase = self.normalization(A_ ) if self.activation is not None: UpperCamelCase = self.activation(A_ ) return features class SCREAMING_SNAKE_CASE__ ( snake_case_): lowerCAmelCase_ = MobileNetVaConfig lowerCAmelCase_ = load_tf_weights_in_mobilenet_va lowerCAmelCase_ = """mobilenet_v1""" lowerCAmelCase_ = """pixel_values""" lowerCAmelCase_ = False def UpperCAmelCase_ ( self , A_ )-> None: '''simple docstring''' if isinstance(A_ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(A_ , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) lowerCAmelCase : Union[str, Any] = r'\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it\n as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n' lowerCAmelCase : Union[str, Any] = r'\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`MobileNetV1ImageProcessor.__call__`] for details.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( """The bare MobileNetV1 model outputting raw hidden-states without any specific head on top.""" , snake_case_ , ) class SCREAMING_SNAKE_CASE__ ( snake_case_): def __init__( self , A_ , A_ = True )-> Union[str, Any]: '''simple docstring''' super().__init__(A_ ) UpperCamelCase = config UpperCamelCase = 32 UpperCamelCase = max(int(depth * config.depth_multiplier ) , config.min_depth ) UpperCamelCase = MobileNetVaConvLayer( A_ , in_channels=config.num_channels , out_channels=A_ , kernel_size=3 , stride=2 , ) UpperCamelCase = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] UpperCamelCase = nn.ModuleList() for i in range(13 ): UpperCamelCase = out_channels if strides[i] == 2 or i == 0: depth *= 2 UpperCamelCase = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( A_ , in_channels=A_ , out_channels=A_ , kernel_size=3 , stride=strides[i] , groups=A_ , ) ) self.layer.append( MobileNetVaConvLayer( A_ , in_channels=A_ , out_channels=A_ , kernel_size=1 , ) ) UpperCamelCase = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def UpperCAmelCase_ ( self , A_ )-> Tuple: '''simple docstring''' raise NotImplementedError @add_start_docstrings_to_model_forward(A_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=A_ , config_class=_CONFIG_FOR_DOC , modality='vision' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def UpperCAmelCase_ ( self , A_ = None , A_ = None , A_ = None , )-> Union[tuple, BaseModelOutputWithPoolingAndNoAttention]: '''simple docstring''' UpperCamelCase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) UpperCamelCase = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('You have to specify pixel_values' ) UpperCamelCase = self.conv_stem(A_ ) UpperCamelCase = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): UpperCamelCase = layer_module(A_ ) if output_hidden_states: UpperCamelCase = all_hidden_states + (hidden_states,) UpperCamelCase = hidden_states if self.pooler is not None: UpperCamelCase = torch.flatten(self.pooler(A_ ) , start_dim=1 ) else: UpperCamelCase = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=A_ , pooler_output=A_ , hidden_states=A_ , ) @add_start_docstrings( """ MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. """ , snake_case_ , ) class SCREAMING_SNAKE_CASE__ ( snake_case_): def __init__( self , A_ )-> None: '''simple docstring''' super().__init__(A_ ) UpperCamelCase = config.num_labels UpperCamelCase = MobileNetVaModel(A_ ) UpperCamelCase = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head UpperCamelCase = nn.Dropout(config.classifier_dropout_prob , inplace=A_ ) UpperCamelCase = nn.Linear(A_ , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(A_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=A_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def UpperCAmelCase_ ( self , A_ = None , A_ = None , A_ = None , A_ = None , )-> Union[tuple, ImageClassifierOutputWithNoAttention]: '''simple docstring''' UpperCamelCase = return_dict if return_dict is not None else self.config.use_return_dict UpperCamelCase = self.mobilenet_va(A_ , output_hidden_states=A_ , return_dict=A_ ) UpperCamelCase = outputs.pooler_output if return_dict else outputs[1] UpperCamelCase = self.classifier(self.dropout(A_ ) ) UpperCamelCase = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: UpperCamelCase = 'regression' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): UpperCamelCase = 'single_label_classification' else: UpperCamelCase = 'multi_label_classification' if self.config.problem_type == "regression": UpperCamelCase = MSELoss() if self.num_labels == 1: UpperCamelCase = loss_fct(logits.squeeze() , labels.squeeze() ) else: UpperCamelCase = loss_fct(A_ , A_ ) elif self.config.problem_type == "single_label_classification": UpperCamelCase = CrossEntropyLoss() UpperCamelCase = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": UpperCamelCase = BCEWithLogitsLoss() UpperCamelCase = loss_fct(A_ , A_ ) if not return_dict: UpperCamelCase = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=A_ , logits=A_ , hidden_states=outputs.hidden_states , )
251
1
import argparse import os import shutil from pathlib import Path import onnx import torch from packaging import version from torch.onnx import export from diffusers import OnnxRuntimeModel, OnnxStableDiffusionPipeline, StableDiffusionPipeline lowerCAmelCase = version.parse(version.parse(torch.__version__).base_version) < version.parse('1.11') def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=False , ): """simple docstring""" output_path.parent.mkdir(parents=SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , f=output_path.as_posix() , input_names=SCREAMING_SNAKE_CASE , output_names=SCREAMING_SNAKE_CASE , dynamic_axes=SCREAMING_SNAKE_CASE , do_constant_folding=SCREAMING_SNAKE_CASE , use_external_data_format=SCREAMING_SNAKE_CASE , enable_onnx_checker=SCREAMING_SNAKE_CASE , opset_version=SCREAMING_SNAKE_CASE , ) else: export( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , f=output_path.as_posix() , input_names=SCREAMING_SNAKE_CASE , output_names=SCREAMING_SNAKE_CASE , dynamic_axes=SCREAMING_SNAKE_CASE , do_constant_folding=SCREAMING_SNAKE_CASE , opset_version=SCREAMING_SNAKE_CASE , ) @torch.no_grad() def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False ): """simple docstring""" lowercase__ = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): lowercase__ = '''cuda''' elif fpaa and not torch.cuda.is_available(): raise ValueError('''`float16` model export is only supported on GPUs with CUDA''' ) else: lowercase__ = '''cpu''' lowercase__ = StableDiffusionPipeline.from_pretrained(SCREAMING_SNAKE_CASE , torch_dtype=SCREAMING_SNAKE_CASE ).to(SCREAMING_SNAKE_CASE ) lowercase__ = Path(SCREAMING_SNAKE_CASE ) # TEXT ENCODER lowercase__ = pipeline.text_encoder.config.max_position_embeddings lowercase__ = pipeline.text_encoder.config.hidden_size lowercase__ = pipeline.tokenizer( '''A sample prompt''' , padding='''max_length''' , max_length=pipeline.tokenizer.model_max_length , truncation=SCREAMING_SNAKE_CASE , return_tensors='''pt''' , ) onnx_export( pipeline.text_encoder , model_args=(text_input.input_ids.to(device=SCREAMING_SNAKE_CASE , dtype=torch.intaa )) , output_path=output_path / '''text_encoder''' / '''model.onnx''' , ordered_input_names=['''input_ids'''] , output_names=['''last_hidden_state''', '''pooler_output'''] , dynamic_axes={ '''input_ids''': {0: '''batch''', 1: '''sequence'''}, } , opset=SCREAMING_SNAKE_CASE , ) del pipeline.text_encoder # UNET lowercase__ = pipeline.unet.config.in_channels lowercase__ = pipeline.unet.config.sample_size lowercase__ = output_path / '''unet''' / '''model.onnx''' onnx_export( pipeline.unet , model_args=( torch.randn(2 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).to(device=SCREAMING_SNAKE_CASE , dtype=SCREAMING_SNAKE_CASE ), torch.randn(2 ).to(device=SCREAMING_SNAKE_CASE , dtype=SCREAMING_SNAKE_CASE ), torch.randn(2 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).to(device=SCREAMING_SNAKE_CASE , dtype=SCREAMING_SNAKE_CASE ), False, ) , output_path=SCREAMING_SNAKE_CASE , ordered_input_names=['''sample''', '''timestep''', '''encoder_hidden_states''', '''return_dict'''] , output_names=['''out_sample'''] , dynamic_axes={ '''sample''': {0: '''batch''', 1: '''channels''', 2: '''height''', 3: '''width'''}, '''timestep''': {0: '''batch'''}, '''encoder_hidden_states''': {0: '''batch''', 1: '''sequence'''}, } , opset=SCREAMING_SNAKE_CASE , use_external_data_format=SCREAMING_SNAKE_CASE , ) lowercase__ = str(unet_path.absolute().as_posix() ) lowercase__ = os.path.dirname(SCREAMING_SNAKE_CASE ) lowercase__ = onnx.load(SCREAMING_SNAKE_CASE ) # clean up existing tensor files shutil.rmtree(SCREAMING_SNAKE_CASE ) os.mkdir(SCREAMING_SNAKE_CASE ) # collate external tensor files into one onnx.save_model( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , save_as_external_data=SCREAMING_SNAKE_CASE , all_tensors_to_one_file=SCREAMING_SNAKE_CASE , location='''weights.pb''' , convert_attribute=SCREAMING_SNAKE_CASE , ) del pipeline.unet # VAE ENCODER lowercase__ = pipeline.vae lowercase__ = vae_encoder.config.in_channels lowercase__ = vae_encoder.config.sample_size # need to get the raw tensor output (sample) from the encoder lowercase__ = lambda SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : vae_encoder.encode(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )[0].sample() onnx_export( SCREAMING_SNAKE_CASE , model_args=( torch.randn(1 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).to(device=SCREAMING_SNAKE_CASE , dtype=SCREAMING_SNAKE_CASE ), False, ) , output_path=output_path / '''vae_encoder''' / '''model.onnx''' , ordered_input_names=['''sample''', '''return_dict'''] , output_names=['''latent_sample'''] , dynamic_axes={ '''sample''': {0: '''batch''', 1: '''channels''', 2: '''height''', 3: '''width'''}, } , opset=SCREAMING_SNAKE_CASE , ) # VAE DECODER lowercase__ = pipeline.vae lowercase__ = vae_decoder.config.latent_channels lowercase__ = vae_decoder.config.out_channels # forward only through the decoder part lowercase__ = vae_encoder.decode onnx_export( SCREAMING_SNAKE_CASE , model_args=( torch.randn(1 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).to(device=SCREAMING_SNAKE_CASE , dtype=SCREAMING_SNAKE_CASE ), False, ) , output_path=output_path / '''vae_decoder''' / '''model.onnx''' , ordered_input_names=['''latent_sample''', '''return_dict'''] , output_names=['''sample'''] , dynamic_axes={ '''latent_sample''': {0: '''batch''', 1: '''channels''', 2: '''height''', 3: '''width'''}, } , opset=SCREAMING_SNAKE_CASE , ) del pipeline.vae # SAFETY CHECKER if pipeline.safety_checker is not None: lowercase__ = pipeline.safety_checker lowercase__ = safety_checker.config.vision_config.num_channels lowercase__ = safety_checker.config.vision_config.image_size lowercase__ = safety_checker.forward_onnx onnx_export( pipeline.safety_checker , model_args=( torch.randn( 1 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ).to(device=SCREAMING_SNAKE_CASE , dtype=SCREAMING_SNAKE_CASE ), torch.randn(1 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).to(device=SCREAMING_SNAKE_CASE , dtype=SCREAMING_SNAKE_CASE ), ) , output_path=output_path / '''safety_checker''' / '''model.onnx''' , ordered_input_names=['''clip_input''', '''images'''] , output_names=['''out_images''', '''has_nsfw_concepts'''] , dynamic_axes={ '''clip_input''': {0: '''batch''', 1: '''channels''', 2: '''height''', 3: '''width'''}, '''images''': {0: '''batch''', 1: '''height''', 2: '''width''', 3: '''channels'''}, } , opset=SCREAMING_SNAKE_CASE , ) del pipeline.safety_checker lowercase__ = OnnxRuntimeModel.from_pretrained(output_path / '''safety_checker''' ) lowercase__ = pipeline.feature_extractor else: lowercase__ = None lowercase__ = None lowercase__ = OnnxStableDiffusionPipeline( vae_encoder=OnnxRuntimeModel.from_pretrained(output_path / '''vae_encoder''' ) , vae_decoder=OnnxRuntimeModel.from_pretrained(output_path / '''vae_decoder''' ) , text_encoder=OnnxRuntimeModel.from_pretrained(output_path / '''text_encoder''' ) , tokenizer=pipeline.tokenizer , unet=OnnxRuntimeModel.from_pretrained(output_path / '''unet''' ) , scheduler=pipeline.scheduler , safety_checker=SCREAMING_SNAKE_CASE , feature_extractor=SCREAMING_SNAKE_CASE , requires_safety_checker=safety_checker is not None , ) onnx_pipeline.save_pretrained(SCREAMING_SNAKE_CASE ) print('''ONNX pipeline saved to''' , SCREAMING_SNAKE_CASE ) del pipeline del onnx_pipeline lowercase__ = OnnxStableDiffusionPipeline.from_pretrained(SCREAMING_SNAKE_CASE , provider='''CPUExecutionProvider''' ) print('''ONNX pipeline is loadable''' ) if __name__ == "__main__": lowerCAmelCase = argparse.ArgumentParser() parser.add_argument( '--model_path', type=str, required=True, help='Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).', ) parser.add_argument('--output_path', type=str, required=True, help='Path to the output model.') parser.add_argument( '--opset', default=14, type=int, help='The version of the ONNX operator set to use.', ) parser.add_argument('--fp16', action='store_true', default=False, help='Export the models in `float16` mode') lowerCAmelCase = parser.parse_args() convert_models(args.model_path, args.output_path, args.opset, args.fpaa)
110
import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process lowerCAmelCase = logging.getLogger(__name__) lowerCAmelCase = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) lowerCAmelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class _a : _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={ '''help''': ( '''The model checkpoint for weights initialization. Leave None if you want to train a model from''' ''' scratch.''' ) } , ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''If training from scratch, pass a model type from the list: ''' + ''', '''.join(UpperCamelCase__ )} , ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''} ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) @dataclass class _a : _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''The input training data file (a text file).'''} ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={ '''help''': ( '''The input training data files (multiple files in glob format). ''' '''Very often splitting large files to smaller files can prevent tokenizer going out of memory''' ) } , ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''An optional input evaluation data file to evaluate the perplexity on (a text file).'''} , ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''An optional input train ref data file for whole word mask in Chinese.'''} , ) _lowercase : Optional[str] = field( default=UpperCamelCase__ , metadata={'''help''': '''An optional input eval ref data file for whole word mask in Chinese.'''} , ) _lowercase : bool = field( default=UpperCamelCase__ , metadata={'''help''': '''Whether distinct lines of text in the dataset are to be handled as distinct sequences.'''} , ) _lowercase : bool = field( default=UpperCamelCase__ , metadata={'''help''': '''Train with masked-language modeling loss instead of language modeling.'''} ) _lowercase : bool = field(default=UpperCamelCase__ , metadata={'''help''': '''Whether ot not to use whole word mask.'''} ) _lowercase : float = field( default=0.15 , metadata={'''help''': '''Ratio of tokens to mask for masked language modeling loss'''} ) _lowercase : float = field( default=1 / 6 , metadata={ '''help''': ( '''Ratio of length of a span of masked tokens to surrounding context length for permutation language''' ''' modeling.''' ) } , ) _lowercase : int = field( default=5 , metadata={'''help''': '''Maximum length of a span of masked tokens for permutation language modeling.'''} ) _lowercase : int = field( default=-1 , metadata={ '''help''': ( '''Optional input sequence length after tokenization.''' '''The training dataset will be truncated in block of this size for training.''' '''Default to the model max input length for single sentence inputs (take into account special tokens).''' ) } , ) _lowercase : bool = field( default=UpperCamelCase__ , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = None , ): """simple docstring""" def _dataset(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError('''You need to set world whole masking and mlm to True for Chinese Whole Word Mask''' ) return LineByLineWithRefDataset( tokenizer=SCREAMING_SNAKE_CASE , file_path=SCREAMING_SNAKE_CASE , block_size=args.block_size , ref_path=SCREAMING_SNAKE_CASE , ) return LineByLineTextDataset(tokenizer=SCREAMING_SNAKE_CASE , file_path=SCREAMING_SNAKE_CASE , block_size=args.block_size ) else: return TextDataset( tokenizer=SCREAMING_SNAKE_CASE , file_path=SCREAMING_SNAKE_CASE , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=SCREAMING_SNAKE_CASE , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(SCREAMING_SNAKE_CASE ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def _a ( ): """simple docstring""" lowercase__ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowercase__ , lowercase__ , lowercase__ = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( '''Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file ''' '''or remove the --do_eval argument.''' ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. Use' ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , SCREAMING_SNAKE_CASE ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: lowercase__ = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: lowercase__ = CONFIG_MAPPING[model_args.model_type]() logger.warning('''You are instantiating a new config instance from scratch.''' ) if model_args.tokenizer_name: lowercase__ = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( '''You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another''' ''' script, save it,and load it from here, using --tokenizer_name''' ) if model_args.model_name_or_path: lowercase__ = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=SCREAMING_SNAKE_CASE , cache_dir=model_args.cache_dir , ) else: logger.info('''Training new model from scratch''' ) lowercase__ = AutoModelWithLMHead.from_config(SCREAMING_SNAKE_CASE ) model.resize_token_embeddings(len(SCREAMING_SNAKE_CASE ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( '''BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the''' '''--mlm flag (masked language modeling).''' ) if data_args.block_size <= 0: lowercase__ = tokenizer.max_len # Our input block size will be the max possible for the model else: lowercase__ = min(data_args.block_size , tokenizer.max_len ) # Get datasets lowercase__ = ( get_dataset(SCREAMING_SNAKE_CASE , tokenizer=SCREAMING_SNAKE_CASE , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) lowercase__ = ( get_dataset(SCREAMING_SNAKE_CASE , tokenizer=SCREAMING_SNAKE_CASE , evaluate=SCREAMING_SNAKE_CASE , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": lowercase__ = DataCollatorForPermutationLanguageModeling( tokenizer=SCREAMING_SNAKE_CASE , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: lowercase__ = DataCollatorForWholeWordMask( tokenizer=SCREAMING_SNAKE_CASE , mlm_probability=data_args.mlm_probability ) else: lowercase__ = DataCollatorForLanguageModeling( tokenizer=SCREAMING_SNAKE_CASE , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer lowercase__ = Trainer( model=SCREAMING_SNAKE_CASE , args=SCREAMING_SNAKE_CASE , data_collator=SCREAMING_SNAKE_CASE , train_dataset=SCREAMING_SNAKE_CASE , eval_dataset=SCREAMING_SNAKE_CASE , prediction_loss_only=SCREAMING_SNAKE_CASE , ) # Training if training_args.do_train: lowercase__ = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=SCREAMING_SNAKE_CASE ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowercase__ = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) lowercase__ = trainer.evaluate() lowercase__ = math.exp(eval_output['''eval_loss'''] ) lowercase__ = {'''perplexity''': perplexity} lowercase__ = os.path.join(training_args.output_dir , '''eval_results_lm.txt''' ) if trainer.is_world_master(): with open(SCREAMING_SNAKE_CASE , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key in sorted(result.keys() ): logger.info(''' %s = %s''' , SCREAMING_SNAKE_CASE , str(result[key] ) ) writer.write('''%s = %s\n''' % (key, str(result[key] )) ) results.update(SCREAMING_SNAKE_CASE ) return results def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" main() if __name__ == "__main__": main()
110
1
import numpy as np from PIL import Image def a_ ( __lowercase : np.ndarray , __lowercase : int , __lowercase : int ) -> np.ndarray: _snake_case = np.array(__lowercase ) if arr.shape[0] != arr.shape[1]: raise ValueError('The input array is not a square matrix' ) _snake_case = 0 _snake_case = 0 _snake_case = 0 _snake_case = 0 # compute the shape of the output matrix _snake_case = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape maxpool_shape _snake_case = np.zeros((maxpool_shape, maxpool_shape) ) while i < arr.shape[0]: if i + size > arr.shape[0]: # if the end of the matrix is reached, break break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the maximum of the pooling matrix _snake_case = np.max(arr[i : i + size, j : j + size] ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 _snake_case = 0 _snake_case = 0 return updated_arr def a_ ( __lowercase : np.ndarray , __lowercase : int , __lowercase : int ) -> np.ndarray: _snake_case = np.array(__lowercase ) if arr.shape[0] != arr.shape[1]: raise ValueError('The input array is not a square matrix' ) _snake_case = 0 _snake_case = 0 _snake_case = 0 _snake_case = 0 # compute the shape of the output matrix _snake_case = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape avgpool_shape _snake_case = np.zeros((avgpool_shape, avgpool_shape) ) while i < arr.shape[0]: # if the end of the matrix is reached, break if i + size > arr.shape[0]: break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the average of the pooling matrix _snake_case = int(np.average(arr[i : i + size, j : j + size] ) ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 _snake_case = 0 _snake_case = 0 return updated_arr # Main Function if __name__ == "__main__": from doctest import testmod testmod(name='''avgpooling''', verbose=True) # Loading the image _lowerCamelCase : Dict = Image.open('''path_to_image''') # Converting the image to numpy array and maxpooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(maxpooling(np.array(image), size=3, stride=2)).show() # Converting the image to numpy array and averagepooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(avgpooling(np.array(image), size=3, stride=2)).show()
130
import logging import os import sys import warnings from dataclasses import dataclass, field from random import randint from typing import Optional import datasets import evaluate import numpy as np from datasets import DatasetDict, load_dataset import transformers from transformers import ( AutoConfig, AutoFeatureExtractor, AutoModelForAudioClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version _lowerCamelCase : Dict = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') require_version('''datasets>=1.14.0''', '''To fix: pip install -r examples/pytorch/audio-classification/requirements.txt''') def a_ ( __lowercase : np.ndarray , __lowercase : float , __lowercase : int = 16_000 ) -> List[str]: _snake_case = int(round(sample_rate * max_length ) ) if len(__lowercase ) <= sample_length: return wav _snake_case = randint(0 , len(__lowercase ) - sample_length - 1 ) return wav[random_offset : random_offset + sample_length] @dataclass class SCREAMING_SNAKE_CASE__ : '''simple docstring''' _UpperCAmelCase : Optional[str] = field(default=UpperCAmelCase ,metadata={"help": "Name of a dataset from the datasets package"} ) _UpperCAmelCase : Optional[str] = field( default=UpperCAmelCase ,metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} ) _UpperCAmelCase : Optional[str] = field( default=UpperCAmelCase ,metadata={"help": "A file containing the training audio paths and labels."} ) _UpperCAmelCase : Optional[str] = field( default=UpperCAmelCase ,metadata={"help": "A file containing the validation audio paths and labels."} ) _UpperCAmelCase : str = field( default="train" ,metadata={ "help": "The name of the training data set split to use (via the datasets library). Defaults to 'train'" } ,) _UpperCAmelCase : str = field( default="validation" ,metadata={ "help": ( "The name of the training data set split to use (via the datasets library). Defaults to 'validation'" ) } ,) _UpperCAmelCase : str = field( default="audio" ,metadata={"help": "The name of the dataset column containing the audio data. Defaults to 'audio'"} ,) _UpperCAmelCase : str = field( default="label" ,metadata={"help": "The name of the dataset column containing the labels. Defaults to 'label'"} ) _UpperCAmelCase : Optional[int] = field( default=UpperCAmelCase ,metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } ,) _UpperCAmelCase : Optional[int] = field( default=UpperCAmelCase ,metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } ,) _UpperCAmelCase : float = field( default=2_0 ,metadata={"help": "Audio clips will be randomly cut to this length during training if the value is set."} ,) @dataclass class SCREAMING_SNAKE_CASE__ : '''simple docstring''' _UpperCAmelCase : str = field( default="facebook/wav2vec2-base" ,metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ,) _UpperCAmelCase : Optional[str] = field( default=UpperCAmelCase ,metadata={"help": "Pretrained config name or path if not the same as model_name"} ) _UpperCAmelCase : Optional[str] = field( default=UpperCAmelCase ,metadata={"help": "Where do you want to store the pretrained models downloaded from the Hub"} ) _UpperCAmelCase : str = field( default="main" ,metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} ,) _UpperCAmelCase : Optional[str] = field( default=UpperCAmelCase ,metadata={"help": "Name or path of preprocessor config."} ) _UpperCAmelCase : bool = field( default=UpperCAmelCase ,metadata={"help": "Whether to freeze the feature encoder layers of the model."} ) _UpperCAmelCase : bool = field( default=UpperCAmelCase ,metadata={"help": "Whether to generate an attention mask in the feature extractor."} ) _UpperCAmelCase : bool = field( default=UpperCAmelCase ,metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } ,) _UpperCAmelCase : Optional[bool] = field( default=UpperCAmelCase ,metadata={"help": "Whether to freeze the feature extractor layers of the model."} ) _UpperCAmelCase : bool = field( default=UpperCAmelCase ,metadata={"help": "Will enable to load a pretrained model whose head dimensions are different."} ,) def A ( self : List[Any] ): '''simple docstring''' if not self.freeze_feature_extractor and self.freeze_feature_encoder: warnings.warn( 'The argument `--freeze_feature_extractor` is deprecated and ' 'will be removed in a future version. Use `--freeze_feature_encoder`' 'instead. Setting `freeze_feature_encoder==True`.' , lowercase , ) if self.freeze_feature_extractor and not self.freeze_feature_encoder: raise ValueError( 'The argument `--freeze_feature_extractor` is deprecated and ' 'should not be used in combination with `--freeze_feature_encoder`.' 'Only make use of `--freeze_feature_encoder`.' ) def a_ ( ) -> Optional[Any]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _snake_case = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _snake_case , _snake_case , _snake_case = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _snake_case , _snake_case , _snake_case = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_audio_classification' , __lowercase , __lowercase ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _snake_case = training_args.get_process_log_level() logger.setLevel(__lowercase ) transformers.utils.logging.set_verbosity(__lowercase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu} ''' + f'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(f'''Training/evaluation parameters {training_args}''' ) # Set seed before initializing model. set_seed(training_args.seed ) # Detecting last checkpoint. _snake_case = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _snake_case = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' 'Use --overwrite_output_dir to train from scratch.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Initialize our dataset and prepare it for the audio classification task. _snake_case = DatasetDict() _snake_case = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.train_split_name , use_auth_token=True if model_args.use_auth_token else None , ) _snake_case = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.eval_split_name , use_auth_token=True if model_args.use_auth_token else None , ) if data_args.audio_column_name not in raw_datasets["train"].column_names: raise ValueError( f'''--audio_column_name {data_args.audio_column_name} not found in dataset \'{data_args.dataset_name}\'. ''' 'Make sure to set `--audio_column_name` to the correct audio column - one of ' f'''{', '.join(raw_datasets['train'].column_names )}.''' ) if data_args.label_column_name not in raw_datasets["train"].column_names: raise ValueError( f'''--label_column_name {data_args.label_column_name} not found in dataset \'{data_args.dataset_name}\'. ''' 'Make sure to set `--label_column_name` to the correct text column - one of ' f'''{', '.join(raw_datasets['train'].column_names )}.''' ) # Setting `return_attention_mask=True` is the way to get a correctly masked mean-pooling over # transformer outputs in the classifier, but it doesn't always lead to better accuracy _snake_case = AutoFeatureExtractor.from_pretrained( model_args.feature_extractor_name or model_args.model_name_or_path , return_attention_mask=model_args.attention_mask , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # `datasets` takes care of automatically loading and resampling the audio, # so we just need to set the correct target sampling rate. _snake_case = raw_datasets.cast_column( data_args.audio_column_name , datasets.features.Audio(sampling_rate=feature_extractor.sampling_rate ) ) _snake_case = feature_extractor.model_input_names[0] def train_transforms(__lowercase : Union[str, Any] ): _snake_case = [] for audio in batch[data_args.audio_column_name]: _snake_case = random_subsample( audio['array'] , max_length=data_args.max_length_seconds , sample_rate=feature_extractor.sampling_rate ) subsampled_wavs.append(__lowercase ) _snake_case = feature_extractor(__lowercase , sampling_rate=feature_extractor.sampling_rate ) _snake_case = {model_input_name: inputs.get(__lowercase )} _snake_case = list(batch[data_args.label_column_name] ) return output_batch def val_transforms(__lowercase : Optional[Any] ): _snake_case = [audio['array'] for audio in batch[data_args.audio_column_name]] _snake_case = feature_extractor(__lowercase , sampling_rate=feature_extractor.sampling_rate ) _snake_case = {model_input_name: inputs.get(__lowercase )} _snake_case = list(batch[data_args.label_column_name] ) return output_batch # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _snake_case = raw_datasets['train'].features[data_args.label_column_name].names _snake_case , _snake_case = {}, {} for i, label in enumerate(__lowercase ): _snake_case = str(__lowercase ) _snake_case = label # Load the accuracy metric from the datasets package _snake_case = evaluate.load('accuracy' ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with # `predictions` and `label_ids` fields) and has to return a dictionary string to float. def compute_metrics(__lowercase : Dict ): _snake_case = np.argmax(eval_pred.predictions , axis=1 ) return metric.compute(predictions=__lowercase , references=eval_pred.label_ids ) _snake_case = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(__lowercase ) , labelaid=__lowercase , idalabel=__lowercase , finetuning_task='audio-classification' , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _snake_case = AutoModelForAudioClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=__lowercase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) # freeze the convolutional waveform encoder if model_args.freeze_feature_encoder: model.freeze_feature_encoder() if training_args.do_train: if data_args.max_train_samples is not None: _snake_case = ( raw_datasets['train'].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms raw_datasets["train"].set_transform(__lowercase , output_all_columns=__lowercase ) if training_args.do_eval: if data_args.max_eval_samples is not None: _snake_case = ( raw_datasets['eval'].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms raw_datasets["eval"].set_transform(__lowercase , output_all_columns=__lowercase ) # Initialize our trainer _snake_case = Trainer( model=__lowercase , args=__lowercase , train_dataset=raw_datasets['train'] if training_args.do_train else None , eval_dataset=raw_datasets['eval'] if training_args.do_eval else None , compute_metrics=__lowercase , tokenizer=__lowercase , ) # Training if training_args.do_train: _snake_case = None if training_args.resume_from_checkpoint is not None: _snake_case = training_args.resume_from_checkpoint elif last_checkpoint is not None: _snake_case = last_checkpoint _snake_case = trainer.train(resume_from_checkpoint=__lowercase ) trainer.save_model() trainer.log_metrics('train' , train_result.metrics ) trainer.save_metrics('train' , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _snake_case = trainer.evaluate() trainer.log_metrics('eval' , __lowercase ) trainer.save_metrics('eval' , __lowercase ) # Write model card and (optionally) push to hub _snake_case = { 'finetuned_from': model_args.model_name_or_path, 'tasks': 'audio-classification', 'dataset': data_args.dataset_name, 'tags': ['audio-classification'], } if training_args.push_to_hub: trainer.push_to_hub(**__lowercase ) else: trainer.create_model_card(**__lowercase ) if __name__ == "__main__": main()
130
1
"""simple docstring""" import unittest from datasets import load_dataset from transformers.pipelines import pipeline from transformers.testing_utils import is_pipeline_test, nested_simplify, require_torch, slow @is_pipeline_test @require_torch class _UpperCAmelCase ( unittest.TestCase): @require_torch def __snake_case ( self ) -> List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = pipeline( task="""zero-shot-audio-classification""" , model="""hf-internal-testing/tiny-clap-htsat-unfused""" ) _UpperCAmelCase : Dict = load_dataset("""ashraq/esc50""" ) _UpperCAmelCase : Dict = dataset["""train"""]["""audio"""][-1]["""array"""] _UpperCAmelCase : Tuple = audio_classifier(__lowerCAmelCase , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [{"""score""": 0.501, """label""": """Sound of a dog"""}, {"""score""": 0.499, """label""": """Sound of vaccum cleaner"""}] , ) @unittest.skip("""No models are available in TF""" ) def __snake_case ( self ) -> List[Any]: '''simple docstring''' pass @slow @require_torch def __snake_case ( self ) -> Dict: '''simple docstring''' _UpperCAmelCase : List[str] = pipeline( task="""zero-shot-audio-classification""" , model="""laion/clap-htsat-unfused""" , ) # This is an audio of a dog _UpperCAmelCase : Any = load_dataset("""ashraq/esc50""" ) _UpperCAmelCase : Any = dataset["""train"""]["""audio"""][-1]["""array"""] _UpperCAmelCase : Optional[int] = audio_classifier(__lowerCAmelCase , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [ {"""score""": 0.999, """label""": """Sound of a dog"""}, {"""score""": 0.001, """label""": """Sound of vaccum cleaner"""}, ] , ) _UpperCAmelCase : int = audio_classifier([audio] * 5 , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [ [ {"""score""": 0.999, """label""": """Sound of a dog"""}, {"""score""": 0.001, """label""": """Sound of vaccum cleaner"""}, ], ] * 5 , ) _UpperCAmelCase : int = audio_classifier( [audio] * 5 , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] , batch_size=5 ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [ [ {"""score""": 0.999, """label""": """Sound of a dog"""}, {"""score""": 0.001, """label""": """Sound of vaccum cleaner"""}, ], ] * 5 , ) @unittest.skip("""No models are available in TF""" ) def __snake_case ( self ) -> List[Any]: '''simple docstring''' pass
246
"""simple docstring""" from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, CustomDtype, DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, DynamoBackend, FPaRecipeKwargs, FullyShardedDataParallelPlugin, GradientAccumulationPlugin, GradScalerKwargs, InitProcessGroupKwargs, KwargsHandler, LoggerType, MegatronLMPlugin, PrecisionType, ProjectConfiguration, RNGType, SageMakerDistributedType, TensorInformation, TorchDynamoPlugin, ) from .environment import get_int_from_env, parse_choice_from_env, parse_flag_from_env from .imports import ( get_ccl_version, is_abit_bnb_available, is_abit_bnb_available, is_aim_available, is_bfaa_available, is_bnb_available, is_botoa_available, is_ccl_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_fpa_available, is_ipex_available, is_megatron_lm_available, is_mlflow_available, is_mps_available, is_npu_available, is_rich_available, is_safetensors_available, is_sagemaker_available, is_tensorboard_available, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) from .modeling import ( check_device_map, check_tied_parameters_in_config, check_tied_parameters_on_same_device, compute_module_sizes, convert_file_size_to_int, dtype_byte_size, find_tied_parameters, get_balanced_memory, get_max_layer_size, get_max_memory, get_mixed_precision_context_manager, id_tensor_storage, infer_auto_device_map, load_checkpoint_in_model, load_offloaded_weights, load_state_dict, named_module_tensors, retie_parameters, set_module_tensor_to_device, shard_checkpoint, ) from .offload import ( OffloadedWeightsLoader, PrefixedDataset, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, save_offload_index, ) from .operations import ( broadcast, broadcast_object_list, concatenate, convert_outputs_to_fpaa, convert_to_fpaa, find_batch_size, find_device, gather, gather_object, get_data_structure, honor_type, initialize_tensors, is_namedtuple, is_tensor_information, is_torch_tensor, listify, pad_across_processes, recursively_apply, reduce, send_to_device, slice_tensors, ) from .versions import compare_versions, is_torch_version if is_deepspeed_available(): from .deepspeed import ( DeepSpeedEngineWrapper, DeepSpeedOptimizerWrapper, DeepSpeedSchedulerWrapper, DummyOptim, DummyScheduler, HfDeepSpeedConfig, ) from .bnb import has_abit_bnb_layers, load_and_quantize_model from .fsdp_utils import load_fsdp_model, load_fsdp_optimizer, save_fsdp_model, save_fsdp_optimizer from .launch import ( PrepareForLaunch, _filter_args, prepare_deepspeed_cmd_env, prepare_multi_gpu_env, prepare_sagemager_args_inputs, prepare_simple_launcher_cmd_env, prepare_tpu, ) from .megatron_lm import ( AbstractTrainStep, BertTrainStep, GPTTrainStep, MegatronEngine, MegatronLMDummyDataLoader, MegatronLMDummyScheduler, MegatronLMOptimizerWrapper, MegatronLMSchedulerWrapper, TaTrainStep, avg_losses_across_data_parallel_group, gather_across_data_parallel_groups, ) from .megatron_lm import initialize as megatron_lm_initialize from .megatron_lm import prepare_data_loader as megatron_lm_prepare_data_loader from .megatron_lm import prepare_model as megatron_lm_prepare_model from .megatron_lm import prepare_optimizer as megatron_lm_prepare_optimizer from .megatron_lm import prepare_scheduler as megatron_lm_prepare_scheduler from .memory import find_executable_batch_size, release_memory from .other import ( extract_model_from_parallel, get_pretty_name, is_port_in_use, merge_dicts, patch_environment, save, wait_for_everyone, write_basic_config, ) from .random import set_seed, synchronize_rng_state, synchronize_rng_states from .torch_xla import install_xla from .tqdm import tqdm from .transformer_engine import convert_model, has_transformer_engine_layers
289
0
import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC a =parse(importlib.metadata.version("""torch""")) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> Any: if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(F"`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}" ) __lowerCamelCase : Union[str, Any] = STR_OPERATION_TO_FUNC[operation] if isinstance(lowerCamelCase__ , lowerCamelCase__ ): __lowerCamelCase : Optional[int] = parse(importlib.metadata.version(lowerCamelCase__ ) ) return operation(lowerCamelCase__ , parse(lowerCamelCase__ ) ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> List[str]: return compare_versions(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ )
113
import os import random import sys from . import cryptomath_module as cryptomath from . import rabin_miller a =3 def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> int: print('Generating primitive root of p' ) while True: __lowerCamelCase : Tuple = random.randrange(3 , lowerCamelCase__ ) if pow(lowerCamelCase__ , 2 , lowerCamelCase__ ) == 1: continue if pow(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) == 1: continue return g def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> tuple[tuple[int, int, int, int], tuple[int, int]]: print('Generating prime p...' ) __lowerCamelCase : List[str] = rabin_miller.generate_large_prime(lowerCamelCase__ ) # select large prime number. __lowerCamelCase : Dict = primitive_root(lowerCamelCase__ ) # one primitive root on modulo p. __lowerCamelCase : Optional[int] = random.randrange(3 , lowerCamelCase__ ) # private_key -> have to be greater than 2 for safety. __lowerCamelCase : List[Any] = cryptomath.find_mod_inverse(pow(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) , lowerCamelCase__ ) __lowerCamelCase : int = (key_size, e_a, e_a, p) __lowerCamelCase : str = (key_size, d) return public_key, private_key def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> None: if os.path.exists(F"{name}_pubkey.txt" ) or os.path.exists(F"{name}_privkey.txt" ): print('\nWARNING:' ) print( F"\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n" 'Use a different name or delete these files and re-run this program.' ) sys.exit() __lowerCamelCase , __lowerCamelCase : List[Any] = generate_key(lowerCamelCase__ ) print(F"\nWriting public key to file {name}_pubkey.txt..." ) with open(F"{name}_pubkey.txt" , 'w' ) as fo: fo.write(F"{public_key[0]},{public_key[1]},{public_key[2]},{public_key[3]}" ) print(F"Writing private key to file {name}_privkey.txt..." ) with open(F"{name}_privkey.txt" , 'w' ) as fo: fo.write(F"{private_key[0]},{private_key[1]}" ) def SCREAMING_SNAKE_CASE__ ( ) -> None: print('Making key files...' ) make_key_files('elgamal' , 2_0_4_8 ) print('Key files generation successful' ) if __name__ == "__main__": main()
113
1
import argparse import math import os import torch from neural_compressor.utils.pytorch import load from PIL import Image from transformers import CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, StableDiffusionPipeline, UNetaDConditionModel def SCREAMING_SNAKE_CASE__ ( ) -> Optional[Any]: __lowerCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument( '-m' , '--pretrained_model_name_or_path' , type=lowerCamelCase__ , default=lowerCamelCase__ , required=lowerCamelCase__ , help='Path to pretrained model or model identifier from huggingface.co/models.' , ) parser.add_argument( '-c' , '--caption' , type=lowerCamelCase__ , default='robotic cat with wings' , help='Text used to generate images.' , ) parser.add_argument( '-n' , '--images_num' , type=lowerCamelCase__ , default=4 , help='How much images to generate.' , ) parser.add_argument( '-s' , '--seed' , type=lowerCamelCase__ , default=4_2 , help='Seed for random process.' , ) parser.add_argument( '-ci' , '--cuda_id' , type=lowerCamelCase__ , default=0 , help='cuda_id.' , ) __lowerCamelCase : Any = parser.parse_args() return args def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> Any: if not len(lowerCamelCase__ ) == rows * cols: raise ValueError('The specified number of rows and columns are not correct.' ) __lowerCamelCase , __lowerCamelCase : Optional[int] = imgs[0].size __lowerCamelCase : List[Any] = Image.new('RGB' , size=(cols * w, rows * h) ) __lowerCamelCase , __lowerCamelCase : List[str] = grid.size for i, img in enumerate(lowerCamelCase__ ): grid.paste(lowerCamelCase__ , box=(i % cols * w, i // cols * h) ) return grid def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__="robotic cat with wings" , lowerCamelCase__=7.5 , lowerCamelCase__=5_0 , lowerCamelCase__=1 , lowerCamelCase__=4_2 , ) -> int: __lowerCamelCase : Any = torch.Generator(pipeline.device ).manual_seed(lowerCamelCase__ ) __lowerCamelCase : int = pipeline( lowerCamelCase__ , guidance_scale=lowerCamelCase__ , num_inference_steps=lowerCamelCase__ , generator=lowerCamelCase__ , num_images_per_prompt=lowerCamelCase__ , ).images __lowerCamelCase : Union[str, Any] = int(math.sqrt(lowerCamelCase__ ) ) __lowerCamelCase : str = image_grid(lowerCamelCase__ , rows=_rows , cols=num_images_per_prompt // _rows ) return grid, images a =parse_args() # Load models and create wrapper for stable diffusion a =CLIPTokenizer.from_pretrained(args.pretrained_model_name_or_path, subfolder="""tokenizer""") a =CLIPTextModel.from_pretrained(args.pretrained_model_name_or_path, subfolder="""text_encoder""") a =AutoencoderKL.from_pretrained(args.pretrained_model_name_or_path, subfolder="""vae""") a =UNetaDConditionModel.from_pretrained(args.pretrained_model_name_or_path, subfolder="""unet""") a =StableDiffusionPipeline.from_pretrained( args.pretrained_model_name_or_path, text_encoder=text_encoder, vae=vae, unet=unet, tokenizer=tokenizer ) a =lambda images, clip_input: (images, False) if os.path.exists(os.path.join(args.pretrained_model_name_or_path, """best_model.pt""")): a =load(args.pretrained_model_name_or_path, model=unet) unet.eval() setattr(pipeline, """unet""", unet) else: a =unet.to(torch.device("""cuda""", args.cuda_id)) a =pipeline.to(unet.device) a , a =generate_images(pipeline, prompt=args.caption, num_images_per_prompt=args.images_num, seed=args.seed) grid.save(os.path.join(args.pretrained_model_name_or_path, """{}.png""".format("""_""".join(args.caption.split())))) a =os.path.join(args.pretrained_model_name_or_path, """_""".join(args.caption.split())) os.makedirs(dirname, exist_ok=True) for idx, image in enumerate(images): image.save(os.path.join(dirname, """{}.png""".format(idx + 1)))
73
"""simple docstring""" import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) lowerCAmelCase__ : Optional[Any] = logging.getLogger() def a_ ( ): UpperCAmelCase__ = argparse.ArgumentParser() parser.add_argument('-f' ) UpperCAmelCase__ = parser.parse_args() return args.f class snake_case ( __UpperCAmelCase ): """simple docstring""" def __lowerCAmelCase ( self : int ): UpperCAmelCase__ = logging.StreamHandler(sys.stdout ) logger.addHandler(lowerCamelCase__ ) def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ): UpperCAmelCase__ = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 ,'run_glue_deebert.py' ) with patch.object(lowerCamelCase__ ,'argv' ,lowerCamelCase__ ): UpperCAmelCase__ = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(lowerCamelCase__ ,0.6_6_6 ) @slow @require_torch_non_multi_gpu def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = '\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n '.split() self.run_and_check(lowerCamelCase__ ) UpperCAmelCase__ = '\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n '.split() self.run_and_check(lowerCamelCase__ ) UpperCAmelCase__ = '\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n '.split() self.run_and_check(lowerCamelCase__ )
98
0
import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCamelCase__ ( lowerCamelCase__ , unittest.TestCase): '''simple docstring''' snake_case_ =CodeGenTokenizer snake_case_ =CodeGenTokenizerFast snake_case_ =True snake_case_ ={"""add_prefix_space""": True} snake_case_ =False def lowerCAmelCase__ (self ) -> Union[str, Any]: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase__ : Optional[Any] = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', '''<|endoftext|>''', ] lowerCAmelCase__ : Optional[Any] = dict(zip(__lowerCamelCase ,range(len(__lowerCamelCase ) ) ) ) lowerCAmelCase__ : Optional[Any] = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] lowerCAmelCase__ : Dict = {'''unk_token''': '''<unk>'''} lowerCAmelCase__ : List[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''vocab_file'''] ) lowerCAmelCase__ : Tuple = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file ,'''w''' ,encoding='''utf-8''' ) as fp: fp.write(json.dumps(__lowerCamelCase ) + '''\n''' ) with open(self.merges_file ,'''w''' ,encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(__lowerCamelCase ) ) def lowerCAmelCase__ (self ,**__lowerCamelCase ) -> List[str]: """simple docstring""" kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname ,**__lowerCamelCase ) def lowerCAmelCase__ (self ,**__lowerCamelCase ) -> int: """simple docstring""" kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname ,**__lowerCamelCase ) def lowerCAmelCase__ (self ,__lowerCamelCase ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ : Any = '''lower newer''' lowerCAmelCase__ : Tuple = '''lower newer''' return input_text, output_text def lowerCAmelCase__ (self ) -> Union[str, Any]: """simple docstring""" lowerCAmelCase__ : Tuple = CodeGenTokenizer(self.vocab_file ,self.merges_file ,**self.special_tokens_map ) lowerCAmelCase__ : List[Any] = '''lower newer''' lowerCAmelCase__ : Union[str, Any] = ['''\u0120low''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] lowerCAmelCase__ : Optional[Any] = tokenizer.tokenize(__lowerCamelCase ,add_prefix_space=__lowerCamelCase ) self.assertListEqual(__lowerCamelCase ,__lowerCamelCase ) lowerCAmelCase__ : Optional[Any] = tokens + [tokenizer.unk_token] lowerCAmelCase__ : Dict = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) ,__lowerCamelCase ) def lowerCAmelCase__ (self ) -> int: """simple docstring""" if not self.test_rust_tokenizer: return lowerCAmelCase__ : List[str] = self.get_tokenizer() lowerCAmelCase__ : List[Any] = self.get_rust_tokenizer(add_prefix_space=__lowerCamelCase ) lowerCAmelCase__ : int = '''lower newer''' # Testing tokenization lowerCAmelCase__ : int = tokenizer.tokenize(__lowerCamelCase ,add_prefix_space=__lowerCamelCase ) lowerCAmelCase__ : Optional[Any] = rust_tokenizer.tokenize(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase ,__lowerCamelCase ) # Testing conversion to ids without special tokens lowerCAmelCase__ : int = tokenizer.encode(__lowerCamelCase ,add_special_tokens=__lowerCamelCase ,add_prefix_space=__lowerCamelCase ) lowerCAmelCase__ : List[str] = rust_tokenizer.encode(__lowerCamelCase ,add_special_tokens=__lowerCamelCase ) self.assertListEqual(__lowerCamelCase ,__lowerCamelCase ) # Testing conversion to ids with special tokens lowerCAmelCase__ : List[str] = self.get_rust_tokenizer(add_prefix_space=__lowerCamelCase ) lowerCAmelCase__ : Union[str, Any] = tokenizer.encode(__lowerCamelCase ,add_prefix_space=__lowerCamelCase ) lowerCAmelCase__ : Tuple = rust_tokenizer.encode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase ,__lowerCamelCase ) # Testing the unknown token lowerCAmelCase__ : str = tokens + [rust_tokenizer.unk_token] lowerCAmelCase__ : Dict = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(__lowerCamelCase ) ,__lowerCamelCase ) def lowerCAmelCase__ (self ,*__lowerCamelCase ,**__lowerCamelCase ) -> Any: """simple docstring""" pass def lowerCAmelCase__ (self ,__lowerCamelCase=15 ) -> Optional[Any]: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase__ : str = self.rust_tokenizer_class.from_pretrained(__lowerCamelCase ,**__lowerCamelCase ) # Simple input lowerCAmelCase__ : Dict = '''This is a simple input''' lowerCAmelCase__ : Optional[Any] = ['''This is a simple input 1''', '''This is a simple input 2'''] lowerCAmelCase__ : int = ('''This is a simple input''', '''This is a pair''') lowerCAmelCase__ : List[Any] = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(__lowerCamelCase ,tokenizer_r.encode ,__lowerCamelCase ,max_length=__lowerCamelCase ,padding='''max_length''' ) # Simple input self.assertRaises(__lowerCamelCase ,tokenizer_r.encode_plus ,__lowerCamelCase ,max_length=__lowerCamelCase ,padding='''max_length''' ) # Simple input self.assertRaises( __lowerCamelCase ,tokenizer_r.batch_encode_plus ,__lowerCamelCase ,max_length=__lowerCamelCase ,padding='''max_length''' ,) # Pair input self.assertRaises(__lowerCamelCase ,tokenizer_r.encode ,__lowerCamelCase ,max_length=__lowerCamelCase ,padding='''max_length''' ) # Pair input self.assertRaises(__lowerCamelCase ,tokenizer_r.encode_plus ,__lowerCamelCase ,max_length=__lowerCamelCase ,padding='''max_length''' ) # Pair input self.assertRaises( __lowerCamelCase ,tokenizer_r.batch_encode_plus ,__lowerCamelCase ,max_length=__lowerCamelCase ,padding='''max_length''' ,) def lowerCAmelCase__ (self ) -> List[str]: """simple docstring""" lowerCAmelCase__ : Tuple = CodeGenTokenizer.from_pretrained(self.tmpdirname ,pad_token='''<pad>''' ) # Simple input lowerCAmelCase__ : str = '''This is a simple input''' lowerCAmelCase__ : Dict = ['''This is a simple input looooooooong''', '''This is a simple input'''] lowerCAmelCase__ : Dict = ('''This is a simple input''', '''This is a pair''') lowerCAmelCase__ : str = [ ('''This is a simple input loooooong''', '''This is a simple input'''), ('''This is a simple pair loooooong''', '''This is a simple pair'''), ] lowerCAmelCase__ : List[Any] = tokenizer.pad_token_id lowerCAmelCase__ : Optional[int] = tokenizer(__lowerCamelCase ,padding='''max_length''' ,max_length=30 ,return_tensors='''np''' ) lowerCAmelCase__ : List[Any] = tokenizer(__lowerCamelCase ,padding=__lowerCamelCase ,truncate=__lowerCamelCase ,return_tensors='''np''' ) lowerCAmelCase__ : Union[str, Any] = tokenizer(*__lowerCamelCase ,padding='''max_length''' ,max_length=60 ,return_tensors='''np''' ) lowerCAmelCase__ : Any = tokenizer(__lowerCamelCase ,padding=__lowerCamelCase ,truncate=__lowerCamelCase ,return_tensors='''np''' ) # s # test single string max_length padding self.assertEqual(out_s['''input_ids'''].shape[-1] ,30 ) self.assertTrue(pad_token_id in out_s['''input_ids'''] ) self.assertTrue(0 in out_s['''attention_mask'''] ) # s2 # test automatic padding self.assertEqual(out_sa['''input_ids'''].shape[-1] ,33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['''input_ids'''][0] ) self.assertFalse(0 in out_sa['''attention_mask'''][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['''input_ids'''][1] ) self.assertTrue(0 in out_sa['''attention_mask'''][1] ) # p # test single pair max_length padding self.assertEqual(out_p['''input_ids'''].shape[-1] ,60 ) self.assertTrue(pad_token_id in out_p['''input_ids'''] ) self.assertTrue(0 in out_p['''attention_mask'''] ) # p2 # test automatic padding pair self.assertEqual(out_pa['''input_ids'''].shape[-1] ,52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['''input_ids'''][0] ) self.assertFalse(0 in out_pa['''attention_mask'''][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['''input_ids'''][1] ) self.assertTrue(0 in out_pa['''attention_mask'''][1] ) def lowerCAmelCase__ (self ) -> Optional[Any]: """simple docstring""" lowerCAmelCase__ : Any = '''$$$''' lowerCAmelCase__ : Dict = CodeGenTokenizer.from_pretrained(self.tmpdirname ,bos_token=__lowerCamelCase ,add_bos_token=__lowerCamelCase ) lowerCAmelCase__ : List[Any] = '''This is a simple input''' lowerCAmelCase__ : int = ['''This is a simple input 1''', '''This is a simple input 2'''] lowerCAmelCase__ : str = tokenizer.bos_token_id lowerCAmelCase__ : List[Any] = tokenizer(__lowerCamelCase ) lowerCAmelCase__ : Optional[int] = tokenizer(__lowerCamelCase ) self.assertEqual(out_s.input_ids[0] ,__lowerCamelCase ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) lowerCAmelCase__ : str = tokenizer.decode(out_s.input_ids ) lowerCAmelCase__ : List[Any] = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] ,__lowerCamelCase ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def lowerCAmelCase__ (self ) -> str: """simple docstring""" lowerCAmelCase__ : List[Any] = CodeGenTokenizer.from_pretrained('''Salesforce/codegen-350M-mono''' ) lowerCAmelCase__ : str = '''\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#''' lowerCAmelCase__ : Dict = '''\nif len_a > len_b: result = a\nelse: result = b''' lowerCAmelCase__ : Tuple = tokenizer.encode(__lowerCamelCase ) lowerCAmelCase__ : int = ['''^#''', re.escape('''<|endoftext|>''' ), '''^\'\'\'''', '''^"""''', '''\n\n\n'''] lowerCAmelCase__ : Dict = tokenizer.decode(__lowerCamelCase ,truncate_before_pattern=__lowerCamelCase ) self.assertEqual(__lowerCamelCase ,__lowerCamelCase ) def lowerCAmelCase__ (self ) -> Tuple: """simple docstring""" pass
94
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCamelCase__ ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase): '''simple docstring''' snake_case_ =IFInpaintingSuperResolutionPipeline snake_case_ =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} snake_case_ =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"""original_image"""}) snake_case_ =PipelineTesterMixin.required_optional_params - {"""latents"""} def lowerCAmelCase__ (self ) -> Union[str, Any]: """simple docstring""" return self._get_superresolution_dummy_components() def lowerCAmelCase__ (self ,__lowerCamelCase ,__lowerCamelCase=0 ) -> Dict: """simple docstring""" if str(__lowerCamelCase ).startswith('''mps''' ): lowerCAmelCase__ : Union[str, Any] = torch.manual_seed(__lowerCamelCase ) else: lowerCAmelCase__ : int = torch.Generator(device=__lowerCamelCase ).manual_seed(__lowerCamelCase ) lowerCAmelCase__ : str = floats_tensor((1, 3, 16, 16) ,rng=random.Random(__lowerCamelCase ) ).to(__lowerCamelCase ) lowerCAmelCase__ : Optional[Any] = floats_tensor((1, 3, 32, 32) ,rng=random.Random(__lowerCamelCase ) ).to(__lowerCamelCase ) lowerCAmelCase__ : Union[str, Any] = floats_tensor((1, 3, 32, 32) ,rng=random.Random(__lowerCamelCase ) ).to(__lowerCamelCase ) lowerCAmelCase__ : Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() ,reason='''XFormers attention is only available with CUDA and `xformers` installed''' ,) def lowerCAmelCase__ (self ) -> int: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def lowerCAmelCase__ (self ) -> int: """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' ,reason='''float16 requires CUDA''' ) def lowerCAmelCase__ (self ) -> str: """simple docstring""" super().test_save_load_floataa(expected_max_diff=1e-1 ) def lowerCAmelCase__ (self ) -> Any: """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def lowerCAmelCase__ (self ) -> Optional[Any]: """simple docstring""" self._test_save_load_local() def lowerCAmelCase__ (self ) -> Any: """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1e-2 ,)
94
1
from __future__ import annotations from collections.abc import Callable from typing import Generic, TypeVar SCREAMING_SNAKE_CASE__ = TypeVar("""T""") SCREAMING_SNAKE_CASE__ = TypeVar("""U""") class A__ ( Generic[T, U] ): def __init__( self : int , _UpperCAmelCase : List[str] , _UpperCAmelCase : List[Any] ) -> Dict: """simple docstring""" __lowercase = key __lowercase = val __lowercase = None __lowercase = None def __repr__( self : Optional[Any] ) -> str: """simple docstring""" return ( f"""Node: key: {self.key}, val: {self.val}, """ f"""has next: {bool(self.next )}, has prev: {bool(self.prev )}""" ) class A__ ( Generic[T, U] ): def __init__( self : int ) -> None: """simple docstring""" __lowercase = DoubleLinkedListNode(lowercase_ , lowercase_ ) __lowercase = DoubleLinkedListNode(lowercase_ , lowercase_ ) __lowercase , __lowercase = self.rear, self.head def __repr__( self : Tuple ) -> str: """simple docstring""" __lowercase = ['DoubleLinkedList'] __lowercase = self.head while node.next is not None: rep.append(str(lowercase_ ) ) __lowercase = node.next rep.append(str(self.rear ) ) return ",\n ".join(lowercase_ ) def a__ ( self : Any , _UpperCAmelCase : int ) -> None: """simple docstring""" __lowercase = self.rear.prev # All nodes other than self.head are guaranteed to have non-None previous assert previous is not None __lowercase = node __lowercase = previous __lowercase = node __lowercase = self.rear def a__ ( self : Tuple , _UpperCAmelCase : Tuple ) -> DoubleLinkedListNode[T, U] | None: """simple docstring""" if node.prev is None or node.next is None: return None __lowercase = node.next __lowercase = node.prev __lowercase = None __lowercase = None return node class A__ ( Generic[T, U] ): lowerCAmelCase__ : dict[Callable[[T], U], LRUCache[T, U]] = {} def __init__( self : List[Any] , _UpperCAmelCase : Union[str, Any] ) -> List[str]: """simple docstring""" __lowercase = DoubleLinkedList() __lowercase = capacity __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = {} def __repr__( self : Union[str, Any] ) -> str: """simple docstring""" return ( f"""CacheInfo(hits={self.hits}, misses={self.miss}, """ f"""capacity={self.capacity}, current size={self.num_keys})""" ) def __contains__( self : List[Any] , _UpperCAmelCase : Union[str, Any] ) -> bool: """simple docstring""" return key in self.cache def a__ ( self : str , _UpperCAmelCase : Union[str, Any] ) -> U | None: """simple docstring""" if key in self.cache: self.hits += 1 __lowercase = self.cache[key] __lowercase = self.list.remove(self.cache[key] ) assert node == value_node # node is guaranteed not None because it is in self.cache assert node is not None self.list.add(lowercase_ ) return node.val self.miss += 1 return None def a__ ( self : Any , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Union[str, Any] ) -> None: """simple docstring""" if key not in self.cache: if self.num_keys >= self.capacity: # delete first node (oldest) when over capacity __lowercase = self.list.head.next # guaranteed to have a non-None first node when num_keys > 0 # explain to type checker via assertions assert first_node is not None assert first_node.key is not None assert ( self.list.remove(lowercase_ ) is not None ) # node guaranteed to be in list assert node.key is not None del self.cache[first_node.key] self.num_keys -= 1 __lowercase = DoubleLinkedListNode(lowercase_ , lowercase_ ) self.list.add(self.cache[key] ) self.num_keys += 1 else: # bump node to the end of the list, update value __lowercase = self.list.remove(self.cache[key] ) assert node is not None # node guaranteed to be in list __lowercase = value self.list.add(lowercase_ ) @classmethod def a__ ( cls : Any , _UpperCAmelCase : List[Any] = 1_28 ) -> Callable[[Callable[[T], U]], Callable[..., U]]: """simple docstring""" def cache_decorator_inner(_UpperCAmelCase : List[str] ) -> Callable[..., U]: def cache_decorator_wrapper(*_UpperCAmelCase : Any ) -> U: if func not in cls.decorator_function_to_instance_map: __lowercase = LRUCache(lowercase_ ) __lowercase = cls.decorator_function_to_instance_map[func].get(args[0] ) if result is None: __lowercase = func(*lowercase_ ) cls.decorator_function_to_instance_map[func].put(args[0] , lowercase_ ) return result def cache_info() -> LRUCache[T, U]: return cls.decorator_function_to_instance_map[func] setattr(lowercase_ , 'cache_info' , lowercase_ ) # noqa: B010 return cache_decorator_wrapper return cache_decorator_inner if __name__ == "__main__": import doctest doctest.testmod()
325
from __future__ import annotations def UpperCAmelCase__ ( _A : float , _A : float , _A : float , ): '''simple docstring''' if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif electron_conc < 0: raise ValueError('''Electron concentration cannot be negative in a semiconductor''' ) elif hole_conc < 0: raise ValueError('''Hole concentration cannot be negative in a semiconductor''' ) elif intrinsic_conc < 0: raise ValueError( '''Intrinsic concentration cannot be negative in a semiconductor''' ) elif electron_conc == 0: return ( "electron_conc", intrinsic_conc**2 / hole_conc, ) elif hole_conc == 0: return ( "hole_conc", intrinsic_conc**2 / electron_conc, ) elif intrinsic_conc == 0: return ( "intrinsic_conc", (electron_conc * hole_conc) ** 0.5, ) else: return (-1, -1) if __name__ == "__main__": import doctest doctest.testmod()
188
0
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} UpperCAmelCase_ = { 'tokenizer_file': { 'EleutherAI/gpt-neox-20b': 'https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/tokenizer.json', }, } UpperCAmelCase_ = { 'gpt-neox-20b': 2_048, } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Tuple = VOCAB_FILES_NAMES UpperCAmelCase__ : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : Optional[Any] = ['input_ids', 'attention_mask'] def __init__( self: List[str] , UpperCamelCase_: int=None , UpperCamelCase_: Union[str, Any]=None , UpperCamelCase_: Dict=None , UpperCamelCase_: str="<|endoftext|>" , UpperCamelCase_: Any="<|endoftext|>" , UpperCamelCase_: Optional[Any]="<|endoftext|>" , UpperCamelCase_: int=False , **UpperCamelCase_: Optional[Any] , ): super().__init__( UpperCamelCase_ , UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , unk_token=UpperCamelCase_ , bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , add_prefix_space=UpperCamelCase_ , **UpperCamelCase_ , ) __lowerCamelCase = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , UpperCamelCase_ ) != add_prefix_space: __lowerCamelCase = getattr(UpperCamelCase_ , pre_tok_state.pop("""type""" ) ) __lowerCamelCase = add_prefix_space __lowerCamelCase = pre_tok_class(**UpperCamelCase_ ) __lowerCamelCase = add_prefix_space def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: str , UpperCamelCase_: Optional[str] = None ): __lowerCamelCase = self._tokenizer.model.save(UpperCamelCase_ , name=UpperCamelCase_ ) return tuple(UpperCamelCase_ ) def lowerCAmelCase__ ( self: str , UpperCamelCase_: "Conversation" ): __lowerCamelCase = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) + [self.eos_token_id] ) if len(UpperCamelCase_ ) > self.model_max_length: __lowerCamelCase = input_ids[-self.model_max_length :] return input_ids
29
# A Bipartite Graph is a graph whose vertices can be divided into two independent sets, # U and V such that every edge (u, v) either connects a vertex from U to V or a vertex # from V to U. In other words, for every edge (u, v), either u belongs to U and v to V, # or u belongs to V and v to U. We can also say that there is no edge that connects # vertices of same set. def lowerCamelCase__ ( A__ : Tuple ): '''simple docstring''' __lowerCamelCase = [False] * len(A__ ) __lowerCamelCase = [-1] * len(A__ ) def dfs(A__ : Optional[int] , A__ : Optional[int] ): __lowerCamelCase = True __lowerCamelCase = c for u in graph[v]: if not visited[u]: dfs(A__ , 1 - c ) for i in range(len(A__ ) ): if not visited[i]: dfs(A__ , 0 ) for i in range(len(A__ ) ): for j in graph[i]: if color[i] == color[j]: return False return True # Adjacency list of graph UpperCAmelCase_ = {0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 4: []} print(check_bipartite_dfs(graph))
29
1
"""simple docstring""" import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated __lowerCAmelCase : List[str] =collections.namedtuple("""_Datasets""", ["""train""", """validation""", """test"""]) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ __lowerCAmelCase : Union[str, Any] ="""https://storage.googleapis.com/cvdf-datasets/mnist/""" def UpperCAmelCase__ ( lowerCAmelCase__ :Union[str, Any] ) -> List[str]: '''simple docstring''' lowercase = numpy.dtype(numpy.uintaa ).newbyteorder(""">""" ) return numpy.frombuffer(bytestream.read(4 ) , dtype=lowerCAmelCase__ )[0] @deprecated(lowerCAmelCase__ , """Please use tf.data to implement this functionality.""" ) def UpperCAmelCase__ ( lowerCAmelCase__ :List[str] ) -> Dict: '''simple docstring''' print("""Extracting""" , f.name ) with gzip.GzipFile(fileobj=lowerCAmelCase__ ) as bytestream: lowercase = _readaa(lowerCAmelCase__ ) if magic != 2_0_5_1: raise ValueError( """Invalid magic number %d in MNIST image file: %s""" % (magic, f.name) ) lowercase = _readaa(lowerCAmelCase__ ) lowercase = _readaa(lowerCAmelCase__ ) lowercase = _readaa(lowerCAmelCase__ ) lowercase = bytestream.read(rows * cols * num_images ) lowercase = numpy.frombuffer(lowerCAmelCase__ , dtype=numpy.uinta ) lowercase = data.reshape(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , 1 ) return data @deprecated(lowerCAmelCase__ , """Please use tf.one_hot on tensors.""" ) def UpperCAmelCase__ ( lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Any ) -> Optional[int]: '''simple docstring''' lowercase = labels_dense.shape[0] lowercase = numpy.arange(lowerCAmelCase__ ) * num_classes lowercase = numpy.zeros((num_labels, num_classes) ) lowercase = 1 return labels_one_hot @deprecated(lowerCAmelCase__ , """Please use tf.data to implement this functionality.""" ) def UpperCAmelCase__ ( lowerCAmelCase__ :List[str] , lowerCAmelCase__ :Optional[int]=False , lowerCAmelCase__ :Dict=1_0 ) -> Tuple: '''simple docstring''' print("""Extracting""" , f.name ) with gzip.GzipFile(fileobj=lowerCAmelCase__ ) as bytestream: lowercase = _readaa(lowerCAmelCase__ ) if magic != 2_0_4_9: raise ValueError( """Invalid magic number %d in MNIST label file: %s""" % (magic, f.name) ) lowercase = _readaa(lowerCAmelCase__ ) lowercase = bytestream.read(lowerCAmelCase__ ) lowercase = numpy.frombuffer(lowerCAmelCase__ , dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(lowerCAmelCase__ , lowerCAmelCase__ ) return labels class _A : @deprecated( __lowerCAmelCase , """Please use alternatives such as official/mnist/_DataSet.py""" """ from tensorflow/models.""" , ) def __init__( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False , __lowerCAmelCase=False , __lowerCAmelCase=dtypes.floataa , __lowerCAmelCase=True , __lowerCAmelCase=None , ): """simple docstring""" lowercase , lowercase = random_seed.get_seed(__lowerCAmelCase ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) lowercase = dtypes.as_dtype(__lowerCAmelCase ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError("""Invalid image dtype %r, expected uint8 or float32""" % dtype ) if fake_data: lowercase = 1_0000 lowercase = one_hot else: assert ( images.shape[0] == labels.shape[0] ), f'images.shape: {images.shape} labels.shape: {labels.shape}' lowercase = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 lowercase = images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. lowercase = images.astype(numpy.floataa ) lowercase = numpy.multiply(__lowerCAmelCase , 1.0 / 2_5_5.0 ) lowercase = images lowercase = labels lowercase = 0 lowercase = 0 @property def A__ ( self ): """simple docstring""" return self._images @property def A__ ( self ): """simple docstring""" return self._labels @property def A__ ( self ): """simple docstring""" return self._num_examples @property def A__ ( self ): """simple docstring""" return self._epochs_completed def A__ ( self , __lowerCAmelCase , __lowerCAmelCase=False , __lowerCAmelCase=True ): """simple docstring""" if fake_data: lowercase = [1] * 784 lowercase = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(__lowerCAmelCase )], [fake_label for _ in range(__lowerCAmelCase )], ) lowercase = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: lowercase = numpy.arange(self._num_examples ) numpy.random.shuffle(__lowerCAmelCase ) lowercase = self.images[perma] lowercase = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch lowercase = self._num_examples - start lowercase = self._images[start : self._num_examples] lowercase = self._labels[start : self._num_examples] # Shuffle the data if shuffle: lowercase = numpy.arange(self._num_examples ) numpy.random.shuffle(__lowerCAmelCase ) lowercase = self.images[perm] lowercase = self.labels[perm] # Start next epoch lowercase = 0 lowercase = batch_size - rest_num_examples lowercase = self._index_in_epoch lowercase = self._images[start:end] lowercase = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size lowercase = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(lowerCAmelCase__ , """Please write your own downloading logic.""" ) def UpperCAmelCase__ ( lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :List[str] ) -> List[Any]: '''simple docstring''' if not gfile.Exists(lowerCAmelCase__ ): gfile.MakeDirs(lowerCAmelCase__ ) lowercase = os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) if not gfile.Exists(lowerCAmelCase__ ): urllib.request.urlretrieve(lowerCAmelCase__ , lowerCAmelCase__ ) # noqa: S310 with gfile.GFile(lowerCAmelCase__ ) as f: lowercase = f.size() print("""Successfully downloaded""" , lowerCAmelCase__ , lowerCAmelCase__ , """bytes.""" ) return filepath @deprecated( lowerCAmelCase__ , """Please use alternatives such as:""" """ tensorflow_datasets.load('mnist')""" ) def UpperCAmelCase__ ( lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :List[Any]=False , lowerCAmelCase__ :Optional[Any]=False , lowerCAmelCase__ :Union[str, Any]=dtypes.floataa , lowerCAmelCase__ :str=True , lowerCAmelCase__ :Optional[int]=5_0_0_0 , lowerCAmelCase__ :int=None , lowerCAmelCase__ :Optional[int]=DEFAULT_SOURCE_URL , ) -> str: '''simple docstring''' if fake_data: def fake(): return _DataSet( [] , [] , fake_data=lowerCAmelCase__ , one_hot=lowerCAmelCase__ , dtype=lowerCAmelCase__ , seed=lowerCAmelCase__ ) lowercase = fake() lowercase = fake() lowercase = fake() return _Datasets(train=lowerCAmelCase__ , validation=lowerCAmelCase__ , test=lowerCAmelCase__ ) if not source_url: # empty string check lowercase = DEFAULT_SOURCE_URL lowercase = """train-images-idx3-ubyte.gz""" lowercase = """train-labels-idx1-ubyte.gz""" lowercase = """t10k-images-idx3-ubyte.gz""" lowercase = """t10k-labels-idx1-ubyte.gz""" lowercase = _maybe_download( lowerCAmelCase__ , lowerCAmelCase__ , source_url + train_images_file ) with gfile.Open(lowerCAmelCase__ , """rb""" ) as f: lowercase = _extract_images(lowerCAmelCase__ ) lowercase = _maybe_download( lowerCAmelCase__ , lowerCAmelCase__ , source_url + train_labels_file ) with gfile.Open(lowerCAmelCase__ , """rb""" ) as f: lowercase = _extract_labels(lowerCAmelCase__ , one_hot=lowerCAmelCase__ ) lowercase = _maybe_download( lowerCAmelCase__ , lowerCAmelCase__ , source_url + test_images_file ) with gfile.Open(lowerCAmelCase__ , """rb""" ) as f: lowercase = _extract_images(lowerCAmelCase__ ) lowercase = _maybe_download( lowerCAmelCase__ , lowerCAmelCase__ , source_url + test_labels_file ) with gfile.Open(lowerCAmelCase__ , """rb""" ) as f: lowercase = _extract_labels(lowerCAmelCase__ , one_hot=lowerCAmelCase__ ) if not 0 <= validation_size <= len(lowerCAmelCase__ ): lowercase = ( """Validation size should be between 0 and """ f'{len(lowerCAmelCase__ )}. Received: {validation_size}.' ) raise ValueError(lowerCAmelCase__ ) lowercase = train_images[:validation_size] lowercase = train_labels[:validation_size] lowercase = train_images[validation_size:] lowercase = train_labels[validation_size:] lowercase = {"""dtype""": dtype, """reshape""": reshape, """seed""": seed} lowercase = _DataSet(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) lowercase = _DataSet(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) lowercase = _DataSet(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) return _Datasets(train=lowerCAmelCase__ , validation=lowerCAmelCase__ , test=lowerCAmelCase__ )
197
"""simple docstring""" import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class _A : def A__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowercase = TaEncoderModel.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) lowercase = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) lowercase = UNetaDConditionModel( sample_size=32 , layers_per_block=1 , block_out_channels=[32, 64] , down_block_types=[ """ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D""", ] , mid_block_type="""UNetMidBlock2DSimpleCrossAttn""" , up_block_types=["""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""] , in_channels=3 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="""text""" , addition_embed_type_num_heads=2 , cross_attention_norm="""group_norm""" , resnet_time_scale_shift="""scale_shift""" , act_fn="""gelu""" , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase = DDPMScheduler( num_train_timesteps=1000 , beta_schedule="""squaredcos_cap_v2""" , beta_start=0.0_0_0_1 , beta_end=0.0_2 , thresholding=__lowerCAmelCase , dynamic_thresholding_ratio=0.9_5 , sample_max_value=1.0 , prediction_type="""epsilon""" , variance_type="""learned_range""" , ) torch.manual_seed(0 ) lowercase = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def A__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowercase = TaEncoderModel.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) lowercase = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) lowercase = UNetaDConditionModel( sample_size=32 , layers_per_block=[1, 2] , block_out_channels=[32, 64] , down_block_types=[ """ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D""", ] , mid_block_type="""UNetMidBlock2DSimpleCrossAttn""" , up_block_types=["""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""] , in_channels=6 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="""text""" , addition_embed_type_num_heads=2 , cross_attention_norm="""group_norm""" , resnet_time_scale_shift="""scale_shift""" , act_fn="""gelu""" , class_embed_type="""timestep""" , mid_block_scale_factor=1.4_1_4 , time_embedding_act_fn="""gelu""" , time_embedding_dim=32 , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase = DDPMScheduler( num_train_timesteps=1000 , beta_schedule="""squaredcos_cap_v2""" , beta_start=0.0_0_0_1 , beta_end=0.0_2 , thresholding=__lowerCAmelCase , dynamic_thresholding_ratio=0.9_5 , sample_max_value=1.0 , prediction_type="""epsilon""" , variance_type="""learned_range""" , ) torch.manual_seed(0 ) lowercase = DDPMScheduler( num_train_timesteps=1000 , beta_schedule="""squaredcos_cap_v2""" , beta_start=0.0_0_0_1 , beta_end=0.0_2 , ) torch.manual_seed(0 ) lowercase = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def A__ ( self ): """simple docstring""" lowercase = self.get_dummy_components() lowercase = self.pipeline_class(**__lowerCAmelCase ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) lowercase = self.get_dummy_inputs(__lowerCAmelCase ) lowercase = inputs["""prompt"""] lowercase = inputs["""generator"""] lowercase = inputs["""num_inference_steps"""] lowercase = inputs["""output_type"""] if "image" in inputs: lowercase = inputs["""image"""] else: lowercase = None if "mask_image" in inputs: lowercase = inputs["""mask_image"""] else: lowercase = None if "original_image" in inputs: lowercase = inputs["""original_image"""] else: lowercase = None lowercase , lowercase = pipe.encode_prompt(__lowerCAmelCase ) # inputs with prompt converted to embeddings lowercase = { """prompt_embeds""": prompt_embeds, """negative_prompt_embeds""": negative_prompt_embeds, """generator""": generator, """num_inference_steps""": num_inference_steps, """output_type""": output_type, } if image is not None: lowercase = image if mask_image is not None: lowercase = mask_image if original_image is not None: lowercase = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) lowercase = pipe(**__lowerCAmelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__lowerCAmelCase ) lowercase = self.pipeline_class.from_pretrained(__lowerCAmelCase ) pipe_loaded.to(__lowerCAmelCase ) pipe_loaded.set_progress_bar_config(disable=__lowerCAmelCase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(__lowerCAmelCase , __lowerCAmelCase ) is None , f'`{optional_component}` did not stay set to None after loading.' , ) lowercase = self.get_dummy_inputs(__lowerCAmelCase ) lowercase = inputs["""generator"""] lowercase = inputs["""num_inference_steps"""] lowercase = inputs["""output_type"""] # inputs with prompt converted to embeddings lowercase = { """prompt_embeds""": prompt_embeds, """negative_prompt_embeds""": negative_prompt_embeds, """generator""": generator, """num_inference_steps""": num_inference_steps, """output_type""": output_type, } if image is not None: lowercase = image if mask_image is not None: lowercase = mask_image if original_image is not None: lowercase = original_image lowercase = pipe_loaded(**__lowerCAmelCase )[0] lowercase = np.abs(to_np(__lowerCAmelCase ) - to_np(__lowerCAmelCase ) ).max() self.assertLess(__lowerCAmelCase , 1E-4 ) def A__ ( self ): """simple docstring""" lowercase = self.get_dummy_components() lowercase = self.pipeline_class(**__lowerCAmelCase ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) lowercase = self.get_dummy_inputs(__lowerCAmelCase ) lowercase = pipe(**__lowerCAmelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__lowerCAmelCase ) lowercase = self.pipeline_class.from_pretrained(__lowerCAmelCase ) pipe_loaded.to(__lowerCAmelCase ) pipe_loaded.set_progress_bar_config(disable=__lowerCAmelCase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests lowercase = self.get_dummy_inputs(__lowerCAmelCase ) lowercase = pipe_loaded(**__lowerCAmelCase )[0] lowercase = np.abs(to_np(__lowerCAmelCase ) - to_np(__lowerCAmelCase ) ).max() self.assertLess(__lowerCAmelCase , 1E-4 )
197
1
import os import unittest from huggingface_hub.utils import are_progress_bars_disabled import transformers.models.bart.tokenization_bart from transformers import logging from transformers.testing_utils import CaptureLogger, mockenv, mockenv_context from transformers.utils.logging import disable_progress_bar, enable_progress_bar class __lowercase ( unittest.TestCase ): """simple docstring""" def __A ( self ) -> Optional[int]: '''simple docstring''' lowerCamelCase = logging.get_logger() # the current default level is logging.WARNING lowerCamelCase = logging.get_verbosity() logging.set_verbosity_error() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_warning() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_info() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_debug() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) # restore to the original level logging.set_verbosity(A ) def __A ( self ) -> int: '''simple docstring''' lowerCamelCase = logging.get_verbosity() lowerCamelCase = logging.get_logger("""transformers.models.bart.tokenization_bart""" ) lowerCamelCase = """Testing 1, 2, 3""" # should be able to log warnings (if default settings weren't overridden by `pytest --log-level-all`) if level_origin <= logging.WARNING: with CaptureLogger(A ) as cl: logger.warning(A ) self.assertEqual(cl.out , msg + """\n""" ) # this is setting the level for all of `transformers.*` loggers logging.set_verbosity_error() # should not be able to log warnings with CaptureLogger(A ) as cl: logger.warning(A ) self.assertEqual(cl.out , """""" ) # should be able to log warnings again logging.set_verbosity_warning() with CaptureLogger(A ) as cl: logger.warning(A ) self.assertEqual(cl.out , msg + """\n""" ) # restore to the original level logging.set_verbosity(A ) @mockenv(TRANSFORMERS_VERBOSITY="""error""" ) def __A ( self ) -> List[str]: '''simple docstring''' transformers.utils.logging._reset_library_root_logger() # this action activates the env var lowerCamelCase = logging.get_logger("""transformers.models.bart.tokenization_bart""" ) lowerCamelCase = os.getenv("""TRANSFORMERS_VERBOSITY""" , A ) lowerCamelCase = logging.log_levels[env_level_str] lowerCamelCase = logging.get_verbosity() self.assertEqual( A , A , F'TRANSFORMERS_VERBOSITY={env_level_str}/{env_level}, but internal verbosity is {current_level}' , ) # restore to the original level lowerCamelCase = """""" transformers.utils.logging._reset_library_root_logger() @mockenv(TRANSFORMERS_VERBOSITY="""super-error""" ) def __A ( self ) -> Dict: '''simple docstring''' transformers.utils.logging._reset_library_root_logger() lowerCamelCase = logging.logging.getLogger() with CaptureLogger(A ) as cl: # this action activates the env var logging.get_logger("""transformers.models.bart.tokenization_bart""" ) self.assertIn("""Unknown option TRANSFORMERS_VERBOSITY=super-error""" , cl.out ) # no need to restore as nothing was changed def __A ( self ) -> List[Any]: '''simple docstring''' transformers.utils.logging._reset_library_root_logger() lowerCamelCase = logging.get_logger("""transformers.models.bart.tokenization_bart""" ) lowerCamelCase = """Testing 1, 2, 3""" with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS="""1""" ): # nothing should be logged as env var disables this method with CaptureLogger(A ) as cl: logger.warning_advice(A ) self.assertEqual(cl.out , """""" ) with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS="""""" ): # should log normally as TRANSFORMERS_NO_ADVISORY_WARNINGS is unset with CaptureLogger(A ) as cl: logger.warning_advice(A ) self.assertEqual(cl.out , msg + """\n""" ) def __lowerCamelCase ( ): '''simple docstring''' disable_progress_bar() assert are_progress_bars_disabled() enable_progress_bar() assert not are_progress_bars_disabled()
66
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase : Union[str, Any] = { "configuration_mctct": ["MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP", "MCTCTConfig"], "feature_extraction_mctct": ["MCTCTFeatureExtractor"], "processing_mctct": ["MCTCTProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : List[Any] = [ "MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST", "MCTCTForCTC", "MCTCTModel", "MCTCTPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys UpperCAmelCase : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: __snake_case = None __snake_case = logging.get_logger(__name__) __snake_case = {"""vocab_file""": """sentencepiece.model""", """tokenizer_file""": """tokenizer.json"""} __snake_case = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, """tokenizer_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/tokenizer.json""", }, } __snake_case = { """google/rembert""": 256, } __snake_case = """▁""" class _lowerCAmelCase ( snake_case_ ): __UpperCAmelCase : List[str] = VOCAB_FILES_NAMES __UpperCAmelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : str = RemBertTokenizer def __init__( self , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=True , UpperCamelCase__=True , UpperCamelCase__=False , UpperCamelCase__="[CLS]" , UpperCamelCase__="[SEP]" , UpperCamelCase__="<unk>" , UpperCamelCase__="[SEP]" , UpperCamelCase__="<pad>" , UpperCamelCase__="[CLS]" , UpperCamelCase__="[MASK]" , **UpperCamelCase__ , ) -> Tuple: '''simple docstring''' snake_case : Dict = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token super().__init__( UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , do_lower_case=UpperCamelCase__ , remove_space=UpperCamelCase__ , keep_accents=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , **UpperCamelCase__ , ) snake_case : int = do_lower_case snake_case : Union[str, Any] = remove_space snake_case : Optional[int] = keep_accents snake_case : Any = vocab_file snake_case : Any = False if not self.vocab_file else True def lowerCamelCase ( self , UpperCamelCase__ , UpperCamelCase__ = None ) -> List[int]: '''simple docstring''' snake_case : Optional[Any] = [self.sep_token_id] snake_case : Optional[int] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def lowerCamelCase ( self , UpperCamelCase__ , UpperCamelCase__ = None , UpperCamelCase__ = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(UpperCamelCase__ )) + [1] + ([0] * len(UpperCamelCase__ )) + [1] return [1] + ([0] * len(UpperCamelCase__ )) + [1] def lowerCamelCase ( self , UpperCamelCase__ , UpperCamelCase__ = None ) -> List[int]: '''simple docstring''' snake_case : Union[str, Any] = [self.sep_token_id] snake_case : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase ( self , UpperCamelCase__ , UpperCamelCase__ = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(UpperCamelCase__ ): logger.error("Vocabulary path ({}) should be a directory".format(UpperCamelCase__ ) ) return snake_case : str = os.path.join( UpperCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ): copyfile(self.vocab_file , UpperCamelCase__ ) return (out_vocab_file,)
203
"""simple docstring""" import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch.nn import CrossEntropyLoss from torch.utils.data import DataLoader, TensorDataset from utils_ner import TokenClassificationTask __snake_case = logging.getLogger(__name__) class _lowerCAmelCase ( snake_case_ ): __UpperCAmelCase : Optional[int] = '''token-classification''' def __init__( self , UpperCamelCase__ ) -> List[Any]: '''simple docstring''' if type(UpperCamelCase__ ) == dict: snake_case : Optional[int] = Namespace(**UpperCamelCase__ ) snake_case : Optional[int] = import_module("tasks" ) try: snake_case : Optional[int] = getattr(UpperCamelCase__ , hparams.task_type ) snake_case : TokenClassificationTask = token_classification_task_clazz() except AttributeError: raise ValueError( F'Task {hparams.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. ' F'Available tasks classes are: {TokenClassificationTask.__subclasses__()}' ) snake_case : str = self.token_classification_task.get_labels(hparams.labels ) snake_case : Union[str, Any] = CrossEntropyLoss().ignore_index super().__init__(UpperCamelCase__ , len(self.labels ) , self.mode ) def lowerCamelCase ( self , **UpperCamelCase__ ) -> Any: '''simple docstring''' return self.model(**UpperCamelCase__ ) def lowerCamelCase ( self , UpperCamelCase__ , UpperCamelCase__ ) -> Tuple: '''simple docstring''' snake_case : Optional[Any] = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": snake_case : Optional[int] = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids snake_case : List[Any] = self(**UpperCamelCase__ ) snake_case : Union[str, Any] = outputs[0] # tensorboard_logs = {"loss": loss, "rate": self.lr_scheduler.get_last_lr()[-1]} return {"loss": loss} def lowerCamelCase ( self ) -> Optional[int]: '''simple docstring''' snake_case : Optional[Any] = self.hparams for mode in ["train", "dev", "test"]: snake_case : Optional[Any] = self._feature_file(UpperCamelCase__ ) if os.path.exists(UpperCamelCase__ ) and not args.overwrite_cache: logger.info("Loading features from cached file %s" , UpperCamelCase__ ) snake_case : List[str] = torch.load(UpperCamelCase__ ) else: logger.info("Creating features from dataset file at %s" , args.data_dir ) snake_case : Optional[Any] = self.token_classification_task.read_examples_from_file(args.data_dir , UpperCamelCase__ ) snake_case : Dict = self.token_classification_task.convert_examples_to_features( UpperCamelCase__ , self.labels , args.max_seq_length , self.tokenizer , cls_token_at_end=bool(self.config.model_type in ["xlnet"] ) , cls_token=self.tokenizer.cls_token , cls_token_segment_id=2 if self.config.model_type in ["xlnet"] else 0 , sep_token=self.tokenizer.sep_token , sep_token_extra=UpperCamelCase__ , pad_on_left=bool(self.config.model_type in ["xlnet"] ) , pad_token=self.tokenizer.pad_token_id , pad_token_segment_id=self.tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info("Saving features into cached file %s" , UpperCamelCase__ ) torch.save(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = False ) -> DataLoader: '''simple docstring''' snake_case : Optional[Any] = self._feature_file(UpperCamelCase__ ) logger.info("Loading features from cached file %s" , UpperCamelCase__ ) snake_case : Any = torch.load(UpperCamelCase__ ) snake_case : Tuple = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) snake_case : List[str] = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) if features[0].token_type_ids is not None: snake_case : Tuple = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) else: snake_case : Union[str, Any] = torch.tensor([0 for f in features] , dtype=torch.long ) # HACK(we will not use this anymore soon) snake_case : Optional[int] = torch.tensor([f.label_ids for f in features] , dtype=torch.long ) return DataLoader( TensorDataset(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) , batch_size=UpperCamelCase__ ) def lowerCamelCase ( self , UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: '''simple docstring''' """Compute validation""" "" snake_case : Any = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": snake_case : Optional[int] = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids snake_case : Optional[int] = self(**UpperCamelCase__ ) snake_case ,snake_case : str = outputs[:2] snake_case : Optional[int] = logits.detach().cpu().numpy() snake_case : str = inputs["labels"].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def lowerCamelCase ( self , UpperCamelCase__ ) -> Any: '''simple docstring''' snake_case : Dict = torch.stack([x["val_loss"] for x in outputs] ).mean() snake_case : List[str] = np.concatenate([x["pred"] for x in outputs] , axis=0 ) snake_case : Any = np.argmax(UpperCamelCase__ , axis=2 ) snake_case : Dict = np.concatenate([x["target"] for x in outputs] , axis=0 ) snake_case : Tuple = dict(enumerate(self.labels ) ) snake_case : str = [[] for _ in range(out_label_ids.shape[0] )] snake_case : List[Any] = [[] for _ in range(out_label_ids.shape[0] )] for i in range(out_label_ids.shape[0] ): for j in range(out_label_ids.shape[1] ): if out_label_ids[i, j] != self.pad_token_label_id: out_label_list[i].append(label_map[out_label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) snake_case : Union[str, Any] = { "val_loss": val_loss_mean, "accuracy_score": accuracy_score(UpperCamelCase__ , UpperCamelCase__ ), "precision": precision_score(UpperCamelCase__ , UpperCamelCase__ ), "recall": recall_score(UpperCamelCase__ , UpperCamelCase__ ), "f1": fa_score(UpperCamelCase__ , UpperCamelCase__ ), } snake_case : int = dict(results.items() ) snake_case : Union[str, Any] = results return ret, preds_list, out_label_list def lowerCamelCase ( self , UpperCamelCase__ ) -> List[str]: '''simple docstring''' snake_case ,snake_case ,snake_case : Optional[Any] = self._eval_end(UpperCamelCase__ ) snake_case : Tuple = ret["log"] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def lowerCamelCase ( self , UpperCamelCase__ ) -> Any: '''simple docstring''' snake_case ,snake_case ,snake_case : List[Any] = self._eval_end(UpperCamelCase__ ) # Converting to the dict required by pl # https://github.com/PyTorchLightning/pytorch-lightning/blob/master/\ # pytorch_lightning/trainer/logging.py#L139 snake_case : Optional[Any] = ret["log"] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ) -> str: '''simple docstring''' BaseTransformer.add_model_specific_args(UpperCamelCase__ , UpperCamelCase__ ) parser.add_argument( "--task_type" , default="NER" , type=UpperCamelCase__ , help="Task type to fine tune in training (e.g. NER, POS, etc)" ) parser.add_argument( "--max_seq_length" , default=128 , type=UpperCamelCase__ , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--labels" , default="" , type=UpperCamelCase__ , help="Path to a file containing all labels. If not specified, CoNLL-2003 labels are used." , ) parser.add_argument( "--gpus" , default=0 , type=UpperCamelCase__ , help="The number of GPUs allocated for this, it is by default 0 meaning none" , ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets" ) return parser if __name__ == "__main__": __snake_case = argparse.ArgumentParser() add_generic_args(parser, os.getcwd()) __snake_case = NERTransformer.add_model_specific_args(parser, os.getcwd()) __snake_case = parser.parse_args() __snake_case = NERTransformer(args) __snake_case = generic_train(model, args) if args.do_predict: # See https://github.com/huggingface/transformers/issues/3159 # pl use this default format to create a checkpoint: # https://github.com/PyTorchLightning/pytorch-lightning/blob/master\ # /pytorch_lightning/callbacks/model_checkpoint.py#L322 __snake_case = sorted(glob.glob(os.path.join(args.output_dir, """checkpoint-epoch=*.ckpt"""), recursive=True)) __snake_case = model.load_from_checkpoint(checkpoints[-1]) trainer.test(model)
203
1
"""simple docstring""" from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class UpperCAmelCase__ ( _a ): """simple docstring""" def __init__( self : Union[str, Any] , __lowerCamelCase : Callable , __lowerCamelCase : Optional[Features] = None , __lowerCamelCase : str = None , __lowerCamelCase : bool = False , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[dict] = None , __lowerCamelCase : Optional[int] = None , **__lowerCamelCase : str , ) -> Dict: super().__init__( features=__lowerCamelCase , cache_dir=__lowerCamelCase , keep_in_memory=__lowerCamelCase , streaming=__lowerCamelCase , num_proc=__lowerCamelCase , **__lowerCamelCase , ) SCREAMING_SNAKE_CASE__ = Generator( cache_dir=__lowerCamelCase , features=__lowerCamelCase , generator=__lowerCamelCase , gen_kwargs=__lowerCamelCase , **__lowerCamelCase , ) def lowercase_ ( self : List[str] ) -> int: # Build iterable dataset if self.streaming: SCREAMING_SNAKE_CASE__ = self.builder.as_streaming_dataset(split='''train''' ) # Build regular (map-style) dataset else: SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None self.builder.download_and_prepare( download_config=__lowerCamelCase , download_mode=__lowerCamelCase , verification_mode=__lowerCamelCase , base_path=__lowerCamelCase , num_proc=self.num_proc , ) SCREAMING_SNAKE_CASE__ = self.builder.as_dataset( split='''train''' , verification_mode=__lowerCamelCase , in_memory=self.keep_in_memory ) return dataset
368
def UpperCAmelCase_ ( _A ): '''simple docstring''' if divisor % 5 == 0 or divisor % 2 == 0: return 0 SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = 1 while repunit: SCREAMING_SNAKE_CASE__ = (10 * repunit + 1) % divisor repunit_index += 1 return repunit_index def UpperCAmelCase_ ( _A = 1_00_00_00 ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = limit - 1 if divisor % 2 == 0: divisor += 1 while least_divisible_repunit(_A ) <= limit: divisor += 2 return divisor if __name__ == "__main__": print(F"{solution() = }")
218
0
'''simple docstring''' import re import string import numpy as np import datasets _UpperCamelCase : List[str] = '\nReturns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list.\n' _UpperCamelCase : Union[str, Any] = '\nArgs:\n predictions: List of predicted texts.\n references: List of reference texts.\n regexes_to_ignore: List, defaults to None. Regex expressions of characters to\n ignore when calculating the exact matches. Note: these regexes are removed\n from the input data before the changes based on the options below (e.g. ignore_case,\n ignore_punctuation, ignore_numbers) are applied.\n ignore_case: Boolean, defaults to False. If true, turns everything\n to lowercase so that capitalization differences are ignored.\n ignore_punctuation: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\n ignore_numbers: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\nReturns:\n exact_match: Dictionary containing exact_match rate. Possible values are between 0.0 and 100.0, inclusive.\nExamples:\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results["exact_match"], 1))\n 25.0\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results["exact_match"], 1))\n 50.0\n\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell", "YELL"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results["exact_match"], 1))\n 75.0\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell", "YELL"], ignore_case=True, ignore_punctuation=True, ignore_numbers=True)\n >>> print(round(results["exact_match"], 1))\n 100.0\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["The cat sat on the mat.", "Theaters are great.", "It\'s like comparing oranges and apples."]\n >>> preds = ["The cat sat on the mat?", "Theaters are great.", "It\'s like comparing apples and oranges."]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results["exact_match"], 1))\n 33.3\n\n' _UpperCamelCase : Optional[int] = '\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class snake_case__ ( datasets.Metric): def A ( self : Tuple ) -> Dict: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , reference_urls=[] , ) def A ( self : int , _A : List[Any] , _A : Union[str, Any] , _A : List[str]=None , _A : List[str]=False , _A : int=False , _A : Optional[Any]=False , ) -> Union[str, Any]: if regexes_to_ignore is not None: for s in regexes_to_ignore: UpperCAmelCase_ : str = np.array([re.sub(_A , '''''' , _A ) for x in predictions] ) UpperCAmelCase_ : int = np.array([re.sub(_A , '''''' , _A ) for x in references] ) else: UpperCAmelCase_ : Union[str, Any] = np.asarray(_A ) UpperCAmelCase_ : Any = np.asarray(_A ) if ignore_case: UpperCAmelCase_ : Tuple = np.char.lower(_A ) UpperCAmelCase_ : Any = np.char.lower(_A ) if ignore_punctuation: UpperCAmelCase_ : Tuple = string.punctuation.maketrans('''''' , '''''' , string.punctuation ) UpperCAmelCase_ : int = np.char.translate(_A , table=_A ) UpperCAmelCase_ : List[str] = np.char.translate(_A , table=_A ) if ignore_numbers: UpperCAmelCase_ : List[str] = string.digits.maketrans('''''' , '''''' , string.digits ) UpperCAmelCase_ : Dict = np.char.translate(_A , table=_A ) UpperCAmelCase_ : List[str] = np.char.translate(_A , table=_A ) UpperCAmelCase_ : Union[str, Any] = predictions == references return {"exact_match": np.mean(_A ) * 1_00}
304
'''simple docstring''' from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class snake_case__ : a_ = 42 # [batch_size x 3] a_ = 42 # [batch_size x 3] a_ = 42 # [batch_size x 3] a_ = 42 # [batch_size x 3] a_ = 42 a_ = 42 a_ = 42 a_ = 42 a_ = 42 def A ( self : Tuple ) -> Optional[int]: assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def A ( self : List[Any] ) -> Union[str, Any]: return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def A ( self : Any ) -> Optional[Any]: return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def A ( self : Optional[int] ) -> torch.Tensor: UpperCAmelCase_ : Dict = torch.arange(self.height * self.width ) UpperCAmelCase_ : int = torch.stack( [ pixel_indices % self.width, torch.div(_A , self.width , rounding_mode='''trunc''' ), ] , axis=1 , ) return coords @property def A ( self : Optional[Any] ) -> Optional[Any]: UpperCAmelCase_ , *UpperCAmelCase_ : Union[str, Any] = self.shape UpperCAmelCase_ : Optional[Any] = int(np.prod(_A ) ) UpperCAmelCase_ : Any = self.get_image_coords() UpperCAmelCase_ : Any = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) UpperCAmelCase_ : Union[str, Any] = self.get_camera_rays(_A ) UpperCAmelCase_ : str = rays.view(_A , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def A ( self : Optional[int] , _A : torch.Tensor ) -> torch.Tensor: UpperCAmelCase_ , *UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] UpperCAmelCase_ : Dict = coords.view(_A , -1 , 2 ) UpperCAmelCase_ : Union[str, Any] = self.resolution() UpperCAmelCase_ : int = self.fov() UpperCAmelCase_ : Dict = (flat.float() / (res - 1)) * 2 - 1 UpperCAmelCase_ : Optional[int] = fracs * torch.tan(fov / 2 ) UpperCAmelCase_ : Any = fracs.view(_A , -1 , 2 ) UpperCAmelCase_ : List[Any] = ( self.z.view(_A , 1 , 3 ) + self.x.view(_A , 1 , 3 ) * fracs[:, :, :1] + self.y.view(_A , 1 , 3 ) * fracs[:, :, 1:] ) UpperCAmelCase_ : Optional[Any] = directions / directions.norm(dim=-1 , keepdim=_A ) UpperCAmelCase_ : Union[str, Any] = torch.stack( [ torch.broadcast_to(self.origin.view(_A , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(_A , *_A , 2 , 3 ) def A ( self : Tuple , _A : int , _A : int ) -> "DifferentiableProjectiveCamera": assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=_A , height=_A , x_fov=self.x_fov , y_fov=self.y_fov , ) def __UpperCAmelCase ( A : int ) -> DifferentiableProjectiveCamera: UpperCAmelCase_ : List[str] = [] UpperCAmelCase_ : Optional[int] = [] UpperCAmelCase_ : Optional[Any] = [] UpperCAmelCase_ : str = [] for theta in np.linspace(0 , 2 * np.pi , num=2_0 ): UpperCAmelCase_ : str = np.array([np.sin(A ), np.cos(A ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) UpperCAmelCase_ : Optional[int] = -z * 4 UpperCAmelCase_ : Optional[int] = np.array([np.cos(A ), -np.sin(A ), 0.0] ) UpperCAmelCase_ : List[Any] = np.cross(A , A ) origins.append(A ) xs.append(A ) ys.append(A ) zs.append(A ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(A , axis=0 ) ).float() , x=torch.from_numpy(np.stack(A , axis=0 ) ).float() , y=torch.from_numpy(np.stack(A , axis=0 ) ).float() , z=torch.from_numpy(np.stack(A , axis=0 ) ).float() , width=A , height=A , x_fov=0.7 , y_fov=0.7 , shape=(1, len(A )) , )
304
1
"""simple docstring""" import math A_ = 10 A_ = 7 A_ = BALLS_PER_COLOUR * NUM_COLOURS def UpperCAmelCase__ (snake_case__ : int = 20 ): """simple docstring""" _snake_case : List[Any] = math.comb(snake_case__ , snake_case__ ) _snake_case : Optional[Any] = math.comb(NUM_BALLS - BALLS_PER_COLOUR , snake_case__ ) _snake_case : Union[str, Any] = NUM_COLOURS * (1 - missing_colour / total) return F"{result:.9f}" if __name__ == "__main__": print(solution(20))
132
"""simple docstring""" import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging A_ = ( '''https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py''' ) A_ = logging.get_logger(__name__) # pylint: disable=invalid-name def UpperCAmelCase__ (): """simple docstring""" _snake_case : Tuple = """https://pypi.org/pypi/diffusers/json""" _snake_case : Optional[int] = json.loads(request.urlopen(snake_case__ ).read() )["""releases"""].keys() return sorted(snake_case__ , key=lambda snake_case__ : version.Version(snake_case__ ) ) def UpperCAmelCase__ (): """simple docstring""" if HF_MODULES_CACHE in sys.path: return sys.path.append(snake_case__ ) os.makedirs(snake_case__ , exist_ok=snake_case__ ) _snake_case : str = Path(snake_case__ ) / """__init__.py""" if not init_path.exists(): init_path.touch() def UpperCAmelCase__ (snake_case__ : Union[str, os.PathLike] ): """simple docstring""" init_hf_modules() _snake_case : List[Any] = Path(snake_case__ ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(snake_case__ , exist_ok=snake_case__ ) _snake_case : List[Any] = dynamic_module_path / """__init__.py""" if not init_path.exists(): init_path.touch() def UpperCAmelCase__ (snake_case__ : Optional[Any] ): """simple docstring""" with open(snake_case__ , """r""" , encoding="""utf-8""" ) as f: _snake_case : Union[str, Any] = f.read() # Imports of the form `import .xxx` _snake_case : Tuple = re.findall("""^\s*import\s+\.(\S+)\s*$""" , snake_case__ , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall("""^\s*from\s+\.(\S+)\s+import""" , snake_case__ , flags=re.MULTILINE ) # Unique-ify return list(set(snake_case__ ) ) def UpperCAmelCase__ (snake_case__ : Tuple ): """simple docstring""" _snake_case : Tuple = False _snake_case : Any = [module_file] _snake_case : str = [] # Let's recurse through all relative imports while not no_change: _snake_case : Dict = [] for f in files_to_check: new_imports.extend(get_relative_imports(snake_case__ ) ) _snake_case : Dict = Path(snake_case__ ).parent _snake_case : Dict = [str(module_path / m ) for m in new_imports] _snake_case : Tuple = [f for f in new_import_files if f not in all_relative_imports] _snake_case : str = [F"{f}.py" for f in new_import_files] _snake_case : Dict = len(snake_case__ ) == 0 all_relative_imports.extend(snake_case__ ) return all_relative_imports def UpperCAmelCase__ (snake_case__ : Tuple ): """simple docstring""" with open(snake_case__ , """r""" , encoding="""utf-8""" ) as f: _snake_case : int = f.read() # Imports of the form `import xxx` _snake_case : Tuple = re.findall("""^\s*import\s+(\S+)\s*$""" , snake_case__ , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall("""^\s*from\s+(\S+)\s+import""" , snake_case__ , flags=re.MULTILINE ) # Only keep the top-level module _snake_case : Tuple = [imp.split(""".""" )[0] for imp in imports if not imp.startswith(""".""" )] # Unique-ify and test we got them all _snake_case : Any = list(set(snake_case__ ) ) _snake_case : int = [] for imp in imports: try: importlib.import_module(snake_case__ ) except ImportError: missing_packages.append(snake_case__ ) if len(snake_case__ ) > 0: raise ImportError( """This modeling file requires the following packages that were not found in your environment: """ F"{', '.join(snake_case__ )}. Run `pip install {' '.join(snake_case__ )}`" ) return get_relative_imports(snake_case__ ) def UpperCAmelCase__ (snake_case__ : int , snake_case__ : Tuple ): """simple docstring""" _snake_case : List[Any] = module_path.replace(os.path.sep , """.""" ) _snake_case : int = importlib.import_module(snake_case__ ) if class_name is None: return find_pipeline_class(snake_case__ ) return getattr(snake_case__ , snake_case__ ) def UpperCAmelCase__ (snake_case__ : List[Any] ): """simple docstring""" from ..pipelines import DiffusionPipeline _snake_case : Tuple = dict(inspect.getmembers(snake_case__ , inspect.isclass ) ) _snake_case : Dict = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , snake_case__ ) and cls.__module__.split(""".""" )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( F"Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:" F" {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in" F" {loaded_module}." ) _snake_case : List[str] = cls return pipeline_class def UpperCAmelCase__ (snake_case__ : Union[str, os.PathLike] , snake_case__ : str , snake_case__ : Optional[Union[str, os.PathLike]] = None , snake_case__ : bool = False , snake_case__ : bool = False , snake_case__ : Optional[Dict[str, str]] = None , snake_case__ : Optional[Union[bool, str]] = None , snake_case__ : Optional[str] = None , snake_case__ : bool = False , ): """simple docstring""" _snake_case : List[str] = str(snake_case__ ) _snake_case : Optional[Any] = os.path.join(snake_case__ , snake_case__ ) if os.path.isfile(snake_case__ ): _snake_case : List[str] = module_file_or_url _snake_case : Optional[Any] = """local""" elif pretrained_model_name_or_path.count("""/""" ) == 0: _snake_case : Tuple = get_diffusers_versions() # cut ".dev0" _snake_case : Union[str, Any] = """v""" + """.""".join(__version__.split(""".""" )[:3] ) # retrieve github version that matches if revision is None: _snake_case : int = latest_version if latest_version[1:] in available_versions else """main""" logger.info(F"Defaulting to latest_version: {revision}." ) elif revision in available_versions: _snake_case : Optional[Any] = F"v{revision}" elif revision == "main": _snake_case : int = revision else: raise ValueError( F"`custom_revision`: {revision} does not exist. Please make sure to choose one of" F" {', '.join(available_versions + ['main'] )}." ) # community pipeline on GitHub _snake_case : List[str] = COMMUNITY_PIPELINES_URL.format(revision=snake_case__ , pipeline=snake_case__ ) try: _snake_case : Dict = cached_download( snake_case__ , cache_dir=snake_case__ , force_download=snake_case__ , proxies=snake_case__ , resume_download=snake_case__ , local_files_only=snake_case__ , use_auth_token=snake_case__ , ) _snake_case : Union[str, Any] = """git""" _snake_case : str = pretrained_model_name_or_path + """.py""" except EnvironmentError: logger.error(F"Could not locate the {module_file} inside {pretrained_model_name_or_path}." ) raise else: try: # Load from URL or cache if already cached _snake_case : str = hf_hub_download( snake_case__ , snake_case__ , cache_dir=snake_case__ , force_download=snake_case__ , proxies=snake_case__ , resume_download=snake_case__ , local_files_only=snake_case__ , use_auth_token=snake_case__ , ) _snake_case : Optional[Any] = os.path.join("""local""" , """--""".join(pretrained_model_name_or_path.split("""/""" ) ) ) except EnvironmentError: logger.error(F"Could not locate the {module_file} inside {pretrained_model_name_or_path}." ) raise # Check we have all the requirements in our environment _snake_case : int = check_imports(snake_case__ ) # Now we move the module inside our cached dynamic modules. _snake_case : Optional[int] = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(snake_case__ ) _snake_case : Any = Path(snake_case__ ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(snake_case__ , submodule_path / module_file ) for module_needed in modules_needed: _snake_case : Any = F"{module_needed}.py" shutil.copy(os.path.join(snake_case__ , snake_case__ ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(snake_case__ , snake_case__ ): _snake_case : Any = use_auth_token elif use_auth_token is True: _snake_case : int = HfFolder.get_token() else: _snake_case : Optional[int] = None _snake_case : int = model_info(snake_case__ , revision=snake_case__ , token=snake_case__ ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. _snake_case : int = submodule_path / commit_hash _snake_case : Optional[int] = full_submodule + os.path.sep + commit_hash create_dynamic_module(snake_case__ ) if not (submodule_path / module_file).exists(): shutil.copy(snake_case__ , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( snake_case__ , F"{module_needed}.py" , cache_dir=snake_case__ , force_download=snake_case__ , resume_download=snake_case__ , proxies=snake_case__ , use_auth_token=snake_case__ , revision=snake_case__ , local_files_only=snake_case__ , ) return os.path.join(snake_case__ , snake_case__ ) def UpperCAmelCase__ (snake_case__ : Union[str, os.PathLike] , snake_case__ : str , snake_case__ : Optional[str] = None , snake_case__ : Optional[Union[str, os.PathLike]] = None , snake_case__ : bool = False , snake_case__ : bool = False , snake_case__ : Optional[Dict[str, str]] = None , snake_case__ : Optional[Union[bool, str]] = None , snake_case__ : Optional[str] = None , snake_case__ : bool = False , **snake_case__ : Tuple , ): """simple docstring""" _snake_case : Union[str, Any] = get_cached_module_file( snake_case__ , snake_case__ , cache_dir=snake_case__ , force_download=snake_case__ , resume_download=snake_case__ , proxies=snake_case__ , use_auth_token=snake_case__ , revision=snake_case__ , local_files_only=snake_case__ , ) return get_class_in_module(snake_case__ , final_module.replace(""".py""" , """""" ) )
132
1
import argparse import json import os from collections import OrderedDict import torch from transformers import LukeConfig, LukeForMaskedLM, MLukeTokenizer, XLMRobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" with open(lowerCamelCase__ ) as metadata_file: lowercase__ : Any = json.load(lowerCamelCase__ ) lowercase__ : List[Any] = LukeConfig(use_entity_aware_attention=lowerCamelCase__ , **metadata["model_config"] ) # Load in the weights from the checkpoint_path lowercase__ : Optional[int] = torch.load(lowerCamelCase__ , map_location="cpu" )["module"] # Load the entity vocab file lowercase__ : Optional[int] = load_original_entity_vocab(lowerCamelCase__ ) # add an entry for [MASK2] lowercase__ : Optional[Any] = max(entity_vocab.values() ) + 1 config.entity_vocab_size += 1 lowercase__ : Union[str, Any] = XLMRobertaTokenizer.from_pretrained(metadata["model_config"]["bert_model_name"] ) # Add special tokens to the token vocabulary for downstream tasks lowercase__ : Optional[Any] = AddedToken("<ent>" , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) lowercase__ : Tuple = AddedToken("<ent2>" , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) tokenizer.add_special_tokens({"additional_special_tokens": [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(lowerCamelCase__ ) with open(os.path.join(lowerCamelCase__ , "tokenizer_config.json" ) , "r" ) as f: lowercase__ : List[str] = json.load(lowerCamelCase__ ) lowercase__ : Tuple = "MLukeTokenizer" with open(os.path.join(lowerCamelCase__ , "tokenizer_config.json" ) , "w" ) as f: json.dump(lowerCamelCase__ , lowerCamelCase__ ) with open(os.path.join(lowerCamelCase__ , MLukeTokenizer.vocab_files_names["entity_vocab_file"] ) , "w" ) as f: json.dump(lowerCamelCase__ , lowerCamelCase__ ) lowercase__ : Dict = MLukeTokenizer.from_pretrained(lowerCamelCase__ ) # Initialize the embeddings of the special tokens lowercase__ : Dict = tokenizer.convert_tokens_to_ids(["@"] )[0] lowercase__ : Any = tokenizer.convert_tokens_to_ids(["#"] )[0] lowercase__ : Optional[int] = state_dict["embeddings.word_embeddings.weight"] lowercase__ : Any = word_emb[ent_init_index].unsqueeze(0 ) lowercase__ : Optional[int] = word_emb[enta_init_index].unsqueeze(0 ) lowercase__ : str = torch.cat([word_emb, ent_emb, enta_emb] ) # add special tokens for 'entity_predictions.bias' for bias_name in ["lm_head.decoder.bias", "lm_head.bias"]: lowercase__ : Dict = state_dict[bias_name] lowercase__ : Optional[Any] = decoder_bias[ent_init_index].unsqueeze(0 ) lowercase__ : List[str] = decoder_bias[enta_init_index].unsqueeze(0 ) lowercase__ : int = torch.cat([decoder_bias, ent_decoder_bias, enta_decoder_bias] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: lowercase__ : Optional[Any] = F"""encoder.layer.{layer_index}.attention.self.""" lowercase__ : Optional[int] = state_dict[prefix + matrix_name] lowercase__ : Optional[int] = state_dict[prefix + matrix_name] lowercase__ : Optional[Any] = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks lowercase__ : Dict = state_dict["entity_embeddings.entity_embeddings.weight"] lowercase__ : str = entity_emb[entity_vocab["[MASK]"]].unsqueeze(0 ) lowercase__ : Tuple = torch.cat([entity_emb, entity_mask_emb] ) # add [MASK2] for 'entity_predictions.bias' lowercase__ : Dict = state_dict["entity_predictions.bias"] lowercase__ : int = entity_prediction_bias[entity_vocab["[MASK]"]].unsqueeze(0 ) lowercase__ : Tuple = torch.cat([entity_prediction_bias, entity_mask_bias] ) lowercase__ : List[Any] = LukeForMaskedLM(config=lowerCamelCase__ ).eval() state_dict.pop("entity_predictions.decoder.weight" ) state_dict.pop("lm_head.decoder.weight" ) state_dict.pop("lm_head.decoder.bias" ) lowercase__ : Optional[Any] = OrderedDict() for key, value in state_dict.items(): if not (key.startswith("lm_head" ) or key.startswith("entity_predictions" )): lowercase__ : Tuple = state_dict[key] else: lowercase__ : Tuple = state_dict[key] lowercase__ , lowercase__ : List[Any] = model.load_state_dict(lowerCamelCase__ , strict=lowerCamelCase__ ) if set(lowerCamelCase__ ) != {"luke.embeddings.position_ids"}: raise ValueError(F"""Unexpected unexpected_keys: {unexpected_keys}""" ) if set(lowerCamelCase__ ) != { "lm_head.decoder.weight", "lm_head.decoder.bias", "entity_predictions.decoder.weight", }: raise ValueError(F"""Unexpected missing_keys: {missing_keys}""" ) model.tie_weights() assert (model.luke.embeddings.word_embeddings.weight == model.lm_head.decoder.weight).all() assert (model.luke.entity_embeddings.entity_embeddings.weight == model.entity_predictions.decoder.weight).all() # Check outputs lowercase__ : Optional[int] = MLukeTokenizer.from_pretrained(lowerCamelCase__ , task="entity_classification" ) lowercase__ : List[Any] = "ISO 639-3 uses the code fas for the dialects spoken across Iran and アフガニスタン (Afghanistan)." lowercase__ : Tuple = (0, 9) lowercase__ : Dict = tokenizer(lowerCamelCase__ , entity_spans=[span] , return_tensors="pt" ) lowercase__ : Any = model(**lowerCamelCase__ ) # Verify word hidden states if model_size == "large": raise NotImplementedError else: # base lowercase__ : Union[str, Any] = torch.Size((1, 33, 768) ) lowercase__ : Union[str, Any] = torch.tensor([[0.0892, 0.0596, -0.2819], [0.0134, 0.1199, 0.0573], [-0.0169, 0.0927, 0.0644]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowerCamelCase__ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": raise NotImplementedError else: # base lowercase__ : int = torch.Size((1, 1, 768) ) lowercase__ : Tuple = torch.tensor([[-0.1482, 0.0609, 0.0322]] ) if not (outputs.entity_last_hidden_state.shape == expected_shape): raise ValueError( F"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" F""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , lowerCamelCase__ , atol=1e-4 ): raise ValueError # Verify masked word/entity prediction lowercase__ : str = MLukeTokenizer.from_pretrained(lowerCamelCase__ ) lowercase__ : str = "Tokyo is the capital of <mask>." lowercase__ : Tuple = (24, 30) lowercase__ : Tuple = tokenizer(lowerCamelCase__ , entity_spans=[span] , return_tensors="pt" ) lowercase__ : str = model(**lowerCamelCase__ ) lowercase__ : Union[str, Any] = encoding["input_ids"][0].tolist() lowercase__ : Tuple = input_ids.index(tokenizer.convert_tokens_to_ids("<mask>" ) ) lowercase__ : int = outputs.logits[0][mask_position_id].argmax(dim=-1 ) assert "Japan" == tokenizer.decode(lowerCamelCase__ ) lowercase__ : str = outputs.entity_logits[0][0].argmax().item() lowercase__ : Optional[Any] = [ entity for entity, entity_id in tokenizer.entity_vocab.items() if entity_id == predicted_entity_id ] assert [e for e in multilingual_predicted_entities if e.startswith("en:" )][0] == "en:Japan" # Finally, save our PyTorch model and tokenizer print("Saving PyTorch model to {}".format(lowerCamelCase__ ) ) model.save_pretrained(lowerCamelCase__ ) def __lowerCamelCase ( lowerCamelCase__ ): """simple docstring""" lowercase__ : Union[str, Any] = ["[MASK]", "[PAD]", "[UNK]"] lowercase__ : int = [json.loads(lowerCamelCase__ ) for line in open(lowerCamelCase__ )] lowercase__ : int = {} for entry in data: lowercase__ : Tuple = entry["id"] for entity_name, language in entry["entities"]: if entity_name in SPECIAL_TOKENS: lowercase__ : List[str] = entity_id break lowercase__ : List[str] = F"""{language}:{entity_name}""" lowercase__ : List[str] = entity_id return new_mapping if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) lowerCAmelCase__ = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
130
from unittest.mock import patch import pyspark from datasets.packaged_modules.spark.spark import ( Spark, SparkExamplesIterable, _generate_iterable_examples, ) from ..utils import ( require_dill_gt_0_3_2, require_not_windows, ) def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" lowercase__ : Optional[int] = [] for part_id in partition_order: lowercase__ : str = df.where(F"""SPARK_PARTITION_ID() = {part_id}""" ).collect() for row_idx, row in enumerate(lowerCamelCase__ ): expected_row_ids_and_row_dicts.append((F"""{part_id}_{row_idx}""", row.asDict()) ) return expected_row_ids_and_row_dicts @require_not_windows @require_dill_gt_0_3_2 def __lowerCamelCase ( ): """simple docstring""" lowercase__ : int = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() lowercase__ : Tuple = spark.range(100 ).repartition(1 ) lowercase__ : Tuple = Spark(lowerCamelCase__ ) # The id ints will be converted to Pyarrow int64s, so each row will be 8 bytes. Setting a max_shard_size of 16 means # that each partition can hold 2 rows. spark_builder._repartition_df_if_needed(max_shard_size=16 ) # Given that the dataframe has 100 rows and each partition has 2 rows, we expect 50 partitions. assert spark_builder.df.rdd.getNumPartitions() == 50 @require_not_windows @require_dill_gt_0_3_2 def __lowerCamelCase ( ): """simple docstring""" lowercase__ : Optional[Any] = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() lowercase__ : Tuple = spark.range(10 ).repartition(2 ) lowercase__ : Any = [1, 0] lowercase__ : Optional[int] = _generate_iterable_examples(lowerCamelCase__ , lowerCamelCase__ ) # Reverse the partitions. lowercase__ : str = _get_expected_row_ids_and_row_dicts_for_partition_order(lowerCamelCase__ , lowerCamelCase__ ) for i, (row_id, row_dict) in enumerate(generate_fn() ): lowercase__ , lowercase__ : List[Any] = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def __lowerCamelCase ( ): """simple docstring""" lowercase__ : Union[str, Any] = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() lowercase__ : int = spark.range(10 ).repartition(1 ) lowercase__ : Optional[int] = SparkExamplesIterable(lowerCamelCase__ ) assert it.n_shards == 1 for i, (row_id, row_dict) in enumerate(lowerCamelCase__ ): assert row_id == F"""0_{i}""" assert row_dict == {"id": i} @require_not_windows @require_dill_gt_0_3_2 def __lowerCamelCase ( ): """simple docstring""" lowercase__ : Any = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() lowercase__ : Optional[Any] = spark.range(30 ).repartition(3 ) # Mock the generator so that shuffle reverses the partition indices. with patch("numpy.random.Generator" ) as generator_mock: lowercase__ : int = lambda lowerCamelCase__ : x.reverse() lowercase__ : str = _get_expected_row_ids_and_row_dicts_for_partition_order(lowerCamelCase__ , [2, 1, 0] ) lowercase__ : int = SparkExamplesIterable(lowerCamelCase__ ).shuffle_data_sources(lowerCamelCase__ ) assert shuffled_it.n_shards == 3 for i, (row_id, row_dict) in enumerate(lowerCamelCase__ ): lowercase__ , lowercase__ : Tuple = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def __lowerCamelCase ( ): """simple docstring""" lowercase__ : Union[str, Any] = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() lowercase__ : Optional[Any] = spark.range(20 ).repartition(4 ) # Partitions 0 and 2 lowercase__ : Optional[Any] = SparkExamplesIterable(lowerCamelCase__ ).shard_data_sources(worker_id=0 , num_workers=2 ) assert shard_it_a.n_shards == 2 lowercase__ : List[str] = _get_expected_row_ids_and_row_dicts_for_partition_order(lowerCamelCase__ , [0, 2] ) for i, (row_id, row_dict) in enumerate(lowerCamelCase__ ): lowercase__ , lowercase__ : Optional[int] = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict # Partitions 1 and 3 lowercase__ : int = SparkExamplesIterable(lowerCamelCase__ ).shard_data_sources(worker_id=1 , num_workers=2 ) assert shard_it_a.n_shards == 2 lowercase__ : Optional[int] = _get_expected_row_ids_and_row_dicts_for_partition_order(lowerCamelCase__ , [1, 3] ) for i, (row_id, row_dict) in enumerate(lowerCamelCase__ ): lowercase__ , lowercase__ : Optional[int] = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def __lowerCamelCase ( ): """simple docstring""" lowercase__ : int = pyspark.sql.SparkSession.builder.master("local[*]" ).appName("pyspark" ).getOrCreate() lowercase__ : int = spark.range(100 ).repartition(1 ) lowercase__ : Tuple = Spark(lowerCamelCase__ ) # Choose a small max_shard_size for maximum partitioning. spark_builder._repartition_df_if_needed(max_shard_size=1 ) # The new number of partitions should not be greater than the number of rows. assert spark_builder.df.rdd.getNumPartitions() == 100
130
1
"""simple docstring""" import unittest from transformers import AutoTokenizer, NystromformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, NystromformerModel, ) from transformers.models.nystromformer.modeling_nystromformer import NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase : def __init__( self ,__UpperCamelCase ,__UpperCamelCase=13 ,__UpperCamelCase=7 ,__UpperCamelCase=True ,__UpperCamelCase=True ,__UpperCamelCase=True ,__UpperCamelCase=True ,__UpperCamelCase=99 ,__UpperCamelCase=32 ,__UpperCamelCase=5 ,__UpperCamelCase=4 ,__UpperCamelCase=37 ,__UpperCamelCase="gelu" ,__UpperCamelCase=0.1 ,__UpperCamelCase=0.1 ,__UpperCamelCase=512 ,__UpperCamelCase=16 ,__UpperCamelCase=2 ,__UpperCamelCase=0.02 ,__UpperCamelCase=3 ,__UpperCamelCase=4 ,__UpperCamelCase=None ,) -> Tuple: '''simple docstring''' lowercase_ : Optional[Any] = parent lowercase_ : List[str] = batch_size lowercase_ : Dict = seq_length lowercase_ : int = is_training lowercase_ : Union[str, Any] = use_input_mask lowercase_ : Tuple = use_token_type_ids lowercase_ : int = use_labels lowercase_ : Optional[int] = vocab_size lowercase_ : str = hidden_size lowercase_ : Optional[Any] = num_hidden_layers lowercase_ : List[Any] = num_attention_heads lowercase_ : Optional[int] = intermediate_size lowercase_ : Optional[int] = hidden_act lowercase_ : List[str] = hidden_dropout_prob lowercase_ : str = attention_probs_dropout_prob lowercase_ : Any = max_position_embeddings lowercase_ : List[str] = type_vocab_size lowercase_ : int = type_sequence_label_size lowercase_ : List[Any] = initializer_range lowercase_ : Any = num_labels lowercase_ : Union[str, Any] = num_choices lowercase_ : Tuple = scope def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ : int = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) lowercase_ : Dict = None if self.use_input_mask: lowercase_ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowercase_ : Dict = None if self.use_token_type_ids: lowercase_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) lowercase_ : str = None lowercase_ : Union[str, Any] = None lowercase_ : List[Any] = None if self.use_labels: lowercase_ : Any = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) lowercase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) lowercase_ : Optional[int] = ids_tensor([self.batch_size] ,self.num_choices ) lowercase_ : Dict = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCAmelCase ( self ) -> int: '''simple docstring''' return NystromformerConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=lowercase_ ,initializer_range=self.initializer_range ,) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Dict: '''simple docstring''' lowercase_ : Optional[Any] = NystromformerModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() lowercase_ : List[str] = model(lowercase_ ,attention_mask=lowercase_ ,token_type_ids=lowercase_ ) lowercase_ : str = model(lowercase_ ,token_type_ids=lowercase_ ) lowercase_ : Optional[int] = model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Optional[Any]: '''simple docstring''' lowercase_ : List[Any] = NystromformerForMaskedLM(config=lowercase_ ) model.to(lowercase_ ) model.eval() lowercase_ : List[Any] = model(lowercase_ ,attention_mask=lowercase_ ,token_type_ids=lowercase_ ,labels=lowercase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> List[str]: '''simple docstring''' lowercase_ : Optional[int] = NystromformerForQuestionAnswering(config=lowercase_ ) model.to(lowercase_ ) model.eval() lowercase_ : Optional[int] = model( lowercase_ ,attention_mask=lowercase_ ,token_type_ids=lowercase_ ,start_positions=lowercase_ ,end_positions=lowercase_ ,) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> List[str]: '''simple docstring''' lowercase_ : Optional[Any] = self.num_labels lowercase_ : Optional[int] = NystromformerForSequenceClassification(lowercase_ ) model.to(lowercase_ ) model.eval() lowercase_ : Optional[int] = model(lowercase_ ,attention_mask=lowercase_ ,token_type_ids=lowercase_ ,labels=lowercase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Any: '''simple docstring''' lowercase_ : Union[str, Any] = self.num_labels lowercase_ : Tuple = NystromformerForTokenClassification(config=lowercase_ ) model.to(lowercase_ ) model.eval() lowercase_ : Optional[int] = model(lowercase_ ,attention_mask=lowercase_ ,token_type_ids=lowercase_ ,labels=lowercase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> List[str]: '''simple docstring''' lowercase_ : List[str] = self.num_choices lowercase_ : Dict = NystromformerForMultipleChoice(config=lowercase_ ) model.to(lowercase_ ) model.eval() lowercase_ : Optional[Any] = input_ids.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() lowercase_ : Union[str, Any] = token_type_ids.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() lowercase_ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() lowercase_ : Optional[int] = model( lowercase_ ,attention_mask=lowercase_ ,token_type_ids=lowercase_ ,labels=lowercase_ ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def _UpperCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' lowercase_ : Tuple = self.prepare_config_and_inputs() ( ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ( lowercase_ ) , ) : Tuple = config_and_inputs lowercase_ : List[str] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class UpperCamelCase ( _lowerCamelCase , _lowerCamelCase , unittest.TestCase ): lowercase = ( ( NystromformerModel, NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, ) if is_torch_available() else () ) lowercase = ( { 'feature-extraction': NystromformerModel, 'fill-mask': NystromformerForMaskedLM, 'question-answering': NystromformerForQuestionAnswering, 'text-classification': NystromformerForSequenceClassification, 'token-classification': NystromformerForTokenClassification, 'zero-shot': NystromformerForSequenceClassification, } if is_torch_available() else {} ) lowercase = False lowercase = False def _UpperCAmelCase ( self ) -> str: '''simple docstring''' lowercase_ : List[str] = NystromformerModelTester(self ) lowercase_ : List[Any] = ConfigTester(self ,config_class=lowercase_ ,hidden_size=37 ) def _UpperCAmelCase ( self ) -> List[str]: '''simple docstring''' self.config_tester.run_common_tests() def _UpperCAmelCase ( self ) -> int: '''simple docstring''' lowercase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def _UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' lowercase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowercase_ : Tuple = type self.model_tester.create_and_check_model(*lowercase_ ) def _UpperCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' lowercase_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase_ ) def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' lowercase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase_ ) def _UpperCAmelCase ( self ) -> Dict: '''simple docstring''' lowercase_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase_ ) def _UpperCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' lowercase_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase_ ) def _UpperCAmelCase ( self ) -> Dict: '''simple docstring''' lowercase_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase_ ) @slow def _UpperCAmelCase ( self ) -> List[Any]: '''simple docstring''' for model_name in NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ : List[str] = NystromformerModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) @require_torch class UpperCamelCase ( unittest.TestCase ): @slow def _UpperCAmelCase ( self ) -> Optional[int]: '''simple docstring''' lowercase_ : int = NystromformerModel.from_pretrained('uw-madison/nystromformer-512' ) lowercase_ : Union[str, Any] = torch.tensor([[0, 1, 2, 3, 4, 5]] ) with torch.no_grad(): lowercase_ : int = model(lowercase_ )[0] lowercase_ : str = torch.Size((1, 6, 768) ) self.assertEqual(output.shape ,lowercase_ ) lowercase_ : int = torch.tensor( [[[-0.4532, -0.0936, 0.5137], [-0.2676, 0.0628, 0.6186], [-0.3629, -0.1726, 0.4716]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,lowercase_ ,atol=1e-4 ) ) @slow def _UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' lowercase_ : Any = 'the [MASK] of Belgium is Brussels' lowercase_ : Dict = AutoTokenizer.from_pretrained('uw-madison/nystromformer-512' ) lowercase_ : str = NystromformerForMaskedLM.from_pretrained('uw-madison/nystromformer-512' ) lowercase_ : Tuple = tokenizer(lowercase_ ,return_tensors='pt' ) with torch.no_grad(): lowercase_ : Tuple = model(encoding.input_ids ).logits lowercase_ : List[Any] = token_logits[:, 2, :].argmax(-1 )[0] self.assertEqual(tokenizer.decode(lowercase_ ) ,'capital' )
356
"""simple docstring""" __SCREAMING_SNAKE_CASE ={ "a": "AAAAA", "b": "AAAAB", "c": "AAABA", "d": "AAABB", "e": "AABAA", "f": "AABAB", "g": "AABBA", "h": "AABBB", "i": "ABAAA", "j": "BBBAA", "k": "ABAAB", "l": "ABABA", "m": "ABABB", "n": "ABBAA", "o": "ABBAB", "p": "ABBBA", "q": "ABBBB", "r": "BAAAA", "s": "BAAAB", "t": "BAABA", "u": "BAABB", "v": "BBBAB", "w": "BABAA", "x": "BABAB", "y": "BABBA", "z": "BABBB", " ": " ", } __SCREAMING_SNAKE_CASE ={value: key for key, value in encode_dict.items()} def lowercase__( __SCREAMING_SNAKE_CASE : str ): lowercase_ : Union[str, Any] = '' for letter in word.lower(): if letter.isalpha() or letter == " ": encoded += encode_dict[letter] else: raise Exception('encode() accepts only letters of the alphabet and spaces' ) return encoded def lowercase__( __SCREAMING_SNAKE_CASE : str ): if set(__SCREAMING_SNAKE_CASE ) - {"A", "B", " "} != set(): raise Exception('decode() accepts only \'A\', \'B\' and spaces' ) lowercase_ : Dict = '' for word in coded.split(): while len(__SCREAMING_SNAKE_CASE ) != 0: decoded += decode_dict[word[:5]] lowercase_ : Any = word[5:] decoded += " " return decoded.strip() if __name__ == "__main__": from doctest import testmod testmod()
321
0
'''simple docstring''' def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' return credit_card_number.startswith(('''34''', '''35''', '''37''', '''4''', '''5''', '''6''') ) def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' A : Union[str, Any] = credit_card_number A : Optional[Any] = 0 A : Tuple = len(snake_case__ ) - 2 for i in range(snake_case__ , -1 , -2 ): # double the value of every second digit A : Dict = int(cc_number[i] ) digit *= 2 # If doubling of a number results in a two digit number # i.e greater than 9(e.g., 6 × 2 = 12), # then add the digits of the product (e.g., 12: 1 + 2 = 3, 15: 1 + 5 = 6), # to get a single digit number. if digit > 9: digit %= 10 digit += 1 A : Union[str, Any] = cc_number[:i] + str(snake_case__ ) + cc_number[i + 1 :] total += digit # Sum up the remaining digits for i in range(len(snake_case__ ) - 1 , -1 , -2 ): total += int(cc_number[i] ) return total % 10 == 0 def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' A : Optional[int] = F'{credit_card_number} is an invalid credit card number because' if not credit_card_number.isdigit(): print(F'{error_message} it has nonnumerical characters.' ) return False if not 13 <= len(snake_case__ ) <= 16: print(F'{error_message} of its length.' ) return False if not validate_initial_digits(snake_case__ ): print(F'{error_message} of its first two digits.' ) return False if not luhn_validation(snake_case__ ): print(F'{error_message} it fails the Luhn check.' ) return False print(F'{credit_card_number} is a valid credit card number.' ) return True if __name__ == "__main__": import doctest doctest.testmod() validate_credit_card_number('4111111111111111') validate_credit_card_number('32323')
3
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy as np import tensorflow as tf from transformers import ( TF_FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST, FlaubertConfig, TFFlaubertForMultipleChoice, TFFlaubertForQuestionAnsweringSimple, TFFlaubertForSequenceClassification, TFFlaubertForTokenClassification, TFFlaubertModel, TFFlaubertWithLMHeadModel, ) class A: '''simple docstring''' def __init__( self : str , A_ : Optional[Any] , ) -> str: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = 13 lowerCamelCase_ = 7 lowerCamelCase_ = True lowerCamelCase_ = True lowerCamelCase_ = True lowerCamelCase_ = True lowerCamelCase_ = True lowerCamelCase_ = False lowerCamelCase_ = False lowerCamelCase_ = False lowerCamelCase_ = 2 lowerCamelCase_ = 99 lowerCamelCase_ = 0 lowerCamelCase_ = 32 lowerCamelCase_ = 2 lowerCamelCase_ = 4 lowerCamelCase_ = 0.1 lowerCamelCase_ = 0.1 lowerCamelCase_ = 512 lowerCamelCase_ = 16 lowerCamelCase_ = 2 lowerCamelCase_ = 0.02 lowerCamelCase_ = 3 lowerCamelCase_ = 4 lowerCamelCase_ = 'last' lowerCamelCase_ = True lowerCamelCase_ = None lowerCamelCase_ = 0 def a__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase_ = random_attention_mask([self.batch_size, self.seq_length] , dtype=tf.floataa ) lowerCamelCase_ = None if self.use_input_lengths: lowerCamelCase_ = ( ids_tensor([self.batch_size] , vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length lowerCamelCase_ = None if self.use_token_type_ids: lowerCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.n_langs ) lowerCamelCase_ = None lowerCamelCase_ = None lowerCamelCase_ = None if self.use_labels: lowerCamelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase_ = ids_tensor([self.batch_size] , 2 , dtype=tf.floataa ) lowerCamelCase_ = ids_tensor([self.batch_size] , self.num_choices ) lowerCamelCase_ = FlaubertConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , bos_token_id=self.bos_token_id , ) return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def a__ ( self : int , A_ : List[str] , A_ : List[Any] , A_ : str , A_ : List[Any] , A_ : int , A_ : Tuple , A_ : Optional[int] , A_ : Optional[int] , A_ : str , ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = TFFlaubertModel(config=A_ ) lowerCamelCase_ = {'input_ids': input_ids, 'lengths': input_lengths, 'langs': token_type_ids} lowerCamelCase_ = model(A_ ) lowerCamelCase_ = [input_ids, input_mask] lowerCamelCase_ = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self : Tuple , A_ : List[str] , A_ : int , A_ : List[Any] , A_ : Any , A_ : Any , A_ : Dict , A_ : str , A_ : List[Any] , A_ : Union[str, Any] , ) -> List[str]: """simple docstring""" lowerCamelCase_ = TFFlaubertWithLMHeadModel(A_ ) lowerCamelCase_ = {'input_ids': input_ids, 'lengths': input_lengths, 'langs': token_type_ids} lowerCamelCase_ = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a__ ( self : str , A_ : Tuple , A_ : Any , A_ : Any , A_ : List[Any] , A_ : Dict , A_ : List[Any] , A_ : Union[str, Any] , A_ : Optional[int] , A_ : List[Any] , ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = TFFlaubertForQuestionAnsweringSimple(A_ ) lowerCamelCase_ = {'input_ids': input_ids, 'lengths': input_lengths} lowerCamelCase_ = model(A_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def a__ ( self : Optional[int] , A_ : List[Any] , A_ : str , A_ : List[str] , A_ : Dict , A_ : Optional[Any] , A_ : Tuple , A_ : str , A_ : Optional[int] , A_ : Tuple , ) -> List[Any]: """simple docstring""" lowerCamelCase_ = TFFlaubertForSequenceClassification(A_ ) lowerCamelCase_ = {'input_ids': input_ids, 'lengths': input_lengths} lowerCamelCase_ = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a__ ( self : Dict , A_ : Optional[Any] , A_ : List[Any] , A_ : int , A_ : Any , A_ : Union[str, Any] , A_ : str , A_ : Any , A_ : Union[str, Any] , A_ : List[str] , ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.num_labels lowerCamelCase_ = TFFlaubertForTokenClassification(config=A_ ) lowerCamelCase_ = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} lowerCamelCase_ = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def a__ ( self : List[Any] , A_ : Optional[int] , A_ : List[Any] , A_ : Optional[int] , A_ : Tuple , A_ : Union[str, Any] , A_ : int , A_ : str , A_ : Tuple , A_ : str , ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.num_choices lowerCamelCase_ = TFFlaubertForMultipleChoice(config=A_ ) lowerCamelCase_ = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) lowerCamelCase_ = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) lowerCamelCase_ = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) lowerCamelCase_ = { 'input_ids': multiple_choice_inputs_ids, 'attention_mask': multiple_choice_input_mask, 'token_type_ids': multiple_choice_token_type_ids, } lowerCamelCase_ = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def a__ ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.prepare_config_and_inputs() ( ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ) = config_and_inputs lowerCamelCase_ = { 'input_ids': input_ids, 'token_type_ids': token_type_ids, 'langs': token_type_ids, 'lengths': input_lengths, } return config, inputs_dict @require_tf class A( UpperCamelCase , UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = ( ( TFFlaubertModel, TFFlaubertWithLMHeadModel, TFFlaubertForSequenceClassification, TFFlaubertForQuestionAnsweringSimple, TFFlaubertForTokenClassification, TFFlaubertForMultipleChoice, ) if is_tf_available() else () ) UpperCamelCase = ( (TFFlaubertWithLMHeadModel,) if is_tf_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable UpperCamelCase = ( { '''feature-extraction''': TFFlaubertModel, '''fill-mask''': TFFlaubertWithLMHeadModel, '''question-answering''': TFFlaubertForQuestionAnsweringSimple, '''text-classification''': TFFlaubertForSequenceClassification, '''token-classification''': TFFlaubertForTokenClassification, '''zero-shot''': TFFlaubertForSequenceClassification, } if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False def a__ ( self : Union[str, Any] , A_ : Any , A_ : List[Any] , A_ : Union[str, Any] , A_ : str , A_ : List[str] ) -> Optional[Any]: """simple docstring""" if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith('Fast' ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def a__ ( self : List[str] ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = TFFlaubertModelTester(self ) lowerCamelCase_ = ConfigTester(self , config_class=A_ , emb_dim=37 ) def a__ ( self : List[str] ) -> str: """simple docstring""" self.config_tester.run_common_tests() def a__ ( self : int ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*A_ ) def a__ ( self : List[str] ) -> Any: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*A_ ) def a__ ( self : Dict ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*A_ ) def a__ ( self : Union[str, Any] ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*A_ ) def a__ ( self : List[Any] ) -> Dict: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_for_token_classification(*A_ ) def a__ ( self : int ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_for_multiple_choice(*A_ ) @slow def a__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" for model_name in TF_FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase_ = TFFlaubertModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) @require_tf @require_sentencepiece @require_tokenizers class A( unittest.TestCase ): '''simple docstring''' @slow def a__ ( self : Dict ) -> str: """simple docstring""" lowerCamelCase_ = TFFlaubertModel.from_pretrained('jplu/tf-flaubert-small-cased' ) lowerCamelCase_ = tf.convert_to_tensor( [[0, 158, 735, 2592, 1424, 6727, 82, 1]] , dtype=tf.intaa , ) # "J'aime flaubert !" lowerCamelCase_ = model(A_ )[0] lowerCamelCase_ = tf.TensorShape((1, 8, 512) ) self.assertEqual(output.shape , A_ ) # compare the actual values for a slice. lowerCamelCase_ = tf.convert_to_tensor( [ [ [-1.8768773, -1.566555, 0.27072418], [-1.6920038, -0.5873505, 1.9329599], [-2.9563985, -1.6993835, 1.7972052], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
204
0
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING a : Union[str, Any] = logging.get_logger(__name__) class a ( _lowerCamelCase ): snake_case_ = "upernet" def __init__( self : Optional[int] , lowercase_ : int=None , lowercase_ : Tuple=512 , lowercase_ : Dict=0.02 , lowercase_ : Optional[int]=[1, 2, 3, 6] , lowercase_ : str=True , lowercase_ : Tuple=0.4 , lowercase_ : List[str]=384 , lowercase_ : Optional[int]=256 , lowercase_ : Optional[Any]=1 , lowercase_ : Any=False , lowercase_ : Dict=255 , **lowercase_ : List[Any] , ): super().__init__(**lowercase_ ) if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) snake_case_ = CONFIG_MAPPING['''resnet'''](out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] ) elif isinstance(lowercase_ , lowercase_ ): snake_case_ = backbone_config.get('''model_type''' ) snake_case_ = CONFIG_MAPPING[backbone_model_type] snake_case_ = config_class.from_dict(lowercase_ ) snake_case_ = backbone_config snake_case_ = hidden_size snake_case_ = initializer_range snake_case_ = pool_scales snake_case_ = use_auxiliary_head snake_case_ = auxiliary_loss_weight snake_case_ = auxiliary_in_channels snake_case_ = auxiliary_channels snake_case_ = auxiliary_num_convs snake_case_ = auxiliary_concat_input snake_case_ = loss_ignore_index def A_ ( self : int ): snake_case_ = copy.deepcopy(self.__dict__ ) snake_case_ = self.backbone_config.to_dict() snake_case_ = self.__class__.model_type return output
72
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a : int = { 'configuration_xlm': ['XLM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMConfig', 'XLMOnnxConfig'], 'tokenization_xlm': ['XLMTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = [ 'XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMForMultipleChoice', 'XLMForQuestionAnswering', 'XLMForQuestionAnsweringSimple', 'XLMForSequenceClassification', 'XLMForTokenClassification', 'XLMModel', 'XLMPreTrainedModel', 'XLMWithLMHeadModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[int] = [ 'TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLMForMultipleChoice', 'TFXLMForQuestionAnsweringSimple', 'TFXLMForSequenceClassification', 'TFXLMForTokenClassification', 'TFXLMMainLayer', 'TFXLMModel', 'TFXLMPreTrainedModel', 'TFXLMWithLMHeadModel', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys a : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
72
1
'''simple docstring''' from math import factorial def a__ ( a__ = 1_00 ): """simple docstring""" return sum(map(a__ , str(factorial(a__ ) ) ) ) if __name__ == "__main__": print(solution(int(input('Enter the Number: ').strip())))
267
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase : Union[str, Any] = {'configuration_sew': ['SEW_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SEWConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Union[str, Any] = [ 'SEW_PRETRAINED_MODEL_ARCHIVE_LIST', 'SEWForCTC', 'SEWForSequenceClassification', 'SEWModel', 'SEWPreTrainedModel', ] if TYPE_CHECKING: from .configuration_sew import SEW_PRETRAINED_CONFIG_ARCHIVE_MAP, SEWConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_sew import ( SEW_PRETRAINED_MODEL_ARCHIVE_LIST, SEWForCTC, SEWForSequenceClassification, SEWModel, SEWPreTrainedModel, ) else: import sys UpperCAmelCase : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
267
1
from .imports import is_rich_available if is_rich_available(): from rich.traceback import install install(show_locals=False) else: raise ModuleNotFoundError("""To use the rich extension, install rich with `pip install rich`""")
113
from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class A_ ( unittest.TestCase ): @slow def lowerCAmelCase ( self : List[str]): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: __lowerCamelCase : Optional[int] = AutoConfig.from_pretrained(SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Tuple = TFAutoModel.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_pt=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : str = AutoModel.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_tf=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) @slow def lowerCAmelCase ( self : Optional[int]): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: __lowerCamelCase : str = AutoConfig.from_pretrained(SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Tuple = TFAutoModelForPreTraining.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_pt=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : str = AutoModelForPreTraining.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_tf=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) @slow def lowerCAmelCase ( self : Optional[int]): for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Union[str, Any] = TFAutoModelForCausalLM.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_pt=SCREAMING_SNAKE_CASE__) __lowerCamelCase , __lowerCamelCase : Union[str, Any] = TFAutoModelForCausalLM.from_pretrained( SCREAMING_SNAKE_CASE__ ,output_loading_info=SCREAMING_SNAKE_CASE__ ,from_pt=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : List[Any] = AutoModelForCausalLM.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_tf=SCREAMING_SNAKE_CASE__) __lowerCamelCase , __lowerCamelCase : Any = AutoModelForCausalLM.from_pretrained( SCREAMING_SNAKE_CASE__ ,output_loading_info=SCREAMING_SNAKE_CASE__ ,from_tf=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) @slow def lowerCAmelCase ( self : List[Any]): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase : List[str] = AutoConfig.from_pretrained(SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : List[str] = TFAutoModelWithLMHead.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_pt=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Optional[Any] = AutoModelWithLMHead.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_tf=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) @slow def lowerCAmelCase ( self : Tuple): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase : Optional[Any] = AutoConfig.from_pretrained(SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Optional[int] = TFAutoModelForMaskedLM.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_pt=SCREAMING_SNAKE_CASE__) __lowerCamelCase , __lowerCamelCase : Dict = TFAutoModelForMaskedLM.from_pretrained( SCREAMING_SNAKE_CASE__ ,output_loading_info=SCREAMING_SNAKE_CASE__ ,from_pt=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : List[Any] = AutoModelForMaskedLM.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_tf=SCREAMING_SNAKE_CASE__) __lowerCamelCase , __lowerCamelCase : Optional[Any] = AutoModelForMaskedLM.from_pretrained( SCREAMING_SNAKE_CASE__ ,output_loading_info=SCREAMING_SNAKE_CASE__ ,from_tf=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) @slow def lowerCAmelCase ( self : str): for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase : Dict = AutoConfig.from_pretrained(SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Optional[Any] = TFAutoModelForSeqaSeqLM.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_pt=SCREAMING_SNAKE_CASE__) __lowerCamelCase , __lowerCamelCase : List[str] = TFAutoModelForSeqaSeqLM.from_pretrained( SCREAMING_SNAKE_CASE__ ,output_loading_info=SCREAMING_SNAKE_CASE__ ,from_pt=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Any = AutoModelForSeqaSeqLM.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_tf=SCREAMING_SNAKE_CASE__) __lowerCamelCase , __lowerCamelCase : List[str] = AutoModelForSeqaSeqLM.from_pretrained( SCREAMING_SNAKE_CASE__ ,output_loading_info=SCREAMING_SNAKE_CASE__ ,from_tf=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) @slow def lowerCAmelCase ( self : Optional[Any]): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: __lowerCamelCase : List[Any] = AutoConfig.from_pretrained(SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Optional[int] = TFAutoModelForSequenceClassification.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_pt=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : int = AutoModelForSequenceClassification.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_tf=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) @slow def lowerCAmelCase ( self : Any): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: __lowerCamelCase : List[Any] = AutoConfig.from_pretrained(SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : str = TFAutoModelForQuestionAnswering.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_pt=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Optional[int] = AutoModelForQuestionAnswering.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_tf=SCREAMING_SNAKE_CASE__) self.assertIsNotNone(SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) def lowerCAmelCase ( self : int): __lowerCamelCase : List[Any] = TFAutoModelWithLMHead.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_pt=SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) self.assertEqual(model.num_parameters() ,1_4_4_1_0) self.assertEqual(model.num_parameters(only_trainable=SCREAMING_SNAKE_CASE__) ,1_4_4_1_0) __lowerCamelCase : Union[str, Any] = AutoModelWithLMHead.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_tf=SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) self.assertEqual(model.num_parameters() ,1_4_4_1_0) self.assertEqual(model.num_parameters(only_trainable=SCREAMING_SNAKE_CASE__) ,1_4_4_1_0) def lowerCAmelCase ( self : Tuple): __lowerCamelCase : str = TFAutoModelWithLMHead.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_pt=SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) self.assertEqual(model.num_parameters() ,1_4_4_1_0) self.assertEqual(model.num_parameters(only_trainable=SCREAMING_SNAKE_CASE__) ,1_4_4_1_0) __lowerCamelCase : Optional[int] = AutoModelWithLMHead.from_pretrained(SCREAMING_SNAKE_CASE__ ,from_tf=SCREAMING_SNAKE_CASE__) self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) self.assertEqual(model.num_parameters() ,1_4_4_1_0) self.assertEqual(model.num_parameters(only_trainable=SCREAMING_SNAKE_CASE__) ,1_4_4_1_0)
113
1
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCamelCase (_snake_case , _snake_case , unittest.TestCase ): '''simple docstring''' _snake_case : Union[str, Any] = IFImgaImgSuperResolutionPipeline _snake_case : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''width''', '''height'''} _snake_case : List[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'''original_image'''} ) _snake_case : List[str] = PipelineTesterMixin.required_optional_params - {'''latents'''} def __UpperCAmelCase ( self ) -> Optional[Any]: return self._get_superresolution_dummy_components() def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase=0 ) -> Any: if str(_UpperCamelCase ).startswith('mps' ): UpperCAmelCase_ : List[Any] = torch.manual_seed(_UpperCamelCase ) else: UpperCAmelCase_ : int = torch.Generator(device=_UpperCamelCase ).manual_seed(_UpperCamelCase ) UpperCAmelCase_ : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_UpperCamelCase ) ).to(_UpperCamelCase ) UpperCAmelCase_ : Dict = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(_UpperCamelCase ) ).to(_UpperCamelCase ) UpperCAmelCase_ : Tuple = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __UpperCAmelCase ( self ) -> Any: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __UpperCAmelCase ( self ) -> Dict: self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def __UpperCAmelCase ( self ) -> str: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __UpperCAmelCase ( self ) -> List[Any]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __UpperCAmelCase ( self ) -> Union[str, Any]: self._test_save_load_local() def __UpperCAmelCase ( self ) -> Dict: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
29
def lowercase__ ( __snake_case : int , __snake_case : int ): '''simple docstring''' if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) UpperCAmelCase_ : Tuple = str(bin(__snake_case ) )[2:] # remove the leading "0b" UpperCAmelCase_ : Union[str, Any] = str(bin(__snake_case ) )[2:] # remove the leading "0b" UpperCAmelCase_ : List[Any] = max(len(__snake_case ) , len(__snake_case ) ) return "0b" + "".join( str(int(char_a == '1' and char_b == '1' ) ) for char_a, char_b in zip(a_binary.zfill(__snake_case ) , b_binary.zfill(__snake_case ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
29
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging UpperCAmelCase: List[str] = logging.get_logger(__name__) UpperCAmelCase: Union[str, Any] = """▁""" UpperCAmelCase: List[Any] = {"""vocab_file""": """sentencepiece.bpe.model"""} UpperCAmelCase: Optional[Any] = { """vocab_file""": { """facebook/nllb-200-distilled-600M""": ( """https://huggingface.co/facebook/nllb-200-distilled-600M/blob/main/sentencepiece.bpe.model""" ), } } UpperCAmelCase: int = { """facebook/nllb-200-distilled-600M""": 1_024, } # fmt: off UpperCAmelCase: Optional[int] = ["""ace_Arab""", """ace_Latn""", """acm_Arab""", """acq_Arab""", """aeb_Arab""", """afr_Latn""", """ajp_Arab""", """aka_Latn""", """amh_Ethi""", """apc_Arab""", """arb_Arab""", """ars_Arab""", """ary_Arab""", """arz_Arab""", """asm_Beng""", """ast_Latn""", """awa_Deva""", """ayr_Latn""", """azb_Arab""", """azj_Latn""", """bak_Cyrl""", """bam_Latn""", """ban_Latn""", """bel_Cyrl""", """bem_Latn""", """ben_Beng""", """bho_Deva""", """bjn_Arab""", """bjn_Latn""", """bod_Tibt""", """bos_Latn""", """bug_Latn""", """bul_Cyrl""", """cat_Latn""", """ceb_Latn""", """ces_Latn""", """cjk_Latn""", """ckb_Arab""", """crh_Latn""", """cym_Latn""", """dan_Latn""", """deu_Latn""", """dik_Latn""", """dyu_Latn""", """dzo_Tibt""", """ell_Grek""", """eng_Latn""", """epo_Latn""", """est_Latn""", """eus_Latn""", """ewe_Latn""", """fao_Latn""", """pes_Arab""", """fij_Latn""", """fin_Latn""", """fon_Latn""", """fra_Latn""", """fur_Latn""", """fuv_Latn""", """gla_Latn""", """gle_Latn""", """glg_Latn""", """grn_Latn""", """guj_Gujr""", """hat_Latn""", """hau_Latn""", """heb_Hebr""", """hin_Deva""", """hne_Deva""", """hrv_Latn""", """hun_Latn""", """hye_Armn""", """ibo_Latn""", """ilo_Latn""", """ind_Latn""", """isl_Latn""", """ita_Latn""", """jav_Latn""", """jpn_Jpan""", """kab_Latn""", """kac_Latn""", """kam_Latn""", """kan_Knda""", """kas_Arab""", """kas_Deva""", """kat_Geor""", """knc_Arab""", """knc_Latn""", """kaz_Cyrl""", """kbp_Latn""", """kea_Latn""", """khm_Khmr""", """kik_Latn""", """kin_Latn""", """kir_Cyrl""", """kmb_Latn""", """kon_Latn""", """kor_Hang""", """kmr_Latn""", """lao_Laoo""", """lvs_Latn""", """lij_Latn""", """lim_Latn""", """lin_Latn""", """lit_Latn""", """lmo_Latn""", """ltg_Latn""", """ltz_Latn""", """lua_Latn""", """lug_Latn""", """luo_Latn""", """lus_Latn""", """mag_Deva""", """mai_Deva""", """mal_Mlym""", """mar_Deva""", """min_Latn""", """mkd_Cyrl""", """plt_Latn""", """mlt_Latn""", """mni_Beng""", """khk_Cyrl""", """mos_Latn""", """mri_Latn""", """zsm_Latn""", """mya_Mymr""", """nld_Latn""", """nno_Latn""", """nob_Latn""", """npi_Deva""", """nso_Latn""", """nus_Latn""", """nya_Latn""", """oci_Latn""", """gaz_Latn""", """ory_Orya""", """pag_Latn""", """pan_Guru""", """pap_Latn""", """pol_Latn""", """por_Latn""", """prs_Arab""", """pbt_Arab""", """quy_Latn""", """ron_Latn""", """run_Latn""", """rus_Cyrl""", """sag_Latn""", """san_Deva""", """sat_Beng""", """scn_Latn""", """shn_Mymr""", """sin_Sinh""", """slk_Latn""", """slv_Latn""", """smo_Latn""", """sna_Latn""", """snd_Arab""", """som_Latn""", """sot_Latn""", """spa_Latn""", """als_Latn""", """srd_Latn""", """srp_Cyrl""", """ssw_Latn""", """sun_Latn""", """swe_Latn""", """swh_Latn""", """szl_Latn""", """tam_Taml""", """tat_Cyrl""", """tel_Telu""", """tgk_Cyrl""", """tgl_Latn""", """tha_Thai""", """tir_Ethi""", """taq_Latn""", """taq_Tfng""", """tpi_Latn""", """tsn_Latn""", """tso_Latn""", """tuk_Latn""", """tum_Latn""", """tur_Latn""", """twi_Latn""", """tzm_Tfng""", """uig_Arab""", """ukr_Cyrl""", """umb_Latn""", """urd_Arab""", """uzn_Latn""", """vec_Latn""", """vie_Latn""", """war_Latn""", """wol_Latn""", """xho_Latn""", """ydd_Hebr""", """yor_Latn""", """yue_Hant""", """zho_Hans""", """zho_Hant""", """zul_Latn"""] class UpperCamelCase ( snake_case ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ : Any = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ : Union[str, Any] = ["input_ids", "attention_mask"] SCREAMING_SNAKE_CASE_ : List[int] = [] SCREAMING_SNAKE_CASE_ : List[int] = [] def __init__( self ,UpperCAmelCase_ ,UpperCAmelCase_="<s>" ,UpperCAmelCase_="</s>" ,UpperCAmelCase_="</s>" ,UpperCAmelCase_="<s>" ,UpperCAmelCase_="<unk>" ,UpperCAmelCase_="<pad>" ,UpperCAmelCase_="<mask>" ,UpperCAmelCase_=None ,UpperCAmelCase_=None ,UpperCAmelCase_=None ,UpperCAmelCase_ = None ,UpperCAmelCase_=None ,UpperCAmelCase_=False ,**UpperCAmelCase_ ,): # Mask token behave like a normal word, i.e. include the space before it _lowercase : List[str] = AddedToken(UpperCAmelCase_ ,lstrip=UpperCAmelCase_ ,rstrip=UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ ,UpperCAmelCase_ ) else mask_token _lowercase : Dict = {} if sp_model_kwargs is None else sp_model_kwargs _lowercase : str = legacy_behaviour super().__init__( bos_token=UpperCAmelCase_ ,eos_token=UpperCAmelCase_ ,unk_token=UpperCAmelCase_ ,sep_token=UpperCAmelCase_ ,cls_token=UpperCAmelCase_ ,pad_token=UpperCAmelCase_ ,mask_token=UpperCAmelCase_ ,tokenizer_file=UpperCAmelCase_ ,src_lang=UpperCAmelCase_ ,tgt_lang=UpperCAmelCase_ ,additional_special_tokens=UpperCAmelCase_ ,sp_model_kwargs=self.sp_model_kwargs ,legacy_behaviour=UpperCAmelCase_ ,**UpperCAmelCase_ ,) _lowercase : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(UpperCAmelCase_ ) ) _lowercase : Dict = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | ---- | ---- | ---- | ---- | ---- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' # spm | '<unk>' | '<s>' | '</s>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' | '▁s' # Mimic fairseq token-to-id alignment for the first 4 token _lowercase : Optional[int] = {"""<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab _lowercase : Tuple = 1 _lowercase : List[Any] = len(self.sp_model ) _lowercase : Any = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(UpperCAmelCase_ ) } _lowercase : List[str] = {v: k for k, v in self.lang_code_to_id.items()} _lowercase : Union[str, Any] = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) _lowercase : List[Any] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} _lowercase : int = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) _lowercase : Optional[int] = src_lang if src_lang is not None else """eng_Latn""" _lowercase : Dict = self.lang_code_to_id[self._src_lang] _lowercase : List[Any] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self ): _lowercase : Optional[int] = self.__dict__.copy() _lowercase : List[str] = None _lowercase : Optional[int] = self.sp_model.serialized_model_proto() return state def __setstate__( self ,UpperCAmelCase_ ): _lowercase : Union[str, Any] = d # for backward compatibility if not hasattr(self ,"""sp_model_kwargs""" ): _lowercase : Optional[int] = {} _lowercase : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def lowerCamelCase__ ( self ): return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def lowerCamelCase__ ( self ): return self._src_lang @src_lang.setter def lowerCamelCase__ ( self ,UpperCAmelCase_ ): _lowercase : Optional[int] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowerCamelCase__ ( self ,UpperCAmelCase_ ,UpperCAmelCase_ = None ,UpperCAmelCase_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase_ ,token_ids_a=UpperCAmelCase_ ,already_has_special_tokens=UpperCAmelCase_ ) _lowercase : Any = [1] * len(self.prefix_tokens ) _lowercase : List[Any] = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(UpperCAmelCase_ )) + suffix_ones return prefix_ones + ([0] * len(UpperCAmelCase_ )) + ([0] * len(UpperCAmelCase_ )) + suffix_ones def lowerCamelCase__ ( self ,UpperCAmelCase_ ,UpperCAmelCase_ = None ): if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowerCamelCase__ ( self ,UpperCAmelCase_ ,UpperCAmelCase_ = None ): _lowercase : List[Any] = [self.sep_token_id] _lowercase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowerCamelCase__ ( self ,UpperCAmelCase_ ,UpperCAmelCase_ ,UpperCAmelCase_ ,UpperCAmelCase_ ,**UpperCAmelCase_ ): if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) _lowercase : List[Any] = src_lang _lowercase : List[str] = self(UpperCAmelCase_ ,add_special_tokens=UpperCAmelCase_ ,return_tensors=UpperCAmelCase_ ,**UpperCAmelCase_ ) _lowercase : int = self.convert_tokens_to_ids(UpperCAmelCase_ ) _lowercase : int = tgt_lang_id return inputs def lowerCamelCase__ ( self ): _lowercase : Optional[int] = {self.convert_ids_to_tokens(UpperCAmelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowerCamelCase__ ( self ,UpperCAmelCase_ ): return self.sp_model.encode(UpperCAmelCase_ ,out_type=UpperCAmelCase_ ) def lowerCamelCase__ ( self ,UpperCAmelCase_ ): if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _lowercase : List[Any] = self.sp_model.PieceToId(UpperCAmelCase_ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def lowerCamelCase__ ( self ,UpperCAmelCase_ ): if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def lowerCamelCase__ ( self ,UpperCAmelCase_ ): _lowercase : Dict = """""".join(UpperCAmelCase_ ).replace(UpperCAmelCase_ ,""" """ ).strip() return out_string def lowerCamelCase__ ( self ,UpperCAmelCase_ ,UpperCAmelCase_ = None ): if not os.path.isdir(UpperCAmelCase_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return _lowercase : int = os.path.join( UpperCAmelCase_ ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,UpperCAmelCase_ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCAmelCase_ ,"""wb""" ) as fi: _lowercase : Optional[int] = self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase_ ) return (out_vocab_file,) def lowerCamelCase__ ( self ,UpperCAmelCase_ ,UpperCAmelCase_ = "eng_Latn" ,UpperCAmelCase_ = None ,UpperCAmelCase_ = "fra_Latn" ,**UpperCAmelCase_ ,): _lowercase : Optional[Any] = src_lang _lowercase : List[str] = tgt_lang return super().prepare_seqaseq_batch(UpperCAmelCase_ ,UpperCAmelCase_ ,**UpperCAmelCase_ ) def lowerCamelCase__ ( self ): return self.set_src_lang_special_tokens(self.src_lang ) def lowerCamelCase__ ( self ): return self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowerCamelCase__ ( self ,UpperCAmelCase_ ): _lowercase : Dict = self.lang_code_to_id[src_lang] if self.legacy_behaviour: _lowercase : Optional[Any] = [] _lowercase : List[Any] = [self.eos_token_id, self.cur_lang_code] else: _lowercase : Optional[Any] = [self.cur_lang_code] _lowercase : int = [self.eos_token_id] def lowerCamelCase__ ( self ,UpperCAmelCase_ ): _lowercase : Dict = self.lang_code_to_id[lang] if self.legacy_behaviour: _lowercase : Optional[int] = [] _lowercase : Dict = [self.eos_token_id, self.cur_lang_code] else: _lowercase : int = [self.cur_lang_code] _lowercase : List[str] = [self.eos_token_id]
336
"""simple docstring""" def __SCREAMING_SNAKE_CASE ( ): return [list(range(1000 - i , -1000 - i , -1 ) ) for i in range(1000 )] UpperCAmelCase: Any = generate_large_matrix() UpperCAmelCase: Dict = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): assert all(row == sorted(__UpperCAmelCase , reverse=__UpperCAmelCase ) for row in grid ) assert all(list(__UpperCAmelCase ) == sorted(__UpperCAmelCase , reverse=__UpperCAmelCase ) for col in zip(*__UpperCAmelCase ) ) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): _lowercase : Tuple = 0 _lowercase : List[Any] = len(__UpperCAmelCase ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: _lowercase : Tuple = (left + right) // 2 _lowercase : List[Any] = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: _lowercase : Dict = mid + 1 else: _lowercase : Dict = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(__UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): _lowercase : Any = 0 _lowercase : Optional[int] = len(grid[0] ) for i in range(len(__UpperCAmelCase ) ): _lowercase : Union[str, Any] = find_negative_index(grid[i][:bound] ) total += bound return (len(__UpperCAmelCase ) * len(grid[0] )) - total def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): return len([number for row in grid for number in row if number < 0] ) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): _lowercase : Tuple = 0 for row in grid: for i, number in enumerate(__UpperCAmelCase ): if number < 0: total += len(__UpperCAmelCase ) - i break return total def __SCREAMING_SNAKE_CASE ( ): from timeit import timeit print("""Running benchmarks""" ) _lowercase : Tuple = ( """from __main__ import count_negatives_binary_search, """ """count_negatives_brute_force, count_negatives_brute_force_with_break, grid""" ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): _lowercase : Dict = timeit(F"""{func}(grid=grid)""" , setup=__UpperCAmelCase , number=500 ) print(F"""{func}() took {time:0.4f} seconds""" ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
336
1
"""simple docstring""" import math import time from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class lowerCamelCase ( _lowerCAmelCase ): '''simple docstring''' def __init__( self: Optional[int] , *snake_case: Union[str, Any] , snake_case: Dict=None , snake_case: Optional[Any]=None , **snake_case: Tuple ) -> str: super().__init__(*snake_case , **snake_case ) snake_case_ :List[Any] = eval_examples snake_case_ :Optional[Any] = post_process_function def lowerCAmelCase_ ( self: List[str] , snake_case: str=None , snake_case: List[Any]=None , snake_case: Tuple=None , snake_case: str = "eval" ) -> Union[str, Any]: snake_case_ :Dict = self.eval_dataset if eval_dataset is None else eval_dataset snake_case_ :str = self.get_eval_dataloader(snake_case ) snake_case_ :Union[str, Any] = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. snake_case_ :str = self.compute_metrics snake_case_ :List[str] = None snake_case_ :Any = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop snake_case_ :Optional[int] = time.time() try: snake_case_ :Optional[Any] = eval_loop( snake_case , description="""Evaluation""" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=snake_case , metric_key_prefix=snake_case , ) finally: snake_case_ :Optional[int] = compute_metrics snake_case_ :Dict = self.args.eval_batch_size * self.args.world_size if f"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[f"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( snake_case , snake_case , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default snake_case_ :str = self.post_process_function(snake_case , snake_case , output.predictions ) snake_case_ :str = self.compute_metrics(snake_case ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"""{metric_key_prefix}_""" ): snake_case_ :Tuple = metrics.pop(snake_case ) metrics.update(output.metrics ) else: snake_case_ :int = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(snake_case ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) snake_case_ :Union[str, Any] = self.callback_handler.on_evaluate(self.args , self.state , self.control , snake_case ) return metrics def lowerCAmelCase_ ( self: int , snake_case: Optional[Any] , snake_case: int , snake_case: Optional[Any]=None , snake_case: str = "test" ) -> Dict: snake_case_ :List[str] = self.get_test_dataloader(snake_case ) # Temporarily disable metric computation, we will do it in the loop here. snake_case_ :Optional[Any] = self.compute_metrics snake_case_ :int = None snake_case_ :List[Any] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop snake_case_ :Optional[Any] = time.time() try: snake_case_ :Dict = eval_loop( snake_case , description="""Prediction""" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=snake_case , metric_key_prefix=snake_case , ) finally: snake_case_ :Any = compute_metrics snake_case_ :int = self.args.eval_batch_size * self.args.world_size if f"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[f"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( snake_case , snake_case , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output snake_case_ :Optional[Any] = self.post_process_function(snake_case , snake_case , output.predictions , """predict""" ) snake_case_ :Optional[Any] = self.compute_metrics(snake_case ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"""{metric_key_prefix}_""" ): snake_case_ :Union[str, Any] = metrics.pop(snake_case ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=snake_case )
66
"""simple docstring""" from __future__ import annotations from collections import Counter from random import random class lowerCamelCase : '''simple docstring''' def __init__( self: Tuple ) -> Optional[Any]: snake_case_ :Optional[int] = {} def lowerCAmelCase_ ( self: Dict , snake_case: str ) -> None: snake_case_ :str = {} def lowerCAmelCase_ ( self: Optional[int] , snake_case: str , snake_case: str , snake_case: float ) -> None: if nodea not in self.connections: self.add_node(snake_case ) if nodea not in self.connections: self.add_node(snake_case ) snake_case_ :Dict = probability def lowerCAmelCase_ ( self: List[Any] ) -> list[str]: return list(self.connections ) def lowerCAmelCase_ ( self: Any , snake_case: str ) -> str: snake_case_ :Optional[Any] = 0 snake_case_ :List[str] = random() for dest in self.connections[node]: current_probability += self.connections[node][dest] if current_probability > random_value: return dest return "" def A_ ( _lowercase, _lowercase, _lowercase ): '''simple docstring''' snake_case_ :List[str] = MarkovChainGraphUndirectedUnweighted() for nodea, nodea, probability in transitions: graph.add_transition_probability(_lowercase, _lowercase, _lowercase ) snake_case_ :int = Counter(graph.get_nodes() ) snake_case_ :Optional[Any] = start for _ in range(_lowercase ): snake_case_ :Tuple = graph.transition(_lowercase ) visited[node] += 1 return visited if __name__ == "__main__": import doctest doctest.testmod()
66
1
"""simple docstring""" from typing import Tuple, Union from ...modeling_outputs import BackboneOutput from ...modeling_utils import PreTrainedModel from ...utils import is_timm_available, is_torch_available, requires_backends from ...utils.backbone_utils import BackboneMixin from .configuration_timm_backbone import TimmBackboneConfig if is_timm_available(): import timm if is_torch_available(): from torch import Tensor class UpperCamelCase ( lowercase , lowercase ): """simple docstring""" UpperCAmelCase : Dict = """pixel_values""" UpperCAmelCase : int = False UpperCAmelCase : Optional[Any] = TimmBackboneConfig def __init__(self : Union[str, Any] , _A : Tuple , **_A : Optional[int]) -> Tuple: requires_backends(self , 'timm') super().__init__(_A) __snake_case : str = config if config.backbone is None: raise ValueError('backbone is not set in the config. Please set it to a timm model name.') if config.backbone not in timm.list_models(): raise ValueError(f"backbone {config.backbone} is not supported by timm.") if hasattr(_A , 'out_features') and config.out_features is not None: raise ValueError('out_features is not supported by TimmBackbone. Please use out_indices instead.') __snake_case : str = getattr(_A , 'use_pretrained_backbone' , _A) if pretrained is None: raise ValueError('use_pretrained_backbone is not set in the config. Please set it to True or False.') # We just take the final layer by default. This matches the default for the transformers models. __snake_case : int = config.out_indices if getattr(_A , 'out_indices' , _A) is not None else (-1,) __snake_case : List[str] = timm.create_model( config.backbone , pretrained=_A , features_only=config.features_only , in_chans=config.num_channels , out_indices=_A , **_A , ) # These are used to control the output of the model when called. If output_hidden_states is True, then # return_layers is modified to include all layers. __snake_case : Tuple = self._backbone.return_layers __snake_case : str = {layer['module']: str(_A) for i, layer in enumerate(self._backbone.feature_info.info)} super()._init_backbone(_A) @classmethod def _lowercase (cls : Any , _A : Optional[int] , *_A : int , **_A : Optional[int]) -> str: requires_backends(cls , ['vision', 'timm']) from ...models.timm_backbone import TimmBackboneConfig __snake_case : Union[str, Any] = kwargs.pop('config' , TimmBackboneConfig()) __snake_case : Any = kwargs.pop('use_timm_backbone' , _A) if not use_timm: raise ValueError('use_timm_backbone must be True for timm backbones') __snake_case : str = kwargs.pop('num_channels' , config.num_channels) __snake_case : int = kwargs.pop('features_only' , config.features_only) __snake_case : List[Any] = kwargs.pop('use_pretrained_backbone' , config.use_pretrained_backbone) __snake_case : Optional[Any] = kwargs.pop('out_indices' , config.out_indices) __snake_case : Optional[Any] = TimmBackboneConfig( backbone=_A , num_channels=_A , features_only=_A , use_pretrained_backbone=_A , out_indices=_A , ) return super()._from_config(_A , **_A) def _lowercase (self : Optional[Any] , _A : Tuple) -> str: pass def _lowercase (self : Tuple , _A : List[Any] , _A : Union[str, Any]=None , _A : str=None , _A : List[Any]=None , **_A : str) -> Union[BackboneOutput, Tuple[Tensor, ...]]: __snake_case : List[str] = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : List[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : Tuple = output_attentions if output_attentions is not None else self.config.output_attentions if output_attentions: raise ValueError('Cannot output attentions for timm backbones at the moment') if output_hidden_states: # We modify the return layers to include all the stages of the backbone __snake_case : int = self._all_layers __snake_case : Optional[Any] = self._backbone(_A , **_A) __snake_case : Dict = self._return_layers __snake_case : str = tuple(hidden_states[i] for i in self.out_indices) else: __snake_case : str = self._backbone(_A , **_A) __snake_case : int = None __snake_case : Dict = tuple(_A) __snake_case : List[str] = tuple(_A) if hidden_states is not None else None if not return_dict: __snake_case : Any = (feature_maps,) if output_hidden_states: __snake_case : int = output + (hidden_states,) return output return BackboneOutput(feature_maps=_A , hidden_states=_A , attentions=_A)
365
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCamelCase ( unittest.TestCase ): def _lowercase (self : Union[str, Any]) -> Optional[int]: __snake_case : Optional[Any] = 0 def _lowercase (self : Tuple) -> int: __snake_case : Optional[Any] = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32') self.assertIsInstance(_A , _A) def _lowercase (self : str) -> List[str]: with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : List[str] = Path(_A) / 'preprocessor_config.json' __snake_case : Optional[Any] = Path(_A) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) __snake_case : Optional[int] = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) def _lowercase (self : Any) -> Optional[int]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : Any = Path(_A) / 'preprocessor_config.json' __snake_case : List[Any] = Path(_A) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) __snake_case : Tuple = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) def _lowercase (self : List[Any]) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : str = CLIPConfig() # Create a dummy config file with image_proceesor_type __snake_case : List[Any] = Path(_A) / 'preprocessor_config.json' __snake_case : Optional[Any] = Path(_A) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) # remove image_processor_type to make sure config.json alone is enough to load image processor locally __snake_case : List[str] = AutoImageProcessor.from_pretrained(_A).to_dict() config_dict.pop('image_processor_type') __snake_case : Optional[int] = CLIPImageProcessor(**_A) # save in new folder model_config.save_pretrained(_A) config.save_pretrained(_A) __snake_case : Optional[int] = AutoImageProcessor.from_pretrained(_A) # make sure private variable is not incorrectly saved __snake_case : int = json.loads(config.to_json_string()) self.assertTrue('_processor_class' not in dict_as_saved) self.assertIsInstance(_A , _A) def _lowercase (self : Union[str, Any]) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : int = Path(_A) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) __snake_case : List[str] = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) def _lowercase (self : Optional[int]) -> Dict: with self.assertRaisesRegex( _A , 'clip-base is not a local folder and is not a valid model identifier'): __snake_case : Tuple = AutoImageProcessor.from_pretrained('clip-base') def _lowercase (self : str) -> int: with self.assertRaisesRegex( _A , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)'): __snake_case : str = AutoImageProcessor.from_pretrained(_A , revision='aaaaaa') def _lowercase (self : List[Any]) -> str: with self.assertRaisesRegex( _A , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): __snake_case : List[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model') def _lowercase (self : Optional[int]) -> List[str]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(_A): __snake_case : Any = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor') # If remote code is disabled, we can't load this config. with self.assertRaises(_A): __snake_case : Tuple = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) __snake_case : Union[str, Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_A) __snake_case : Optional[int] = AutoImageProcessor.from_pretrained(_A , trust_remote_code=_A) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor') def _lowercase (self : int) -> Optional[int]: try: AutoConfig.register('custom' , _A) AutoImageProcessor.register(_A , _A) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_A): AutoImageProcessor.register(_A , _A) with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : Tuple = Path(_A) / 'preprocessor_config.json' __snake_case : Dict = Path(_A) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(_A , 'w') , ) json.dump({'model_type': 'clip'} , open(_A , 'w')) __snake_case : Tuple = CustomImageProcessor.from_pretrained(_A) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_A) __snake_case : Tuple = AutoImageProcessor.from_pretrained(_A) self.assertIsInstance(_A , _A) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _lowercase (self : List[Any]) -> Tuple: class UpperCamelCase ( lowercase ): UpperCAmelCase : str = True try: AutoConfig.register('custom' , _A) AutoImageProcessor.register(_A , _A) # If remote code is not set, the default is to use local __snake_case : Tuple = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor') self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') self.assertTrue(image_processor.is_local) # If remote code is disabled, we load the local one. __snake_case : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') self.assertTrue(image_processor.is_local) # If remote is enabled, we load from the Hub __snake_case : List[Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_A) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor') self.assertTrue(not hasattr(_A , 'is_local')) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
95
0
import numpy as np def lowerCamelCase_ ( _a ): """simple docstring""" return (2 / (1 + np.exp(-2 * vector ))) - 1 if __name__ == "__main__": import doctest doctest.testmod()
131
import os import time import warnings from dataclasses import dataclass, field from enum import Enum from typing import List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import logging from ..processors.glue import glue_convert_examples_to_features, glue_output_modes, glue_processors from ..processors.utils import InputFeatures lowerCamelCase = logging.get_logger(__name__) @dataclass class _a : _a : str = field(metadata={'''help''': '''The name of the task to train on: ''' + ''', '''.join(glue_processors.keys())}) _a : str = field( metadata={'''help''': '''The input data dir. Should contain the .tsv files (or other data files) for the task.'''}) _a : int = field( default=128 , metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) _a : bool = field( default=_lowercase , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''}) def UpperCAmelCase__( self : str )-> List[Any]: lowerCAmelCase__ : int = self.task_name.lower() class _a ( _lowercase): _a : Tuple = '''train''' _a : str = '''dev''' _a : List[str] = '''test''' class _a ( _lowercase): _a : GlueDataTrainingArguments _a : str _a : List[InputFeatures] def __init__( self : List[Any] , _SCREAMING_SNAKE_CASE : GlueDataTrainingArguments , _SCREAMING_SNAKE_CASE : PreTrainedTokenizerBase , _SCREAMING_SNAKE_CASE : Optional[int] = None , _SCREAMING_SNAKE_CASE : Union[str, Split] = Split.train , _SCREAMING_SNAKE_CASE : Optional[str] = None , )-> Optional[Any]: warnings.warn( '''This dataset will be removed from the library soon, preprocessing should be handled with the 🤗 Datasets ''' '''library. You can have a look at this example script for pointers: ''' '''https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py''' , _SCREAMING_SNAKE_CASE , ) lowerCAmelCase__ : Optional[Any] = args lowerCAmelCase__ : str = glue_processors[args.task_name]() lowerCAmelCase__ : Tuple = glue_output_modes[args.task_name] if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): try: lowerCAmelCase__ : str = Split[mode] except KeyError: raise KeyError('''mode is not a valid split name''' ) # Load data features from cache or dataset file lowerCAmelCase__ : Any = os.path.join( cache_dir if cache_dir is not None else args.data_dir , F'cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{args.task_name}' , ) lowerCAmelCase__ : Union[str, Any] = self.processor.get_labels() if args.task_name in ["mnli", "mnli-mm"] and tokenizer.__class__.__name__ in ( "RobertaTokenizer", "RobertaTokenizerFast", "XLMRobertaTokenizer", "BartTokenizer", "BartTokenizerFast", ): # HACK(label indices are swapped in RoBERTa pretrained model) lowerCAmelCase__ , lowerCAmelCase__ : str = label_list[2], label_list[1] lowerCAmelCase__ : Tuple = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. lowerCAmelCase__ : List[Any] = cached_features_file + '''.lock''' with FileLock(_SCREAMING_SNAKE_CASE ): if os.path.exists(_SCREAMING_SNAKE_CASE ) and not args.overwrite_cache: lowerCAmelCase__ : List[str] = time.time() lowerCAmelCase__ : Optional[int] = torch.load(_SCREAMING_SNAKE_CASE ) logger.info( F'Loading features from cached file {cached_features_file} [took %.3f s]' , time.time() - start ) else: logger.info(F'Creating features from dataset file at {args.data_dir}' ) if mode == Split.dev: lowerCAmelCase__ : int = self.processor.get_dev_examples(args.data_dir ) elif mode == Split.test: lowerCAmelCase__ : Optional[Any] = self.processor.get_test_examples(args.data_dir ) else: lowerCAmelCase__ : str = self.processor.get_train_examples(args.data_dir ) if limit_length is not None: lowerCAmelCase__ : List[str] = examples[:limit_length] lowerCAmelCase__ : Union[str, Any] = glue_convert_examples_to_features( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , max_length=args.max_seq_length , label_list=_SCREAMING_SNAKE_CASE , output_mode=self.output_mode , ) lowerCAmelCase__ : Optional[int] = time.time() torch.save(self.features , _SCREAMING_SNAKE_CASE ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( F'Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]' ) def __len__( self : Optional[Any] )-> Dict: return len(self.features ) def __getitem__( self : Tuple , _SCREAMING_SNAKE_CASE : str )-> InputFeatures: return self.features[i] def UpperCAmelCase__( self : int )-> int: return self.label_list
131
1
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _lowerCAmelCase : Union[str, Any] = logging.get_logger(__name__) def __snake_case ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any]=False ) -> Union[str, Any]: A_ : Optional[Any] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"blocks.{i}.norm1.weight", f"deit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((f"blocks.{i}.norm1.bias", f"deit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append((f"blocks.{i}.attn.proj.weight", f"deit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((f"blocks.{i}.attn.proj.bias", f"deit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((f"blocks.{i}.norm2.weight", f"deit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((f"blocks.{i}.norm2.bias", f"deit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((f"blocks.{i}.mlp.fc1.weight", f"deit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((f"blocks.{i}.mlp.fc1.bias", f"deit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((f"blocks.{i}.mlp.fc2.weight", f"deit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((f"blocks.{i}.mlp.fc2.bias", f"deit.encoder.layer.{i}.output.dense.bias") ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "deit.embeddings.cls_token"), ("dist_token", "deit.embeddings.distillation_token"), ("patch_embed.proj.weight", "deit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "deit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "deit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" A_ : Dict = [(pair[0], pair[1][4:]) if pair[1].startswith("deit" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("norm.weight", "deit.layernorm.weight"), ("norm.bias", "deit.layernorm.bias"), ("head.weight", "cls_classifier.weight"), ("head.bias", "cls_classifier.bias"), ("head_dist.weight", "distillation_classifier.weight"), ("head_dist.bias", "distillation_classifier.bias"), ] ) return rename_keys def __snake_case ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]=False ) -> Optional[int]: for i in range(config.num_hidden_layers ): if base_model: A_ : List[Any] = "" else: A_ : Any = "deit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A_ : List[Any] = state_dict.pop(f"blocks.{i}.attn.qkv.weight" ) A_ : List[Any] = state_dict.pop(f"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict A_ : List[Any] = in_proj_weight[ : config.hidden_size, : ] A_ : List[str] = in_proj_bias[: config.hidden_size] A_ : Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A_ : Optional[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A_ : Optional[int] = in_proj_weight[ -config.hidden_size :, : ] A_ : Tuple = in_proj_bias[-config.hidden_size :] def __snake_case ( _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any ) -> Dict: A_ : List[Any] = dct.pop(_lowerCAmelCase ) A_ : Dict = val def __snake_case ( ) -> Dict: A_ : Tuple = "http://images.cocodataset.org/val2017/000000039769.jpg" A_ : int = Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ) return im @torch.no_grad() def __snake_case ( _lowerCAmelCase : str , _lowerCAmelCase : List[str] ) -> str: A_ : Optional[Any] = DeiTConfig() # all deit models have fine-tuned heads A_ : Union[str, Any] = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size A_ : int = 1000 A_ : List[Any] = "huggingface/label-files" A_ : Tuple = "imagenet-1k-id2label.json" A_ : str = json.load(open(hf_hub_download(_lowerCAmelCase , _lowerCAmelCase , repo_type="dataset" ) , "r" ) ) A_ : Optional[int] = {int(_lowerCAmelCase ): v for k, v in idalabel.items()} A_ : List[Any] = idalabel A_ : Union[str, Any] = {v: k for k, v in idalabel.items()} A_ : Any = int(deit_name[-6:-4] ) A_ : Dict = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("tiny" ): A_ : Optional[Any] = 192 A_ : Optional[Any] = 768 A_ : Any = 12 A_ : Optional[Any] = 3 elif deit_name[9:].startswith("small" ): A_ : List[Any] = 384 A_ : Dict = 1536 A_ : Optional[int] = 12 A_ : Dict = 6 if deit_name[9:].startswith("base" ): pass elif deit_name[4:].startswith("large" ): A_ : List[Any] = 1024 A_ : Optional[Any] = 4096 A_ : Tuple = 24 A_ : Optional[Any] = 16 # load original model from timm A_ : str = timm.create_model(_lowerCAmelCase , pretrained=_lowerCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys A_ : int = timm_model.state_dict() A_ : Optional[Any] = create_rename_keys(_lowerCAmelCase , _lowerCAmelCase ) for src, dest in rename_keys: rename_key(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) read_in_q_k_v(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # load HuggingFace model A_ : Dict = DeiTForImageClassificationWithTeacher(_lowerCAmelCase ).eval() model.load_state_dict(_lowerCAmelCase ) # Check outputs on an image, prepared by DeiTImageProcessor A_ : Optional[int] = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 A_ : Optional[Any] = DeiTImageProcessor(size=_lowerCAmelCase , crop_size=config.image_size ) A_ : Union[str, Any] = image_processor(images=prepare_img() , return_tensors="pt" ) A_ : List[str] = encoding["pixel_values"] A_ : List[Any] = model(_lowerCAmelCase ) A_ : Any = timm_model(_lowerCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_lowerCAmelCase , outputs.logits , atol=1e-3 ) Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) print(f"Saving model {deit_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(_lowerCAmelCase ) print(f"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": _lowerCAmelCase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--deit_name''', default='''vit_deit_base_distilled_patch16_224''', type=str, help='''Name of the DeiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) _lowerCAmelCase : str = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
354
from collections.abc import Sequence def __snake_case ( _lowerCAmelCase : Sequence[int] | None = None ) -> int: if nums is None or not nums: raise ValueError("Input sequence should not be empty" ) A_ : Any = nums[0] for i in range(1 , len(_lowerCAmelCase ) ): A_ : Any = nums[i] A_ : List[str] = max(_lowerCAmelCase , ans + num , _lowerCAmelCase ) return ans if __name__ == "__main__": import doctest doctest.testmod() # Try on a sample input from the user _lowerCAmelCase : List[Any] = int(input('''Enter number of elements : ''').strip()) _lowerCAmelCase : Dict = list(map(int, input('''\nEnter the numbers : ''').strip().split()))[:n] print(max_subsequence_sum(array))
70
0
"""simple docstring""" import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def _lowerCamelCase( ): with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(a ): requests.request("GET" , "https://huggingface.co" ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request("GET" , "https://huggingface.co" , timeout=1.0 ) @pytest.mark.integration def _lowerCamelCase( ): with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request("GET" , "https://huggingface.co" ) def _lowerCamelCase( ): with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(a ): http_head("https://huggingface.co" )
261
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class snake_case__ ( snake_case_, snake_case_, snake_case_, unittest.TestCase ): _snake_case : str = StableUnCLIPImgaImgPipeline _snake_case : Optional[Any] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS _snake_case : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _snake_case : Optional[Any] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _snake_case : List[Any] = frozenset([] ) def a__ ( self ): __a = 32 __a = embedder_hidden_size # image encoding components __a = CLIPImageProcessor(crop_size=32 , size=32 ) torch.manual_seed(0 ) __a = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=lowerCamelCase , projection_dim=lowerCamelCase , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , ) ) # regular denoising components torch.manual_seed(0 ) __a = StableUnCLIPImageNormalizer(embedding_dim=lowerCamelCase ) __a = DDPMScheduler(beta_schedule="squaredcos_cap_v2" ) torch.manual_seed(0 ) __a = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) torch.manual_seed(0 ) __a = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=lowerCamelCase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) __a = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "CrossAttnUpBlock2D") , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type="projection" , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=lowerCamelCase , layers_per_block=1 , upcast_attention=lowerCamelCase , use_linear_projection=lowerCamelCase , ) torch.manual_seed(0 ) __a = DDIMScheduler( beta_schedule="scaled_linear" , beta_start=0.0_0085 , beta_end=0.012 , prediction_type="v_prediction" , set_alpha_to_one=lowerCamelCase , steps_offset=1 , ) torch.manual_seed(0 ) __a = AutoencoderKL() __a = { # image encoding components "feature_extractor": feature_extractor, "image_encoder": image_encoder.eval(), # image noising components "image_normalizer": image_normalizer.eval(), "image_noising_scheduler": image_noising_scheduler, # regular denoising components "tokenizer": tokenizer, "text_encoder": text_encoder.eval(), "unet": unet.eval(), "scheduler": scheduler, "vae": vae.eval(), } return components def a__ ( self , lowerCamelCase , lowerCamelCase=0 , lowerCamelCase=True ): if str(lowerCamelCase ).startswith("mps" ): __a = torch.manual_seed(lowerCamelCase ) else: __a = torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) __a = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase ) ).to(lowerCamelCase ) if pil_image: __a = input_image * 0.5 + 0.5 __a = input_image.clamp(0 , 1 ) __a = input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() __a = DiffusionPipeline.numpy_to_pil(lowerCamelCase )[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def a__ ( self ): __a = "cpu" # ensure determinism for the device-dependent torch.Generator __a = self.get_dummy_components() __a = StableUnCLIPImgaImgPipeline(**lowerCamelCase ) __a = sd_pipe.to(lowerCamelCase ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase ) __a = self.get_dummy_inputs(lowerCamelCase ) inputs.update({"image_embeds": None} ) __a = sd_pipe(**lowerCamelCase ).images __a = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __a = np.array([0.3872, 0.7224, 0.5601, 0.4741, 0.6872, 0.5814, 0.4636, 0.3867, 0.5078] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def a__ ( self ): __a = torch_device in ["cpu", "mps"] self._test_attention_slicing_forward_pass(test_max_difference=lowerCamelCase ) def a__ ( self ): __a = torch_device in ["cpu", "mps"] self._test_inference_batch_single_identical(test_max_difference=lowerCamelCase ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def a__ ( self ): self._test_xformers_attention_forwardGenerator_pass(test_max_difference=lowerCamelCase ) @slow @require_torch_gpu class snake_case__ ( unittest.TestCase ): def a__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def a__ ( self ): __a = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png" ) __a = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy" ) __a = StableUnCLIPImgaImgPipeline.from_pretrained( "fusing/stable-unclip-2-1-l-img2img" , torch_dtype=torch.floataa ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __a = torch.Generator(device="cpu" ).manual_seed(0 ) __a = pipe(lowerCamelCase , "anime turle" , generator=lowerCamelCase , output_type="np" ) __a = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(lowerCamelCase , lowerCamelCase ) def a__ ( self ): __a = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png" ) __a = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy" ) __a = StableUnCLIPImgaImgPipeline.from_pretrained( "fusing/stable-unclip-2-1-h-img2img" , torch_dtype=torch.floataa ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __a = torch.Generator(device="cpu" ).manual_seed(0 ) __a = pipe(lowerCamelCase , "anime turle" , generator=lowerCamelCase , output_type="np" ) __a = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(lowerCamelCase , lowerCamelCase ) def a__ ( self ): __a = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png" ) torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __a = StableUnCLIPImgaImgPipeline.from_pretrained( "fusing/stable-unclip-2-1-h-img2img" , torch_dtype=torch.floataa ) __a = pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __a = pipe( lowerCamelCase , "anime turtle" , num_inference_steps=2 , output_type="np" , ) __a = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
261
1
'''simple docstring''' from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer _lowercase : List[Any] = logging.get_logger(__name__) _lowercase : Optional[int] = { "vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_config_file": "tokenizer_config.json", } _lowercase : Optional[int] = { "vocab_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json" }, "merges_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt" }, "tokenizer_config_file": { "facebook/blenderbot_small-90M": ( "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json" ) }, } _lowercase : List[Any] = { "facebook/blenderbot_small-90M": 512, } class __magic_name__ ( _UpperCAmelCase): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = BlenderbotSmallTokenizer def __init__( self : Any , lowercase_ : List[str]=None , lowercase_ : List[Any]=None , lowercase_ : int="<|endoftext|>" , lowercase_ : int="<|endoftext|>" , lowercase_ : Optional[int]="<|endoftext|>" , lowercase_ : Tuple=False , lowercase_ : Any=True , **lowercase_ : Any , ): super().__init__( ByteLevelBPETokenizer( vocab=lowercase_ , merges=lowercase_ , add_prefix_space=lowercase_ , trim_offsets=lowercase_ , ) , bos_token=lowercase_ , eos_token=lowercase_ , unk_token=lowercase_ , **lowercase_ , ) lowercase_ : Union[str, Any] = add_prefix_space def SCREAMING_SNAKE_CASE_ ( self : Optional[int] , lowercase_ : Any , lowercase_ : Any=None ): lowercase_ : List[Any] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def SCREAMING_SNAKE_CASE_ ( self : List[Any] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None ): lowercase_ : Dict = [self.sep_token_id] lowercase_ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
21
'''simple docstring''' from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_VISION_2_SEQ_MAPPING if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_VISION_2_SEQ_MAPPING _lowercase : str = logging.get_logger(__name__) @add_end_docstrings(_UpperCAmelCase) class __magic_name__ ( _UpperCAmelCase): def __init__( self : str , *lowercase_ : Dict , **lowercase_ : List[Any] ): super().__init__(*lowercase_ , **lowercase_ ) requires_backends(self , """vision""" ) self.check_model_type( TF_MODEL_FOR_VISION_2_SEQ_MAPPING if self.framework == """tf""" else MODEL_FOR_VISION_2_SEQ_MAPPING ) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] , lowercase_ : str=None , lowercase_ : List[Any]=None , lowercase_ : Dict=None ): lowercase_ : Optional[Any] = {} lowercase_ : Tuple = {} if prompt is not None: lowercase_ : Tuple = prompt if generate_kwargs is not None: lowercase_ : List[str] = generate_kwargs if max_new_tokens is not None: if "generate_kwargs" not in forward_kwargs: lowercase_ : List[Any] = {} if "max_new_tokens" in forward_kwargs["generate_kwargs"]: raise ValueError( """'max_new_tokens' is defined twice, once in 'generate_kwargs' and once as a direct parameter,""" """ please use only one""" ) lowercase_ : str = max_new_tokens return preprocess_params, forward_kwargs, {} def __call__( self : List[Any] , lowercase_ : Union[str, List[str], "Image.Image", List["Image.Image"]] , **lowercase_ : Optional[int] ): return super().__call__(lowercase_ , **lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : str , lowercase_ : List[Any] , lowercase_ : Tuple=None ): lowercase_ : List[Any] = load_image(lowercase_ ) if prompt is not None: if not isinstance(lowercase_ , lowercase_ ): raise ValueError( f'''Received an invalid text input, got - {type(lowercase_ )} - but expected a single string. ''' """Note also that one single text can be provided for conditional image to text generation.""" ) lowercase_ : List[Any] = self.model.config.model_type if model_type == "git": lowercase_ : Dict = self.image_processor(images=lowercase_ , return_tensors=self.framework ) lowercase_ : Union[str, Any] = self.tokenizer(text=lowercase_ , add_special_tokens=lowercase_ ).input_ids lowercase_ : int = [self.tokenizer.cls_token_id] + input_ids lowercase_ : List[Any] = torch.tensor(lowercase_ ).unsqueeze(0 ) model_inputs.update({"""input_ids""": input_ids} ) elif model_type == "pix2struct": lowercase_ : Union[str, Any] = self.image_processor(images=lowercase_ , header_text=lowercase_ , return_tensors=self.framework ) elif model_type != "vision-encoder-decoder": # vision-encoder-decoder does not support conditional generation lowercase_ : Dict = self.image_processor(images=lowercase_ , return_tensors=self.framework ) lowercase_ : List[str] = self.tokenizer(lowercase_ , return_tensors=self.framework ) model_inputs.update(lowercase_ ) else: raise ValueError(f'''Model type {model_type} does not support conditional text generation''' ) else: lowercase_ : List[str] = self.image_processor(images=lowercase_ , return_tensors=self.framework ) if self.model.config.model_type == "git" and prompt is None: lowercase_ : str = None return model_inputs def SCREAMING_SNAKE_CASE_ ( self : Tuple , lowercase_ : Dict , lowercase_ : Optional[Any]=None ): # Git model sets `model_inputs["input_ids"] = None` in `preprocess` (when `prompt=None`). In batch model, the # pipeline will group them into a list of `None`, which fail `_forward`. Avoid this by checking it first. if ( "input_ids" in model_inputs and isinstance(model_inputs["""input_ids"""] , lowercase_ ) and all(x is None for x in model_inputs["""input_ids"""] ) ): lowercase_ : Any = None if generate_kwargs is None: lowercase_ : Optional[Any] = {} # FIXME: We need to pop here due to a difference in how `generation.py` and `generation.tf_utils.py` # parse inputs. In the Tensorflow version, `generate` raises an error if we don't use `input_ids` whereas # the PyTorch version matches it with `self.model.main_input_name` or `self.model.encoder.main_input_name` # in the `_prepare_model_inputs` method. lowercase_ : Dict = model_inputs.pop(self.model.main_input_name ) lowercase_ : Any = self.model.generate(lowercase_ , **lowercase_ , **lowercase_ ) return model_outputs def SCREAMING_SNAKE_CASE_ ( self : List[str] , lowercase_ : List[Any] ): lowercase_ : List[str] = [] for output_ids in model_outputs: lowercase_ : Union[str, Any] = { """generated_text""": self.tokenizer.decode( lowercase_ , skip_special_tokens=lowercase_ , ) } records.append(lowercase_ ) return records
21
1
"""simple docstring""" import torch from torch import nn class snake_case ( nn.Module ): '''simple docstring''' def __init__( self : List[Any], _lowerCamelCase : Tuple, _lowerCamelCase : int, _lowerCamelCase : List[Any], _lowerCamelCase : Any, _lowerCamelCase : Any=1, _lowerCamelCase : Tuple=False ): '''simple docstring''' super().__init__() __A = n_token __A = d_embed __A = d_proj __A = cutoffs + [n_token] __A = [0] + self.cutoffs __A = div_val __A = self.cutoffs[0] __A = len(self.cutoffs ) - 1 __A = self.shortlist_size + self.n_clusters if self.n_clusters > 0: __A = nn.Parameter(torch.zeros(self.n_clusters, self.d_embed ) ) __A = nn.Parameter(torch.zeros(self.n_clusters ) ) __A = nn.ModuleList() __A = nn.ParameterList() if div_val == 1: for i in range(len(self.cutoffs ) ): if d_proj != d_embed: self.out_projs.append(nn.Parameter(torch.FloatTensor(_lowerCamelCase, _lowerCamelCase ) ) ) else: self.out_projs.append(_lowerCamelCase ) self.out_layers.append(nn.Linear(_lowerCamelCase, _lowerCamelCase ) ) else: for i in range(len(self.cutoffs ) ): __A , __A = self.cutoff_ends[i], self.cutoff_ends[i + 1] __A = d_embed // (div_val**i) self.out_projs.append(nn.Parameter(torch.FloatTensor(_lowerCamelCase, _lowerCamelCase ) ) ) self.out_layers.append(nn.Linear(_lowerCamelCase, r_idx - l_idx ) ) __A = keep_order def _SCREAMING_SNAKE_CASE ( self : List[str], _lowerCamelCase : Optional[int], _lowerCamelCase : int, _lowerCamelCase : Optional[Any], _lowerCamelCase : List[str] ): '''simple docstring''' if proj is None: __A = nn.functional.linear(_lowerCamelCase, _lowerCamelCase, bias=_lowerCamelCase ) else: # if CUDA_MAJOR <= 9 and CUDA_MINOR <= 1: __A = nn.functional.linear(_lowerCamelCase, proj.t().contiguous() ) __A = nn.functional.linear(_lowerCamelCase, _lowerCamelCase, bias=_lowerCamelCase ) # else: # logit = torch.einsum('bd,de,ev->bv', (hidden, proj, weight.t())) # if bias is not None: # logit = logit + bias return logit def _SCREAMING_SNAKE_CASE ( self : Tuple, _lowerCamelCase : List[str], _lowerCamelCase : Tuple=None, _lowerCamelCase : Tuple=False ): '''simple docstring''' if labels is not None: # Shift so that tokens < n predict n __A = hidden[..., :-1, :].contiguous() __A = labels[..., 1:].contiguous() __A = hidden.view(-1, hidden.size(-1 ) ) __A = labels.view(-1 ) if hidden.size(0 ) != labels.size(0 ): raise RuntimeError('''Input and labels should have the same size in the batch dimension.''' ) else: __A = hidden.view(-1, hidden.size(-1 ) ) if self.n_clusters == 0: __A = self._compute_logit(_lowerCamelCase, self.out_layers[0].weight, self.out_layers[0].bias, self.out_projs[0] ) if labels is not None: __A = labels != -1_00 __A = torch.zeros_like(_lowerCamelCase, dtype=hidden.dtype, device=hidden.device ) __A = ( -nn.functional.log_softmax(_lowerCamelCase, dim=-1 )[mask].gather(1, labels[mask].unsqueeze(1 ) ).squeeze(1 ) ) else: __A = nn.functional.log_softmax(_lowerCamelCase, dim=-1 ) else: # construct weights and biases __A , __A = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: __A , __A = self.cutoff_ends[i], self.cutoff_ends[i + 1] __A = self.out_layers[0].weight[l_idx:r_idx] __A = self.out_layers[0].bias[l_idx:r_idx] else: __A = self.out_layers[i].weight __A = self.out_layers[i].bias if i == 0: __A = torch.cat([weight_i, self.cluster_weight], dim=0 ) __A = torch.cat([bias_i, self.cluster_bias], dim=0 ) weights.append(_lowerCamelCase ) biases.append(_lowerCamelCase ) __A , __A , __A = weights[0], biases[0], self.out_projs[0] __A = self._compute_logit(_lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase ) __A = nn.functional.log_softmax(_lowerCamelCase, dim=1 ) if labels is None: __A = hidden.new_empty((head_logit.size(0 ), self.n_token) ) else: __A = torch.zeros_like(_lowerCamelCase, dtype=hidden.dtype, device=hidden.device ) __A = 0 __A = [0] + self.cutoffs for i in range(len(_lowerCamelCase ) - 1 ): __A , __A = cutoff_values[i], cutoff_values[i + 1] if labels is not None: __A = (labels >= l_idx) & (labels < r_idx) __A = mask_i.nonzero().squeeze() if indices_i.numel() == 0: continue __A = labels.index_select(0, _lowerCamelCase ) - l_idx __A = head_logprob.index_select(0, _lowerCamelCase ) __A = hidden.index_select(0, _lowerCamelCase ) else: __A = hidden if i == 0: if labels is not None: __A = head_logprob_i.gather(1, target_i[:, None] ).squeeze(1 ) else: __A = head_logprob[:, : self.cutoffs[0]] else: __A , __A , __A = weights[i], biases[i], self.out_projs[i] __A = self._compute_logit(_lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase ) __A = nn.functional.log_softmax(_lowerCamelCase, dim=1 ) __A = self.cutoffs[0] + i - 1 # No probability for the head cluster if labels is not None: __A = head_logprob_i[:, cluster_prob_idx] + tail_logprob_i.gather( 1, target_i[:, None] ).squeeze(1 ) else: __A = head_logprob[:, cluster_prob_idx, None] + tail_logprob_i __A = logprob_i if labels is not None: if (hasattr(self, '''keep_order''' ) and self.keep_order) or keep_order: out.index_copy_(0, _lowerCamelCase, -logprob_i ) else: out[offset : offset + logprob_i.size(0 )].copy_(-logprob_i ) offset += logprob_i.size(0 ) return out def _SCREAMING_SNAKE_CASE ( self : Any, _lowerCamelCase : Optional[int] ): '''simple docstring''' if self.n_clusters == 0: __A = self._compute_logit(_lowerCamelCase, self.out_layers[0].weight, self.out_layers[0].bias, self.out_projs[0] ) return nn.functional.log_softmax(_lowerCamelCase, dim=-1 ) else: # construct weights and biases __A , __A = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: __A , __A = self.cutoff_ends[i], self.cutoff_ends[i + 1] __A = self.out_layers[0].weight[l_idx:r_idx] __A = self.out_layers[0].bias[l_idx:r_idx] else: __A = self.out_layers[i].weight __A = self.out_layers[i].bias if i == 0: __A = torch.cat([weight_i, self.cluster_weight], dim=0 ) __A = torch.cat([bias_i, self.cluster_bias], dim=0 ) weights.append(_lowerCamelCase ) biases.append(_lowerCamelCase ) __A , __A , __A = weights[0], biases[0], self.out_projs[0] __A = self._compute_logit(_lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase ) __A = hidden.new_empty((head_logit.size(0 ), self.n_token) ) __A = nn.functional.log_softmax(_lowerCamelCase, dim=1 ) __A = [0] + self.cutoffs for i in range(len(_lowerCamelCase ) - 1 ): __A , __A = cutoff_values[i], cutoff_values[i + 1] if i == 0: __A = head_logprob[:, : self.cutoffs[0]] else: __A , __A , __A = weights[i], biases[i], self.out_projs[i] __A = self._compute_logit(_lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase ) __A = nn.functional.log_softmax(_lowerCamelCase, dim=1 ) __A = head_logprob[:, -i] + tail_logprob_i __A = logprob_i return out
266
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import ASTFeatureExtractor from transformers.testing_utils import require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin lowercase_ = random.Random() if is_torch_available(): import torch def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase=1.0 , __UpperCamelCase=None , __UpperCamelCase=None ): """simple docstring""" if rng is None: __A = global_rng __A = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class snake_case ( unittest.TestCase ): '''simple docstring''' def __init__( self : Any, _lowerCamelCase : List[str], _lowerCamelCase : Any=7, _lowerCamelCase : Optional[int]=4_00, _lowerCamelCase : Optional[int]=20_00, _lowerCamelCase : Dict=1, _lowerCamelCase : Optional[Any]=0.0, _lowerCamelCase : int=1_60_00, _lowerCamelCase : Optional[int]=True, _lowerCamelCase : Dict=True, ): '''simple docstring''' __A = parent __A = batch_size __A = min_seq_length __A = max_seq_length __A = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __A = feature_size __A = padding_value __A = sampling_rate __A = return_attention_mask __A = do_normalize def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _SCREAMING_SNAKE_CASE ( self : Any, _lowerCamelCase : Optional[Any]=False, _lowerCamelCase : int=False ): '''simple docstring''' def _flatten(_lowerCamelCase : List[str] ): return list(itertools.chain(*_lowerCamelCase ) ) if equal_length: __A = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __A = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length, self.max_seq_length, self.seq_length_diff ) ] if numpify: __A = [np.asarray(_lowerCamelCase ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class snake_case ( _lowerCAmelCase , unittest.TestCase ): '''simple docstring''' A_ : int = ASTFeatureExtractor def _SCREAMING_SNAKE_CASE ( self : Tuple ): '''simple docstring''' __A = ASTFeatureExtractionTester(self ) def _SCREAMING_SNAKE_CASE ( self : Dict ): '''simple docstring''' # Tests that all call wrap to encode_plus and batch_encode_plus __A = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __A = [floats_list((1, x) )[0] for x in range(8_00, 14_00, 2_00 )] __A = [np.asarray(_lowerCamelCase ) for speech_input in speech_inputs] # Test not batched input __A = feat_extract(speech_inputs[0], return_tensors='''np''' ).input_values __A = feat_extract(np_speech_inputs[0], return_tensors='''np''' ).input_values self.assertTrue(np.allclose(_lowerCamelCase, _lowerCamelCase, atol=1e-3 ) ) # Test batched __A = feat_extract(_lowerCamelCase, padding=_lowerCamelCase, return_tensors='''np''' ).input_values __A = feat_extract(_lowerCamelCase, padding=_lowerCamelCase, return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(_lowerCamelCase, _lowerCamelCase ): self.assertTrue(np.allclose(_lowerCamelCase, _lowerCamelCase, atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __A = [floats_list((1, x) )[0] for x in (8_00, 8_00, 8_00)] __A = np.asarray(_lowerCamelCase ) __A = feat_extract(_lowerCamelCase, return_tensors='''np''' ).input_values __A = feat_extract(_lowerCamelCase, return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(_lowerCamelCase, _lowerCamelCase ): self.assertTrue(np.allclose(_lowerCamelCase, _lowerCamelCase, atol=1e-3 ) ) @require_torch def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): '''simple docstring''' import torch __A = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __A = np.random.rand(1_00 ).astype(np.floataa ) __A = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __A = feature_extractor.pad([{'''input_values''': inputs}], return_tensors='''np''' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __A = feature_extractor.pad([{'''input_values''': inputs}], return_tensors='''pt''' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def _SCREAMING_SNAKE_CASE ( self : Optional[int], _lowerCamelCase : Union[str, Any] ): '''simple docstring''' from datasets import load_dataset __A = load_dataset('''hf-internal-testing/librispeech_asr_dummy''', '''clean''', split='''validation''' ) # automatic decoding with librispeech __A = ds.sort('''id''' ).select(range(_lowerCamelCase ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] @require_torch def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ): '''simple docstring''' # fmt: off __A = torch.tensor( [-0.98_94, -1.27_76, -0.90_66, -1.27_76, -0.93_49, -1.26_09, -1.03_86, -1.27_76, -1.15_61, -1.27_76, -1.20_52, -1.27_23, -1.21_90, -1.21_32, -1.27_76, -1.11_33, -1.19_53, -1.13_43, -1.15_84, -1.22_03, -1.17_70, -1.24_74, -1.23_81, -1.19_36, -0.92_70, -0.83_17, -0.80_49, -0.77_06, -0.75_65, -0.78_69] ) # fmt: on __A = self._load_datasamples(1 ) __A = ASTFeatureExtractor() __A = feature_extractor(_lowerCamelCase, return_tensors='''pt''' ).input_values self.assertEquals(input_values.shape, (1, 10_24, 1_28) ) self.assertTrue(torch.allclose(input_values[0, 0, :30], _lowerCamelCase, atol=1e-4 ) )
266
1
from queue import Queue from typing import TYPE_CHECKING, Optional if TYPE_CHECKING: from ..models.auto import AutoTokenizer class _lowerCAmelCase : def lowerCamelCase ( self , UpperCamelCase__ ) -> Dict: '''simple docstring''' raise NotImplementedError() def lowerCamelCase ( self ) -> int: '''simple docstring''' raise NotImplementedError() class _lowerCAmelCase ( snake_case_ ): def __init__( self , UpperCamelCase__ , UpperCamelCase__ = False , **UpperCamelCase__ ) -> Optional[int]: '''simple docstring''' snake_case : int = tokenizer snake_case : str = skip_prompt snake_case : List[str] = decode_kwargs # variables used in the streaming process snake_case : int = [] snake_case : Optional[int] = 0 snake_case : Optional[int] = True def lowerCamelCase ( self , UpperCamelCase__ ) -> Any: '''simple docstring''' if len(value.shape ) > 1 and value.shape[0] > 1: raise ValueError("TextStreamer only supports batch size 1" ) elif len(value.shape ) > 1: snake_case : Union[str, Any] = value[0] if self.skip_prompt and self.next_tokens_are_prompt: snake_case : Union[str, Any] = False return # Add the new token to the cache and decodes the entire thing. self.token_cache.extend(value.tolist() ) snake_case : List[str] = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) # After the symbol for a new line, we flush the cache. if text.endswith("\n" ): snake_case : Optional[Any] = text[self.print_len :] snake_case : int = [] snake_case : List[Any] = 0 # If the last token is a CJK character, we print the characters. elif len(UpperCamelCase__ ) > 0 and self._is_chinese_char(ord(text[-1] ) ): snake_case : Dict = text[self.print_len :] self.print_len += len(UpperCamelCase__ ) # Otherwise, prints until the last space char (simple heuristic to avoid printing incomplete words, # which may change with the subsequent token -- there are probably smarter ways to do this!) else: snake_case : Dict = text[self.print_len : text.rfind(" " ) + 1] self.print_len += len(UpperCamelCase__ ) self.on_finalized_text(UpperCamelCase__ ) def lowerCamelCase ( self ) -> int: '''simple docstring''' if len(self.token_cache ) > 0: snake_case : Tuple = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) snake_case : List[str] = text[self.print_len :] snake_case : List[Any] = [] snake_case : str = 0 else: snake_case : Tuple = "" snake_case : Tuple = True self.on_finalized_text(UpperCamelCase__ , stream_end=UpperCamelCase__ ) def lowerCamelCase ( self , UpperCamelCase__ , UpperCamelCase__ = False ) -> List[str]: '''simple docstring''' print(UpperCamelCase__ , flush=UpperCamelCase__ , end="" if not stream_end else None ) def lowerCamelCase ( self , UpperCamelCase__ ) -> str: '''simple docstring''' if ( (cp >= 0x4_e_0_0 and cp <= 0x9_f_f_f) or (cp >= 0x3_4_0_0 and cp <= 0x4_d_b_f) # or (cp >= 0x2_0_0_0_0 and cp <= 0x2_a_6_d_f) # or (cp >= 0x2_a_7_0_0 and cp <= 0x2_b_7_3_f) # or (cp >= 0x2_b_7_4_0 and cp <= 0x2_b_8_1_f) # or (cp >= 0x2_b_8_2_0 and cp <= 0x2_c_e_a_f) # or (cp >= 0xf_9_0_0 and cp <= 0xf_a_f_f) or (cp >= 0x2_f_8_0_0 and cp <= 0x2_f_a_1_f) # ): # return True return False class _lowerCAmelCase ( snake_case_ ): def __init__( self , UpperCamelCase__ , UpperCamelCase__ = False , UpperCamelCase__ = None , **UpperCamelCase__ ) -> Tuple: '''simple docstring''' super().__init__(UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ) snake_case : List[str] = Queue() snake_case : Union[str, Any] = None snake_case : Union[str, Any] = timeout def lowerCamelCase ( self , UpperCamelCase__ , UpperCamelCase__ = False ) -> Any: '''simple docstring''' self.text_queue.put(UpperCamelCase__ , timeout=self.timeout ) if stream_end: self.text_queue.put(self.stop_signal , timeout=self.timeout ) def __iter__( self ) -> List[Any]: '''simple docstring''' return self def lowerCamelCase ( self ) -> Optional[int]: '''simple docstring''' snake_case : Tuple = self.text_queue.get(timeout=self.timeout ) if value == self.stop_signal: raise StopIteration() else: return value
356
"""simple docstring""" import argparse import os import re import numpy as np import PIL import torch from timm import create_model from torch.optim.lr_scheduler import OneCycleLR from torch.utils.data import DataLoader, Dataset from torchvision.transforms import Compose, RandomResizedCrop, Resize, ToTensor from accelerate import Accelerator def __lowerCAmelCase ( lowercase : int ) -> Tuple: """simple docstring""" snake_case : Any = fname.split(os.path.sep )[-1] return re.search(R"^(.*)_\d+\.jpg$" , lowercase ).groups()[0] class _lowerCAmelCase ( snake_case_ ): def __init__( self , UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None ) -> Union[str, Any]: '''simple docstring''' snake_case : Union[str, Any] = file_names snake_case : Optional[Any] = image_transform snake_case : Optional[int] = label_to_id def __len__( self ) -> Tuple: '''simple docstring''' return len(self.file_names ) def __getitem__( self , UpperCamelCase__ ) -> int: '''simple docstring''' snake_case : str = self.file_names[idx] snake_case : Any = PIL.Image.open(UpperCamelCase__ ) snake_case : Optional[int] = raw_image.convert("RGB" ) if self.image_transform is not None: snake_case : Optional[Any] = self.image_transform(UpperCamelCase__ ) snake_case : Optional[Any] = extract_label(UpperCamelCase__ ) if self.label_to_id is not None: snake_case : Optional[Any] = self.label_to_id[label] return {"image": image, "label": label} def __lowerCAmelCase ( lowercase : Any , lowercase : List[Any] ) -> List[str]: """simple docstring""" if args.with_tracking: snake_case : List[str] = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , log_with="all" , project_dir=args.project_dir ) else: snake_case : Optional[int] = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs snake_case : str = config["lr"] snake_case : Union[str, Any] = int(config["num_epochs"] ) snake_case : str = int(config["seed"] ) snake_case : str = int(config["batch_size"] ) snake_case : Any = config["image_size"] if not isinstance(lowercase , (list, tuple) ): snake_case : str = (image_size, image_size) # Parse out whether we are saving every epoch or after a certain number of batches if hasattr(args.checkpointing_steps , "isdigit" ): if args.checkpointing_steps == "epoch": snake_case : List[str] = args.checkpointing_steps elif args.checkpointing_steps.isdigit(): snake_case : Any = int(args.checkpointing_steps ) else: raise ValueError( F'Argument `checkpointing_steps` must be either a number or `epoch`. `{args.checkpointing_steps}` passed.' ) else: snake_case : List[str] = None # We need to initialize the trackers we use, and also store our configuration if args.with_tracking: snake_case : Union[str, Any] = os.path.split(lowercase )[-1].split("." )[0] accelerator.init_trackers(lowercase , lowercase ) # Grab all the image filenames snake_case : int = [os.path.join(args.data_dir , lowercase ) for fname in os.listdir(args.data_dir ) if fname.endswith(".jpg" )] # Build the label correspondences snake_case : Union[str, Any] = [extract_label(lowercase ) for fname in file_names] snake_case : Any = list(set(lowercase ) ) id_to_label.sort() snake_case : int = {lbl: i for i, lbl in enumerate(lowercase )} # Set the seed before splitting the data. np.random.seed(lowercase ) torch.manual_seed(lowercase ) torch.cuda.manual_seed_all(lowercase ) # Split our filenames between train and validation snake_case : Optional[Any] = np.random.permutation(len(lowercase ) ) snake_case : int = int(0.8 * len(lowercase ) ) snake_case : int = random_perm[:cut] snake_case : int = random_perm[cut:] # For training we use a simple RandomResizedCrop snake_case : List[Any] = Compose([RandomResizedCrop(lowercase , scale=(0.5, 1.0) ), ToTensor()] ) snake_case : List[str] = PetsDataset( [file_names[i] for i in train_split] , image_transform=lowercase , label_to_id=lowercase ) # For evaluation, we use a deterministic Resize snake_case : Optional[Any] = Compose([Resize(lowercase ), ToTensor()] ) snake_case : Optional[Any] = PetsDataset([file_names[i] for i in eval_split] , image_transform=lowercase , label_to_id=lowercase ) # Instantiate dataloaders. snake_case : Optional[Any] = DataLoader(lowercase , shuffle=lowercase , batch_size=lowercase , num_workers=4 ) snake_case : Tuple = DataLoader(lowercase , shuffle=lowercase , batch_size=lowercase , num_workers=4 ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) snake_case : Optional[int] = create_model("resnet50d" , pretrained=lowercase , num_classes=len(lowercase ) ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). snake_case : Any = model.to(accelerator.device ) # Freezing the base model for param in model.parameters(): snake_case : Dict = False for param in model.get_classifier().parameters(): snake_case : List[Any] = True # We normalize the batches of images to be a bit faster. snake_case : Dict = torch.tensor(model.default_cfg["mean"] )[None, :, None, None].to(accelerator.device ) snake_case : Union[str, Any] = torch.tensor(model.default_cfg["std"] )[None, :, None, None].to(accelerator.device ) # Instantiate optimizer snake_case : int = torch.optim.Adam(params=model.parameters() , lr=lr / 25 ) # Instantiate learning rate scheduler snake_case : Dict = OneCycleLR(optimizer=lowercase , max_lr=lowercase , epochs=lowercase , steps_per_epoch=len(lowercase ) ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. snake_case ,snake_case ,snake_case ,snake_case ,snake_case : List[str] = accelerator.prepare( lowercase , lowercase , lowercase , lowercase , lowercase ) # We need to keep track of how many total steps we have iterated over snake_case : List[Any] = 0 # We also need to keep track of the starting epoch so files are named properly snake_case : Optional[int] = 0 # Potentially load in the weights and states from a previous save if args.resume_from_checkpoint: if args.resume_from_checkpoint is not None or args.resume_from_checkpoint != "": accelerator.print(F'Resumed from checkpoint: {args.resume_from_checkpoint}' ) accelerator.load_state(args.resume_from_checkpoint ) snake_case : List[str] = os.path.basename(args.resume_from_checkpoint ) else: # Get the most recent checkpoint snake_case : List[Any] = [f.name for f in os.scandir(os.getcwd() ) if f.is_dir()] dirs.sort(key=os.path.getctime ) snake_case : int = dirs[-1] # Sorts folders by date modified, most recent checkpoint is the last # Extract `epoch_{i}` or `step_{i}` snake_case : Union[str, Any] = os.path.splitext(lowercase )[0] if "epoch" in training_difference: snake_case : Any = int(training_difference.replace("epoch_" , "" ) ) + 1 snake_case : int = None else: snake_case : Any = int(training_difference.replace("step_" , "" ) ) snake_case : Optional[int] = resume_step // len(lowercase ) resume_step -= starting_epoch * len(lowercase ) # Now we train the model for epoch in range(lowercase , lowercase ): model.train() if args.with_tracking: snake_case : Union[str, Any] = 0 if args.resume_from_checkpoint and epoch == starting_epoch and resume_step is not None: # We need to skip steps until we reach the resumed step snake_case : List[str] = accelerator.skip_first_batches(lowercase , lowercase ) overall_step += resume_step else: # After the first iteration though, we need to go back to the original dataloader snake_case : Any = train_dataloader for batch in active_dataloader: # We could avoid this line since we set the accelerator with `device_placement=True`. snake_case : List[Any] = {k: v.to(accelerator.device ) for k, v in batch.items()} snake_case : Optional[int] = (batch["image"] - mean) / std snake_case : str = model(lowercase ) snake_case : Dict = torch.nn.functional.cross_entropy(lowercase , batch["label"] ) # We keep track of the loss at each epoch if args.with_tracking: total_loss += loss.detach().float() accelerator.backward(lowercase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 if isinstance(lowercase , lowercase ): snake_case : Any = F'step_{overall_step}' if overall_step % checkpointing_steps == 0: if args.output_dir is not None: snake_case : List[str] = os.path.join(args.output_dir , lowercase ) accelerator.save_state(lowercase ) model.eval() snake_case : List[str] = 0 snake_case : List[str] = 0 for step, batch in enumerate(lowercase ): # We could avoid this line since we set the accelerator with `device_placement=True`. snake_case : int = {k: v.to(accelerator.device ) for k, v in batch.items()} snake_case : Tuple = (batch["image"] - mean) / std with torch.no_grad(): snake_case : Optional[int] = model(lowercase ) snake_case : List[Any] = outputs.argmax(dim=-1 ) snake_case ,snake_case : int = accelerator.gather_for_metrics((predictions, batch["label"]) ) snake_case : Union[str, Any] = predictions == references num_elems += accurate_preds.shape[0] accurate += accurate_preds.long().sum() snake_case : List[Any] = accurate.item() / num_elems # Use accelerator.print to print only on the main process. accelerator.print(F'epoch {epoch}: {100 * eval_metric:.2f}' ) if args.with_tracking: accelerator.log( { "accuracy": 100 * eval_metric, "train_loss": total_loss.item() / len(lowercase ), "epoch": epoch, } , step=lowercase , ) if checkpointing_steps == "epoch": snake_case : Optional[Any] = F'epoch_{epoch}' if args.output_dir is not None: snake_case : Union[str, Any] = os.path.join(args.output_dir , lowercase ) accelerator.save_state(lowercase ) if args.with_tracking: accelerator.end_training() def __lowerCAmelCase ( ) -> str: """simple docstring""" snake_case : Optional[Any] = argparse.ArgumentParser(description="Simple example of training script." ) parser.add_argument("--data_dir" , required=lowercase , help="The data folder on disk." ) parser.add_argument("--fp16" , action="store_true" , help="If passed, will use FP16 training." ) parser.add_argument( "--mixed_precision" , type=lowercase , default=lowercase , choices=["no", "fp16", "bf16", "fp8"] , help="Whether to use mixed precision. Choose" "between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10." "and an Nvidia Ampere GPU." , ) parser.add_argument("--cpu" , action="store_true" , help="If passed, will train on the CPU." ) parser.add_argument( "--checkpointing_steps" , type=lowercase , default=lowercase , help="Whether the various states should be saved at the end of every n steps, or 'epoch' for each epoch." , ) parser.add_argument( "--output_dir" , type=lowercase , default="." , help="Optional save directory where all checkpoint folders will be stored. Default is the current working directory." , ) parser.add_argument( "--resume_from_checkpoint" , type=lowercase , default=lowercase , help="If the training should continue from a checkpoint folder." , ) parser.add_argument( "--with_tracking" , action="store_true" , help="Whether to load in all available experiment trackers from the environment and use them for logging." , ) parser.add_argument( "--project_dir" , type=lowercase , default="logs" , help="Location on where to store experiment tracking logs` and relevent project information" , ) snake_case : Optional[Any] = parser.parse_args() snake_case : List[str] = {"lr": 3e-2, "num_epochs": 3, "seed": 42, "batch_size": 64, "image_size": 224} training_function(lowercase , lowercase ) if __name__ == "__main__": main()
112
0
"""simple docstring""" from ..utils import DummyObject, requires_backends class __snake_case ( metaclass=_lowercase): snake_case__ : Union[str, Any] = ["torch", "torchsde"] def __init__( self : Dict , *__lowerCAmelCase : str , **__lowerCAmelCase : Optional[int] ): """simple docstring""" requires_backends(self , ['''torch''', '''torchsde'''] ) @classmethod def SCREAMING_SNAKE_CASE ( cls : Optional[int] , *__lowerCAmelCase : List[Any] , **__lowerCAmelCase : Optional[Any] ): """simple docstring""" requires_backends(cls , ['''torch''', '''torchsde'''] ) @classmethod def SCREAMING_SNAKE_CASE ( cls : int , *__lowerCAmelCase : str , **__lowerCAmelCase : str ): """simple docstring""" requires_backends(cls , ['''torch''', '''torchsde'''] )
72
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __snake_case ( unittest.TestCase): def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : Union[str, Any] = [ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertTrue(is_safetensors_compatible(__lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase : List[Any] = [ '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertTrue(is_safetensors_compatible(__lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" _lowerCamelCase : Optional[int] = [ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', '''unet/diffusion_pytorch_model.bin''', # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(__lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" _lowerCamelCase : Optional[Any] = [ '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', ] self.assertTrue(is_safetensors_compatible(__lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" _lowerCamelCase : int = [ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', # Removed: 'text_encoder/model.safetensors', '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertFalse(is_safetensors_compatible(__lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" _lowerCamelCase : int = [ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] _lowerCamelCase : Optional[int] = '''fp16''' self.assertTrue(is_safetensors_compatible(__lowerCAmelCase , variant=__lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" _lowerCamelCase : Optional[Any] = [ '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] _lowerCamelCase : Union[str, Any] = '''fp16''' self.assertTrue(is_safetensors_compatible(__lowerCAmelCase , variant=__lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): """simple docstring""" _lowerCamelCase : str = [ '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] _lowerCamelCase : Optional[Any] = '''fp16''' self.assertTrue(is_safetensors_compatible(__lowerCAmelCase , variant=__lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" _lowerCamelCase : Tuple = [ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', '''unet/diffusion_pytorch_model.fp16.bin''', # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] _lowerCamelCase : Any = '''fp16''' self.assertFalse(is_safetensors_compatible(__lowerCAmelCase , variant=__lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" _lowerCamelCase : Optional[Any] = [ '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', ] _lowerCamelCase : str = '''fp16''' self.assertTrue(is_safetensors_compatible(__lowerCAmelCase , variant=__lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" _lowerCamelCase : Optional[Any] = [ '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', ] _lowerCamelCase : Union[str, Any] = '''fp16''' self.assertTrue(is_safetensors_compatible(__lowerCAmelCase , variant=__lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase : int = [ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', # 'text_encoder/model.fp16.safetensors', '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] _lowerCamelCase : int = '''fp16''' self.assertFalse(is_safetensors_compatible(__lowerCAmelCase , variant=__lowerCAmelCase ) )
72
1
import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def UpperCamelCase ( __lowerCamelCase : Any ): snake_case : Any = [ "encoder.version", "decoder.version", "model.encoder.version", "model.decoder.version", "decoder.output_projection.weight", "_float_tensor", "encoder.embed_positions._float_tensor", "decoder.embed_positions._float_tensor", ] for k in ignore_keys: state_dict.pop(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def UpperCamelCase ( __lowerCamelCase : str ): snake_case : Optional[Any] = list(s_dict.keys() ) for key in keys: if "transformer_layers" in key: snake_case : Union[str, Any] = s_dict.pop(__SCREAMING_SNAKE_CASE ) elif "subsample" in key: snake_case : List[Any] = s_dict.pop(__SCREAMING_SNAKE_CASE ) def UpperCamelCase ( __lowerCamelCase : Union[str, Any] ): snake_case : int = emb.weight.shape snake_case : List[Any] = nn.Linear(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , bias=__SCREAMING_SNAKE_CASE ) snake_case : Dict = emb.weight.data return lin_layer def UpperCamelCase ( __lowerCamelCase : Dict , __lowerCamelCase : List[Any] ): snake_case : List[Any] = torch.load(__SCREAMING_SNAKE_CASE , map_location="cpu" ) snake_case : int = mam_aaa["args"] snake_case : Tuple = mam_aaa["model"] snake_case : Dict = state_dict["decoder.output_projection.weight"] remove_ignore_keys_(__SCREAMING_SNAKE_CASE ) rename_keys(__SCREAMING_SNAKE_CASE ) snake_case : Optional[Any] = state_dict["decoder.embed_tokens.weight"].shape[0] snake_case : List[Any] = args.share_decoder_input_output_embed snake_case : Optional[int] = [int(__SCREAMING_SNAKE_CASE ) for i in args.conv_kernel_sizes.split("," )] snake_case : Optional[int] = SpeechaTextConfig( vocab_size=__SCREAMING_SNAKE_CASE , max_source_positions=args.max_source_positions , max_target_positions=args.max_target_positions , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function="relu" , num_conv_layers=len(__SCREAMING_SNAKE_CASE ) , conv_channels=args.conv_channels , conv_kernel_sizes=__SCREAMING_SNAKE_CASE , input_feat_per_channel=args.input_feat_per_channel , input_channels=args.input_channels , tie_word_embeddings=__SCREAMING_SNAKE_CASE , num_beams=5 , max_length=200 , use_cache=__SCREAMING_SNAKE_CASE , decoder_start_token_id=2 , early_stopping=__SCREAMING_SNAKE_CASE , ) snake_case : int = SpeechaTextForConditionalGeneration(__SCREAMING_SNAKE_CASE ) snake_case : Dict = model.model.load_state_dict(__SCREAMING_SNAKE_CASE , strict=__SCREAMING_SNAKE_CASE ) if len(__SCREAMING_SNAKE_CASE ) > 0 and not set(__SCREAMING_SNAKE_CASE ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( "Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing," f""" but all the following weights are missing {missing}""" ) if tie_embeds: snake_case : Any = make_linear_from_emb(model.model.decoder.embed_tokens ) else: snake_case : Any = lm_head_weights model.save_pretrained(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument("""--fairseq_path""", type=str, help="""Path to the fairseq model (.pt) file.""") parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") __lowerCamelCase = parser.parse_args() convert_fairseq_sat_checkpoint_to_tfms(args.fairseq_path, args.pytorch_dump_folder_path)
369
import fire from utils import calculate_rouge, save_json def UpperCamelCase ( __lowerCamelCase : Dict , __lowerCamelCase : Dict , __lowerCamelCase : Tuple=None , **__lowerCamelCase : Tuple ): snake_case : Optional[Any] = [x.strip() for x in open(__lowerCamelCase ).readlines()] snake_case : Union[str, Any] = [x.strip() for x in open(__lowerCamelCase ).readlines()][: len(__lowerCamelCase )] snake_case : List[Any] = calculate_rouge(__lowerCamelCase , __lowerCamelCase , **__lowerCamelCase ) if save_path is not None: save_json(__lowerCamelCase , __lowerCamelCase , indent=__lowerCamelCase ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
10
0
def lowerCamelCase_ ( _a ): """simple docstring""" assert column_title.isupper() lowerCAmelCase__ : Dict = 0 lowerCAmelCase__ : int = len(_a ) - 1 lowerCAmelCase__ : Any = 0 while index >= 0: lowerCAmelCase__ : Optional[int] = (ord(column_title[index] ) - 64) * pow(26 , _a ) answer += value power += 1 index -= 1 return answer if __name__ == "__main__": from doctest import testmod testmod()
131
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCamelCase = { '''configuration_bloom''': ['''BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BloomConfig''', '''BloomOnnxConfig'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = ['''BloomTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ '''BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BloomForCausalLM''', '''BloomModel''', '''BloomPreTrainedModel''', '''BloomForSequenceClassification''', '''BloomForTokenClassification''', '''BloomForQuestionAnswering''', ] if TYPE_CHECKING: from .configuration_bloom import BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP, BloomConfig, BloomOnnxConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bloom_fast import BloomTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bloom import ( BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST, BloomForCausalLM, BloomForQuestionAnswering, BloomForSequenceClassification, BloomForTokenClassification, BloomModel, BloomPreTrainedModel, ) else: import sys lowerCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
131
1
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class UpperCamelCase_ ( UpperCamelCase , UpperCamelCase , unittest.TestCase): """simple docstring""" snake_case__ : str = IFInpaintingSuperResolutionPipeline snake_case__ : Optional[Any] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} snake_case__ : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"original_image"}) snake_case__ : Dict = PipelineTesterMixin.required_optional_params - {"latents"} def UpperCAmelCase_ ( self : List[Any] ) -> Optional[int]: return self._get_superresolution_dummy_components() def UpperCAmelCase_ ( self : List[str] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[str]=0 ) -> List[str]: if str(UpperCAmelCase__ ).startswith("mps" ): __SCREAMING_SNAKE_CASE = torch.manual_seed(UpperCAmelCase__ ) else: __SCREAMING_SNAKE_CASE = torch.Generator(device=UpperCAmelCase__ ).manual_seed(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(UpperCAmelCase__ ) ).to(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(UpperCAmelCase__ ) ).to(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(UpperCAmelCase__ ) ).to(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def UpperCAmelCase_ ( self : str ) -> Dict: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def UpperCAmelCase_ ( self : str ) -> Tuple: self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def UpperCAmelCase_ ( self : Tuple ) -> Tuple: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def UpperCAmelCase_ ( self : Optional[Any] ) -> List[Any]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def UpperCAmelCase_ ( self : Any ) -> Union[str, Any]: self._test_save_load_local() def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[Any]: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
195
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' if not nums: return 0 __SCREAMING_SNAKE_CASE = nums[0] __SCREAMING_SNAKE_CASE = 0 for num in nums[1:]: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = ( max_excluding + num, max(lowerCAmelCase_ , lowerCAmelCase_ ), ) return max(lowerCAmelCase_ , lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
195
1
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class A__ : def __init__( self : List[Any] , _a : Dict , _a : int=13 , _a : Optional[Any]=7 , _a : int=6 , _a : Tuple=17 , _a : str=23 , _a : Optional[Any]=11 , _a : List[Any]=True , ) -> Tuple: '''simple docstring''' _SCREAMING_SNAKE_CASE =parent _SCREAMING_SNAKE_CASE =batch_size _SCREAMING_SNAKE_CASE =seq_length _SCREAMING_SNAKE_CASE =act_dim _SCREAMING_SNAKE_CASE =state_dim _SCREAMING_SNAKE_CASE =hidden_size _SCREAMING_SNAKE_CASE =max_length _SCREAMING_SNAKE_CASE =is_training def A ( self : Optional[Any] ) -> List[str]: '''simple docstring''' _SCREAMING_SNAKE_CASE =floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) _SCREAMING_SNAKE_CASE =floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) _SCREAMING_SNAKE_CASE =floats_tensor((self.batch_size, self.seq_length, 1) ) _SCREAMING_SNAKE_CASE =floats_tensor((self.batch_size, self.seq_length, 1) ) _SCREAMING_SNAKE_CASE =ids_tensor((self.batch_size, self.seq_length) , vocab_size=1000 ) _SCREAMING_SNAKE_CASE =random_attention_mask((self.batch_size, self.seq_length) ) _SCREAMING_SNAKE_CASE =self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def A ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def A ( self : Optional[int] , _a : Any , _a : Optional[int] , _a : Tuple , _a : Optional[int] , _a : List[str] , _a : Dict , _a : Optional[int] , ) -> List[str]: '''simple docstring''' _SCREAMING_SNAKE_CASE =DecisionTransformerModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() _SCREAMING_SNAKE_CASE =model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def A ( self : Any ) -> Union[str, Any]: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) =config_and_inputs _SCREAMING_SNAKE_CASE ={ 'states': states, 'actions': actions, 'rewards': rewards, 'returns_to_go': returns_to_go, 'timesteps': timesteps, 'attention_mask': attention_mask, } return config, inputs_dict @require_torch class A__ ( A__ , A__ , A__ , unittest.TestCase ): A__ = (DecisionTransformerModel,) if is_torch_available() else () A__ = () A__ = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids A__ = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features A__ = False A__ = False A__ = False A__ = False A__ = False A__ = False A__ = False A__ = False A__ = False def A ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' _SCREAMING_SNAKE_CASE =DecisionTransformerModelTester(self ) _SCREAMING_SNAKE_CASE =ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , hidden_size=37 ) def A ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' self.config_tester.run_common_tests() def A ( self : List[Any] ) -> int: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) @slow def A ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE =DecisionTransformerModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def A ( self : Optional[Any] ) -> Optional[Any]: '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _SCREAMING_SNAKE_CASE =model_class(_SCREAMING_SNAKE_CASE ) _SCREAMING_SNAKE_CASE =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _SCREAMING_SNAKE_CASE =[*signature.parameters.keys()] _SCREAMING_SNAKE_CASE =[ 'states', 'actions', 'rewards', 'returns_to_go', 'timesteps', 'attention_mask', ] self.assertListEqual(arg_names[: len(_SCREAMING_SNAKE_CASE )] , _SCREAMING_SNAKE_CASE ) @require_torch class A__ ( unittest.TestCase ): @slow def A ( self : int ) -> Tuple: '''simple docstring''' _SCREAMING_SNAKE_CASE =2 # number of steps of autoregressive prediction we will perform _SCREAMING_SNAKE_CASE =10 # defined by the RL environment, may be normalized _SCREAMING_SNAKE_CASE =DecisionTransformerModel.from_pretrained('edbeeching/decision-transformer-gym-hopper-expert' ) _SCREAMING_SNAKE_CASE =model.to(_SCREAMING_SNAKE_CASE ) _SCREAMING_SNAKE_CASE =model.config torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE =torch.randn(1 , 1 , config.state_dim ).to(device=_SCREAMING_SNAKE_CASE , dtype=torch.floataa ) # env.reset() _SCREAMING_SNAKE_CASE =torch.tensor( [[0.24_27_93, -0.28_69_30_74, 0.8_74_26_13], [0.67_81_52_74, -0.08_10_10_85, -0.12_95_21_47]] , device=_SCREAMING_SNAKE_CASE ) _SCREAMING_SNAKE_CASE =torch.tensor(_SCREAMING_SNAKE_CASE , device=_SCREAMING_SNAKE_CASE , dtype=torch.floataa ).reshape(1 , 1 , 1 ) _SCREAMING_SNAKE_CASE =state _SCREAMING_SNAKE_CASE =torch.zeros(1 , 0 , config.act_dim , device=_SCREAMING_SNAKE_CASE , dtype=torch.floataa ) _SCREAMING_SNAKE_CASE =torch.zeros(1 , 0 , device=_SCREAMING_SNAKE_CASE , dtype=torch.floataa ) _SCREAMING_SNAKE_CASE =torch.tensor(0 , device=_SCREAMING_SNAKE_CASE , dtype=torch.long ).reshape(1 , 1 ) for step in range(_SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE =torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=_SCREAMING_SNAKE_CASE )] , dim=1 ) _SCREAMING_SNAKE_CASE =torch.cat([rewards, torch.zeros(1 , 1 , device=_SCREAMING_SNAKE_CASE )] , dim=1 ) _SCREAMING_SNAKE_CASE =torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE =model( states=_SCREAMING_SNAKE_CASE , actions=_SCREAMING_SNAKE_CASE , rewards=_SCREAMING_SNAKE_CASE , returns_to_go=_SCREAMING_SNAKE_CASE , timesteps=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , return_dict=_SCREAMING_SNAKE_CASE , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1e-4 ) ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE =( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=_SCREAMING_SNAKE_CASE , dtype=torch.floataa ), 1.0, False, {}, ) _SCREAMING_SNAKE_CASE =action_pred[0, -1] _SCREAMING_SNAKE_CASE =torch.cat([states, state] , dim=1 ) _SCREAMING_SNAKE_CASE =returns_to_go[0, -1] - reward _SCREAMING_SNAKE_CASE =torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) _SCREAMING_SNAKE_CASE =torch.cat( [timesteps, torch.ones((1, 1) , device=_SCREAMING_SNAKE_CASE , dtype=torch.long ) * (step + 1)] , dim=1 )
47
'''simple docstring''' def lowercase__ ( __UpperCamelCase = 1000 )-> int: UpperCamelCase = -1 UpperCamelCase = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c UpperCamelCase = (n * n - 2 * a * n) // (2 * n - 2 * a) UpperCamelCase = n - a - b if c * c == (a * a + b * b): UpperCamelCase = a * b * c if candidate >= product: UpperCamelCase = candidate return product if __name__ == "__main__": print(f'{solution() = }')
321
0
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, BlipProcessor, PreTrainedTokenizerFast @require_vision class A_ ( unittest.TestCase ): def lowerCAmelCase ( self : Optional[Any]): __lowerCamelCase : Optional[int] = tempfile.mkdtemp() __lowerCamelCase : Optional[int] = BlipImageProcessor() __lowerCamelCase : List[Any] = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-BertModel') __lowerCamelCase : Optional[int] = BlipProcessor(_a ,_a) processor.save_pretrained(self.tmpdirname) def lowerCAmelCase ( self : List[Any] ,**SCREAMING_SNAKE_CASE__ : Any): return AutoProcessor.from_pretrained(self.tmpdirname ,**_a).tokenizer def lowerCAmelCase ( self : Dict ,**SCREAMING_SNAKE_CASE__ : Tuple): return AutoProcessor.from_pretrained(self.tmpdirname ,**_a).image_processor def lowerCAmelCase ( self : int): shutil.rmtree(self.tmpdirname) def lowerCAmelCase ( self : Dict): __lowerCamelCase : Tuple = [np.random.randint(2_5_5 ,size=(3, 3_0, 4_0_0) ,dtype=np.uinta)] __lowerCamelCase : Dict = [Image.fromarray(np.moveaxis(_a ,0 ,-1)) for x in image_inputs] return image_inputs def lowerCAmelCase ( self : Any): __lowerCamelCase : Union[str, Any] = BlipProcessor(tokenizer=self.get_tokenizer() ,image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) __lowerCamelCase : Tuple = self.get_tokenizer(bos_token='(BOS)' ,eos_token='(EOS)') __lowerCamelCase : Dict = self.get_image_processor(do_normalize=_a ,padding_value=1.0) __lowerCamelCase : Dict = BlipProcessor.from_pretrained( self.tmpdirname ,bos_token='(BOS)' ,eos_token='(EOS)' ,do_normalize=_a ,padding_value=1.0) self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer_add_kwargs.get_vocab()) self.assertIsInstance(processor.tokenizer ,_a) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor ,_a) def lowerCAmelCase ( self : str): __lowerCamelCase : Optional[int] = self.get_image_processor() __lowerCamelCase : Any = self.get_tokenizer() __lowerCamelCase : List[Any] = BlipProcessor(tokenizer=_a ,image_processor=_a) __lowerCamelCase : List[str] = self.prepare_image_inputs() __lowerCamelCase : Optional[int] = image_processor(_a ,return_tensors='np') __lowerCamelCase : Any = processor(images=_a ,return_tensors='np') for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1E-2) def lowerCAmelCase ( self : Any): __lowerCamelCase : List[Any] = self.get_image_processor() __lowerCamelCase : str = self.get_tokenizer() __lowerCamelCase : Optional[int] = BlipProcessor(tokenizer=_a ,image_processor=_a) __lowerCamelCase : str = 'lower newer' __lowerCamelCase : List[str] = processor(text=_a) __lowerCamelCase : Union[str, Any] = tokenizer(_a ,return_token_type_ids=_a) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key]) def lowerCAmelCase ( self : List[str]): __lowerCamelCase : Optional[Any] = self.get_image_processor() __lowerCamelCase : int = self.get_tokenizer() __lowerCamelCase : Union[str, Any] = BlipProcessor(tokenizer=_a ,image_processor=_a) __lowerCamelCase : Union[str, Any] = 'lower newer' __lowerCamelCase : str = self.prepare_image_inputs() __lowerCamelCase : int = processor(text=_a ,images=_a) self.assertListEqual(list(inputs.keys()) ,['pixel_values', 'input_ids', 'attention_mask']) # test if it raises when no input is passed with pytest.raises(_a): processor() def lowerCAmelCase ( self : List[Any]): __lowerCamelCase : Optional[Any] = self.get_image_processor() __lowerCamelCase : Optional[Any] = self.get_tokenizer() __lowerCamelCase : int = BlipProcessor(tokenizer=_a ,image_processor=_a) __lowerCamelCase : List[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __lowerCamelCase : Optional[Any] = processor.batch_decode(_a) __lowerCamelCase : Optional[Any] = tokenizer.batch_decode(_a) self.assertListEqual(_a ,_a) def lowerCAmelCase ( self : Tuple): __lowerCamelCase : int = self.get_image_processor() __lowerCamelCase : Any = self.get_tokenizer() __lowerCamelCase : str = BlipProcessor(tokenizer=_a ,image_processor=_a) __lowerCamelCase : Any = 'lower newer' __lowerCamelCase : Optional[int] = self.prepare_image_inputs() __lowerCamelCase : List[str] = processor(text=_a ,images=_a) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys()) ,['pixel_values', 'input_ids', 'attention_mask'])
350
import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class A_ ( unittest.TestCase ): def lowerCAmelCase ( self : Optional[Any]): # A mock response for an HTTP head request to emulate server down __lowerCamelCase : List[Any] = mock.Mock() __lowerCamelCase : Tuple = 5_0_0 __lowerCamelCase : Dict = {} __lowerCamelCase : List[Any] = HTTPError __lowerCamelCase : int = {} # Download this model to make sure it's in the cache. __lowerCamelCase : Optional[int] = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert') # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' ,return_value=SCREAMING_SNAKE_CASE__) as mock_head: __lowerCamelCase : Tuple = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert') # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def lowerCAmelCase ( self : Any): # A mock response for an HTTP head request to emulate server down __lowerCamelCase : int = mock.Mock() __lowerCamelCase : List[str] = 5_0_0 __lowerCamelCase : Tuple = {} __lowerCamelCase : List[str] = HTTPError __lowerCamelCase : int = {} # Download this model to make sure it's in the cache. __lowerCamelCase : Dict = GPTaTokenizerFast.from_pretrained('gpt2') # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' ,return_value=SCREAMING_SNAKE_CASE__) as mock_head: __lowerCamelCase : Optional[int] = GPTaTokenizerFast.from_pretrained('gpt2') # This check we did call the fake head request mock_head.assert_called() def lowerCAmelCase ( self : str): # This test is for deprecated behavior and can be removed in v5 try: __lowerCamelCase : Tuple = tempfile.mktemp() with open(SCREAMING_SNAKE_CASE__ ,'wb') as f: http_get('https://huggingface.co/albert-base-v1/resolve/main/spiece.model' ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Dict = AlbertTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__) finally: os.remove(SCREAMING_SNAKE_CASE__) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile('tokenizer.json'): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open('tokenizer.json' ,'wb') as f: http_get('https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json' ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Any = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2') # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size ,1_0_0_0) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove('tokenizer.json') def lowerCAmelCase ( self : Optional[Any]): # This test is for deprecated behavior and can be removed in v5 __lowerCamelCase : str = AlbertTokenizer.from_pretrained('https://huggingface.co/albert-base-v1/resolve/main/spiece.model') @is_staging_test class A_ ( unittest.TestCase ): _UpperCAmelCase : Any = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''bla''', '''blou'''] @classmethod def lowerCAmelCase ( cls : Optional[int]): __lowerCamelCase : Optional[int] = TOKEN HfFolder.save_token(SCREAMING_SNAKE_CASE__) @classmethod def lowerCAmelCase ( cls : str): try: delete_repo(token=cls._token ,repo_id='test-tokenizer') except HTTPError: pass try: delete_repo(token=cls._token ,repo_id='valid_org/test-tokenizer-org') except HTTPError: pass try: delete_repo(token=cls._token ,repo_id='test-dynamic-tokenizer') except HTTPError: pass def lowerCAmelCase ( self : List[str]): with tempfile.TemporaryDirectory() as tmp_dir: __lowerCamelCase : Dict = os.path.join(SCREAMING_SNAKE_CASE__ ,'vocab.txt') with open(SCREAMING_SNAKE_CASE__ ,'w' ,encoding='utf-8') as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens])) __lowerCamelCase : Union[str, Any] = BertTokenizer(SCREAMING_SNAKE_CASE__) tokenizer.push_to_hub('test-tokenizer' ,use_auth_token=self._token) __lowerCamelCase : List[str] = BertTokenizer.from_pretrained(F"{USER}/test-tokenizer") self.assertDictEqual(new_tokenizer.vocab ,tokenizer.vocab) # Reset repo delete_repo(token=self._token ,repo_id='test-tokenizer') # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(SCREAMING_SNAKE_CASE__ ,repo_id='test-tokenizer' ,push_to_hub=SCREAMING_SNAKE_CASE__ ,use_auth_token=self._token) __lowerCamelCase : str = BertTokenizer.from_pretrained(F"{USER}/test-tokenizer") self.assertDictEqual(new_tokenizer.vocab ,tokenizer.vocab) def lowerCAmelCase ( self : Any): with tempfile.TemporaryDirectory() as tmp_dir: __lowerCamelCase : Dict = os.path.join(SCREAMING_SNAKE_CASE__ ,'vocab.txt') with open(SCREAMING_SNAKE_CASE__ ,'w' ,encoding='utf-8') as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens])) __lowerCamelCase : Optional[Any] = BertTokenizer(SCREAMING_SNAKE_CASE__) tokenizer.push_to_hub('valid_org/test-tokenizer-org' ,use_auth_token=self._token) __lowerCamelCase : List[Any] = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org') self.assertDictEqual(new_tokenizer.vocab ,tokenizer.vocab) # Reset repo delete_repo(token=self._token ,repo_id='valid_org/test-tokenizer-org') # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( SCREAMING_SNAKE_CASE__ ,repo_id='valid_org/test-tokenizer-org' ,push_to_hub=SCREAMING_SNAKE_CASE__ ,use_auth_token=self._token) __lowerCamelCase : List[Any] = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org') self.assertDictEqual(new_tokenizer.vocab ,tokenizer.vocab) @require_tokenizers def lowerCAmelCase ( self : int): CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __lowerCamelCase : Dict = os.path.join(SCREAMING_SNAKE_CASE__ ,'vocab.txt') with open(SCREAMING_SNAKE_CASE__ ,'w' ,encoding='utf-8') as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens])) __lowerCamelCase : Any = CustomTokenizer(SCREAMING_SNAKE_CASE__) # No fast custom tokenizer tokenizer.push_to_hub('test-dynamic-tokenizer' ,use_auth_token=self._token) __lowerCamelCase : Optional[int] = AutoTokenizer.from_pretrained(F"{USER}/test-dynamic-tokenizer" ,trust_remote_code=SCREAMING_SNAKE_CASE__) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ ,'CustomTokenizer') # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __lowerCamelCase : List[str] = os.path.join(SCREAMING_SNAKE_CASE__ ,'vocab.txt') with open(SCREAMING_SNAKE_CASE__ ,'w' ,encoding='utf-8') as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens])) __lowerCamelCase : Dict = BertTokenizerFast.from_pretrained(SCREAMING_SNAKE_CASE__) bert_tokenizer.save_pretrained(SCREAMING_SNAKE_CASE__) __lowerCamelCase : List[Any] = CustomTokenizerFast.from_pretrained(SCREAMING_SNAKE_CASE__) tokenizer.push_to_hub('test-dynamic-tokenizer' ,use_auth_token=self._token) __lowerCamelCase : Optional[Any] = AutoTokenizer.from_pretrained(F"{USER}/test-dynamic-tokenizer" ,trust_remote_code=SCREAMING_SNAKE_CASE__) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ ,'CustomTokenizerFast') __lowerCamelCase : Optional[int] = AutoTokenizer.from_pretrained( F"{USER}/test-dynamic-tokenizer" ,use_fast=SCREAMING_SNAKE_CASE__ ,trust_remote_code=SCREAMING_SNAKE_CASE__) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ ,'CustomTokenizer') class A_ ( unittest.TestCase ): def lowerCAmelCase ( self : List[Any]): __lowerCamelCase : List[str] = Trie() trie.add('Hello 友達') self.assertEqual(trie.data ,{'H': {'e': {'l': {'l': {'o': {' ': {'友': {'達': {'': 1}}}}}}}}}) trie.add('Hello') trie.data self.assertEqual(trie.data ,{'H': {'e': {'l': {'l': {'o': {'': 1, ' ': {'友': {'達': {'': 1}}}}}}}}}) def lowerCAmelCase ( self : Tuple): __lowerCamelCase : str = Trie() self.assertEqual(trie.split('[CLS] This is a extra_id_100') ,['[CLS] This is a extra_id_100']) trie.add('[CLS]') trie.add('extra_id_1') trie.add('extra_id_100') self.assertEqual(trie.split('[CLS] This is a extra_id_100') ,['[CLS]', ' This is a ', 'extra_id_100']) def lowerCAmelCase ( self : Optional[Any]): __lowerCamelCase : Dict = Trie() trie.add('A') self.assertEqual(trie.split('ABC') ,['A', 'BC']) self.assertEqual(trie.split('BCA') ,['BC', 'A']) def lowerCAmelCase ( self : Dict): __lowerCamelCase : Any = Trie() trie.add('TOKEN]') trie.add('[SPECIAL_TOKEN]') self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]') ,['This is something ', '[SPECIAL_TOKEN]']) def lowerCAmelCase ( self : int): __lowerCamelCase : List[Any] = Trie() trie.add('A') trie.add('P') trie.add('[SPECIAL_TOKEN]') self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]') ,['This is something ', '[SPECIAL_TOKEN]']) def lowerCAmelCase ( self : Optional[int]): __lowerCamelCase : Any = Trie() trie.add('AB') trie.add('B') trie.add('C') self.assertEqual(trie.split('ABC') ,['AB', 'C']) def lowerCAmelCase ( self : Optional[int]): __lowerCamelCase : Dict = Trie() trie.add('ABC') trie.add('B') trie.add('CD') self.assertEqual(trie.split('ABCD') ,['ABC', 'D']) def lowerCAmelCase ( self : List[Any]): # Even if the offsets are wrong, we necessarily output correct string # parts. __lowerCamelCase : List[Any] = Trie() __lowerCamelCase : Any = trie.cut_text('ABC' ,[0, 0, 2, 1, 2, 3]) self.assertEqual(SCREAMING_SNAKE_CASE__ ,['AB', 'C'])
113
0
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird import BigBirdTokenizer else: UpperCamelCase__ = None UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} UpperCamelCase__ = { """vocab_file""": { """google/bigbird-roberta-base""": """https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model""", """google/bigbird-roberta-large""": ( """https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model""" ), """google/bigbird-base-trivia-itc""": ( """https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model""" ), }, """tokenizer_file""": { """google/bigbird-roberta-base""": ( """https://huggingface.co/google/bigbird-roberta-base/resolve/main/tokenizer.json""" ), """google/bigbird-roberta-large""": ( """https://huggingface.co/google/bigbird-roberta-large/resolve/main/tokenizer.json""" ), """google/bigbird-base-trivia-itc""": ( """https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/tokenizer.json""" ), }, } UpperCamelCase__ = { """google/bigbird-roberta-base""": 4096, """google/bigbird-roberta-large""": 4096, """google/bigbird-base-trivia-itc""": 4096, } UpperCamelCase__ = """▁""" class a__ ( UpperCamelCase__ ): _a : Union[str, Any] = VOCAB_FILES_NAMES _a : Any = PRETRAINED_VOCAB_FILES_MAP _a : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _a : Tuple = BigBirdTokenizer _a : Any = ["""input_ids""", """attention_mask"""] _a : List[int] = [] def __init__( self , _A=None , _A=None , _A="<unk>" , _A="<s>" , _A="</s>" , _A="<pad>" , _A="[SEP]" , _A="[MASK]" , _A="[CLS]" , **_A , ): """simple docstring""" __lowerCAmelCase = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else bos_token __lowerCAmelCase = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else eos_token __lowerCAmelCase = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else unk_token __lowerCAmelCase = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else pad_token __lowerCAmelCase = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else cls_token __lowerCAmelCase = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it __lowerCAmelCase = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , **lowerCAmelCase__ , ) __lowerCAmelCase = vocab_file __lowerCAmelCase = False if not self.vocab_file else True def __SCREAMING_SNAKE_CASE( self , _A , _A = None ): """simple docstring""" __lowerCAmelCase = [self.sep_token_id] __lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __SCREAMING_SNAKE_CASE( self , _A , _A = None , _A = False ): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase__ )) + [1] return [1] + ([0] * len(lowerCAmelCase__ )) + [1] + ([0] * len(lowerCAmelCase__ )) + [1] def __SCREAMING_SNAKE_CASE( self , _A , _A = None ): """simple docstring""" __lowerCAmelCase = [self.sep_token_id] __lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __SCREAMING_SNAKE_CASE( self , _A , _A = None ): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCAmelCase = os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
92
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase : Dict = logging.get_logger(__name__) UpperCAmelCase : Tuple = { """caidas/swin2sr-classicalsr-x2-64""": ( """https://huggingface.co/caidas/swin2sr-classicalsr-x2-64/resolve/main/config.json""" ), } class __lowerCAmelCase ( UpperCamelCase__): _lowercase : Any = """swin2sr""" _lowercase : Tuple = { """hidden_size""": """embed_dim""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , lowerCAmelCase__=6_4 , lowerCAmelCase__=1 , lowerCAmelCase__=3 , lowerCAmelCase__=1_8_0 , lowerCAmelCase__=[6, 6, 6, 6, 6, 6] , lowerCAmelCase__=[6, 6, 6, 6, 6, 6] , lowerCAmelCase__=8 , lowerCAmelCase__=2.0 , lowerCAmelCase__=True , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.1 , lowerCAmelCase__="gelu" , lowerCAmelCase__=False , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-5 , lowerCAmelCase__=2 , lowerCAmelCase__=1.0 , lowerCAmelCase__="1conv" , lowerCAmelCase__="pixelshuffle" , **lowerCAmelCase__ , ) -> int: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) a__ : Optional[Any] =image_size a__ : Dict =patch_size a__ : Tuple =num_channels a__ : Union[str, Any] =embed_dim a__ : Optional[Any] =depths a__ : List[str] =len(lowerCAmelCase__ ) a__ : Any =num_heads a__ : Any =window_size a__ : str =mlp_ratio a__ : List[str] =qkv_bias a__ : Dict =hidden_dropout_prob a__ : List[str] =attention_probs_dropout_prob a__ : Dict =drop_path_rate a__ : Optional[Any] =hidden_act a__ : Union[str, Any] =use_absolute_embeddings a__ : Optional[Any] =layer_norm_eps a__ : List[Any] =initializer_range a__ : int =upscale a__ : Optional[int] =img_range a__ : Any =resi_connection a__ : Optional[Any] =upsampler
95
0
"""simple docstring""" from math import factorial, pi def UpperCAmelCase__ ( lowerCAmelCase__ :float , lowerCAmelCase__ :int = 3_0 ) -> float: '''simple docstring''' if not isinstance(lowerCAmelCase__ , (int, float) ): raise ValueError("""maclaurin_sin() requires either an int or float for theta""" ) if not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) or accuracy <= 0: raise ValueError("""maclaurin_sin() requires a positive int for accuracy""" ) lowercase = float(lowerCAmelCase__ ) lowercase = theta // (2 * pi) theta -= 2 * div * pi return sum( (-1) ** r * theta ** (2 * r + 1) / factorial(2 * r + 1 ) for r in range(lowerCAmelCase__ ) ) def UpperCAmelCase__ ( lowerCAmelCase__ :float , lowerCAmelCase__ :int = 3_0 ) -> float: '''simple docstring''' if not isinstance(lowerCAmelCase__ , (int, float) ): raise ValueError("""maclaurin_cos() requires either an int or float for theta""" ) if not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) or accuracy <= 0: raise ValueError("""maclaurin_cos() requires a positive int for accuracy""" ) lowercase = float(lowerCAmelCase__ ) lowercase = theta // (2 * pi) theta -= 2 * div * pi return sum((-1) ** r * theta ** (2 * r) / factorial(2 * r ) for r in range(lowerCAmelCase__ ) ) if __name__ == "__main__": import doctest doctest.testmod() print(maclaurin_sin(1_0)) print(maclaurin_sin(-1_0)) print(maclaurin_sin(1_0, 1_5)) print(maclaurin_sin(-1_0, 1_5)) print(maclaurin_cos(5)) print(maclaurin_cos(-5)) print(maclaurin_cos(1_0, 1_5)) print(maclaurin_cos(-1_0, 1_5))
32
"""simple docstring""" import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def UpperCAmelCase__ ( lowerCAmelCase__ :Any , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Dict , lowerCAmelCase__ :Any ) -> Dict: '''simple docstring''' lowercase = s.rsplit(lowerCAmelCase__ , lowerCAmelCase__ ) return new.join(lowerCAmelCase__ ) def UpperCAmelCase__ ( lowerCAmelCase__ :str ) -> List[Any]: '''simple docstring''' return sum(param.float().sum() if """encoder.embeddings""" not in key else 0 for key, param in state_dict.items() ) def UpperCAmelCase__ ( lowerCAmelCase__ :Union[str, Any] ) -> List[str]: '''simple docstring''' lowercase = {} lowercase = ["""group_1""", """group_2""", """group_3""", """group_4"""] for key, value in state_dict.items(): for group_key in group_keys: if group_key in key: lowercase = key.replace(f'{group_key}.' , f'{group_key}.group.' ) if "res_path" in key: lowercase = key.replace("""res_path.""" , """res_path.path.""" ) if key.endswith(""".w""" ): lowercase = rreplace(lowerCAmelCase__ , """.w""" , """.weight""" , 1 ) if key.endswith(""".b""" ): lowercase = rreplace(lowerCAmelCase__ , """.b""" , """.bias""" , 1 ) lowercase = value.float() return upgrade @torch.no_grad() def UpperCAmelCase__ ( lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Union[str, Any]=None , lowerCAmelCase__ :Any=True ) -> Any: '''simple docstring''' from dall_e import Encoder lowercase = Encoder() if os.path.exists(lowerCAmelCase__ ): lowercase = torch.load(lowerCAmelCase__ ) else: lowercase = torch.hub.load_state_dict_from_url(lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): lowercase = ckpt.state_dict() encoder.load_state_dict(lowerCAmelCase__ ) if config_path is not None: lowercase = FlavaImageCodebookConfig.from_pretrained(lowerCAmelCase__ ) else: lowercase = FlavaImageCodebookConfig() lowercase = FlavaImageCodebook(lowerCAmelCase__ ).eval() lowercase = encoder.state_dict() lowercase = upgrade_state_dict(lowerCAmelCase__ ) hf_model.load_state_dict(lowerCAmelCase__ ) lowercase = hf_model.state_dict() lowercase = count_parameters(lowerCAmelCase__ ) lowercase = count_parameters(lowerCAmelCase__ ) assert torch.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1e-3 ) if save_checkpoint: hf_model.save_pretrained(lowerCAmelCase__ ) else: return hf_state_dict if __name__ == "__main__": __lowerCAmelCase : Tuple =argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to flava checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") __lowerCAmelCase : Any =parser.parse_args() convert_dalle_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
32
1
"""simple docstring""" import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class lowerCamelCase : '''simple docstring''' @staticmethod def _a (*_lowerCamelCase , **_lowerCamelCase ): """simple docstring""" pass @is_pipeline_test @require_vision class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @require_torch def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[int] = pipeline( model="""hf-internal-testing/tiny-random-clip-zero-shot-image-classification""" , ) UpperCAmelCase__ : int = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) UpperCAmelCase__ : Any = image_classifier(__snake_case , candidate_labels=["""a""", """b""", """c"""] ) # The floating scores are so close, we enter floating error approximation and the order is not guaranteed across # python and torch versions. self.assertIn( nested_simplify(__snake_case ) , [ [{"""score""": 0.333, """label""": """a"""}, {"""score""": 0.333, """label""": """b"""}, {"""score""": 0.333, """label""": """c"""}], [{"""score""": 0.333, """label""": """a"""}, {"""score""": 0.333, """label""": """c"""}, {"""score""": 0.333, """label""": """b"""}], ] , ) UpperCAmelCase__ : Union[str, Any] = image_classifier([image] * 5 , candidate_labels=["""A""", """B""", """C"""] , batch_size=2 ) self.assertEqual( nested_simplify(__snake_case ) , [ [ {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, ], [ {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, ], [ {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, ], [ {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, ], [ {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, ], ] , ) @require_tf def _a (self ): """simple docstring""" UpperCAmelCase__ : Any = pipeline( model="""hf-internal-testing/tiny-random-clip-zero-shot-image-classification""" , framework="""tf""" ) UpperCAmelCase__ : Tuple = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) UpperCAmelCase__ : Union[str, Any] = image_classifier(__snake_case , candidate_labels=["""a""", """b""", """c"""] ) self.assertEqual( nested_simplify(__snake_case ) , [{"""score""": 0.333, """label""": """a"""}, {"""score""": 0.333, """label""": """b"""}, {"""score""": 0.333, """label""": """c"""}] , ) UpperCAmelCase__ : Tuple = image_classifier([image] * 5 , candidate_labels=["""A""", """B""", """C"""] , batch_size=2 ) self.assertEqual( nested_simplify(__snake_case ) , [ [ {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, ], [ {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, ], [ {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, ], [ {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, ], [ {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, {"""score""": 0.333, """label""": ANY(__snake_case )}, ], ] , ) @slow @require_torch def _a (self ): """simple docstring""" UpperCAmelCase__ : Any = pipeline( task="""zero-shot-image-classification""" , model="""openai/clip-vit-base-patch32""" , ) # This is an image of 2 cats with remotes and no planes UpperCAmelCase__ : Optional[int] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) UpperCAmelCase__ : Tuple = image_classifier(__snake_case , candidate_labels=["""cat""", """plane""", """remote"""] ) self.assertEqual( nested_simplify(__snake_case ) , [ {"""score""": 0.511, """label""": """remote"""}, {"""score""": 0.485, """label""": """cat"""}, {"""score""": 0.004, """label""": """plane"""}, ] , ) UpperCAmelCase__ : Any = image_classifier([image] * 5 , candidate_labels=["""cat""", """plane""", """remote"""] , batch_size=2 ) self.assertEqual( nested_simplify(__snake_case ) , [ [ {"""score""": 0.511, """label""": """remote"""}, {"""score""": 0.485, """label""": """cat"""}, {"""score""": 0.004, """label""": """plane"""}, ], ] * 5 , ) @slow @require_tf def _a (self ): """simple docstring""" UpperCAmelCase__ : List[str] = pipeline( task="""zero-shot-image-classification""" , model="""openai/clip-vit-base-patch32""" , framework="""tf""" ) # This is an image of 2 cats with remotes and no planes UpperCAmelCase__ : str = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) UpperCAmelCase__ : Dict = image_classifier(__snake_case , candidate_labels=["""cat""", """plane""", """remote"""] ) self.assertEqual( nested_simplify(__snake_case ) , [ {"""score""": 0.511, """label""": """remote"""}, {"""score""": 0.485, """label""": """cat"""}, {"""score""": 0.004, """label""": """plane"""}, ] , ) UpperCAmelCase__ : int = image_classifier([image] * 5 , candidate_labels=["""cat""", """plane""", """remote"""] , batch_size=2 ) self.assertEqual( nested_simplify(__snake_case ) , [ [ {"""score""": 0.511, """label""": """remote"""}, {"""score""": 0.485, """label""": """cat"""}, {"""score""": 0.004, """label""": """plane"""}, ], ] * 5 , )
171
'''simple docstring''' import itertools from dataclasses import dataclass from typing import Optional import pandas as pd import pyarrow as pa import datasets from datasets.table import table_cast @dataclass class UpperCAmelCase ( datasets.BuilderConfig ): _lowercase: Optional[datasets.Features] = None class UpperCAmelCase ( datasets.ArrowBasedBuilder ): _lowercase: Tuple = PandasConfig def lowercase__ ( self : Optional[Any] ) -> str: return datasets.DatasetInfo(features=self.config.features ) def lowercase__ ( self : List[str] , __snake_case : Dict ) -> int: if not self.config.data_files: raise ValueError(f"At least one data file must be specified, but got data_files={self.config.data_files}" ) _lowerCAmelCase = dl_manager.download_and_extract(self.config.data_files ) if isinstance(__snake_case , (str, list, tuple) ): _lowerCAmelCase = data_files if isinstance(__snake_case , __snake_case ): _lowerCAmelCase = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive _lowerCAmelCase = [dl_manager.iter_files(__snake_case ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""files""": files} )] _lowerCAmelCase = [] for split_name, files in data_files.items(): if isinstance(__snake_case , __snake_case ): _lowerCAmelCase = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive _lowerCAmelCase = [dl_manager.iter_files(__snake_case ) for file in files] splits.append(datasets.SplitGenerator(name=__snake_case , gen_kwargs={"""files""": files} ) ) return splits def lowercase__ ( self : List[Any] , __snake_case : pa.Table ) -> pa.Table: if self.config.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example _lowerCAmelCase = table_cast(__snake_case , self.config.features.arrow_schema ) return pa_table def lowercase__ ( self : Dict , __snake_case : Optional[Any] ) -> Any: for i, file in enumerate(itertools.chain.from_iterable(__snake_case ) ): with open(__snake_case , """rb""" ) as f: _lowerCAmelCase = pa.Table.from_pandas(pd.read_pickle(__snake_case ) ) yield i, self._cast_table(__snake_case )
70
0
from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig lowerCAmelCase_ = logging.get_logger(__name__) # General docstring lowerCAmelCase_ = 'MobileNetV1Config' # Base docstring lowerCAmelCase_ = 'google/mobilenet_v1_1.0_224' lowerCAmelCase_ = [1, 10_24, 7, 7] # Image classification docstring lowerCAmelCase_ = 'google/mobilenet_v1_1.0_224' lowerCAmelCase_ = 'tabby, tabby cat' lowerCAmelCase_ = [ 'google/mobilenet_v1_1.0_224', 'google/mobilenet_v1_0.75_192', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def snake_case( __magic_name__ , __magic_name__ , __magic_name__=None ) -> List[str]: '''simple docstring''' lowercase : Any = {} if isinstance(__magic_name__ , __magic_name__ ): lowercase : Tuple = model.mobilenet_va else: lowercase : int = model lowercase : str = '''MobilenetV1/Conv2d_0/''' lowercase : Any = backbone.conv_stem.convolution.weight lowercase : List[str] = backbone.conv_stem.normalization.bias lowercase : str = backbone.conv_stem.normalization.weight lowercase : List[Any] = backbone.conv_stem.normalization.running_mean lowercase : Dict = backbone.conv_stem.normalization.running_var for i in range(13 ): lowercase : Optional[int] = i + 1 lowercase : Tuple = i * 2 lowercase : Dict = backbone.layer[pt_index] lowercase : List[str] = F"""MobilenetV1/Conv2d_{tf_index}_depthwise/""" lowercase : str = pointer.convolution.weight lowercase : Optional[int] = pointer.normalization.bias lowercase : List[Any] = pointer.normalization.weight lowercase : Optional[int] = pointer.normalization.running_mean lowercase : Union[str, Any] = pointer.normalization.running_var lowercase : Any = backbone.layer[pt_index + 1] lowercase : Tuple = F"""MobilenetV1/Conv2d_{tf_index}_pointwise/""" lowercase : List[str] = pointer.convolution.weight lowercase : str = pointer.normalization.bias lowercase : int = pointer.normalization.weight lowercase : str = pointer.normalization.running_mean lowercase : Any = pointer.normalization.running_var if isinstance(__magic_name__ , __magic_name__ ): lowercase : List[str] = '''MobilenetV1/Logits/Conv2d_1c_1x1/''' lowercase : Any = model.classifier.weight lowercase : str = model.classifier.bias return tf_to_pt_map def snake_case( __magic_name__ , __magic_name__ , __magic_name__ ) -> Any: '''simple docstring''' try: import numpy as np import tensorflow as tf except ImportError: logger.error( '''Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see ''' '''https://www.tensorflow.org/install/ for installation instructions.''' ) raise # Load weights from TF model lowercase : List[str] = tf.train.list_variables(__magic_name__ ) lowercase : Union[str, Any] = {} for name, shape in init_vars: logger.info(F"""Loading TF weight {name} with shape {shape}""" ) lowercase : int = tf.train.load_variable(__magic_name__ , __magic_name__ ) lowercase : str = array # Build TF to PyTorch weights loading map lowercase : Dict = _build_tf_to_pytorch_map(__magic_name__ , __magic_name__ , __magic_name__ ) for name, pointer in tf_to_pt_map.items(): logger.info(F"""Importing {name}""" ) if name not in tf_weights: logger.info(F"""{name} not in tf pre-trained weights, skipping""" ) continue lowercase : Any = tf_weights[name] if "depthwise_weights" in name: logger.info('''Transposing depthwise''' ) lowercase : str = np.transpose(__magic_name__ , (2, 3, 0, 1) ) elif "weights" in name: logger.info('''Transposing''' ) if len(pointer.shape ) == 2: # copying into linear layer lowercase : Tuple = array.squeeze().transpose() else: lowercase : Optional[Any] = np.transpose(__magic_name__ , (3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F"""Pointer shape {pointer.shape} and array shape {array.shape} mismatched""" ) logger.info(F"""Initialize PyTorch weight {name} {array.shape}""" ) lowercase : Union[str, Any] = torch.from_numpy(__magic_name__ ) tf_weights.pop(__magic_name__ , __magic_name__ ) tf_weights.pop(name + '''/RMSProp''' , __magic_name__ ) tf_weights.pop(name + '''/RMSProp_1''' , __magic_name__ ) tf_weights.pop(name + '''/ExponentialMovingAverage''' , __magic_name__ ) logger.info(F"""Weights not copied to PyTorch model: {', '.join(tf_weights.keys() )}""" ) return model def snake_case( __magic_name__ , __magic_name__ ) -> torch.Tensor: '''simple docstring''' lowercase , lowercase : Dict = features.shape[-2:] lowercase , lowercase : List[Any] = conv_layer.stride lowercase , lowercase : int = conv_layer.kernel_size if in_height % stride_height == 0: lowercase : Tuple = max(kernel_height - stride_height , 0 ) else: lowercase : Any = max(kernel_height - (in_height % stride_height) , 0 ) if in_width % stride_width == 0: lowercase : Union[str, Any] = max(kernel_width - stride_width , 0 ) else: lowercase : str = max(kernel_width - (in_width % stride_width) , 0 ) lowercase : Union[str, Any] = pad_along_width // 2 lowercase : str = pad_along_width - pad_left lowercase : str = pad_along_height // 2 lowercase : Union[str, Any] = pad_along_height - pad_top lowercase : List[str] = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(__magic_name__ , __magic_name__ , '''constant''' , 0.0 ) class _A ( nn.Module ): def __init__( self : Any , _A : MobileNetVaConfig , _A : int , _A : int , _A : int , _A : Optional[int] = 1 , _A : Optional[int] = 1 , _A : bool = False , _A : Optional[bool] = True , _A : Optional[bool or str] = True , ) -> None: """simple docstring""" super().__init__() lowercase : Any = config if in_channels % groups != 0: raise ValueError(f"""Input channels ({in_channels}) are not divisible by {groups} groups.""" ) if out_channels % groups != 0: raise ValueError(f"""Output channels ({out_channels}) are not divisible by {groups} groups.""" ) lowercase : int = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) lowercase : int = nn.Convad( in_channels=_A , out_channels=_A , kernel_size=_A , stride=_A , padding=_A , groups=_A , bias=_A , padding_mode='''zeros''' , ) if use_normalization: lowercase : Any = nn.BatchNormad( num_features=_A , eps=config.layer_norm_eps , momentum=0.9_997 , affine=_A , track_running_stats=_A , ) else: lowercase : Tuple = None if use_activation: if isinstance(_A , _A ): lowercase : Optional[Any] = ACTaFN[use_activation] elif isinstance(config.hidden_act , _A ): lowercase : Tuple = ACTaFN[config.hidden_act] else: lowercase : Any = config.hidden_act else: lowercase : Dict = None def __a ( self : Tuple , _A : torch.Tensor ) -> torch.Tensor: """simple docstring""" if self.config.tf_padding: lowercase : int = apply_tf_padding(_A , self.convolution ) lowercase : str = self.convolution(_A ) if self.normalization is not None: lowercase : str = self.normalization(_A ) if self.activation is not None: lowercase : str = self.activation(_A ) return features class _A ( _lowerCamelCase ): _UpperCamelCase : Dict = MobileNetVaConfig _UpperCamelCase : Tuple = load_tf_weights_in_mobilenet_va _UpperCamelCase : List[Any] = '''mobilenet_v1''' _UpperCamelCase : Optional[Any] = '''pixel_values''' _UpperCamelCase : Dict = False def __a ( self : int , _A : Union[nn.Linear, nn.Convad] ) -> None: """simple docstring""" if isinstance(_A , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(_A , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) lowerCAmelCase_ = R'\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it\n as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n' lowerCAmelCase_ = R'\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`MobileNetV1ImageProcessor.__call__`] for details.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( '''The bare MobileNetV1 model outputting raw hidden-states without any specific head on top.''' , _lowerCamelCase , ) class _A ( _lowerCamelCase ): def __init__( self : Optional[Any] , _A : MobileNetVaConfig , _A : bool = True ) -> Union[str, Any]: """simple docstring""" super().__init__(_A ) lowercase : Optional[Any] = config lowercase : Tuple = 32 lowercase : Union[str, Any] = max(int(depth * config.depth_multiplier ) , config.min_depth ) lowercase : Union[str, Any] = MobileNetVaConvLayer( _A , in_channels=config.num_channels , out_channels=_A , kernel_size=3 , stride=2 , ) lowercase : Any = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] lowercase : Optional[Any] = nn.ModuleList() for i in range(13 ): lowercase : Union[str, Any] = out_channels if strides[i] == 2 or i == 0: depth *= 2 lowercase : Optional[Any] = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( _A , in_channels=_A , out_channels=_A , kernel_size=3 , stride=strides[i] , groups=_A , ) ) self.layer.append( MobileNetVaConvLayer( _A , in_channels=_A , out_channels=_A , kernel_size=1 , ) ) lowercase : str = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def __a ( self : Any , _A : Optional[Any] ) -> Any: """simple docstring""" raise NotImplementedError @add_start_docstrings_to_model_forward(_A ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=_A , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def __a ( self : str , _A : Optional[torch.Tensor] = None , _A : Optional[bool] = None , _A : Optional[bool] = None , ) -> Union[tuple, BaseModelOutputWithPoolingAndNoAttention]: """simple docstring""" lowercase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase : int = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('''You have to specify pixel_values''' ) lowercase : Dict = self.conv_stem(_A ) lowercase : List[Any] = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): lowercase : List[str] = layer_module(_A ) if output_hidden_states: lowercase : Any = all_hidden_states + (hidden_states,) lowercase : Tuple = hidden_states if self.pooler is not None: lowercase : Any = torch.flatten(self.pooler(_A ) , start_dim=1 ) else: lowercase : Optional[int] = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=_A , pooler_output=_A , hidden_states=_A , ) @add_start_docstrings( ''' MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. ''' , _lowerCamelCase , ) class _A ( _lowerCamelCase ): def __init__( self : Optional[Any] , _A : MobileNetVaConfig ) -> None: """simple docstring""" super().__init__(_A ) lowercase : Dict = config.num_labels lowercase : List[Any] = MobileNetVaModel(_A ) lowercase : Dict = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head lowercase : Union[str, Any] = nn.Dropout(config.classifier_dropout_prob , inplace=_A ) lowercase : List[str] = nn.Linear(_A , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_A ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=_A , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def __a ( self : Dict , _A : Optional[torch.Tensor] = None , _A : Optional[bool] = None , _A : Optional[torch.Tensor] = None , _A : Optional[bool] = None , ) -> Union[tuple, ImageClassifierOutputWithNoAttention]: """simple docstring""" lowercase : Dict = return_dict if return_dict is not None else self.config.use_return_dict lowercase : str = self.mobilenet_va(_A , output_hidden_states=_A , return_dict=_A ) lowercase : Any = outputs.pooler_output if return_dict else outputs[1] lowercase : Tuple = self.classifier(self.dropout(_A ) ) lowercase : Dict = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowercase : Union[str, Any] = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowercase : Optional[Any] = '''single_label_classification''' else: lowercase : Dict = '''multi_label_classification''' if self.config.problem_type == "regression": lowercase : int = MSELoss() if self.num_labels == 1: lowercase : int = loss_fct(logits.squeeze() , labels.squeeze() ) else: lowercase : List[str] = loss_fct(_A , _A ) elif self.config.problem_type == "single_label_classification": lowercase : Union[str, Any] = CrossEntropyLoss() lowercase : Tuple = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": lowercase : Optional[int] = BCEWithLogitsLoss() lowercase : Dict = loss_fct(_A , _A ) if not return_dict: lowercase : Optional[int] = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=_A , logits=_A , hidden_states=outputs.hidden_states , )
116
class _A : # Public class to implement a graph def __init__( self : List[Any] , _A : int , _A : int , _A : list[list[bool]] ) -> None: """simple docstring""" lowercase : Tuple = row lowercase : Union[str, Any] = col lowercase : int = graph def __a ( self : List[Any] , _A : int , _A : int , _A : list[list[bool]] ) -> bool: """simple docstring""" return ( 0 <= i < self.ROW and 0 <= j < self.COL and not visited[i][j] and self.graph[i][j] ) def __a ( self : int , _A : int , _A : int , _A : list[list[bool]] ) -> None: """simple docstring""" lowercase : List[str] = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order lowercase : Dict = [-1, 0, 1, -1, 1, -1, 0, 1] lowercase : Dict = True # Make those cells visited for k in range(8 ): if self.is_safe(i + row_nbr[k] , j + col_nbr[k] , _A ): self.diffs(i + row_nbr[k] , j + col_nbr[k] , _A ) def __a ( self : List[str] ) -> int: # And finally, count all islands. """simple docstring""" lowercase : List[str] = [[False for j in range(self.COL )] for i in range(self.ROW )] lowercase : Optional[Any] = 0 for i in range(self.ROW ): for j in range(self.COL ): if visited[i][j] is False and self.graph[i][j] == 1: self.diffs(_A , _A , _A ) count += 1 return count
116
1
import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[Any]: # Initialise PyTorch model _lowercase : Union[str, Any] = FunnelConfig.from_json_file(lowerCamelCase_ ) print(F'''Building PyTorch model from configuration: {config}''' ) _lowercase : Tuple = FunnelBaseModel(lowerCamelCase_ ) if base_model else FunnelModel(lowerCamelCase_ ) # Load weights from tf checkpoint load_tf_weights_in_funnel(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , lowerCamelCase_ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--base_model", action="store_true", help="Whether you want just the base model (no decoder) or not." ) SCREAMING_SNAKE_CASE : Tuple = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path, args.base_model )
21
import random import unittest import numpy as np import torch from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionUpscalePipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class _lowerCamelCase( _a, unittest.TestCase ): # TODO: is there an appropriate internal test set? lowercase_ : int = """ssube/stable-diffusion-x4-upscaler-onnx""" def UpperCamelCase ( self, lowerCamelCase=0) -> Union[str, Any]: """simple docstring""" _lowercase : Dict = floats_tensor((1, 3, 1_28, 1_28), rng=random.Random(lowerCamelCase)) _lowercase : Union[str, Any] = torch.manual_seed(lowerCamelCase) _lowercase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Tuple = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Dict = self.get_dummy_inputs() _lowercase : Optional[int] = pipe(**lowerCamelCase).images _lowercase : Optional[int] = image[0, -3:, -3:, -1].flatten() # started as 128, should now be 512 assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Union[str, Any] = np.array( [0.6_9_7_4_7_8_2, 0.6_8_9_0_2_0_9_3, 0.7_0_1_3_5_8_8_5, 0.7_5_8_3_6_1_8, 0.7_8_0_4_5_4_5, 0.7_8_5_4_9_1_2, 0.7_8_6_6_7_4_2_6, 0.7_8_7_4_3_8_6_3, 0.7_8_0_7_0_2_2_3]) assert np.abs(image_slice - expected_slice).max() < 1E-1 def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : int = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : str = PNDMScheduler.from_config(pipe.scheduler.config, skip_prk_steps=lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[str] = self.get_dummy_inputs() _lowercase : List[Any] = pipe(**lowerCamelCase).images _lowercase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : int = np.array( [0.6_8_9_8_8_9_2, 0.5_9_2_4_0_5_5_6, 0.5_2_4_9_9_5_2_7, 0.5_8_8_6_6_2_1_5, 0.5_2_2_5_8_2_3_5, 0.5_2_5_7_2_7_1_5, 0.6_2_4_1_4_4_7_3, 0.6_1_7_4_3_8_7, 0.6_2_1_4_9_6_4]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-1 def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Dict = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Optional[int] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Optional[int] = self.get_dummy_inputs() _lowercase : Union[str, Any] = pipe(**lowerCamelCase).images _lowercase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Optional[int] = np.array( [0.7_6_5_9_2_7_8, 0.7_6_4_3_7_6_6_4, 0.7_5_5_7_9_1_0_7, 0.7_6_9_1_1_1_6, 0.7_7_6_6_6_9_8_6, 0.7_7_2_7_6_7_2, 0.7_7_5_8_6_6_4, 0.7_8_1_2_2_2_6, 0.7_6_9_4_2_5_1_5]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-1 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : List[str] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : List[str] = EulerDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Dict = self.get_dummy_inputs() _lowercase : Optional[Any] = pipe(**lowerCamelCase).images _lowercase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Union[str, Any] = np.array( [0.6_9_7_4_7_8_2, 0.6_8_9_0_2_0_9_3, 0.7_0_1_3_5_8_8_5, 0.7_5_8_3_6_1_8, 0.7_8_0_4_5_4_5, 0.7_8_5_4_9_1_2, 0.7_8_6_6_7_4_2_6, 0.7_8_7_4_3_8_6_3, 0.7_8_0_7_0_2_2_3]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-1 def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : Tuple = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Any = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Any = self.get_dummy_inputs() _lowercase : List[str] = pipe(**lowerCamelCase).images _lowercase : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Union[str, Any] = np.array( [0.7_7_4_2_4_4_9_6, 0.7_7_3_6_0_1, 0.7_6_4_5_2_8_8, 0.7_7_6_9_5_9_8, 0.7_7_7_2_7_3_9, 0.7_7_3_8_6_8_8, 0.7_8_1_8_7_2_3_3, 0.7_7_8_7_9_5_8_4, 0.7_6_7_0_4_3]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class _lowerCamelCase( unittest.TestCase ): @property def UpperCamelCase ( self) -> List[Any]: """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : Union[str, Any] = ort.SessionOptions() _lowercase : str = False return options def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : Optional[int] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/img2img/sketch-mountains-input.jpg') _lowercase : int = init_image.resize((1_28, 1_28)) # using the PNDM scheduler by default _lowercase : Tuple = OnnxStableDiffusionUpscalePipeline.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx', provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Optional[int] = 'A fantasy landscape, trending on artstation' _lowercase : List[Any] = torch.manual_seed(0) _lowercase : str = pipe( prompt=lowerCamelCase, image=lowerCamelCase, guidance_scale=7.5, num_inference_steps=10, generator=lowerCamelCase, output_type='np', ) _lowercase : List[Any] = output.images _lowercase : List[Any] = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 5_12, 3) _lowercase : List[Any] = np.array([0.4_8_8_3, 0.4_9_4_7, 0.4_9_8_0, 0.4_9_7_5, 0.4_9_8_2, 0.4_9_8_0, 0.5_0_0_0, 0.5_0_0_6, 0.4_9_7_2]) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice).max() < 2E-2 def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : Optional[int] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/img2img/sketch-mountains-input.jpg') _lowercase : int = init_image.resize((1_28, 1_28)) _lowercase : str = LMSDiscreteScheduler.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx', subfolder='scheduler') _lowercase : Dict = OnnxStableDiffusionUpscalePipeline.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx', scheduler=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Optional[int] = 'A fantasy landscape, trending on artstation' _lowercase : List[Any] = torch.manual_seed(0) _lowercase : str = pipe( prompt=lowerCamelCase, image=lowerCamelCase, guidance_scale=7.5, num_inference_steps=20, generator=lowerCamelCase, output_type='np', ) _lowercase : str = output.images _lowercase : str = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 5_12, 3) _lowercase : Union[str, Any] = np.array( [0.5_0_1_7_3_7_5_3, 0.5_0_2_2_3_3_5_6, 0.5_0_2_0_3_9, 0.5_0_2_3_3_0_3_6, 0.5_0_2_3_7_2_5, 0.5_0_2_2_6_0_1, 0.5_0_1_8_7_5_8, 0.5_0_2_3_4_0_8_5, 0.5_0_2_4_1_5_6_6]) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice).max() < 2E-2
21
1
'''simple docstring''' def _lowerCAmelCase ( lowerCamelCase_ : list , lowerCamelCase_ : int , lowerCamelCase_ : int = 0 , lowerCamelCase_ : int = 0 ): __lowercase = right or len(lowerCamelCase_ ) - 1 if left > right: return -1 elif list_data[left] == key: return left elif list_data[right] == key: return right else: return search(lowerCamelCase_ , lowerCamelCase_ , left + 1 , right - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
217
'''simple docstring''' # NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline # noqa: F401 deprecate( '''stable diffusion controlnet''', '''0.22.0''', '''Importing `FlaxStableDiffusionControlNetPipeline` from diffusers.pipelines.stable_diffusion.flax_pipeline_stable_diffusion_controlnet is deprecated. Please import `from diffusers import FlaxStableDiffusionControlNetPipeline` instead.''', standard_warn=False, stacklevel=3, )
217
1
import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class _a : def __init__(self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_=13, SCREAMING_SNAKE_CASE_=7, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=False, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=99, SCREAMING_SNAKE_CASE_=32, SCREAMING_SNAKE_CASE_=5, SCREAMING_SNAKE_CASE_=4, SCREAMING_SNAKE_CASE_=37, SCREAMING_SNAKE_CASE_="gelu", SCREAMING_SNAKE_CASE_=0.1, SCREAMING_SNAKE_CASE_=0.1, SCREAMING_SNAKE_CASE_=512, SCREAMING_SNAKE_CASE_=16, SCREAMING_SNAKE_CASE_=2, SCREAMING_SNAKE_CASE_=0.0_2, SCREAMING_SNAKE_CASE_=3, SCREAMING_SNAKE_CASE_=4, SCREAMING_SNAKE_CASE_=None, ) -> Optional[int]: UpperCAmelCase_: Tuple = parent UpperCAmelCase_: Optional[Any] = batch_size UpperCAmelCase_: Any = seq_length UpperCAmelCase_: Any = is_training UpperCAmelCase_: List[Any] = use_input_mask UpperCAmelCase_: Optional[Any] = use_token_type_ids UpperCAmelCase_: Optional[Any] = use_labels UpperCAmelCase_: Optional[Any] = vocab_size UpperCAmelCase_: Optional[int] = hidden_size UpperCAmelCase_: Tuple = num_hidden_layers UpperCAmelCase_: List[Any] = num_attention_heads UpperCAmelCase_: Dict = intermediate_size UpperCAmelCase_: int = hidden_act UpperCAmelCase_: str = hidden_dropout_prob UpperCAmelCase_: Optional[Any] = attention_probs_dropout_prob UpperCAmelCase_: Tuple = max_position_embeddings UpperCAmelCase_: List[str] = type_vocab_size UpperCAmelCase_: Tuple = type_sequence_label_size UpperCAmelCase_: int = initializer_range UpperCAmelCase_: Optional[Any] = num_labels UpperCAmelCase_: Any = num_choices UpperCAmelCase_: List[Any] = scope def __snake_case (self ) -> Optional[Any]: UpperCAmelCase_: Optional[int] = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) UpperCAmelCase_: Dict = None if self.use_input_mask: UpperCAmelCase_: int = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_: List[str] = None if self.use_token_type_ids: UpperCAmelCase_: str = ids_tensor([self.batch_size, self.seq_length], self.type_vocab_size ) UpperCAmelCase_: List[Any] = None UpperCAmelCase_: Dict = None UpperCAmelCase_: Optional[Any] = None if self.use_labels: UpperCAmelCase_: Dict = ids_tensor([self.batch_size], self.type_sequence_label_size ) UpperCAmelCase_: Any = ids_tensor([self.batch_size, self.seq_length], self.num_labels ) UpperCAmelCase_: Optional[Any] = ids_tensor([self.batch_size], self.num_choices ) UpperCAmelCase_: int = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __snake_case (self ) -> List[Any]: return LlamaConfig( vocab_size=self.vocab_size, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, type_vocab_size=self.type_vocab_size, is_decoder=SCREAMING_SNAKE_CASE_, initializer_range=self.initializer_range, ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> str: UpperCAmelCase_: Union[str, Any] = LlamaModel(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() UpperCAmelCase_: List[Any] = model(SCREAMING_SNAKE_CASE_, attention_mask=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Optional[Any] = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, ) -> Tuple: UpperCAmelCase_: Union[str, Any] = True UpperCAmelCase_: Optional[Any] = LlamaModel(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() UpperCAmelCase_: Tuple = model( SCREAMING_SNAKE_CASE_, attention_mask=SCREAMING_SNAKE_CASE_, encoder_hidden_states=SCREAMING_SNAKE_CASE_, encoder_attention_mask=SCREAMING_SNAKE_CASE_, ) UpperCAmelCase_: str = model( SCREAMING_SNAKE_CASE_, attention_mask=SCREAMING_SNAKE_CASE_, encoder_hidden_states=SCREAMING_SNAKE_CASE_, ) UpperCAmelCase_: List[str] = model(SCREAMING_SNAKE_CASE_, attention_mask=SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, ) -> List[str]: UpperCAmelCase_: str = LlamaForCausalLM(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() UpperCAmelCase_: Dict = model(SCREAMING_SNAKE_CASE_, attention_mask=SCREAMING_SNAKE_CASE_, labels=SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size) ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, ) -> int: UpperCAmelCase_: List[Any] = True UpperCAmelCase_: Optional[int] = True UpperCAmelCase_: int = LlamaForCausalLM(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() # first forward pass UpperCAmelCase_: Tuple = model( SCREAMING_SNAKE_CASE_, attention_mask=SCREAMING_SNAKE_CASE_, encoder_hidden_states=SCREAMING_SNAKE_CASE_, encoder_attention_mask=SCREAMING_SNAKE_CASE_, use_cache=SCREAMING_SNAKE_CASE_, ) UpperCAmelCase_: Any = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids UpperCAmelCase_: Optional[Any] = ids_tensor((self.batch_size, 3), config.vocab_size ) UpperCAmelCase_: Dict = ids_tensor((self.batch_size, 3), vocab_size=2 ) # append to next input_ids and UpperCAmelCase_: int = torch.cat([input_ids, next_tokens], dim=-1 ) UpperCAmelCase_: Optional[Any] = torch.cat([input_mask, next_mask], dim=-1 ) UpperCAmelCase_: Union[str, Any] = model( SCREAMING_SNAKE_CASE_, attention_mask=SCREAMING_SNAKE_CASE_, encoder_hidden_states=SCREAMING_SNAKE_CASE_, encoder_attention_mask=SCREAMING_SNAKE_CASE_, output_hidden_states=SCREAMING_SNAKE_CASE_, )["""hidden_states"""][0] UpperCAmelCase_: Optional[int] = model( SCREAMING_SNAKE_CASE_, attention_mask=SCREAMING_SNAKE_CASE_, encoder_hidden_states=SCREAMING_SNAKE_CASE_, encoder_attention_mask=SCREAMING_SNAKE_CASE_, past_key_values=SCREAMING_SNAKE_CASE_, output_hidden_states=SCREAMING_SNAKE_CASE_, )["""hidden_states"""][0] # select random slice UpperCAmelCase_: Any = ids_tensor((1,), output_from_past.shape[-1] ).item() UpperCAmelCase_: Optional[int] = output_from_no_past[:, -3:, random_slice_idx].detach() UpperCAmelCase_: List[Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, atol=1E-3 ) ) def __snake_case (self ) -> str: UpperCAmelCase_: Optional[Any] = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ): int = config_and_inputs UpperCAmelCase_: List[str] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class _a ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): A = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () A = (LlamaForCausalLM,) if is_torch_available() else () A = ( { '''feature-extraction''': LlamaModel, '''text-classification''': LlamaForSequenceClassification, '''text-generation''': LlamaForCausalLM, '''zero-shot''': LlamaForSequenceClassification, } if is_torch_available() else {} ) A = False A = False def __snake_case (self ) -> Dict: UpperCAmelCase_: List[Any] = LlamaModelTester(self ) UpperCAmelCase_: str = ConfigTester(self, config_class=SCREAMING_SNAKE_CASE_, hidden_size=37 ) def __snake_case (self ) -> List[str]: self.config_tester.run_common_tests() def __snake_case (self ) -> str: UpperCAmelCase_: List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_ ) def __snake_case (self ) -> str: UpperCAmelCase_: List[Any] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase_: List[str] = type self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_ ) def __snake_case (self ) -> Dict: UpperCAmelCase_ , UpperCAmelCase_: Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_: int = 3 UpperCAmelCase_: Optional[Any] = input_dict["""input_ids"""] UpperCAmelCase_: Union[str, Any] = input_ids.ne(1 ).to(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Dict = ids_tensor([self.model_tester.batch_size], self.model_tester.type_sequence_label_size ) UpperCAmelCase_: Union[str, Any] = LlamaForSequenceClassification(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() UpperCAmelCase_: Union[str, Any] = model(SCREAMING_SNAKE_CASE_, attention_mask=SCREAMING_SNAKE_CASE_, labels=SCREAMING_SNAKE_CASE_ ) self.assertEqual(result.logits.shape, (self.model_tester.batch_size, self.model_tester.num_labels) ) def __snake_case (self ) -> List[Any]: UpperCAmelCase_ , UpperCAmelCase_: int = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_: List[Any] = 3 UpperCAmelCase_: List[Any] = """single_label_classification""" UpperCAmelCase_: Optional[Any] = input_dict["""input_ids"""] UpperCAmelCase_: List[str] = input_ids.ne(1 ).to(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Union[str, Any] = ids_tensor([self.model_tester.batch_size], self.model_tester.type_sequence_label_size ) UpperCAmelCase_: Optional[int] = LlamaForSequenceClassification(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() UpperCAmelCase_: Union[str, Any] = model(SCREAMING_SNAKE_CASE_, attention_mask=SCREAMING_SNAKE_CASE_, labels=SCREAMING_SNAKE_CASE_ ) self.assertEqual(result.logits.shape, (self.model_tester.batch_size, self.model_tester.num_labels) ) def __snake_case (self ) -> Optional[Any]: UpperCAmelCase_ , UpperCAmelCase_: Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_: Dict = 3 UpperCAmelCase_: List[Any] = """multi_label_classification""" UpperCAmelCase_: Optional[Any] = input_dict["""input_ids"""] UpperCAmelCase_: Union[str, Any] = input_ids.ne(1 ).to(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: List[str] = ids_tensor( [self.model_tester.batch_size, config.num_labels], self.model_tester.type_sequence_label_size ).to(torch.float ) UpperCAmelCase_: int = LlamaForSequenceClassification(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() UpperCAmelCase_: Dict = model(SCREAMING_SNAKE_CASE_, attention_mask=SCREAMING_SNAKE_CASE_, labels=SCREAMING_SNAKE_CASE_ ) self.assertEqual(result.logits.shape, (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip("""LLaMA buffers include complex numbers, which breaks this test""" ) def __snake_case (self ) -> List[Any]: pass @parameterized.expand([("""linear""",), ("""dynamic""",)] ) def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> Any: UpperCAmelCase_ , UpperCAmelCase_: List[str] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_: Dict = ids_tensor([1, 10], config.vocab_size ) UpperCAmelCase_: str = ids_tensor([1, int(config.max_position_embeddings * 1.5 )], config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights UpperCAmelCase_: List[Any] = LlamaModel(SCREAMING_SNAKE_CASE_ ) original_model.to(SCREAMING_SNAKE_CASE_ ) original_model.eval() UpperCAmelCase_: int = original_model(SCREAMING_SNAKE_CASE_ ).last_hidden_state UpperCAmelCase_: List[Any] = original_model(SCREAMING_SNAKE_CASE_ ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights UpperCAmelCase_: str = {"""type""": scaling_type, """factor""": 1_0.0} UpperCAmelCase_: Dict = LlamaModel(SCREAMING_SNAKE_CASE_ ) scaled_model.to(SCREAMING_SNAKE_CASE_ ) scaled_model.eval() UpperCAmelCase_: Any = scaled_model(SCREAMING_SNAKE_CASE_ ).last_hidden_state UpperCAmelCase_: int = scaled_model(SCREAMING_SNAKE_CASE_ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, atol=1E-5 ) ) else: self.assertFalse(torch.allclose(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, atol=1E-5 ) ) @require_torch class _a ( unittest.TestCase ): @unittest.skip("""Logits are not exactly the same, once we fix the instabalities somehow, will update!""" ) @slow def __snake_case (self ) -> Optional[Any]: UpperCAmelCase_: List[str] = [1, 306, 4658, 278, 6593, 310, 2834, 338] UpperCAmelCase_: Union[str, Any] = LlamaForCausalLM.from_pretrained("""meta-llama/Llama-2-7b-hf""", device_map="""auto""" ) UpperCAmelCase_: Optional[Any] = model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 UpperCAmelCase_: Optional[int] = torch.tensor([[-6.6_5_5_0, -4.1_2_2_7, -4.9_8_5_9, -3.2_4_0_6, 0.8_2_6_2, -3.0_0_3_3, 1.2_9_6_4, -3.3_6_9_9]] ) torch.testing.assert_close(out.mean(-1 ), SCREAMING_SNAKE_CASE_, atol=1E-2, rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off UpperCAmelCase_: Optional[Any] = torch.tensor([-1_2.8_2_8_1, -7.4_4_5_3, -0.4_6_3_9, -8.0_6_2_5, -7.2_5_0_0, -8.0_0_0_0, -6.4_8_8_3, -7.7_6_9_5, -7.8_4_3_8, -7.0_3_1_2, -6.2_1_8_8, -7.1_3_2_8, -1.8_4_9_6, 1.9_9_6_1, -8.6_2_5_0, -6.7_2_2_7, -1_2.8_2_8_1, -6.9_4_9_2, -7.0_7_4_2, -7.7_8_5_2, -7.5_8_2_0, -7.9_0_6_2, -6.9_3_7_5, -7.9_8_0_5, -8.3_4_3_8, -8.1_5_6_2, -8.0_4_6_9, -7.6_2_5_0, -7.7_4_2_2, -7.3_3_9_8,] ) # fmt: on torch.testing.assert_close(out[0, 0, :30], SCREAMING_SNAKE_CASE_, atol=1E-5, rtol=1E-5 ) @unittest.skip("""Logits are not exactly the same, once we fix the instabalities somehow, will update!""" ) @slow def __snake_case (self ) -> Optional[Any]: UpperCAmelCase_: int = [1, 306, 4658, 278, 6593, 310, 2834, 338] UpperCAmelCase_: List[str] = LlamaForCausalLM.from_pretrained("""meta-llama/Llama-2-13b-hf""", device_map="""auto""" ) UpperCAmelCase_: Optional[Any] = model(torch.tensor(SCREAMING_SNAKE_CASE_ ) ) # Expected mean on dim = -1 UpperCAmelCase_: Optional[int] = torch.tensor([[-2.0_6_2_2, -1.2_7_9_4, -1.1_6_3_8, -0.9_7_8_8, -1.4_6_0_3, -1.0_2_3_8, -1.7_8_9_3, -1.4_4_1_1]] ) torch.testing.assert_close(out.mean(-1 ), SCREAMING_SNAKE_CASE_, atol=1E-2, rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off UpperCAmelCase_: Optional[int] = torch.tensor([-8.1_4_0_6, -8.0_5_4_7, 2.7_4_6_1, -1.2_3_4_4, -0.1_4_4_8, -1.8_2_6_2, -1.0_0_2_0, -1.8_1_5_4, -1.6_8_9_5, -1.8_5_1_6, -2.3_5_7_4, -0.9_2_7_7, 3.7_5_9_8, 6.5_7_4_2, -1.2_9_9_8, -0.1_1_7_7, -8.1_4_0_6, -2.9_6_8_8, -2.9_1_9_9, -3.1_6_9_9, -3.5_2_5_4, -2.3_5_5_5, -2.7_9_8_8, -3.4_1_4_1, -2.8_2_6_2, -4.5_1_9_5, -3.3_3_7_9, -3.3_1_6_4, -2.7_8_3_2, -3.0_2_7_3] ) # fmt: on torch.testing.assert_close(out[0, 0, :30], SCREAMING_SNAKE_CASE_, atol=1E-5, rtol=1E-5 ) @unittest.skip("""Logits are not exactly the same, once we fix the instabalities somehow, will update!""" ) @slow def __snake_case (self ) -> Dict: UpperCAmelCase_: Optional[Any] = [1, 306, 4658, 278, 6593, 310, 2834, 338] UpperCAmelCase_: List[Any] = LlamaForCausalLM.from_pretrained("""meta-llama/Llama-2-13b-chat-hf""", device_map="""auto""" ) UpperCAmelCase_: Optional[Any] = model(torch.tensor(SCREAMING_SNAKE_CASE_ ) ) # Expected mean on dim = -1 UpperCAmelCase_: Any = torch.tensor([[-0.8_5_6_2, -1.8_5_2_0, -0.7_5_5_1, -0.4_1_6_2, -1.5_1_6_1, -1.2_0_3_8, -2.4_8_2_3, -2.3_2_5_4]] ) torch.testing.assert_close(out.mean(-1 ), SCREAMING_SNAKE_CASE_, atol=1E-2, rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off UpperCAmelCase_: Dict = torch.tensor([-2.2_2_2_7, 4.8_8_2_8, 0.9_0_2_3, -0.4_5_7_8, -0.7_8_7_1, -0.1_0_3_3, -0.6_2_2_1, -0.5_7_8_6, -0.7_8_0_3, -1.0_6_7_4, -1.2_9_2_0, -0.1_5_7_0, 0.8_0_0_8, 2.0_7_2_3, -0.9_4_9_7, 0.2_7_7_1, -2.2_2_2_7, -0.7_6_1_2, -1.4_3_4_6, -1.2_0_6_1, -1.6_4_2_6, -0.3_0_0_0, -0.7_1_3_9, -1.1_9_3_4, -1.8_6_9_1, -1.6_9_7_3, -1.5_9_4_7, -1.2_7_0_5, -0.3_5_2_3, -0.5_5_1_3] ) # fmt: on torch.testing.assert_close(out.mean(-1 ), SCREAMING_SNAKE_CASE_, atol=1E-2, rtol=1E-2 ) @unittest.skip( """Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test""" ) @slow def __snake_case (self ) -> Any: UpperCAmelCase_: Any = [1, 306, 4658, 278, 6593, 310, 2834, 338] UpperCAmelCase_: Optional[int] = LlamaForCausalLM.from_pretrained("""meta-llama/Llama-2-70b-hf""", device_map="""auto""" ) UpperCAmelCase_: Optional[int] = model(torch.tensor(SCREAMING_SNAKE_CASE_ ) ) UpperCAmelCase_: Optional[Any] = torch.tensor( [[-4.2_3_2_7, -3.3_3_6_0, -4.6_6_6_5, -4.7_6_3_1, -1.8_1_8_0, -3.4_1_7_0, -1.4_2_1_1, -3.1_8_1_0]], dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ), SCREAMING_SNAKE_CASE_, atol=1E-2, rtol=1E-2 ) # fmt: off UpperCAmelCase_: int = torch.tensor([-9.4_9_2_2, -3.9_5_5_1, 1.7_9_9_8, -5.6_7_5_8, -5.1_0_5_5, -5.8_9_8_4, -4.8_3_2_0, -6.8_0_8_6, -6.5_3_9_1, -5.6_1_7_2, -5.5_8_2_0, -5.5_3_5_2, 1.7_8_8_1, 3.6_2_8_9, -6.5_1_1_7, -3.4_7_8_5, -9.5_0_0_0, -6.0_3_5_2, -6.8_1_2_5, -6.0_1_9_5, -6.6_8_3_6, -5.4_7_2_7, -6.2_8_1_2, -6.0_3_9_1, -7.3_3_9_8, -7.4_2_9_7, -7.4_8_4_4, -6.5_8_2_0, -5.8_7_8_9, -5.5_3_1_2] ) # fmt: on torch.testing.assert_close(out[0, 0, :30], SCREAMING_SNAKE_CASE_, atol=1E-5, rtol=1E-5 ) @unittest.skip("""Model is curently gated""" ) @slow def __snake_case (self ) -> Any: UpperCAmelCase_: int = """Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the \"princi""" UpperCAmelCase_: Optional[int] = """Simply put, the theory of relativity states that """ UpperCAmelCase_: List[str] = LlamaTokenizer.from_pretrained("""meta-llama/Llama-2-13b-chat-hf""" ) UpperCAmelCase_: Dict = tokenizer.encode(SCREAMING_SNAKE_CASE_, return_tensors="""pt""" ) UpperCAmelCase_: str = LlamaForCausalLM.from_pretrained( """meta-llama/Llama-2-13b-chat-hf""", device_map="""sequential""", use_safetensors=SCREAMING_SNAKE_CASE_ ) # greedy generation outputs UpperCAmelCase_: Union[str, Any] = model.generate(SCREAMING_SNAKE_CASE_, max_new_tokens=64, top_p=SCREAMING_SNAKE_CASE_, temperature=1, do_sample=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: str = tokenizer.decode(generated_ids[0], skip_special_tokens=SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ )
147
from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers.models.esm.modeling_tf_esm import ( TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, TFEsmModel, ) class _a : def __init__(self, SCREAMING_SNAKE_CASE_, ) -> Optional[Any]: UpperCAmelCase_: Optional[int] = parent UpperCAmelCase_: List[Any] = 13 UpperCAmelCase_: Union[str, Any] = 7 UpperCAmelCase_: Optional[Any] = True UpperCAmelCase_: Tuple = True UpperCAmelCase_: Dict = True UpperCAmelCase_: str = 99 UpperCAmelCase_: Tuple = 32 UpperCAmelCase_: Optional[int] = 2 UpperCAmelCase_: Union[str, Any] = 4 UpperCAmelCase_: List[Any] = 37 UpperCAmelCase_: str = """gelu""" UpperCAmelCase_: Dict = 0.1 UpperCAmelCase_: Optional[Any] = 0.1 UpperCAmelCase_: Optional[Any] = 512 UpperCAmelCase_: List[str] = 16 UpperCAmelCase_: Any = 2 UpperCAmelCase_: Union[str, Any] = 0.0_2 UpperCAmelCase_: List[str] = 3 UpperCAmelCase_: Tuple = 4 UpperCAmelCase_: Any = None def __snake_case (self ) -> str: UpperCAmelCase_: List[Any] = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) UpperCAmelCase_: Optional[int] = None if self.use_input_mask: UpperCAmelCase_: Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_: List[Any] = None UpperCAmelCase_: Optional[int] = None UpperCAmelCase_: Tuple = None if self.use_labels: UpperCAmelCase_: List[Any] = ids_tensor([self.batch_size], self.type_sequence_label_size ) UpperCAmelCase_: int = ids_tensor([self.batch_size, self.seq_length], self.num_labels ) UpperCAmelCase_: List[Any] = ids_tensor([self.batch_size], self.num_choices ) UpperCAmelCase_: List[str] = EsmConfig( vocab_size=self.vocab_size, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, pad_token_id=1, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, type_vocab_size=self.type_vocab_size, initializer_range=self.initializer_range, ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def __snake_case (self ) -> Optional[int]: ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ): Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase_: Dict = True UpperCAmelCase_: Dict = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCAmelCase_: Dict = ids_tensor([self.batch_size, self.seq_length], vocab_size=2 ) return ( config, input_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Tuple: UpperCAmelCase_: Optional[int] = TFEsmModel(config=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: str = {"""input_ids""": input_ids, """attention_mask""": input_mask} UpperCAmelCase_: Dict = model(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Dict = [input_ids, input_mask] UpperCAmelCase_: Dict = model(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: List[Any] = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, ) -> List[str]: UpperCAmelCase_: Tuple = True UpperCAmelCase_: List[Any] = TFEsmModel(config=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Optional[int] = { """input_ids""": input_ids, """attention_mask""": input_mask, """encoder_hidden_states""": encoder_hidden_states, """encoder_attention_mask""": encoder_attention_mask, } UpperCAmelCase_: Optional[Any] = model(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Tuple = [input_ids, input_mask] UpperCAmelCase_: List[Any] = model(SCREAMING_SNAKE_CASE_, encoder_hidden_states=SCREAMING_SNAKE_CASE_ ) # Also check the case where encoder outputs are not passed UpperCAmelCase_: Tuple = model(SCREAMING_SNAKE_CASE_, attention_mask=SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Optional[int]: UpperCAmelCase_: List[Any] = TFEsmForMaskedLM(config=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Optional[int] = model([input_ids, input_mask] ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size) ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: UpperCAmelCase_: int = self.num_labels UpperCAmelCase_: Dict = TFEsmForTokenClassification(config=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: List[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} UpperCAmelCase_: Tuple = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.num_labels) ) def __snake_case (self ) -> Any: UpperCAmelCase_: Dict = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ): str = config_and_inputs UpperCAmelCase_: int = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class _a ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): A = ( ( TFEsmModel, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, ) if is_tf_available() else () ) A = ( { '''feature-extraction''': TFEsmModel, '''fill-mask''': TFEsmForMaskedLM, '''text-classification''': TFEsmForSequenceClassification, '''token-classification''': TFEsmForTokenClassification, '''zero-shot''': TFEsmForSequenceClassification, } if is_tf_available() else {} ) A = False A = False def __snake_case (self ) -> Tuple: UpperCAmelCase_: List[Any] = TFEsmModelTester(self ) UpperCAmelCase_: Union[str, Any] = ConfigTester(self, config_class=SCREAMING_SNAKE_CASE_, hidden_size=37 ) def __snake_case (self ) -> Any: self.config_tester.run_common_tests() def __snake_case (self ) -> List[str]: UpperCAmelCase_: int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_ ) def __snake_case (self ) -> List[Any]: UpperCAmelCase_: Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*SCREAMING_SNAKE_CASE_ ) def __snake_case (self ) -> Tuple: UpperCAmelCase_: Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*SCREAMING_SNAKE_CASE_ ) def __snake_case (self ) -> List[str]: UpperCAmelCase_: Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*SCREAMING_SNAKE_CASE_ ) @slow def __snake_case (self ) -> str: for model_name in TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_: Dict = TFEsmModel.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) @unittest.skip("""Protein models do not support embedding resizing.""" ) def __snake_case (self ) -> Tuple: pass @unittest.skip("""Protein models do not support embedding resizing.""" ) def __snake_case (self ) -> Optional[Any]: pass def __snake_case (self ) -> List[Any]: UpperCAmelCase_ , UpperCAmelCase_: Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_: List[Any] = model_class(SCREAMING_SNAKE_CASE_ ) assert isinstance(model.get_input_embeddings(), tf.keras.layers.Layer ) if model_class is TFEsmForMaskedLM: # Output embedding test differs from the main test because they're a matrix, not a layer UpperCAmelCase_: Any = model.get_bias() assert isinstance(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) for k, v in name.items(): assert isinstance(SCREAMING_SNAKE_CASE_, tf.Variable ) else: UpperCAmelCase_: Union[str, Any] = model.get_output_embeddings() assert x is None UpperCAmelCase_: Optional[int] = model.get_bias() assert name is None @require_tf class _a ( unittest.TestCase ): @slow def __snake_case (self ) -> str: UpperCAmelCase_: str = TFEsmForMaskedLM.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) UpperCAmelCase_: Tuple = tf.constant([[0, 1, 2, 3, 4, 5]] ) UpperCAmelCase_: int = model(SCREAMING_SNAKE_CASE_ )[0] UpperCAmelCase_: Optional[int] = [1, 6, 33] self.assertEqual(list(output.numpy().shape ), SCREAMING_SNAKE_CASE_ ) # compare the actual values for a slice. UpperCAmelCase_: List[str] = tf.constant( [ [ [8.9_2_1_5_1_8, -1_0.5_8_9_8_1_4, -6.4_6_7_1_3_0_7], [-6.3_9_6_7_1_5_6, -1_3.9_1_1_3_7_7, -1.1_2_1_1_9_1_5], [-7.7_8_1_2_4_7, -1_3.9_5_1_5_5_7, -3.7_4_0_5_9_2], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy(), expected_slice.numpy(), atol=1E-2 ) ) @slow def __snake_case (self ) -> Optional[int]: UpperCAmelCase_: List[Any] = TFEsmModel.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) UpperCAmelCase_: Any = tf.constant([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) UpperCAmelCase_: Optional[int] = model(SCREAMING_SNAKE_CASE_ )[0] # compare the actual values for a slice. UpperCAmelCase_: str = tf.constant( [ [ [0.1_4_4_4_3_0_9_2, 0.5_4_1_2_5_3_2_7, 0.3_2_4_7_7_3_9], [0.3_0_3_4_0_4_8_4, 0.0_0_5_2_6_6_7_6, 0.3_1_0_7_7_7_2_2], [0.3_2_2_7_8_0_4_3, -0.2_4_9_8_7_0_9_6, 0.3_4_1_4_6_2_8], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy(), expected_slice.numpy(), atol=1E-4 ) )
147
1
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def UpperCamelCase ( snake_case__ : List[str] , snake_case__ : str ) -> int: # Load checkpoint UpperCamelCase : Union[str, Any] = torch.load(snake_case__ , map_location='cpu' ) UpperCamelCase : Optional[Any] = chkpt['model'] # We have the base model one level deeper than the original XLM repository UpperCamelCase : Union[str, Any] = {} for k, v in state_dict.items(): if "pred_layer" in k: UpperCamelCase : Any = v else: UpperCamelCase : str = v UpperCamelCase : Any = chkpt['params'] UpperCamelCase : Tuple = {n: v for n, v in config.items() if not isinstance(snake_case__ , (torch.FloatTensor, numpy.ndarray) )} UpperCamelCase : str = chkpt['dico_word2id'] UpperCamelCase : Union[str, Any] = {s + '</w>' if s.find('@@' ) == -1 and i > 13 else s.replace('@@' , '' ): i for s, i in vocab.items()} # Save pytorch-model UpperCamelCase : Tuple = pytorch_dump_folder_path + '/' + WEIGHTS_NAME UpperCamelCase : List[str] = pytorch_dump_folder_path + '/' + CONFIG_NAME UpperCamelCase : Any = pytorch_dump_folder_path + '/' + VOCAB_FILES_NAMES['vocab_file'] print(F"""Save PyTorch model to {pytorch_weights_dump_path}""" ) torch.save(snake_case__ , snake_case__ ) print(F"""Save configuration file to {pytorch_config_dump_path}""" ) with open(snake_case__ , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(snake_case__ , indent=2 ) + '\n' ) print(F"""Save vocab file to {pytorch_config_dump_path}""" ) with open(snake_case__ , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(snake_case__ , indent=2 ) + '\n' ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--xlm_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __UpperCAmelCase = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
103
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.local_sgd import LocalSGD ######################################################################## # This is a fully working simple example to use Accelerate # with LocalSGD, which is a method to synchronize model # parameters every K batches. It is different, but complementary # to gradient accumulation. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __UpperCAmelCase = 16 __UpperCAmelCase = 32 def UpperCamelCase ( snake_case__ : Accelerator , snake_case__ : int = 16 ) -> Dict: UpperCamelCase : str = AutoTokenizer.from_pretrained('bert-base-cased' ) UpperCamelCase : str = load_dataset('glue' , 'mrpc' ) def tokenize_function(snake_case__ : Dict ): # max_length=None => use the model max length (it's actually the default) UpperCamelCase : Optional[Any] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=snake_case__ , max_length=snake_case__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): UpperCamelCase : List[str] = datasets.map( snake_case__ , batched=snake_case__ , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library UpperCamelCase : Dict = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(snake_case__ : Tuple ): # On TPU it's best to pad everything to the same length or training will be very slow. UpperCamelCase : List[str] = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": UpperCamelCase : str = 16 elif accelerator.mixed_precision != "no": UpperCamelCase : Dict = 8 else: UpperCamelCase : Union[str, Any] = None return tokenizer.pad( snake_case__ , padding='longest' , max_length=snake_case__ , pad_to_multiple_of=snake_case__ , return_tensors='pt' , ) # Instantiate dataloaders. UpperCamelCase : Optional[int] = DataLoader( tokenized_datasets['train'] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) UpperCamelCase : List[Any] = DataLoader( tokenized_datasets['validation'] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''', None) == "1": from accelerate.test_utils.training import mocked_dataloaders __UpperCAmelCase = mocked_dataloaders # noqa: F811 def UpperCamelCase ( snake_case__ : str , snake_case__ : List[Any] ) -> List[str]: # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS' , snake_case__ ) == "1": UpperCamelCase : List[str] = 2 # New Code # UpperCamelCase : Optional[int] = int(args.gradient_accumulation_steps ) UpperCamelCase : List[str] = int(args.local_sgd_steps ) # Initialize accelerator UpperCamelCase : Optional[Any] = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , gradient_accumulation_steps=snake_case__ ) if accelerator.distributed_type not in [DistributedType.NO, DistributedType.MULTI_CPU, DistributedType.MULTI_GPU]: raise NotImplementedError('LocalSGD is supported only for CPUs and GPUs (no DeepSpeed or MegatronLM)' ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs UpperCamelCase : Union[str, Any] = config['lr'] UpperCamelCase : Optional[int] = int(config['num_epochs'] ) UpperCamelCase : List[Any] = int(config['seed'] ) UpperCamelCase : List[str] = int(config['batch_size'] ) UpperCamelCase : Optional[Any] = evaluate.load('glue' , 'mrpc' ) set_seed(snake_case__ ) UpperCamelCase , UpperCamelCase : int = get_dataloaders(snake_case__ , snake_case__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) UpperCamelCase : Tuple = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=snake_case__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). UpperCamelCase : Dict = model.to(accelerator.device ) # Instantiate optimizer UpperCamelCase : List[str] = AdamW(params=model.parameters() , lr=snake_case__ ) # Instantiate scheduler UpperCamelCase : Any = get_linear_schedule_with_warmup( optimizer=snake_case__ , num_warmup_steps=100 , num_training_steps=(len(snake_case__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase : str = accelerator.prepare( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) # Now we train the model for epoch in range(snake_case__ ): model.train() with LocalSGD( accelerator=snake_case__ , model=snake_case__ , local_sgd_steps=snake_case__ , enabled=local_sgd_steps is not None ) as local_sgd: for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(snake_case__ ): UpperCamelCase : int = model(**snake_case__ ) UpperCamelCase : Union[str, Any] = output.loss accelerator.backward(snake_case__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() # LocalSGD-specific line local_sgd.step() model.eval() for step, batch in enumerate(snake_case__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): UpperCamelCase : Dict = model(**snake_case__ ) UpperCamelCase : Any = outputs.logits.argmax(dim=-1 ) UpperCamelCase , UpperCamelCase : Any = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=snake_case__ , references=snake_case__ , ) UpperCamelCase : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"""epoch {epoch}:""" , snake_case__ ) def UpperCamelCase ( ) -> Dict: UpperCamelCase : Any = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=snake_case__ , default=snake_case__ , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) # New Code # parser.add_argument( '--gradient_accumulation_steps' , type=snake_case__ , default=1 , help='The number of minibatches to be ran before gradients are accumulated.' , ) parser.add_argument( '--local_sgd_steps' , type=snake_case__ , default=8 , help='Number of local SGD steps or None to disable local SGD' ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) UpperCamelCase : str = parser.parse_args() UpperCamelCase : Tuple = {'lr': 2E-5, 'num_epochs': 3, 'seed': 42, 'batch_size': 16} training_function(snake_case__ , snake_case__ ) if __name__ == "__main__": main()
103
1
"""simple docstring""" def a__ ( SCREAMING_SNAKE_CASE : int ): '''simple docstring''' lowerCAmelCase : List[str] = int(SCREAMING_SNAKE_CASE ) if decimal in (0, 1): # Exit cases for the recursion return str(SCREAMING_SNAKE_CASE ) lowerCAmelCase , lowerCAmelCase : Any = divmod(SCREAMING_SNAKE_CASE , 2 ) return binary_recursive(SCREAMING_SNAKE_CASE ) + str(SCREAMING_SNAKE_CASE ) def a__ ( SCREAMING_SNAKE_CASE : str ): '''simple docstring''' lowerCAmelCase : Optional[int] = str(SCREAMING_SNAKE_CASE ).strip() if not number: raise ValueError("No input value was provided" ) lowerCAmelCase : Tuple = "-" if number.startswith("-" ) else "" lowerCAmelCase : Union[str, Any] = number.lstrip("-" ) if not number.isnumeric(): raise ValueError("Input value is not an integer" ) return f"""{negative}0b{binary_recursive(int(SCREAMING_SNAKE_CASE ) )}""" if __name__ == "__main__": from doctest import testmod testmod()
108
import os # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_doctest_list.py __A = "." if __name__ == "__main__": __A = os.path.join(REPO_PATH, "utils/documentation_tests.txt") __A = [] __A = [] with open(doctest_file_path) as fp: for line in fp: __A = line.strip() __A = os.path.join(REPO_PATH, line) if not (os.path.isfile(path) or os.path.isdir(path)): non_existent_paths.append(line) all_paths.append(path) if len(non_existent_paths) > 0: __A = "\n".join(non_existent_paths) raise ValueError(f'`utils/documentation_tests.txt` contains non-existent paths:\n{non_existent_paths}') if all_paths != sorted(all_paths): raise ValueError("Files in `utils/documentation_tests.txt` are not in alphabetical order.")
10
0
__magic_name__ = { "meter": "m", "kilometer": "km", "megametre": "Mm", "gigametre": "Gm", "terametre": "Tm", "petametre": "Pm", "exametre": "Em", "zettametre": "Zm", "yottametre": "Ym", } # Exponent of the factor(meter) __magic_name__ = { "m": 0, "km": 3, "Mm": 6, "Gm": 9, "Tm": 12, "Pm": 15, "Em": 18, "Zm": 21, "Ym": 24, } def _lowerCAmelCase ( A__: float , A__: str , A__: str ): '''simple docstring''' UpperCAmelCase = from_type.lower().strip('''s''' ) UpperCAmelCase = to_type.lower().strip('''s''' ) UpperCAmelCase = UNIT_SYMBOL.get(A__ , A__ ) UpperCAmelCase = UNIT_SYMBOL.get(A__ , A__ ) if from_sanitized not in METRIC_CONVERSION: UpperCAmelCase = ( F"""Invalid 'from_type' value: {from_type!r}.\n""" F"""Conversion abbreviations are: {', '.join(A__ )}""" ) raise ValueError(A__ ) if to_sanitized not in METRIC_CONVERSION: UpperCAmelCase = ( F"""Invalid 'to_type' value: {to_type!r}.\n""" F"""Conversion abbreviations are: {', '.join(A__ )}""" ) raise ValueError(A__ ) UpperCAmelCase = METRIC_CONVERSION[from_sanitized] UpperCAmelCase = METRIC_CONVERSION[to_sanitized] UpperCAmelCase = 1 if from_exponent > to_exponent: UpperCAmelCase = from_exponent - to_exponent else: UpperCAmelCase = -(to_exponent - from_exponent) return value * pow(10 , A__ ) if __name__ == "__main__": from doctest import testmod testmod()
152
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available __magic_name__ = { "configuration_ernie": ["ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP", "ErnieConfig", "ErnieOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ = [ "ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST", "ErnieForCausalLM", "ErnieForMaskedLM", "ErnieForMultipleChoice", "ErnieForNextSentencePrediction", "ErnieForPreTraining", "ErnieForQuestionAnswering", "ErnieForSequenceClassification", "ErnieForTokenClassification", "ErnieModel", "ErniePreTrainedModel", ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys __magic_name__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
152
1
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: str , SCREAMING_SNAKE_CASE_: str , SCREAMING_SNAKE_CASE_: str , SCREAMING_SNAKE_CASE_: PreTrainedTokenizer , SCREAMING_SNAKE_CASE_: int , SCREAMING_SNAKE_CASE_: Optional[int] = None , ) -> Optional[int]: '''simple docstring''' A__ = {} if train_file is not None: A__ = [train_file] if eval_file is not None: A__ = [eval_file] if test_file is not None: A__ = [test_file] A__ = datasets.load_dataset("csv" , data_files=UpperCamelCase_ ) A__ = list(ds[list(files.keys() )[0]].features.keys() ) A__ = features_name.pop(UpperCamelCase_ ) A__ = list(set(ds[list(files.keys() )[0]][label_name] ) ) A__ = {label: i for i, label in enumerate(UpperCamelCase_ )} A__ = tokenizer.model_input_names A__ = {} if len(UpperCamelCase_ ) == 1: for k in files.keys(): A__ = ds[k].map( lambda SCREAMING_SNAKE_CASE_ : tokenizer.batch_encode_plus( example[features_name[0]] , truncation=UpperCamelCase_ , max_length=UpperCamelCase_ , padding="max_length" ) , batched=UpperCamelCase_ , ) elif len(UpperCamelCase_ ) == 2: for k in files.keys(): A__ = ds[k].map( lambda SCREAMING_SNAKE_CASE_ : tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]) , truncation=UpperCamelCase_ , max_length=UpperCamelCase_ , padding="max_length" , ) , batched=UpperCamelCase_ , ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: A__ = {k: v for k, v in ex.items() if k in input_names} A__ = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: A__ = {k: v for k, v in ex.items() if k in input_names} A__ = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: A__ = {k: v for k, v in ex.items() if k in input_names} A__ = labelaid[ex[label_name]] yield (d, label) A__ = ( tf.data.Dataset.from_generator( UpperCamelCase_ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: A__ = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN] ) ) ) A__ = ( tf.data.Dataset.from_generator( UpperCamelCase_ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: A__ = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION] ) ) ) A__ = ( tf.data.Dataset.from_generator( UpperCamelCase_ , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: A__ = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST] ) ) ) return train_ds, val_ds, test_ds, labelaid lowerCAmelCase__ = logging.getLogger(__name__) @dataclass class a__ : """simple docstring""" __lowerCamelCase = field(metadata={'help': 'Which column contains the label'} ) __lowerCamelCase = field(default=_UpperCAmelCase , metadata={'help': 'The path of the training file'} ) __lowerCamelCase = field(default=_UpperCAmelCase , metadata={'help': 'The path of the development file'} ) __lowerCamelCase = field(default=_UpperCAmelCase , metadata={'help': 'The path of the test file'} ) __lowerCamelCase = field( default=128 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __lowerCamelCase = field( default=_UpperCAmelCase , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) @dataclass class a__ : """simple docstring""" __lowerCamelCase = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) __lowerCamelCase = field( default=_UpperCAmelCase , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) __lowerCamelCase = field( default=_UpperCAmelCase , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) __lowerCamelCase = field(default=_UpperCAmelCase , metadata={'help': 'Set this flag to use fast tokenization.'} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. __lowerCamelCase = field( default=_UpperCAmelCase , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) def lowerCAmelCase__ ( ) -> str: '''simple docstring''' A__ = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments) ) A__ , A__ , A__ = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F'Output directory ({training_args.output_dir}) already exists and is not empty. Use' " --overwrite_output_dir to overcome." ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , level=logging.INFO , ) logger.info( F'n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1 )}, ' F'16-bits training: {training_args.fpaa}' ) logger.info(F'Training/evaluation parameters {training_args}' ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. A__ = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) A__ , A__ , A__ , A__ = get_tfds( train_file=data_args.train_file , eval_file=data_args.dev_file , test_file=data_args.test_file , tokenizer=UpperCamelCase_ , label_column_id=data_args.label_column_id , max_seq_length=data_args.max_seq_length , ) A__ = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=len(UpperCamelCase_ ) , labelaid=UpperCamelCase_ , idalabel={id: label for label, id in labelaid.items()} , finetuning_task="text-classification" , cache_dir=model_args.cache_dir , ) with training_args.strategy.scope(): A__ = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_pt=bool(".bin" in model_args.model_name_or_path ) , config=UpperCamelCase_ , cache_dir=model_args.cache_dir , ) def compute_metrics(SCREAMING_SNAKE_CASE_: EvalPrediction ) -> Dict: A__ = np.argmax(p.predictions , axis=1 ) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer A__ = TFTrainer( model=UpperCamelCase_ , args=UpperCamelCase_ , train_dataset=UpperCamelCase_ , eval_dataset=UpperCamelCase_ , compute_metrics=UpperCamelCase_ , ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir ) # Evaluation A__ = {} if training_args.do_eval: logger.info("*** Evaluate ***" ) A__ = trainer.evaluate() A__ = os.path.join(training_args.output_dir , "eval_results.txt" ) with open(UpperCamelCase_ , "w" ) as writer: logger.info("***** Eval results *****" ) for key, value in result.items(): logger.info(F' {key} = {value}' ) writer.write(F'{key} = {value}\n' ) results.update(UpperCamelCase_ ) return results if __name__ == "__main__": main()
68
import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class lowercase__ ( _UpperCAmelCase ): a_ ="""char""" a_ ="""bpe""" a_ ="""wp""" a_ = (DecodeType.CHARACTER, DecodeType.BPE, DecodeType.WORDPIECE) class lowercase__ ( _UpperCAmelCase ): a_ =["""image_processor""", """char_tokenizer"""] a_ ="""ViTImageProcessor""" a_ ="""MgpstrTokenizer""" def __init__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , **__UpperCAmelCase )-> str: '''simple docstring''' lowerCAmelCase__ = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , __UpperCAmelCase , ) lowerCAmelCase__ = kwargs.pop("feature_extractor" ) lowerCAmelCase__ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) lowerCAmelCase__ = tokenizer lowerCAmelCase__ = AutoTokenizer.from_pretrained("gpt2" ) lowerCAmelCase__ = AutoTokenizer.from_pretrained("bert-base-uncased" ) super().__init__(__UpperCAmelCase , __UpperCAmelCase ) def __call__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=None , **__UpperCAmelCase )-> List[Any]: '''simple docstring''' if images is None and text is None: raise ValueError("You need to specify either an `images` or `text` input to process." ) if images is not None: lowerCAmelCase__ = self.image_processor(__UpperCAmelCase , return_tensors=__UpperCAmelCase , **__UpperCAmelCase ) if text is not None: lowerCAmelCase__ = self.char_tokenizer(__UpperCAmelCase , return_tensors=__UpperCAmelCase , **__UpperCAmelCase ) if text is None: return inputs elif images is None: return encodings else: lowerCAmelCase__ = encodings["input_ids"] return inputs def UpperCAmelCase ( self , __UpperCAmelCase )-> Optional[int]: '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = sequences lowerCAmelCase__ = char_preds.size(0 ) lowerCAmelCase__ , lowerCAmelCase__ = self._decode_helper(__UpperCAmelCase , "char" ) lowerCAmelCase__ , lowerCAmelCase__ = self._decode_helper(__UpperCAmelCase , "bpe" ) lowerCAmelCase__ , lowerCAmelCase__ = self._decode_helper(__UpperCAmelCase , "wp" ) lowerCAmelCase__ = [] lowerCAmelCase__ = [] for i in range(__UpperCAmelCase ): lowerCAmelCase__ = [char_scores[i], bpe_scores[i], wp_scores[i]] lowerCAmelCase__ = [char_strs[i], bpe_strs[i], wp_strs[i]] lowerCAmelCase__ = scores.index(max(__UpperCAmelCase ) ) final_strs.append(strs[max_score_index] ) final_scores.append(scores[max_score_index] ) lowerCAmelCase__ = {} lowerCAmelCase__ = final_strs lowerCAmelCase__ = final_scores lowerCAmelCase__ = char_strs lowerCAmelCase__ = bpe_strs lowerCAmelCase__ = wp_strs return out def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase )-> Optional[int]: '''simple docstring''' if format == DecodeType.CHARACTER: lowerCAmelCase__ = self.char_decode lowerCAmelCase__ = 1 lowerCAmelCase__ = "[s]" elif format == DecodeType.BPE: lowerCAmelCase__ = self.bpe_decode lowerCAmelCase__ = 2 lowerCAmelCase__ = "#" elif format == DecodeType.WORDPIECE: lowerCAmelCase__ = self.wp_decode lowerCAmelCase__ = 102 lowerCAmelCase__ = "[SEP]" else: raise ValueError(F"Format {format} is not supported." ) lowerCAmelCase__ , lowerCAmelCase__ = [], [] lowerCAmelCase__ = pred_logits.size(0 ) lowerCAmelCase__ = pred_logits.size(1 ) lowerCAmelCase__ , lowerCAmelCase__ = pred_logits.topk(1 , dim=-1 , largest=__UpperCAmelCase , sorted=__UpperCAmelCase ) lowerCAmelCase__ = preds_index.view(-1 , __UpperCAmelCase )[:, 1:] lowerCAmelCase__ = decoder(__UpperCAmelCase ) lowerCAmelCase__ , lowerCAmelCase__ = torch.nn.functional.softmax(__UpperCAmelCase , dim=2 ).max(dim=2 ) lowerCAmelCase__ = preds_max_prob[:, 1:] for index in range(__UpperCAmelCase ): lowerCAmelCase__ = preds_str[index].find(__UpperCAmelCase ) lowerCAmelCase__ = preds_str[index][:pred_eos] lowerCAmelCase__ = preds_index[index].cpu().tolist() lowerCAmelCase__ = pred_index.index(__UpperCAmelCase ) if eos_token in pred_index else -1 lowerCAmelCase__ = preds_max_prob[index][: pred_eos_index + 1] lowerCAmelCase__ = pred_max_prob.cumprod(dim=0 )[-1] if pred_max_prob.nelement() != 0 else 0.0 dec_strs.append(__UpperCAmelCase ) conf_scores.append(__UpperCAmelCase ) return dec_strs, conf_scores def UpperCAmelCase ( self , __UpperCAmelCase )-> Optional[Any]: '''simple docstring''' lowerCAmelCase__ = [seq.replace(" " , "" ) for seq in self.char_tokenizer.batch_decode(__UpperCAmelCase )] return decode_strs def UpperCAmelCase ( self , __UpperCAmelCase )-> Optional[Any]: '''simple docstring''' return self.bpe_tokenizer.batch_decode(__UpperCAmelCase ) def UpperCAmelCase ( self , __UpperCAmelCase )-> Optional[int]: '''simple docstring''' lowerCAmelCase__ = [seq.replace(" " , "" ) for seq in self.wp_tokenizer.batch_decode(__UpperCAmelCase )] return decode_strs
340
0
"""simple docstring""" def _lowerCAmelCase ( UpperCAmelCase__ : list ) ->bool: if not isinstance(UpperCAmelCase__, UpperCAmelCase__ ): raise ValueError("""Input series is not valid, valid series - [2, 4, 6]""" ) if len(UpperCAmelCase__ ) == 0: raise ValueError("""Input list must be a non empty list""" ) if len(UpperCAmelCase__ ) == 1: return True A__ : Optional[Any] = series[1] - series[0] for index in range(len(UpperCAmelCase__ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def _lowerCAmelCase ( UpperCAmelCase__ : list ) ->float: if not isinstance(UpperCAmelCase__, UpperCAmelCase__ ): raise ValueError("""Input series is not valid, valid series - [2, 4, 6]""" ) if len(UpperCAmelCase__ ) == 0: raise ValueError("""Input list must be a non empty list""" ) A__ : Optional[int] = 0 for val in series: answer += val return answer / len(UpperCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod()
296
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING A_ = logging.get_logger(__name__) A_ = { '''microsoft/table-transformer-detection''': ( '''https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json''' ), } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'table-transformer' snake_case_ = ['past_key_values'] snake_case_ = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self : Dict , snake_case : int=True , snake_case : Dict=None , snake_case : Union[str, Any]=3 , snake_case : Dict=100 , snake_case : Tuple=6 , snake_case : Optional[int]=2048 , snake_case : int=8 , snake_case : Dict=6 , snake_case : Any=2048 , snake_case : str=8 , snake_case : Union[str, Any]=0.0 , snake_case : List[str]=0.0 , snake_case : List[str]=True , snake_case : Any="relu" , snake_case : str=256 , snake_case : int=0.1 , snake_case : Dict=0.0 , snake_case : str=0.0 , snake_case : Union[str, Any]=0.02 , snake_case : Union[str, Any]=1.0 , snake_case : Optional[Any]=False , snake_case : int="sine" , snake_case : Optional[Any]="resnet50" , snake_case : Optional[int]=True , snake_case : Any=False , snake_case : int=1 , snake_case : Tuple=5 , snake_case : Optional[int]=2 , snake_case : Tuple=1 , snake_case : Optional[Any]=1 , snake_case : Optional[Any]=5 , snake_case : Dict=2 , snake_case : Any=0.1 , **snake_case : Any , ): '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) A__ : Optional[Any] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(snake_case , snake_case ): A__ : Optional[int] = backbone_config.get("""model_type""" ) A__ : Optional[int] = CONFIG_MAPPING[backbone_model_type] A__ : List[str] = config_class.from_dict(snake_case ) # set timm attributes to None A__ , A__ , A__ : str = None, None, None A__ : Tuple = use_timm_backbone A__ : str = backbone_config A__ : str = num_channels A__ : List[Any] = num_queries A__ : Optional[Any] = d_model A__ : Tuple = encoder_ffn_dim A__ : Union[str, Any] = encoder_layers A__ : List[Any] = encoder_attention_heads A__ : Optional[int] = decoder_ffn_dim A__ : Any = decoder_layers A__ : int = decoder_attention_heads A__ : Any = dropout A__ : Dict = attention_dropout A__ : Dict = activation_dropout A__ : Tuple = activation_function A__ : List[str] = init_std A__ : List[str] = init_xavier_std A__ : Any = encoder_layerdrop A__ : Optional[Any] = decoder_layerdrop A__ : Union[str, Any] = encoder_layers A__ : Dict = auxiliary_loss A__ : List[Any] = position_embedding_type A__ : Optional[Any] = backbone A__ : str = use_pretrained_backbone A__ : Union[str, Any] = dilation # Hungarian matcher A__ : Tuple = class_cost A__ : Optional[Any] = bbox_cost A__ : Dict = giou_cost # Loss coefficients A__ : Any = mask_loss_coefficient A__ : str = dice_loss_coefficient A__ : str = bbox_loss_coefficient A__ : Union[str, Any] = giou_loss_coefficient A__ : List[str] = eos_coefficient super().__init__(is_encoder_decoder=snake_case , **snake_case ) @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return self.encoder_attention_heads @property def _UpperCamelCase ( self : Dict ): '''simple docstring''' return self.d_model class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = version.parse('1.11' ) @property def _UpperCamelCase ( self : Any ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return 1e-5 @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return 12
296
1
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def lowercase ( SCREAMING_SNAKE_CASE__ : Any ) -> str: _snake_case : Union[str, Any] = [2, 2, 6, 2] if """tiny""" in model_name else [2, 2, 18, 2] _snake_case : List[str] = True if """large""" in model_name or """huge""" in model_name else False _snake_case : str = True if """large""" in model_name or """huge""" in model_name else False _snake_case : Optional[int] = True if """large""" in model_name or """huge""" in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: _snake_case : Optional[Any] = [3, 3, 3, 3] _snake_case : str = [5, 5, 5, 5] elif "fl4" in model_name: _snake_case : List[str] = [4, 4, 4, 4] _snake_case : str = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: _snake_case : Optional[Any] = [3, 3, 3, 3] if "lrf" in model_name: _snake_case : List[str] = [3, 3, 3, 3] else: _snake_case : Dict = [2, 2, 2, 2] if "tiny" in model_name: _snake_case : Tuple = 96 elif "small" in model_name: _snake_case : List[str] = 96 elif "base" in model_name: _snake_case : str = 128 elif "large" in model_name: _snake_case : Dict = 192 elif "xlarge" in model_name: _snake_case : str = 256 elif "huge" in model_name: _snake_case : List[str] = 352 # set label information _snake_case : Optional[int] = """huggingface/label-files""" if "large" in model_name or "huge" in model_name: _snake_case : str = """imagenet-22k-id2label.json""" else: _snake_case : Dict = """imagenet-1k-id2label.json""" _snake_case : Dict = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , repo_type="""dataset""" ) , """r""" ) ) _snake_case : Optional[Any] = {int(SCREAMING_SNAKE_CASE__ ): v for k, v in idalabel.items()} _snake_case : Optional[Any] = {v: k for k, v in idalabel.items()} _snake_case : Dict = FocalNetConfig( embed_dim=SCREAMING_SNAKE_CASE__ , depths=SCREAMING_SNAKE_CASE__ , focal_levels=SCREAMING_SNAKE_CASE__ , focal_windows=SCREAMING_SNAKE_CASE__ , use_conv_embed=SCREAMING_SNAKE_CASE__ , idalabel=SCREAMING_SNAKE_CASE__ , labelaid=SCREAMING_SNAKE_CASE__ , use_post_layernorm=SCREAMING_SNAKE_CASE__ , use_layerscale=SCREAMING_SNAKE_CASE__ , ) return config def lowercase ( SCREAMING_SNAKE_CASE__ : int ) -> Tuple: if "patch_embed.proj" in name: _snake_case : Dict = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "patch_embed.norm" in name: _snake_case : Dict = name.replace("""patch_embed.norm""" , """embeddings.norm""" ) if "layers" in name: _snake_case : Dict = """encoder.""" + name if "encoder.layers" in name: _snake_case : Union[str, Any] = name.replace("""encoder.layers""" , """encoder.stages""" ) if "downsample.proj" in name: _snake_case : str = name.replace("""downsample.proj""" , """downsample.projection""" ) if "blocks" in name: _snake_case : List[Any] = name.replace("""blocks""" , """layers""" ) if "modulation.f.weight" in name or "modulation.f.bias" in name: _snake_case : Union[str, Any] = name.replace("""modulation.f""" , """modulation.projection_in""" ) if "modulation.h.weight" in name or "modulation.h.bias" in name: _snake_case : Union[str, Any] = name.replace("""modulation.h""" , """modulation.projection_context""" ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: _snake_case : Union[str, Any] = name.replace("""modulation.proj""" , """modulation.projection_out""" ) if name == "norm.weight": _snake_case : Tuple = """layernorm.weight""" if name == "norm.bias": _snake_case : Optional[int] = """layernorm.bias""" if "head" in name: _snake_case : Any = name.replace("""head""" , """classifier""" ) else: _snake_case : Any = """focalnet.""" + name return name def lowercase ( SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Tuple=False ) -> str: # fmt: off _snake_case : Optional[Any] = { """focalnet-tiny""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth""", """focalnet-tiny-lrf""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth""", """focalnet-small""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth""", """focalnet-small-lrf""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth""", """focalnet-base""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth""", """focalnet-base-lrf""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth""", """focalnet-large-lrf-fl3""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth""", """focalnet-large-lrf-fl4""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth""", """focalnet-xlarge-lrf-fl3""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth""", """focalnet-xlarge-lrf-fl4""": """https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth""", } # fmt: on _snake_case : Dict = model_name_to_url[model_name] print("""Checkpoint URL: """ , SCREAMING_SNAKE_CASE__ ) _snake_case : Optional[Any] = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE__ , map_location="""cpu""" )["""model"""] # rename keys for key in state_dict.copy().keys(): _snake_case : Dict = state_dict.pop(SCREAMING_SNAKE_CASE__ ) _snake_case : Optional[Any] = val _snake_case : Optional[Any] = get_focalnet_config(SCREAMING_SNAKE_CASE__ ) _snake_case : Optional[Any] = FocalNetForImageClassification(SCREAMING_SNAKE_CASE__ ) model.eval() # load state dict model.load_state_dict(SCREAMING_SNAKE_CASE__ ) # verify conversion _snake_case : Dict = """http://images.cocodataset.org/val2017/000000039769.jpg""" _snake_case : Union[str, Any] = BitImageProcessor( do_resize=SCREAMING_SNAKE_CASE__ , size={"""shortest_edge""": 256} , resample=PILImageResampling.BILINEAR , do_center_crop=SCREAMING_SNAKE_CASE__ , crop_size=224 , do_normalize=SCREAMING_SNAKE_CASE__ , image_mean=SCREAMING_SNAKE_CASE__ , image_std=SCREAMING_SNAKE_CASE__ , ) _snake_case : Union[str, Any] = Image.open(requests.get(SCREAMING_SNAKE_CASE__ , stream=SCREAMING_SNAKE_CASE__ ).raw ) _snake_case : Tuple = processor(images=SCREAMING_SNAKE_CASE__ , return_tensors="""pt""" ) _snake_case : Optional[Any] = transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize(mean=[0.4_8_5, 0.4_5_6, 0.4_0_6] , std=[0.2_2_9, 0.2_2_4, 0.2_2_5] ), ] ) _snake_case : Any = image_transforms(SCREAMING_SNAKE_CASE__ ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , SCREAMING_SNAKE_CASE__ , atol=1e-4 ) _snake_case : str = model(**SCREAMING_SNAKE_CASE__ ) _snake_case : List[Any] = outputs.logits.argmax(-1 ).item() print("""Predicted class:""" , model.config.idalabel[predicted_class_idx] ) print("""First values of logits:""" , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": _snake_case : str = torch.tensor([0.2_1_6_6, -0.4_3_6_8, 0.2_1_9_1] ) elif model_name == "focalnet-tiny-lrf": _snake_case : Optional[Any] = torch.tensor([1.1_6_6_9, 0.0_1_2_5, -0.1_6_9_5] ) elif model_name == "focalnet-small": _snake_case : Optional[Any] = torch.tensor([0.4_9_1_7, -0.0_4_3_0, 0.1_3_4_1] ) elif model_name == "focalnet-small-lrf": _snake_case : List[Any] = torch.tensor([-0.2_5_8_8, -0.5_3_4_2, -0.2_3_3_1] ) elif model_name == "focalnet-base": _snake_case : Dict = torch.tensor([-0.1_6_5_5, -0.4_0_9_0, -0.1_7_3_0] ) elif model_name == "focalnet-base-lrf": _snake_case : Tuple = torch.tensor([0.5_3_0_6, -0.0_4_8_3, -0.3_9_2_8] ) assert torch.allclose(outputs.logits[0, :3] , SCREAMING_SNAKE_CASE__ , atol=1e-4 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: print(F'''Saving model and processor of {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) processor.save_pretrained(SCREAMING_SNAKE_CASE__ ) if push_to_hub: print(F'''Pushing model and processor of {model_name} to the hub...''' ) model.push_to_hub(F'''{model_name}''' ) processor.push_to_hub(F'''{model_name}''' ) if __name__ == "__main__": a__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""focalnet-tiny""", type=str, help="""Name of the FocalNet model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model and processor to the hub.""", ) a__ = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
317
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) a__ = { """configuration_wav2vec2""": ["""WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Wav2Vec2Config"""], """feature_extraction_wav2vec2""": ["""Wav2Vec2FeatureExtractor"""], """processing_wav2vec2""": ["""Wav2Vec2Processor"""], """tokenization_wav2vec2""": ["""Wav2Vec2CTCTokenizer""", """Wav2Vec2Tokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ = [ """WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST""", """Wav2Vec2ForAudioFrameClassification""", """Wav2Vec2ForCTC""", """Wav2Vec2ForMaskedLM""", """Wav2Vec2ForPreTraining""", """Wav2Vec2ForSequenceClassification""", """Wav2Vec2ForXVector""", """Wav2Vec2Model""", """Wav2Vec2PreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ = [ """TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFWav2Vec2ForCTC""", """TFWav2Vec2Model""", """TFWav2Vec2PreTrainedModel""", """TFWav2Vec2ForSequenceClassification""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ = [ """FlaxWav2Vec2ForCTC""", """FlaxWav2Vec2ForPreTraining""", """FlaxWav2Vec2Model""", """FlaxWav2Vec2PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_wavaveca import WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, WavaVecaConfig from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .processing_wavaveca import WavaVecaProcessor from .tokenization_wavaveca import WavaVecaCTCTokenizer, WavaVecaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavaveca import ( WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaForAudioFrameClassification, WavaVecaForCTC, WavaVecaForMaskedLM, WavaVecaForPreTraining, WavaVecaForSequenceClassification, WavaVecaForXVector, WavaVecaModel, WavaVecaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, TFWavaVecaForCTC, TFWavaVecaForSequenceClassification, TFWavaVecaModel, TFWavaVecaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( FlaxWavaVecaForCTC, FlaxWavaVecaForPreTraining, FlaxWavaVecaModel, FlaxWavaVecaPreTrainedModel, ) else: import sys a__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
317
1
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ =logging.get_logger(__name__) UpperCamelCase__ ={ 'caidas/swin2sr-classicalsr-x2-64': ( 'https://huggingface.co/caidas/swin2sr-classicalsr-x2-64/resolve/main/config.json' ), } class lowerCAmelCase__( __lowercase ): '''simple docstring''' __snake_case = 'swin2sr' __snake_case = { 'hidden_size': 'embed_dim', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , __lowerCamelCase=6_4 , __lowerCamelCase=1 , __lowerCamelCase=3 , __lowerCamelCase=1_8_0 , __lowerCamelCase=[6, 6, 6, 6, 6, 6] , __lowerCamelCase=[6, 6, 6, 6, 6, 6] , __lowerCamelCase=8 , __lowerCamelCase=2.0 , __lowerCamelCase=True , __lowerCamelCase=0.0 , __lowerCamelCase=0.0 , __lowerCamelCase=0.1 , __lowerCamelCase="gelu" , __lowerCamelCase=False , __lowerCamelCase=0.02 , __lowerCamelCase=1E-5 , __lowerCamelCase=2 , __lowerCamelCase=1.0 , __lowerCamelCase="1conv" , __lowerCamelCase="pixelshuffle" , **__lowerCamelCase , ) -> Union[str, Any]: super().__init__(**__lowerCamelCase ) _SCREAMING_SNAKE_CASE : int = image_size _SCREAMING_SNAKE_CASE : Any = patch_size _SCREAMING_SNAKE_CASE : Tuple = num_channels _SCREAMING_SNAKE_CASE : List[Any] = embed_dim _SCREAMING_SNAKE_CASE : Dict = depths _SCREAMING_SNAKE_CASE : Optional[int] = len(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : int = num_heads _SCREAMING_SNAKE_CASE : Union[str, Any] = window_size _SCREAMING_SNAKE_CASE : List[str] = mlp_ratio _SCREAMING_SNAKE_CASE : Union[str, Any] = qkv_bias _SCREAMING_SNAKE_CASE : List[Any] = hidden_dropout_prob _SCREAMING_SNAKE_CASE : int = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE : Tuple = drop_path_rate _SCREAMING_SNAKE_CASE : List[str] = hidden_act _SCREAMING_SNAKE_CASE : Optional[Any] = use_absolute_embeddings _SCREAMING_SNAKE_CASE : str = layer_norm_eps _SCREAMING_SNAKE_CASE : Dict = initializer_range _SCREAMING_SNAKE_CASE : List[Any] = upscale _SCREAMING_SNAKE_CASE : int = img_range _SCREAMING_SNAKE_CASE : str = resi_connection _SCREAMING_SNAKE_CASE : Optional[Any] = upsampler
325
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () UpperCamelCase__ =np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). UpperCamelCase__ =[0, 25, 50] UpperCamelCase__ =[25, 50, 75] UpperCamelCase__ =fuzz.membership.trimf(X, abca) UpperCamelCase__ =fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. UpperCamelCase__ =np.ones(75) UpperCamelCase__ =np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) UpperCamelCase__ =fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) UpperCamelCase__ =fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) UpperCamelCase__ =fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) UpperCamelCase__ =fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] UpperCamelCase__ =young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) UpperCamelCase__ =young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] UpperCamelCase__ =fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] UpperCamelCase__ =fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
325
1
from __future__ import annotations import inspect import unittest from math import floor import numpy as np from transformers import CvtConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFCvtForImageClassification, TFCvtModel from transformers.models.cvt.modeling_tf_cvt import TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE__ ( lowercase__ ): def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> List[str]: a_ : Any = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ , 'embed_dim' ) ) self.parent.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ , 'num_heads' ) ) class SCREAMING_SNAKE_CASE__ : def __init__( self : Tuple , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Tuple=1_3 , SCREAMING_SNAKE_CASE__ : Tuple=6_4 , SCREAMING_SNAKE_CASE__ : Dict=3 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=[1_6, 4_8, 9_6] , SCREAMING_SNAKE_CASE__ : str=[1, 3, 6] , SCREAMING_SNAKE_CASE__ : Tuple=[1, 2, 1_0] , SCREAMING_SNAKE_CASE__ : Optional[int]=[7, 3, 3] , SCREAMING_SNAKE_CASE__ : Optional[int]=[4, 2, 2] , SCREAMING_SNAKE_CASE__ : Optional[int]=[2, 1, 1] , SCREAMING_SNAKE_CASE__ : Dict=[2, 2, 2] , SCREAMING_SNAKE_CASE__ : str=[False, False, True] , SCREAMING_SNAKE_CASE__ : Tuple=[0.0, 0.0, 0.0] , SCREAMING_SNAKE_CASE__ : Dict=0.02 , SCREAMING_SNAKE_CASE__ : Dict=1E-12 , SCREAMING_SNAKE_CASE__ : List[str]=True , SCREAMING_SNAKE_CASE__ : Tuple=True , SCREAMING_SNAKE_CASE__ : Optional[int]=2 , ) -> int: a_ : Optional[Any] = parent a_ : List[str] = batch_size a_ : Optional[Any] = image_size a_ : Tuple = patch_sizes a_ : int = patch_stride a_ : Union[str, Any] = patch_padding a_ : List[Any] = is_training a_ : Optional[Any] = use_labels a_ : Tuple = num_labels a_ : int = num_channels a_ : Dict = embed_dim a_ : int = num_heads a_ : Tuple = stride_kv a_ : Any = depth a_ : List[str] = cls_token a_ : Any = attention_drop_rate a_ : List[Any] = initializer_range a_ : List[Any] = layer_norm_eps def SCREAMING_SNAKE_CASE ( self : List[str] ) -> str: a_ : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) a_ : Union[str, Any] = None if self.use_labels: # create a random int32 tensor of given shape a_ : int = ids_tensor([self.batch_size] , self.num_labels ) a_ : int = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Optional[Any]: return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def SCREAMING_SNAKE_CASE ( self : List[Any] , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> Optional[int]: a_ : Dict = TFCvtModel(config=SCREAMING_SNAKE_CASE__ ) a_ : Union[str, Any] = model(SCREAMING_SNAKE_CASE__ , training=SCREAMING_SNAKE_CASE__ ) a_ : List[Any] = (self.image_size, self.image_size) a_ , a_ : Union[str, Any] = image_size[0], image_size[1] for i in range(len(self.depth ) ): a_ : Any = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) a_ : List[Any] = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width) ) def SCREAMING_SNAKE_CASE ( self : Dict , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Any ) -> Optional[Any]: a_ : List[str] = self.num_labels a_ : List[str] = TFCvtForImageClassification(SCREAMING_SNAKE_CASE__ ) a_ : int = model(SCREAMING_SNAKE_CASE__ , labels=SCREAMING_SNAKE_CASE__ , training=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Optional[int]: a_ : Optional[Any] = self.prepare_config_and_inputs() a_ , a_ , a_ : Union[str, Any] = config_and_inputs a_ : Any = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE__ ( lowercase__ , lowercase__ , unittest.TestCase ): snake_case__ : str = (TFCvtModel, TFCvtForImageClassification) if is_tf_available() else () snake_case__ : List[Any] = ( {'''feature-extraction''': TFCvtModel, '''image-classification''': TFCvtForImageClassification} if is_tf_available() else {} ) snake_case__ : List[Any] = False snake_case__ : Dict = False snake_case__ : Optional[Any] = False snake_case__ : Optional[int] = False snake_case__ : List[Any] = False def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> int: a_ : str = TFCvtModelTester(self ) a_ : List[Any] = TFCvtConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ , has_text_modality=SCREAMING_SNAKE_CASE__ , hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> List[Any]: self.config_tester.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() @unittest.skip(reason='Cvt does not output attentions' ) def SCREAMING_SNAKE_CASE ( self : int ) -> str: pass @unittest.skip(reason='Cvt does not use inputs_embeds' ) def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Optional[int]: pass @unittest.skip(reason='Cvt does not support input and output embeddings' ) def SCREAMING_SNAKE_CASE ( self : int ) -> Union[str, Any]: pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('GPU' ) ) == 0 , reason='TF does not support backprop for grouped convolutions on CPU.' , ) def SCREAMING_SNAKE_CASE ( self : int ) -> int: super().test_dataset_conversion() @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('GPU' ) ) == 0 , reason='TF does not support backprop for grouped convolutions on CPU.' , ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ) -> Any: super().test_keras_fit() @unittest.skip(reason='Get `Failed to determine best cudnn convolution algo.` error after using TF 2.12+cuda 11.8' ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Any: a_ : Optional[int] = tf.keras.mixed_precision.Policy('mixed_float16' ) tf.keras.mixed_precision.set_global_policy(SCREAMING_SNAKE_CASE__ ) super().test_keras_fit() tf.keras.mixed_precision.set_global_policy('float32' ) def SCREAMING_SNAKE_CASE ( self : Any ) -> List[Any]: a_ , a_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a_ : int = model_class(SCREAMING_SNAKE_CASE__ ) a_ : int = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic a_ : Optional[int] = [*signature.parameters.keys()] a_ : Union[str, Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] , SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Optional[Any]: def check_hidden_states_output(SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] ): a_ : Union[str, Any] = model_class(SCREAMING_SNAKE_CASE__ ) a_ : Optional[int] = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) a_ : Optional[int] = outputs.hidden_states a_ : Dict = len(self.model_tester.depth ) self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) , SCREAMING_SNAKE_CASE__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) a_ , a_ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a_ : Optional[int] = True check_hidden_states_output(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] a_ : Union[str, Any] = True check_hidden_states_output(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> int: a_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ) -> str: a_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*SCREAMING_SNAKE_CASE__ ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Any: for model_name in TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a_ : Any = TFCvtModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE_ ( ) -> Optional[int]: """simple docstring""" a_ : Optional[int] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @cached_property def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Dict: return AutoImageProcessor.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Optional[int]: a_ : int = TFCvtForImageClassification.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) a_ : Optional[int] = self.default_image_processor a_ : List[Any] = prepare_img() a_ : Optional[Any] = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors='tf' ) # forward pass a_ : Optional[Any] = model(**SCREAMING_SNAKE_CASE__ ) # verify the logits a_ : List[Any] = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , SCREAMING_SNAKE_CASE__ ) a_ : Any = tf.constant([0.9285, 0.9015, -0.3150] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , SCREAMING_SNAKE_CASE__ , atol=1E-4 ) )
32
import flax.linen as nn import jax import jax.numpy as jnp class SCREAMING_SNAKE_CASE__ ( nn.Module ): snake_case__ : int snake_case__ : jnp.dtype = jnp.floataa def SCREAMING_SNAKE_CASE ( self : str ) -> int: a_ : Dict = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : Tuple , SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> Optional[int]: a_ , a_ , a_ , a_ : Union[str, Any] = hidden_states.shape a_ : List[str] = jax.image.resize( SCREAMING_SNAKE_CASE__ , shape=(batch, height * 2, width * 2, channels) , method='nearest' , ) a_ : Any = self.conv(SCREAMING_SNAKE_CASE__ ) return hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): snake_case__ : int snake_case__ : jnp.dtype = jnp.floataa def SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[int]: a_ : Optional[int] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : Any , SCREAMING_SNAKE_CASE__ : int ) -> Optional[int]: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) a_ : str = self.conv(SCREAMING_SNAKE_CASE__ ) return hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): snake_case__ : int snake_case__ : int = None snake_case__ : float = 0.0 snake_case__ : bool = None snake_case__ : jnp.dtype = jnp.floataa def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Dict: a_ : List[str] = self.in_channels if self.out_channels is None else self.out_channels a_ : Optional[int] = nn.GroupNorm(num_groups=3_2 , epsilon=1E-5 ) a_ : Any = nn.Conv( SCREAMING_SNAKE_CASE__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a_ : Optional[int] = nn.Dense(SCREAMING_SNAKE_CASE__ , dtype=self.dtype ) a_ : Union[str, Any] = nn.GroupNorm(num_groups=3_2 , epsilon=1E-5 ) a_ : int = nn.Dropout(self.dropout_prob ) a_ : Optional[Any] = nn.Conv( SCREAMING_SNAKE_CASE__ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) a_ : List[str] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut a_ : List[Any] = None if use_nin_shortcut: a_ : Union[str, Any] = nn.Conv( SCREAMING_SNAKE_CASE__ , kernel_size=(1, 1) , strides=(1, 1) , padding='VALID' , dtype=self.dtype , ) def __call__( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : List[Any]=True ) -> int: a_ : List[Any] = hidden_states a_ : Any = self.norma(SCREAMING_SNAKE_CASE__ ) a_ : Any = nn.swish(SCREAMING_SNAKE_CASE__ ) a_ : Union[str, Any] = self.conva(SCREAMING_SNAKE_CASE__ ) a_ : int = self.time_emb_proj(nn.swish(SCREAMING_SNAKE_CASE__ ) ) a_ : List[str] = jnp.expand_dims(jnp.expand_dims(SCREAMING_SNAKE_CASE__ , 1 ) , 1 ) a_ : Optional[int] = hidden_states + temb a_ : List[str] = self.norma(SCREAMING_SNAKE_CASE__ ) a_ : Tuple = nn.swish(SCREAMING_SNAKE_CASE__ ) a_ : Optional[int] = self.dropout(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) a_ : List[Any] = self.conva(SCREAMING_SNAKE_CASE__ ) if self.conv_shortcut is not None: a_ : List[str] = self.conv_shortcut(SCREAMING_SNAKE_CASE__ ) return hidden_states + residual
32
1
'''simple docstring''' from typing import Dict from .base import GenericTensor, Pipeline class A__ ( A__ ): def A ( self : Optional[int] , _a : int=None , _a : Optional[Any]=None , _a : Tuple=None , **_a : int ) -> Union[str, Any]: '''simple docstring''' if tokenize_kwargs is None: _SCREAMING_SNAKE_CASE ={} if truncation is not None: if "truncation" in tokenize_kwargs: raise ValueError( 'truncation parameter defined twice (given as keyword argument as well as in tokenize_kwargs)' ) _SCREAMING_SNAKE_CASE =truncation _SCREAMING_SNAKE_CASE =tokenize_kwargs _SCREAMING_SNAKE_CASE ={} if return_tensors is not None: _SCREAMING_SNAKE_CASE =return_tensors return preprocess_params, {}, postprocess_params def A ( self : Union[str, Any] , _a : int , **_a : int ) -> Dict[str, GenericTensor]: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.framework _SCREAMING_SNAKE_CASE =self.tokenizer(_a , return_tensors=_a , **_a ) return model_inputs def A ( self : Optional[Any] , _a : Optional[int] ) -> str: '''simple docstring''' _SCREAMING_SNAKE_CASE =self.model(**_a ) return model_outputs def A ( self : Tuple , _a : Optional[int] , _a : Optional[int]=False ) -> str: '''simple docstring''' if return_tensors: return model_outputs[0] if self.framework == "pt": return model_outputs[0].tolist() elif self.framework == "tf": return model_outputs[0].numpy().tolist() def __call__( self : Tuple , *_a : List[Any] , **_a : Optional[int] ) -> List[str]: '''simple docstring''' return super().__call__(*_a , **_a )
114
'''simple docstring''' import os import sys import unittest lowerCamelCase : Optional[Any] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path lowerCamelCase : Optional[int] = os.path.join(git_repo_path, "src", "transformers") lowerCamelCase : Union[str, Any] = "\n{0} = None\n" lowerCamelCase : Optional[Any] = "\nclass {0}(metaclass=DummyObject):\n _backends = {1}\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, {1})\n" lowerCamelCase : List[Any] = "\ndef {0}(*args, **kwargs):\n requires_backends({0}, {1})\n" class A__ ( unittest.TestCase ): def A ( self : Optional[int] ) -> Dict: '''simple docstring''' _SCREAMING_SNAKE_CASE =find_backend(' _import_structure["models.albert"].append("AlbertTokenizerFast")' ) self.assertIsNone(_a ) _SCREAMING_SNAKE_CASE =find_backend(' if not is_tokenizers_available():' ) self.assertEqual(_a , 'tokenizers' ) _SCREAMING_SNAKE_CASE =find_backend(' if not is_tensorflow_text_available():' ) self.assertEqual(_a , 'tensorflow_text' ) _SCREAMING_SNAKE_CASE =find_backend(' if not (is_sentencepiece_available() and is_tokenizers_available()):' ) self.assertEqual(_a , 'sentencepiece_and_tokenizers' ) _SCREAMING_SNAKE_CASE =find_backend( ' if not (is_sentencepiece_available() and is_tensorflow_text_available()):' ) self.assertEqual(_a , 'sentencepiece_and_tensorflow_text' ) _SCREAMING_SNAKE_CASE =find_backend( ' if not (is_sentencepiece_available() and is_tokenizers_available() and is_vision_available()):' ) self.assertEqual(_a , 'sentencepiece_and_tokenizers_and_vision' ) def A ( self : Optional[int] ) -> Any: '''simple docstring''' _SCREAMING_SNAKE_CASE =read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn('torch' , _a ) self.assertIn('tensorflow_text' , _a ) self.assertIn('sentencepiece_and_tokenizers' , _a ) # Likewise, we can't assert on the exact content of a key self.assertIn('BertModel' , objects['torch'] ) self.assertIn('TFBertModel' , objects['tf'] ) self.assertIn('FlaxBertModel' , objects['flax'] ) self.assertIn('BertModel' , objects['torch'] ) self.assertIn('TFBertTokenizer' , objects['tensorflow_text'] ) self.assertIn('convert_slow_tokenizer' , objects['sentencepiece_and_tokenizers'] ) def A ( self : Optional[int] ) -> List[str]: '''simple docstring''' _SCREAMING_SNAKE_CASE =create_dummy_object('CONSTANT' , '\'torch\'' ) self.assertEqual(_a , '\nCONSTANT = None\n' ) _SCREAMING_SNAKE_CASE =create_dummy_object('function' , '\'torch\'' ) self.assertEqual( _a , '\ndef function(*args, **kwargs):\n requires_backends(function, \'torch\')\n' ) _SCREAMING_SNAKE_CASE ='\nclass FakeClass(metaclass=DummyObject):\n _backends = \'torch\'\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, \'torch\')\n' _SCREAMING_SNAKE_CASE =create_dummy_object('FakeClass' , '\'torch\'' ) self.assertEqual(_a , _a ) def A ( self : Optional[int] ) -> Optional[int]: '''simple docstring''' _SCREAMING_SNAKE_CASE ='# This file is autogenerated by the command `make fix-copies`, do not edit.\nfrom ..utils import DummyObject, requires_backends\n\n\nCONSTANT = None\n\n\ndef function(*args, **kwargs):\n requires_backends(function, ["torch"])\n\n\nclass FakeClass(metaclass=DummyObject):\n _backends = ["torch"]\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, ["torch"])\n' _SCREAMING_SNAKE_CASE =create_dummy_files({'torch': ['CONSTANT', 'function', 'FakeClass']} ) self.assertEqual(dummy_files['torch'] , _a )
114
1
import json import os import unittest from transformers import DebertaTokenizer, DebertaTokenizerFast from transformers.models.deberta.tokenization_deberta import VOCAB_FILES_NAMES from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): '''simple docstring''' __lowerCamelCase : Union[str, Any] = DebertaTokenizer __lowerCamelCase : Any = True __lowerCamelCase : Tuple = DebertaTokenizerFast def _lowerCAmelCase ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt A : int = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """[UNK]""", ] A : Optional[Any] = dict(zip(lowerCamelCase__, range(len(lowerCamelCase__ ) ) ) ) A : Tuple = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] A : List[str] = {"""unk_token""": """[UNK]"""} A : Optional[Any] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""vocab_file"""] ) A : str = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file, """w""", encoding="""utf-8""" ) as fp: fp.write(json.dumps(lowerCamelCase__ ) + """\n""" ) with open(self.merges_file, """w""", encoding="""utf-8""" ) as fp: fp.write("""\n""".join(lowerCamelCase__ ) ) def _lowerCAmelCase ( self, **lowerCamelCase__ ): kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname, **lowerCamelCase__ ) def _lowerCAmelCase ( self, lowerCamelCase__ ): A : Any = """lower newer""" A : Dict = """lower newer""" return input_text, output_text def _lowerCAmelCase ( self ): A : Dict = self.get_tokenizer() A : Dict = """lower newer""" A : List[Any] = ["""l""", """o""", """w""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] A : Optional[Any] = tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__, lowerCamelCase__ ) A : Optional[int] = tokens + [tokenizer.unk_token] A : Optional[Any] = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase__ ), lowerCamelCase__ ) def _lowerCAmelCase ( self ): A : Any = self.get_tokenizer() A : Any = tokenizer("""Hello""", """World""" ) A : Optional[int] = [0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1] self.assertListEqual(tokd["""token_type_ids"""], lowerCamelCase__ ) @slow def _lowerCAmelCase ( self ): A : str = self.tokenizer_class.from_pretrained("""microsoft/deberta-base""" ) A : Tuple = tokenizer.encode("""sequence builders""", add_special_tokens=lowerCamelCase__ ) A : Optional[Any] = tokenizer.encode("""multi-sequence build""", add_special_tokens=lowerCamelCase__ ) A : str = tokenizer.encode( """sequence builders""", add_special_tokens=lowerCamelCase__, add_prefix_space=lowerCamelCase__ ) A : Any = tokenizer.encode( """sequence builders""", """multi-sequence build""", add_special_tokens=lowerCamelCase__, add_prefix_space=lowerCamelCase__ ) A : Any = tokenizer.build_inputs_with_special_tokens(lowerCamelCase__ ) A : str = tokenizer.build_inputs_with_special_tokens(lowerCamelCase__, lowerCamelCase__ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode @slow def _lowerCAmelCase ( self ): A : Dict = [self.tokenizer_class] if self.test_rust_tokenizer: tokenizer_classes.append(self.rust_tokenizer_class ) for tokenizer_class in tokenizer_classes: A : List[str] = tokenizer_class.from_pretrained("""microsoft/deberta-base""" ) A : int = [ """ALBERT: A Lite BERT for Self-supervised Learning of Language Representations""", """ALBERT incorporates two parameter reduction techniques""", """The first one is a factorized embedding parameterization. By decomposing the large vocabulary""" """ embedding matrix into two small matrices, we separate the size of the hidden layers from the size of""" """ vocabulary embedding.""", ] A : Tuple = tokenizer(lowerCamelCase__, padding=lowerCamelCase__ ) A : Dict = [tokenizer.decode(lowerCamelCase__, skip_special_tokens=lowerCamelCase__ ) for seq in encoding["""input_ids"""]] # fmt: off A : Optional[int] = { """input_ids""": [ [1, 2118, 1_1126, 565, 35, 83, 2_5191, 163, 1_8854, 13, 1_2156, 12, 1_6101, 2_5376, 1_3807, 9, 2_2205, 2_7893, 1635, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 2118, 1_1126, 565, 2_4536, 80, 4_3797, 4878, 7373, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 133, 78, 65, 16, 10, 3724, 1538, 3_3183, 1_1303, 4_3797, 1938, 4, 870, 2_4165, 2_9105, 5, 739, 3_2644, 3_3183, 1_1303, 3_6173, 88, 80, 650, 7821, 4_5940, 6, 52, 2559, 5, 1836, 9, 5, 7397, 1_3171, 31, 5, 1836, 9, 3_2644, 3_3183, 1_1303, 4, 2] ], """token_type_ids""": [ [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ], """attention_mask""": [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1] ] } # fmt: on A : List[str] = [ """ALBERT: A Lite BERT for Self-supervised Learning of Language Representations""", """ALBERT incorporates two parameter reduction techniques""", """The first one is a factorized embedding parameterization. By decomposing the large vocabulary""" """ embedding matrix into two small matrices, we separate the size of the hidden layers from the size of""" """ vocabulary embedding.""", ] self.assertDictEqual(encoding.data, lowerCamelCase__ ) for expected, decoded in zip(lowerCamelCase__, lowerCamelCase__ ): self.assertEqual(lowerCamelCase__, lowerCamelCase__ )
116
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE_:List[Any] = """▁""" SCREAMING_SNAKE_CASE_:int = {"""vocab_file""": """spiece.model"""} SCREAMING_SNAKE_CASE_:Any = { """vocab_file""": {"""google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"""} } SCREAMING_SNAKE_CASE_:Optional[Any] = { """google/pegasus-xsum""": 512, } SCREAMING_SNAKE_CASE_:Tuple = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' __lowerCamelCase : Union[str, Any] = VOCAB_FILES_NAMES __lowerCamelCase : List[Any] = VOCAB_FILES_NAMES __lowerCamelCase : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase : str = ["input_ids", "attention_mask"] def __init__( self, lowerCamelCase__, lowerCamelCase__="<pad>", lowerCamelCase__="</s>", lowerCamelCase__="<unk>", lowerCamelCase__="<mask_2>", lowerCamelCase__="<mask_1>", lowerCamelCase__=None, lowerCamelCase__=103, lowerCamelCase__ = None, **lowerCamelCase__, ): A : int = offset if additional_special_tokens is not None: if not isinstance(lowerCamelCase__, lowerCamelCase__ ): raise TypeError( f'''additional_special_tokens should be of type {type(lowerCamelCase__ )}, but is''' f''' {type(lowerCamelCase__ )}''' ) A : Dict = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'''<unk_{i}>''' for i in range(len(lowerCamelCase__ ), self.offset - 1 ) ] if len(set(lowerCamelCase__ ) ) != len(lowerCamelCase__ ): raise ValueError( """Please make sure that the provided additional_special_tokens do not contain an incorrectly""" f''' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.''' ) A : int = additional_special_tokens_extended else: A : Optional[int] = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'''<unk_{i}>''' for i in range(2, self.offset )] A : Any = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=lowerCamelCase__, unk_token=lowerCamelCase__, mask_token=lowerCamelCase__, pad_token=lowerCamelCase__, mask_token_sent=lowerCamelCase__, offset=lowerCamelCase__, additional_special_tokens=lowerCamelCase__, sp_model_kwargs=self.sp_model_kwargs, **lowerCamelCase__, ) A : Union[str, Any] = mask_token_sent A : Optional[Any] = vocab_file A : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCamelCase__ ) # add special tokens to encoder dict A : Dict[int, str] = { 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1, self.offset - 1 )} ) A : Dict[str, int] = {v: k for k, v in self.encoder.items()} @property def _lowerCAmelCase ( self ): return len(self.sp_model ) + self.offset def _lowerCAmelCase ( self ): A : Optional[int] = {self.convert_ids_to_tokens(lowerCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): A : List[Any] = self.__dict__.copy() A : Union[str, Any] = None return state def __setstate__( self, lowerCamelCase__ ): A : List[Any] = d # for backward compatibility if not hasattr(self, """sp_model_kwargs""" ): A : int = {} A : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _lowerCAmelCase ( self, lowerCamelCase__ ): return self.sp_model.encode(lowerCamelCase__, out_type=lowerCamelCase__ ) def _lowerCAmelCase ( self, lowerCamelCase__ ): if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] A : List[str] = self.sp_model.piece_to_id(lowerCamelCase__ ) return sp_id + self.offset def _lowerCAmelCase ( self, lowerCamelCase__ ): if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: A : Dict = self.sp_model.IdToPiece(index - self.offset ) return token def _lowerCAmelCase ( self, lowerCamelCase__ ): A : Optional[int] = [] A : Optional[Any] = """""" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(lowerCamelCase__ ) + token A : int = [] else: current_sub_tokens.append(lowerCamelCase__ ) out_string += self.sp_model.decode(lowerCamelCase__ ) return out_string.strip() def _lowerCAmelCase ( self, lowerCamelCase__=False ): return 1 def _lowerCAmelCase ( self, lowerCamelCase__ ): A : Optional[int] = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def _lowerCAmelCase ( self, lowerCamelCase__, lowerCamelCase__ = None, lowerCamelCase__ = False ): if already_has_special_tokens: return self._special_token_mask(lowerCamelCase__ ) elif token_ids_a is None: return self._special_token_mask(lowerCamelCase__ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _lowerCAmelCase ( self, lowerCamelCase__, lowerCamelCase__=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _lowerCAmelCase ( self, lowerCamelCase__, lowerCamelCase__ = None ): if not os.path.isdir(lowerCamelCase__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return A : Any = os.path.join( lowerCamelCase__, (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file, lowerCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase__, """wb""" ) as fi: A : Dict = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase__ ) return (out_vocab_file,)
116
1
__a :Optional[Any] = 'Input must be a string of 8 numbers plus letter' __a :int = 'TRWAGMYFPDXBNJZSQVHLCKE' def __snake_case ( __UpperCamelCase : str ): """simple docstring""" if not isinstance(__UpperCamelCase ,__UpperCamelCase ): A_ = f'''Expected string as input, found {type(__UpperCamelCase ).__name__}''' raise TypeError(__UpperCamelCase ) A_ = spanish_id.replace("-" ,"" ).upper() if len(__UpperCamelCase ) != 9: raise ValueError(__UpperCamelCase ) try: A_ = int(spanish_id_clean[0:8] ) A_ = spanish_id_clean[8] except ValueError as ex: raise ValueError(__UpperCamelCase ) from ex if letter.isdigit(): raise ValueError(__UpperCamelCase ) return letter == LOOKUP_LETTERS[number % 23] if __name__ == "__main__": import doctest doctest.testmod()
329
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, BlipProcessor, PreTrainedTokenizerFast @require_vision class _a ( unittest.TestCase ): """simple docstring""" def __A ( self : int ): A_ = tempfile.mkdtemp() A_ = BlipImageProcessor() A_ = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-BertModel" ) A_ = BlipProcessor(UpperCAmelCase , UpperCAmelCase ) processor.save_pretrained(self.tmpdirname ) def __A ( self : Optional[int] , **UpperCAmelCase : Union[str, Any] ): return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase ).tokenizer def __A ( self : Optional[Any] , **UpperCAmelCase : int ): return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase ).image_processor def __A ( self : Any ): shutil.rmtree(self.tmpdirname ) def __A ( self : Dict ): A_ = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] A_ = [Image.fromarray(np.moveaxis(UpperCAmelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __A ( self : Any ): A_ = BlipProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) A_ = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) A_ = self.get_image_processor(do_normalize=UpperCAmelCase , padding_value=1.0 ) A_ = BlipProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=UpperCAmelCase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase ) def __A ( self : Dict ): A_ = self.get_image_processor() A_ = self.get_tokenizer() A_ = BlipProcessor(tokenizer=UpperCAmelCase , image_processor=UpperCAmelCase ) A_ = self.prepare_image_inputs() A_ = image_processor(UpperCAmelCase , return_tensors="np" ) A_ = processor(images=UpperCAmelCase , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def __A ( self : int ): A_ = self.get_image_processor() A_ = self.get_tokenizer() A_ = BlipProcessor(tokenizer=UpperCAmelCase , image_processor=UpperCAmelCase ) A_ = "lower newer" A_ = processor(text=UpperCAmelCase ) A_ = tokenizer(UpperCAmelCase , return_token_type_ids=UpperCAmelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __A ( self : Tuple ): A_ = self.get_image_processor() A_ = self.get_tokenizer() A_ = BlipProcessor(tokenizer=UpperCAmelCase , image_processor=UpperCAmelCase ) A_ = "lower newer" A_ = self.prepare_image_inputs() A_ = processor(text=UpperCAmelCase , images=UpperCAmelCase ) self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "input_ids", "attention_mask"] ) # test if it raises when no input is passed with pytest.raises(UpperCAmelCase ): processor() def __A ( self : Any ): A_ = self.get_image_processor() A_ = self.get_tokenizer() A_ = BlipProcessor(tokenizer=UpperCAmelCase , image_processor=UpperCAmelCase ) A_ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] A_ = processor.batch_decode(UpperCAmelCase ) A_ = tokenizer.batch_decode(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) def __A ( self : Optional[Any] ): A_ = self.get_image_processor() A_ = self.get_tokenizer() A_ = BlipProcessor(tokenizer=UpperCAmelCase , image_processor=UpperCAmelCase ) A_ = "lower newer" A_ = self.prepare_image_inputs() A_ = processor(text=UpperCAmelCase , images=UpperCAmelCase ) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "input_ids", "attention_mask"] )
329
1
'''simple docstring''' from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class UpperCAmelCase ( UpperCamelCase__ ): __lowercase = 42 __lowercase = 42 try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.26.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version(">=", "0.0.12") ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class UpperCAmelCase ( UpperCamelCase__ ): __lowercase = 42 __lowercase = 42 from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
237
import torch from diffusers import DDIMParallelScheduler from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE__ ( UpperCamelCase__ ): __SCREAMING_SNAKE_CASE = (DDIMParallelScheduler,) __SCREAMING_SNAKE_CASE = (('''eta''', 0.0), ('''num_inference_steps''', 50)) def UpperCamelCase ( self,**__lowerCamelCase ): A__ = { '''num_train_timesteps''': 1000, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''clip_sample''': True, } config.update(**__lowerCamelCase ) return config def UpperCamelCase ( self,**__lowerCamelCase ): A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config(**__lowerCamelCase ) A__ = scheduler_class(**__lowerCamelCase ) A__ , A__ = 10, 0.0 A__ = self.dummy_model() A__ = self.dummy_sample_deter scheduler.set_timesteps(__lowerCamelCase ) for t in scheduler.timesteps: A__ = model(__lowerCamelCase,__lowerCamelCase ) A__ = scheduler.step(__lowerCamelCase,__lowerCamelCase,__lowerCamelCase,__lowerCamelCase ).prev_sample return sample def UpperCamelCase ( self ): for timesteps in [100, 500, 1000]: self.check_over_configs(num_train_timesteps=__lowerCamelCase ) def UpperCamelCase ( self ): for steps_offset in [0, 1]: self.check_over_configs(steps_offset=__lowerCamelCase ) A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config(steps_offset=1 ) A__ = scheduler_class(**__lowerCamelCase ) scheduler.set_timesteps(5 ) assert torch.equal(scheduler.timesteps,torch.LongTensor([801, 601, 401, 201, 1] ) ) def UpperCamelCase ( self ): for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1],[0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=__lowerCamelCase,beta_end=__lowerCamelCase ) def UpperCamelCase ( self ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=__lowerCamelCase ) def UpperCamelCase ( self ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=__lowerCamelCase ) def UpperCamelCase ( self ): for clip_sample in [True, False]: self.check_over_configs(clip_sample=__lowerCamelCase ) def UpperCamelCase ( self ): for timestep_spacing in ["trailing", "leading"]: self.check_over_configs(timestep_spacing=__lowerCamelCase ) def UpperCamelCase ( self ): for rescale_betas_zero_snr in [True, False]: self.check_over_configs(rescale_betas_zero_snr=__lowerCamelCase ) def UpperCamelCase ( self ): self.check_over_configs(thresholding=__lowerCamelCase ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs( thresholding=__lowerCamelCase,prediction_type=__lowerCamelCase,sample_max_value=__lowerCamelCase,) def UpperCamelCase ( self ): for t in [1, 10, 49]: self.check_over_forward(time_step=__lowerCamelCase ) def UpperCamelCase ( self ): for t, num_inference_steps in zip([1, 10, 50],[10, 50, 500] ): self.check_over_forward(time_step=__lowerCamelCase,num_inference_steps=__lowerCamelCase ) def UpperCamelCase ( self ): for t, eta in zip([1, 10, 49],[0.0, 0.5, 1.0] ): self.check_over_forward(time_step=__lowerCamelCase,eta=__lowerCamelCase ) def UpperCamelCase ( self ): A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config() A__ = scheduler_class(**__lowerCamelCase ) assert torch.sum(torch.abs(scheduler._get_variance(0,0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(420,400 ) - 0.14771 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(980,960 ) - 0.32460 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(0,0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487,486 ) - 0.00979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999,998 ) - 0.02 ) ) < 1E-5 def UpperCamelCase ( self ): A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config() A__ = scheduler_class(**__lowerCamelCase ) A__ , A__ = 10, 0.0 scheduler.set_timesteps(__lowerCamelCase ) A__ = self.dummy_model() A__ = self.dummy_sample_deter A__ = self.dummy_sample_deter + 0.1 A__ = self.dummy_sample_deter - 0.1 A__ = samplea.shape[0] A__ = torch.stack([samplea, samplea, samplea],dim=0 ) A__ = torch.arange(__lowerCamelCase )[0:3, None].repeat(1,__lowerCamelCase ) A__ = model(samples.flatten(0,1 ),timesteps.flatten(0,1 ) ) A__ = scheduler.batch_step_no_noise(__lowerCamelCase,timesteps.flatten(0,1 ),samples.flatten(0,1 ),__lowerCamelCase ) A__ = torch.sum(torch.abs(__lowerCamelCase ) ) A__ = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 1147.7904 ) < 1E-2 assert abs(result_mean.item() - 0.4982 ) < 1E-3 def UpperCamelCase ( self ): A__ = self.full_loop() A__ = torch.sum(torch.abs(__lowerCamelCase ) ) A__ = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 172.0067 ) < 1E-2 assert abs(result_mean.item() - 0.223967 ) < 1E-3 def UpperCamelCase ( self ): A__ = self.full_loop(prediction_type='''v_prediction''' ) A__ = torch.sum(torch.abs(__lowerCamelCase ) ) A__ = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 52.5302 ) < 1E-2 assert abs(result_mean.item() - 0.0684 ) < 1E-3 def UpperCamelCase ( self ): # We specify different beta, so that the first alpha is 0.99 A__ = self.full_loop(set_alpha_to_one=__lowerCamelCase,beta_start=0.01 ) A__ = torch.sum(torch.abs(__lowerCamelCase ) ) A__ = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 149.8295 ) < 1E-2 assert abs(result_mean.item() - 0.1951 ) < 1E-3 def UpperCamelCase ( self ): # We specify different beta, so that the first alpha is 0.99 A__ = self.full_loop(set_alpha_to_one=__lowerCamelCase,beta_start=0.01 ) A__ = torch.sum(torch.abs(__lowerCamelCase ) ) A__ = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 149.0784 ) < 1E-2 assert abs(result_mean.item() - 0.1941 ) < 1E-3
193
0
import json import os import shutil import tempfile import unittest from transformers import BatchEncoding, CanineTokenizer from transformers.testing_utils import require_tokenizers, require_torch from transformers.tokenization_utils import AddedToken from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin class _snake_case ( _lowercase , unittest.TestCase ): lowerCamelCase__: List[Any] = CanineTokenizer lowerCamelCase__: Optional[int] = False def _lowerCamelCase ( self: Optional[Any] ) -> Optional[int]: super().setUp() __UpperCAmelCase : Tuple = CanineTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def _lowerCamelCase ( self: Union[str, Any] ) -> List[Any]: return CanineTokenizer.from_pretrained("google/canine-s" ) def _lowerCamelCase ( self: Any , **__lowerCamelCase: List[Any] ) -> CanineTokenizer: __UpperCAmelCase : Optional[int] = self.tokenizer_class.from_pretrained(self.tmpdirname , **__lowerCamelCase ) __UpperCAmelCase : Optional[int] = 10_24 return tokenizer @require_torch def _lowerCamelCase ( self: List[str] ) -> int: __UpperCAmelCase : Union[str, Any] = self.canine_tokenizer __UpperCAmelCase : List[str] = ["Life is like a box of chocolates.", "You never know what you're gonna get."] # fmt: off __UpperCAmelCase : Dict = [5_73_44, 76, 1_05, 1_02, 1_01, 32, 1_05, 1_15, 32, 1_08, 1_05, 1_07, 1_01, 32, 97, 32, 98, 1_11, 1_20, 32, 1_11, 1_02, 32, 99, 1_04, 1_11, 99, 1_11, 1_08, 97, 1_16, 1_01, 1_15, 46, 5_73_45, 0, 0, 0, 0] # fmt: on __UpperCAmelCase : Union[str, Any] = tokenizer(__lowerCamelCase , padding=__lowerCamelCase , return_tensors="pt" ) self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) __UpperCAmelCase : Optional[Any] = list(batch.input_ids.numpy()[0] ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) self.assertEqual((2, 39) , batch.input_ids.shape ) self.assertEqual((2, 39) , batch.attention_mask.shape ) @require_torch def _lowerCamelCase ( self: Optional[Any] ) -> Tuple: __UpperCAmelCase : Optional[Any] = self.canine_tokenizer __UpperCAmelCase : Dict = ["Once there was a man.", "He wrote a test in HuggingFace Tranformers."] __UpperCAmelCase : Union[str, Any] = tokenizer(__lowerCamelCase , padding=__lowerCamelCase , return_tensors="pt" ) # check if input_ids, attention_mask and token_type_ids are returned self.assertIn("input_ids" , __lowerCamelCase ) self.assertIn("attention_mask" , __lowerCamelCase ) self.assertIn("token_type_ids" , __lowerCamelCase ) @require_torch def _lowerCamelCase ( self: Any ) -> List[str]: __UpperCAmelCase : Optional[Any] = self.canine_tokenizer __UpperCAmelCase : int = [ "What's the weater?", "It's about 25 degrees.", ] __UpperCAmelCase : List[Any] = tokenizer( text_target=__lowerCamelCase , max_length=32 , padding="max_length" , truncation=__lowerCamelCase , return_tensors="pt" ) self.assertEqual(32 , targets["input_ids"].shape[1] ) def _lowerCamelCase ( self: List[Any] ) -> Tuple: # safety check on max_len default value so we are sure the test works __UpperCAmelCase : Optional[int] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test __UpperCAmelCase : str = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc __UpperCAmelCase : int = tempfile.mkdtemp() __UpperCAmelCase : List[Any] = " He is very happy, UNwant\u00E9d,running" __UpperCAmelCase : Union[str, Any] = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) tokenizer.save_pretrained(__lowerCamelCase ) __UpperCAmelCase : Tuple = tokenizer.__class__.from_pretrained(__lowerCamelCase ) __UpperCAmelCase : Dict = after_tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) shutil.rmtree(__lowerCamelCase ) __UpperCAmelCase : Optional[Any] = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc __UpperCAmelCase : List[Any] = tempfile.mkdtemp() __UpperCAmelCase : Optional[int] = " He is very happy, UNwant\u00E9d,running" __UpperCAmelCase : str = tokenizer.additional_special_tokens # We can add a new special token for Canine as follows: __UpperCAmelCase : Tuple = chr(0xE_0_0_7 ) additional_special_tokens.append(__lowerCamelCase ) tokenizer.add_special_tokens({"additional_special_tokens": additional_special_tokens} ) __UpperCAmelCase : Optional[int] = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) tokenizer.save_pretrained(__lowerCamelCase ) __UpperCAmelCase : str = tokenizer.__class__.from_pretrained(__lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = after_tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) self.assertIn(__lowerCamelCase , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) __UpperCAmelCase : Optional[Any] = tokenizer.__class__.from_pretrained(__lowerCamelCase , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(__lowerCamelCase ) def _lowerCamelCase ( self: str ) -> Optional[int]: __UpperCAmelCase : List[Any] = self.get_tokenizers(do_lower_case=__lowerCamelCase ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = self.get_clean_sequence(__lowerCamelCase ) # a special token for Canine can be defined as follows: __UpperCAmelCase : int = 0xE_0_0_5 __UpperCAmelCase : Tuple = chr(__lowerCamelCase ) tokenizer.add_special_tokens({"cls_token": special_token} ) __UpperCAmelCase : Union[str, Any] = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(len(__lowerCamelCase ) , 1 ) __UpperCAmelCase : Any = tokenizer.decode(ids + encoded_special_token , clean_up_tokenization_spaces=__lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) __UpperCAmelCase : Dict = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) __UpperCAmelCase : int = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(__lowerCamelCase , input_encoded + special_token_id ) __UpperCAmelCase : Optional[int] = tokenizer.decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) self.assertTrue(special_token not in decoded ) def _lowerCamelCase ( self: Optional[int] ) -> Optional[Any]: __UpperCAmelCase : List[str] = self.get_tokenizers(do_lower_case=__lowerCamelCase ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): __UpperCAmelCase : Optional[int] = chr(0xE_0_0_5 ) __UpperCAmelCase : List[str] = chr(0xE_0_0_6 ) # `add_tokens` method stores special tokens only in `tokenizer.unique_no_split_tokens`. (in tokenization_utils.py) tokenizer.add_tokens([SPECIAL_TOKEN_1] , special_tokens=__lowerCamelCase ) # `add_special_tokens` method stores special tokens in `tokenizer.additional_special_tokens`, # which also occur in `tokenizer.all_special_tokens`. (in tokenization_utils_base.py) tokenizer.add_special_tokens({"additional_special_tokens": [SPECIAL_TOKEN_2]} ) __UpperCAmelCase : Tuple = tokenizer.tokenize(__lowerCamelCase ) __UpperCAmelCase : Optional[Any] = tokenizer.tokenize(__lowerCamelCase ) self.assertEqual(len(__lowerCamelCase ) , 1 ) self.assertEqual(len(__lowerCamelCase ) , 1 ) self.assertEqual(token_a[0] , __lowerCamelCase ) self.assertEqual(token_a[0] , __lowerCamelCase ) @require_tokenizers def _lowerCamelCase ( self: str ) -> Union[str, Any]: __UpperCAmelCase : Any = self.get_tokenizers(do_lower_case=__lowerCamelCase ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): # a special token for Canine can be defined as follows: __UpperCAmelCase : Union[str, Any] = 0xE_0_0_6 __UpperCAmelCase : int = chr(__lowerCamelCase ) __UpperCAmelCase : int = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase ) tokenizer.add_special_tokens({"additional_special_tokens": [new_token]} ) with tempfile.TemporaryDirectory() as tmp_dir_name: tokenizer.save_pretrained(__lowerCamelCase ) tokenizer.from_pretrained(__lowerCamelCase ) def _lowerCamelCase ( self: Dict ) -> List[str]: __UpperCAmelCase : str = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(__lowerCamelCase ) with open(os.path.join(__lowerCamelCase , "special_tokens_map.json" ) , encoding="utf-8" ) as json_file: __UpperCAmelCase : Tuple = json.load(__lowerCamelCase ) with open(os.path.join(__lowerCamelCase , "tokenizer_config.json" ) , encoding="utf-8" ) as json_file: __UpperCAmelCase : Optional[int] = json.load(__lowerCamelCase ) # a special token for Canine can be defined as follows: __UpperCAmelCase : Any = 0xE_0_0_6 __UpperCAmelCase : Union[str, Any] = chr(__lowerCamelCase ) __UpperCAmelCase : Dict = [new_token_a] __UpperCAmelCase : int = [new_token_a] with open(os.path.join(__lowerCamelCase , "special_tokens_map.json" ) , "w" , encoding="utf-8" ) as outfile: json.dump(__lowerCamelCase , __lowerCamelCase ) with open(os.path.join(__lowerCamelCase , "tokenizer_config.json" ) , "w" , encoding="utf-8" ) as outfile: json.dump(__lowerCamelCase , __lowerCamelCase ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files __UpperCAmelCase : List[str] = tokenizer_class.from_pretrained(__lowerCamelCase , extra_ids=0 ) self.assertIn(__lowerCamelCase , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( [new_token_a] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids([new_token_a] ) ) , ) __UpperCAmelCase : List[Any] = 0xE_0_0_7 __UpperCAmelCase : List[Any] = chr(__lowerCamelCase ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained __UpperCAmelCase : str = [AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase )] __UpperCAmelCase : Dict = tokenizer_class.from_pretrained( __lowerCamelCase , additional_special_tokens=__lowerCamelCase , extra_ids=0 ) self.assertIn(__lowerCamelCase , tokenizer.additional_special_tokens ) # self.assertIn(new_token_2,tokenizer.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( [new_token_a] , tokenizer.convert_ids_to_tokens(tokenizer.convert_tokens_to_ids([new_token_a] ) ) ) @require_tokenizers def _lowerCamelCase ( self: Optional[Any] ) -> Optional[int]: __UpperCAmelCase : Optional[int] = self.get_tokenizers(do_lower_case=__lowerCamelCase ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): __UpperCAmelCase : int = "hello world" if self.space_between_special_tokens: __UpperCAmelCase : Any = "[CLS] hello world [SEP]" else: __UpperCAmelCase : Union[str, Any] = input __UpperCAmelCase : List[Any] = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) __UpperCAmelCase : Any = tokenizer.decode(__lowerCamelCase , spaces_between_special_tokens=self.space_between_special_tokens ) self.assertIn(__lowerCamelCase , [output, output.lower()] ) def _lowerCamelCase ( self: Dict ) -> Any: __UpperCAmelCase : Any = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): __UpperCAmelCase : List[str] = [ "bos_token", "eos_token", "unk_token", "sep_token", "pad_token", "cls_token", "mask_token", ] __UpperCAmelCase : List[str] = "a" __UpperCAmelCase : Any = ord(__lowerCamelCase ) for attr in attributes_list: setattr(__lowerCamelCase , attr + "_id" , __lowerCamelCase ) self.assertEqual(getattr(__lowerCamelCase , __lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(getattr(__lowerCamelCase , attr + "_id" ) , __lowerCamelCase ) setattr(__lowerCamelCase , attr + "_id" , __lowerCamelCase ) self.assertEqual(getattr(__lowerCamelCase , __lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(getattr(__lowerCamelCase , attr + "_id" ) , __lowerCamelCase ) setattr(__lowerCamelCase , "additional_special_tokens_ids" , [] ) self.assertListEqual(getattr(__lowerCamelCase , "additional_special_tokens" ) , [] ) self.assertListEqual(getattr(__lowerCamelCase , "additional_special_tokens_ids" ) , [] ) __UpperCAmelCase : Tuple = 0xE_0_0_6 __UpperCAmelCase : Optional[Any] = chr(__lowerCamelCase ) setattr(__lowerCamelCase , "additional_special_tokens_ids" , [additional_special_token_id] ) self.assertListEqual(getattr(__lowerCamelCase , "additional_special_tokens" ) , [additional_special_token] ) self.assertListEqual(getattr(__lowerCamelCase , "additional_special_tokens_ids" ) , [additional_special_token_id] ) def _lowerCamelCase ( self: str ) -> Union[str, Any]: pass def _lowerCamelCase ( self: Any ) -> Any: pass def _lowerCamelCase ( self: Union[str, Any] ) -> Tuple: pass def _lowerCamelCase ( self: Optional[int] ) -> Any: pass def _lowerCamelCase ( self: List[str] ) -> str: pass def _lowerCamelCase ( self: Union[str, Any] ) -> Optional[int]: pass def _lowerCamelCase ( self: Optional[Any] ) -> Tuple: pass def _lowerCamelCase ( self: str ) -> Tuple: pass
342
import inspect import unittest from transformers import ConvNextVaConfig from transformers.models.auto import get_values from transformers.models.auto.modeling_auto import MODEL_FOR_BACKBONE_MAPPING_NAMES, MODEL_MAPPING_NAMES from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextVaBackbone, ConvNextVaForImageClassification, ConvNextVaModel from transformers.models.convnextva.modeling_convnextva import CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _snake_case : def __init__( self: Tuple , __lowerCamelCase: Optional[int] , __lowerCamelCase: Optional[Any]=13 , __lowerCamelCase: Optional[int]=32 , __lowerCamelCase: List[str]=3 , __lowerCamelCase: Dict=4 , __lowerCamelCase: Optional[Any]=[10, 20, 30, 40] , __lowerCamelCase: int=[2, 2, 3, 2] , __lowerCamelCase: Union[str, Any]=True , __lowerCamelCase: Union[str, Any]=True , __lowerCamelCase: Tuple=37 , __lowerCamelCase: Tuple="gelu" , __lowerCamelCase: List[Any]=10 , __lowerCamelCase: Optional[int]=0.02 , __lowerCamelCase: Optional[Any]=["stage2", "stage3", "stage4"] , __lowerCamelCase: Optional[int]=[2, 3, 4] , __lowerCamelCase: int=None , ) -> List[str]: __UpperCAmelCase : Union[str, Any] = parent __UpperCAmelCase : List[str] = batch_size __UpperCAmelCase : Optional[int] = image_size __UpperCAmelCase : List[str] = num_channels __UpperCAmelCase : Union[str, Any] = num_stages __UpperCAmelCase : List[str] = hidden_sizes __UpperCAmelCase : Any = depths __UpperCAmelCase : Optional[int] = is_training __UpperCAmelCase : List[Any] = use_labels __UpperCAmelCase : Optional[int] = intermediate_size __UpperCAmelCase : Optional[Any] = hidden_act __UpperCAmelCase : Union[str, Any] = num_labels __UpperCAmelCase : Any = initializer_range __UpperCAmelCase : List[str] = out_features __UpperCAmelCase : Tuple = out_indices __UpperCAmelCase : List[Any] = scope def _lowerCamelCase ( self: List[Any] ) -> Optional[int]: __UpperCAmelCase : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __UpperCAmelCase : List[str] = None if self.use_labels: __UpperCAmelCase : List[Any] = ids_tensor([self.batch_size] , self.num_labels ) __UpperCAmelCase : Optional[Any] = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self: Tuple ) -> List[Any]: return ConvNextVaConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _lowerCamelCase ( self: List[Any] , __lowerCamelCase: int , __lowerCamelCase: int , __lowerCamelCase: Optional[int] ) -> Union[str, Any]: __UpperCAmelCase : Optional[Any] = ConvNextVaModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : List[str] = model(__lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _lowerCamelCase ( self: Optional[Any] , __lowerCamelCase: Optional[Any] , __lowerCamelCase: Any , __lowerCamelCase: Tuple ) -> Tuple: __UpperCAmelCase : Union[str, Any] = ConvNextVaForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : Optional[int] = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self: int , __lowerCamelCase: Any , __lowerCamelCase: Optional[int] , __lowerCamelCase: Optional[Any] ) -> Optional[int]: __UpperCAmelCase : List[str] = ConvNextVaBackbone(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : Any = model(__lowerCamelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __UpperCAmelCase : List[Any] = None __UpperCAmelCase : List[str] = ConvNextVaBackbone(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : Any = model(__lowerCamelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _lowerCamelCase ( self: int ) -> List[str]: __UpperCAmelCase : int = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = config_and_inputs __UpperCAmelCase : str = {"pixel_values": pixel_values} return config, inputs_dict def _lowerCamelCase ( self: List[Any] ) -> List[Any]: __UpperCAmelCase : Optional[int] = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Tuple = config_and_inputs __UpperCAmelCase : Dict = {"pixel_values": pixel_values, "labels": labels} return config, inputs_dict @require_torch class _snake_case ( _lowercase , _lowercase , unittest.TestCase ): lowerCamelCase__: Dict = ( ( ConvNextVaModel, ConvNextVaForImageClassification, ConvNextVaBackbone, ) if is_torch_available() else () ) lowerCamelCase__: str = ( {"feature-extraction": ConvNextVaModel, "image-classification": ConvNextVaForImageClassification} if is_torch_available() else {} ) lowerCamelCase__: Tuple = False lowerCamelCase__: int = False lowerCamelCase__: Dict = False lowerCamelCase__: int = False lowerCamelCase__: Any = False def _lowerCamelCase ( self: Union[str, Any] ) -> Union[str, Any]: __UpperCAmelCase : Union[str, Any] = ConvNextVaModelTester(self ) __UpperCAmelCase : str = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=37 ) def _lowerCamelCase ( self: Dict ) -> Tuple: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowerCamelCase ( self: List[Any] ) -> int: return @unittest.skip(reason="ConvNextV2 does not use inputs_embeds" ) def _lowerCamelCase ( self: Optional[Any] ) -> Optional[int]: pass @unittest.skip(reason="ConvNextV2 does not support input and output embeddings" ) def _lowerCamelCase ( self: Any ) -> Any: pass @unittest.skip(reason="ConvNextV2 does not use feedforward chunking" ) def _lowerCamelCase ( self: str ) -> Optional[Any]: pass def _lowerCamelCase ( self: List[Any] ) -> int: if not self.model_tester.is_training: return for model_class in self.all_model_classes: __UpperCAmelCase , __UpperCAmelCase : str = self.model_tester.prepare_config_and_inputs_with_labels() __UpperCAmelCase : Optional[Any] = True if model_class.__name__ in [ *get_values(__lowerCamelCase ), *get_values(__lowerCamelCase ), ]: continue __UpperCAmelCase : Optional[Any] = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.train() __UpperCAmelCase : Any = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) __UpperCAmelCase : Any = model(**__lowerCamelCase ).loss loss.backward() def _lowerCamelCase ( self: Optional[int] ) -> Dict: if not self.model_tester.is_training: return for model_class in self.all_model_classes: __UpperCAmelCase , __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_with_labels() __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = True if ( model_class.__name__ in [*get_values(__lowerCamelCase ), *get_values(__lowerCamelCase )] or not model_class.supports_gradient_checkpointing ): continue __UpperCAmelCase : int = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.gradient_checkpointing_enable() model.train() __UpperCAmelCase : List[Any] = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) __UpperCAmelCase : Any = model(**__lowerCamelCase ).loss loss.backward() def _lowerCamelCase ( self: List[str] ) -> Dict: __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : str = model_class(__lowerCamelCase ) __UpperCAmelCase : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase : List[Any] = [*signature.parameters.keys()] __UpperCAmelCase : int = ["pixel_values"] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowerCamelCase ( self: str ) -> List[Any]: __UpperCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowerCamelCase ( self: Union[str, Any] ) -> Dict: def check_hidden_states_output(__lowerCamelCase: Any , __lowerCamelCase: Tuple , __lowerCamelCase: str ): __UpperCAmelCase : Any = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() with torch.no_grad(): __UpperCAmelCase : Tuple = model(**self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) ) __UpperCAmelCase : List[Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __UpperCAmelCase : Optional[int] = self.model_tester.num_stages self.assertEqual(len(__lowerCamelCase ) , expected_num_stages + 1 ) # ConvNextV2's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __UpperCAmelCase , __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : Optional[int] = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase : Any = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def _lowerCamelCase ( self: Optional[Any] ) -> Optional[int]: __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) @slow def _lowerCamelCase ( self: Dict ) -> List[Any]: for model_name in CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCAmelCase : Optional[int] = ConvNextVaModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) -> List[Any]: __UpperCAmelCase : List[str] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class _snake_case ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self: Optional[int] ) -> Dict: return AutoImageProcessor.from_pretrained("facebook/convnextv2-tiny-1k-224" ) if is_vision_available() else None @slow def _lowerCamelCase ( self: List[Any] ) -> Tuple: __UpperCAmelCase : List[Any] = ConvNextVaForImageClassification.from_pretrained("facebook/convnextv2-tiny-1k-224" ).to(__lowerCamelCase ) __UpperCAmelCase : List[str] = self.default_image_processor __UpperCAmelCase : Optional[Any] = prepare_img() __UpperCAmelCase : int = preprocessor(images=__lowerCamelCase , return_tensors="pt" ).to(__lowerCamelCase ) # forward pass with torch.no_grad(): __UpperCAmelCase : str = model(**__lowerCamelCase ) # verify the logits __UpperCAmelCase : Dict = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) __UpperCAmelCase : str = torch.tensor([0.99_96, 0.19_66, -0.43_86] ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
342
1
'''simple docstring''' import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, ClassLabel, Features from .base import TaskTemplate @dataclass(frozen=__SCREAMING_SNAKE_CASE ) class SCREAMING_SNAKE_CASE( __SCREAMING_SNAKE_CASE ): """simple docstring""" lowerCamelCase__ = field(default="""audio-classification""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) lowerCamelCase__ = Features({"""audio""": Audio()} ) lowerCamelCase__ = Features({"""labels""": ClassLabel} ) lowerCamelCase__ = "audio" lowerCamelCase__ = "labels" def A ( self : Optional[Any] , __snake_case : str ) -> List[str]: if self.label_column not in features: raise ValueError(F"""Column {self.label_column} is not present in features.""" ) if not isinstance(features[self.label_column] , __snake_case ): raise ValueError(F"""Column {self.label_column} is not a ClassLabel.""" ) UpperCAmelCase : int = copy.deepcopy(self ) UpperCAmelCase : Tuple = self.label_schema.copy() UpperCAmelCase : Optional[int] = features[self.label_column] UpperCAmelCase : Dict = label_schema return task_template @property def A ( self : Optional[int] ) -> int: return { self.audio_column: "audio", self.label_column: "labels", }
23
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/config.json""", """umberto-commoncrawl-cased-v1""": ( """https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json""" ), """umberto-wikipedia-uncased-v1""": ( """https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json""" ), } class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Tuple = """camembert""" def __init__( self , __magic_name__=3_0_5_2_2 , __magic_name__=7_6_8 , __magic_name__=1_2 , __magic_name__=1_2 , __magic_name__=3_0_7_2 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_1_2 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=1e-12 , __magic_name__=1 , __magic_name__=0 , __magic_name__=2 , __magic_name__="absolute" , __magic_name__=True , __magic_name__=None , **__magic_name__ , ): super().__init__(pad_token_id=__magic_name__ , bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , **__magic_name__ ) lowerCamelCase : int = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : int = num_hidden_layers lowerCamelCase : int = num_attention_heads lowerCamelCase : Optional[int] = hidden_act lowerCamelCase : List[Any] = intermediate_size lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Optional[int] = attention_probs_dropout_prob lowerCamelCase : Optional[int] = max_position_embeddings lowerCamelCase : str = type_vocab_size lowerCamelCase : Optional[Any] = initializer_range lowerCamelCase : int = layer_norm_eps lowerCamelCase : Any = position_embedding_type lowerCamelCase : Optional[int] = use_cache lowerCamelCase : Union[str, Any] = classifier_dropout class A__ ( __SCREAMING_SNAKE_CASE): @property def UpperCamelCase__ ( self ): if self.task == "multiple-choice": lowerCamelCase : List[str] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowerCamelCase : List[str] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
287
0
'''simple docstring''' class lowerCAmelCase_ : '''simple docstring''' def __init__( self : List[Any] , SCREAMING_SNAKE_CASE_ : int ) -> Optional[Any]: '''simple docstring''' A: List[Any] = n A: Tuple = [None] * self.n A: Optional[Any] = 0 # index of the first element A: Optional[int] = 0 A: List[Any] = 0 def __len__( self : Tuple ) -> int: '''simple docstring''' return self.size def _snake_case ( self : Union[str, Any] ) -> bool: '''simple docstring''' return self.size == 0 def _snake_case ( self : str ) -> int: '''simple docstring''' return False if self.is_empty() else self.array[self.front] def _snake_case ( self : Optional[int] , SCREAMING_SNAKE_CASE_ : str ) -> List[Any]: '''simple docstring''' if self.size >= self.n: raise Exception('''QUEUE IS FULL''' ) A: Optional[int] = data A: str = (self.rear + 1) % self.n self.size += 1 return self def _snake_case ( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' if self.size == 0: raise Exception('''UNDERFLOW''' ) A: int = self.array[self.front] A: Union[str, Any] = None A: Tuple = (self.front + 1) % self.n self.size -= 1 return temp
350
'''simple docstring''' import requests UpperCamelCase = '''https://newsapi.org/v1/articles?source=bbc-news&sortBy=top&apiKey=''' def SCREAMING_SNAKE_CASE( __lowercase ) -> None: # fetching a list of articles in json format A: Tuple = requests.get(_NEWS_API + bbc_news_api_key ).json() # each article in the list is a dict for i, article in enumerate(bbc_news_page['''articles'''] , 1 ): print(F"""{i}.) {article['title']}""" ) if __name__ == "__main__": fetch_bbc_news(bbc_news_api_key='''<Your BBC News API key goes here>''')
334
0
"""simple docstring""" import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.local_sgd import LocalSGD ######################################################################## # This is a fully working simple example to use Accelerate # with LocalSGD, which is a method to synchronize model # parameters every K batches. It is different, but complementary # to gradient accumulation. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _lowerCamelCase : Any = 16 _lowerCamelCase : int = 32 def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase = 16 ): """simple docstring""" A_ : Dict = AutoTokenizer.from_pretrained('''bert-base-cased''' ) A_ : str = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(_UpperCAmelCase ): # max_length=None => use the model max length (it's actually the default) A_ : List[Any] = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=_UpperCAmelCase , max_length=_UpperCAmelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): A_ : Dict = datasets.map( _UpperCAmelCase , batched=_UpperCAmelCase , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library A_ : int = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(_UpperCAmelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. A_ : Any = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": A_ : str = 16 elif accelerator.mixed_precision != "no": A_ : Optional[int] = 8 else: A_ : Union[str, Any] = None return tokenizer.pad( _UpperCAmelCase , padding='''longest''' , max_length=_UpperCAmelCase , pad_to_multiple_of=_UpperCAmelCase , return_tensors='''pt''' , ) # Instantiate dataloaders. A_ : Optional[Any] = DataLoader( tokenized_datasets['''train'''] , shuffle=_UpperCAmelCase , collate_fn=_UpperCAmelCase , batch_size=_UpperCAmelCase ) A_ : Union[str, Any] = DataLoader( tokenized_datasets['''validation'''] , shuffle=_UpperCAmelCase , collate_fn=_UpperCAmelCase , batch_size=_UpperCAmelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders _lowerCamelCase : Dict = mocked_dataloaders # noqa: F811 def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , _UpperCAmelCase ) == "1": A_ : Union[str, Any] = 2 # New Code # A_ : str = int(args.gradient_accumulation_steps ) A_ : str = int(args.local_sgd_steps ) # Initialize accelerator A_ : int = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , gradient_accumulation_steps=_UpperCAmelCase ) if accelerator.distributed_type not in [DistributedType.NO, DistributedType.MULTI_CPU, DistributedType.MULTI_GPU]: raise NotImplementedError('''LocalSGD is supported only for CPUs and GPUs (no DeepSpeed or MegatronLM)''' ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs A_ : Union[str, Any] = config['''lr'''] A_ : Any = int(config['''num_epochs'''] ) A_ : Optional[int] = int(config['''seed'''] ) A_ : List[str] = int(config['''batch_size'''] ) A_ : Optional[int] = evaluate.load('''glue''' , '''mrpc''' ) set_seed(_UpperCAmelCase ) A_ , A_ : Optional[Any] = get_dataloaders(_UpperCAmelCase , _UpperCAmelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) A_ : Any = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=_UpperCAmelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). A_ : Tuple = model.to(accelerator.device ) # Instantiate optimizer A_ : Optional[Any] = AdamW(params=model.parameters() , lr=_UpperCAmelCase ) # Instantiate scheduler A_ : Dict = get_linear_schedule_with_warmup( optimizer=_UpperCAmelCase , num_warmup_steps=100 , num_training_steps=(len(_UpperCAmelCase ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. A_ , A_ , A_ , A_ , A_ : Dict = accelerator.prepare( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # Now we train the model for epoch in range(_UpperCAmelCase ): model.train() with LocalSGD( accelerator=_UpperCAmelCase , model=_UpperCAmelCase , local_sgd_steps=_UpperCAmelCase , enabled=local_sgd_steps is not None ) as local_sgd: for step, batch in enumerate(_UpperCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(_UpperCAmelCase ): A_ : str = model(**_UpperCAmelCase ) A_ : Any = output.loss accelerator.backward(_UpperCAmelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() # LocalSGD-specific line local_sgd.step() model.eval() for step, batch in enumerate(_UpperCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): A_ : int = model(**_UpperCAmelCase ) A_ : List[str] = outputs.logits.argmax(dim=-1 ) A_ , A_ : Optional[Any] = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=_UpperCAmelCase , references=_UpperCAmelCase , ) A_ : Any = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , _UpperCAmelCase ) def lowercase_ ( ): """simple docstring""" A_ : Tuple = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=_UpperCAmelCase , default=_UpperCAmelCase , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) # New Code # parser.add_argument( '''--gradient_accumulation_steps''' , type=_UpperCAmelCase , default=1 , help='''The number of minibatches to be ran before gradients are accumulated.''' , ) parser.add_argument( '''--local_sgd_steps''' , type=_UpperCAmelCase , default=8 , help='''Number of local SGD steps or None to disable local SGD''' ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) A_ : List[str] = parser.parse_args() A_ : Optional[Any] = {'''lr''': 2E-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(_UpperCAmelCase , _UpperCAmelCase ) if __name__ == "__main__": main()
167
"""simple docstring""" import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _lowerCamelCase : List[Any] = logging.get_logger(__name__) _lowerCamelCase : List[str] = { 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/config.json', # See all BART models at https://huggingface.co/models?filter=bart } class lowercase ( __UpperCAmelCase): __lowerCAmelCase : Union[str, Any] = """bart""" __lowerCAmelCase : Optional[int] = ["""past_key_values"""] __lowerCAmelCase : List[Any] = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[Any] , _lowerCamelCase : List[Any]=5_02_65 , _lowerCamelCase : Optional[Any]=10_24 , _lowerCamelCase : Dict=12 , _lowerCamelCase : Dict=40_96 , _lowerCamelCase : Tuple=16 , _lowerCamelCase : Optional[Any]=12 , _lowerCamelCase : Tuple=40_96 , _lowerCamelCase : List[str]=16 , _lowerCamelCase : List[str]=0.0 , _lowerCamelCase : List[str]=0.0 , _lowerCamelCase : int="gelu" , _lowerCamelCase : Any=10_24 , _lowerCamelCase : Union[str, Any]=0.1 , _lowerCamelCase : Tuple=0.0 , _lowerCamelCase : List[Any]=0.0 , _lowerCamelCase : Optional[int]=0.02 , _lowerCamelCase : int=0.0 , _lowerCamelCase : Any=False , _lowerCamelCase : List[Any]=True , _lowerCamelCase : int=3 , _lowerCamelCase : Tuple=1 , _lowerCamelCase : int=0 , _lowerCamelCase : Optional[int]=2 , _lowerCamelCase : Any=True , _lowerCamelCase : str=2 , _lowerCamelCase : str=2 , **_lowerCamelCase : str , ): """simple docstring""" A_ : Dict = vocab_size A_ : Union[str, Any] = max_position_embeddings A_ : Union[str, Any] = d_model A_ : Optional[int] = encoder_ffn_dim A_ : Optional[Any] = encoder_layers A_ : Union[str, Any] = encoder_attention_heads A_ : List[str] = decoder_ffn_dim A_ : List[str] = decoder_layers A_ : Any = decoder_attention_heads A_ : List[Any] = dropout A_ : Optional[int] = attention_dropout A_ : List[Any] = activation_dropout A_ : Tuple = activation_function A_ : Any = init_std A_ : Union[str, Any] = encoder_layerdrop A_ : Optional[Any] = decoder_layerdrop A_ : Tuple = classifier_dropout A_ : Tuple = use_cache A_ : List[str] = encoder_layers A_ : Optional[int] = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=_lowerCamelCase , pad_token_id=_lowerCamelCase , bos_token_id=_lowerCamelCase , eos_token_id=_lowerCamelCase , is_encoder_decoder=_lowerCamelCase , decoder_start_token_id=_lowerCamelCase , forced_eos_token_id=_lowerCamelCase , **_lowerCamelCase , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get('''force_bos_token_to_be_generated''' , _lowerCamelCase ): A_ : int = self.bos_token_id warnings.warn( F"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ '''The config can simply be saved and uploaded again to be fixed.''' ) class lowercase ( __UpperCAmelCase): @property def a_ ( self : Optional[Any] ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: A_ : Tuple = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: A_ : Tuple = {0: '''batch'''} A_ : Dict = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: A_ : Optional[Any] = {0: '''batch''', 1: '''decoder_sequence'''} A_ : List[Any] = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(_lowerCamelCase , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. A_ : List[Any] = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: A_ , A_ : Any = self.num_layers for i in range(_lowerCamelCase ): A_ : List[Any] = {0: '''batch''', 2: '''past_sequence + sequence'''} A_ : Optional[Any] = {0: '''batch''', 2: '''past_sequence + sequence'''} else: A_ : List[Any] = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property def a_ ( self : str ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: A_ : Optional[Any] = super().outputs else: A_ : List[Any] = super(_lowerCamelCase , self ).outputs if self.use_past: A_ , A_ : int = self.num_layers for i in range(_lowerCamelCase ): A_ : Any = {0: '''batch''', 2: '''past_sequence + sequence'''} A_ : Dict = {0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def a_ ( self : Union[str, Any] , _lowerCamelCase : PreTrainedTokenizer , _lowerCamelCase : int = -1 , _lowerCamelCase : int = -1 , _lowerCamelCase : bool = False , _lowerCamelCase : Optional[TensorType] = None , ): """simple docstring""" A_ : Optional[int] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # Generate decoder inputs A_ : Tuple = seq_length if not self.use_past else 1 A_ : List[str] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Dict = {F"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} A_ : Dict = dict(**_lowerCamelCase , **_lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch A_ , A_ : Union[str, Any] = common_inputs['''input_ids'''].shape A_ : Any = common_inputs['''decoder_input_ids'''].shape[1] A_ , A_ : Optional[Any] = self.num_attention_heads A_ : Union[str, Any] = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) A_ : Optional[Any] = decoder_seq_length + 3 A_ : Dict = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) A_ : Optional[int] = torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(_lowerCamelCase , _lowerCamelCase )] , dim=1 ) A_ : Optional[Any] = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered A_ , A_ : Optional[Any] = self.num_layers A_ : Optional[Any] = min(_lowerCamelCase , _lowerCamelCase ) A_ : Tuple = max(_lowerCamelCase , _lowerCamelCase ) - min_num_layers A_ : Tuple = '''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(_lowerCamelCase ): common_inputs["past_key_values"].append( ( torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase ), ) ) # TODO: test this. A_ : List[str] = encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(_lowerCamelCase , _lowerCamelCase ): common_inputs["past_key_values"].append((torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase )) ) return common_inputs def a_ ( self : Optional[int] , _lowerCamelCase : PreTrainedTokenizer , _lowerCamelCase : int = -1 , _lowerCamelCase : int = -1 , _lowerCamelCase : bool = False , _lowerCamelCase : Optional[TensorType] = None , ): """simple docstring""" A_ : Union[str, Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch A_ , A_ : Optional[Any] = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values A_ : Union[str, Any] = seqlen + 2 A_ , A_ : Tuple = self.num_layers A_ , A_ : Optional[int] = self.num_attention_heads A_ : List[str] = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) A_ : str = common_inputs['''attention_mask'''].dtype A_ : Optional[Any] = torch.cat( [common_inputs['''attention_mask'''], torch.ones(_lowerCamelCase , _lowerCamelCase , dtype=_lowerCamelCase )] , dim=1 ) A_ : int = [ (torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase )) for _ in range(_lowerCamelCase ) ] return common_inputs def a_ ( self : str , _lowerCamelCase : PreTrainedTokenizer , _lowerCamelCase : int = -1 , _lowerCamelCase : int = -1 , _lowerCamelCase : bool = False , _lowerCamelCase : Optional[TensorType] = None , ): """simple docstring""" A_ : List[Any] = compute_effective_axis_dimension( _lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX A_ : Dict = tokenizer.num_special_tokens_to_add(_lowerCamelCase ) A_ : int = compute_effective_axis_dimension( _lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=_lowerCamelCase ) # Generate dummy inputs according to compute batch and sequence A_ : List[str] = [''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size A_ : List[str] = dict(tokenizer(_lowerCamelCase , return_tensors=_lowerCamelCase ) ) return common_inputs def a_ ( self : Tuple , _lowerCamelCase : PreTrainedTokenizer , _lowerCamelCase : int = -1 , _lowerCamelCase : int = -1 , _lowerCamelCase : bool = False , _lowerCamelCase : Optional[TensorType] = None , ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: A_ : Tuple = self._generate_dummy_inputs_for_default_and_seqaseq_lm( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) elif self.task == "causal-lm": A_ : Optional[Any] = self._generate_dummy_inputs_for_causal_lm( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) else: A_ : Union[str, Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) return common_inputs def a_ ( self : str , _lowerCamelCase : Dict , _lowerCamelCase : Tuple , _lowerCamelCase : Optional[int] , _lowerCamelCase : Union[str, Any] ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: A_ : Tuple = super()._flatten_past_key_values_(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) else: A_ : List[Any] = super(_lowerCamelCase , self )._flatten_past_key_values_( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase )
167
1
import numpy as np from numpy import ndarray from scipy.optimize import Bounds, LinearConstraint, minimize def UpperCamelCase ( _A ): """simple docstring""" return np.dot(_A, _A ) class snake_case__ : def __init__( self , *, lowerCAmelCase__ = np.inf , lowerCAmelCase__ = "linear" , lowerCAmelCase__ = 0.0 , ) -> None: __magic_name__ : Optional[int] = regularization __magic_name__ : Dict = gamma if kernel == "linear": __magic_name__ : str = self.__linear elif kernel == "rbf": if self.gamma == 0: raise ValueError("""rbf kernel requires gamma""" ) if not isinstance(self.gamma , (float, int) ): raise ValueError("""gamma must be float or int""" ) if not self.gamma > 0: raise ValueError("""gamma must be > 0""" ) __magic_name__ : Optional[int] = self.__rbf # in the future, there could be a default value like in sklearn # sklear: def_gamma = 1/(n_features * X.var()) (wiki) # previously it was 1/(n_features) else: __magic_name__ : Tuple = F'Unknown kernel: {kernel}' raise ValueError(lowerCAmelCase__ ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> float: return np.dot(lowerCAmelCase__ , lowerCAmelCase__ ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> float: return np.exp(-(self.gamma * norm_squared(vectora - vectora )) ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: __magic_name__ : Union[str, Any] = observations __magic_name__ : Tuple = classes # using Wolfe's Dual to calculate w. # Primal problem: minimize 1/2*norm_squared(w) # constraint: yn(w . xn + b) >= 1 # # With l a vector # Dual problem: maximize sum_n(ln) - # 1/2 * sum_n(sum_m(ln*lm*yn*ym*xn . xm)) # constraint: self.C >= ln >= 0 # and sum_n(ln*yn) = 0 # Then we get w using w = sum_n(ln*yn*xn) # At the end we can get b ~= mean(yn - w . xn) # # Since we use kernels, we only need l_star to calculate b # and to classify observations (__magic_name__) : int = np.shape(lowerCAmelCase__ ) def to_minimize(lowerCAmelCase__ ) -> float: __magic_name__ : List[Any] = 0 (__magic_name__) : int = np.shape(lowerCAmelCase__ ) for i in range(lowerCAmelCase__ ): for j in range(lowerCAmelCase__ ): s += ( candidate[i] * candidate[j] * classes[i] * classes[j] * self.kernel(observations[i] , observations[j] ) ) return 1 / 2 * s - sum(lowerCAmelCase__ ) __magic_name__ : Tuple = LinearConstraint(lowerCAmelCase__ , 0 , 0 ) __magic_name__ : Optional[Any] = Bounds(0 , self.regularization ) __magic_name__ : Tuple = minimize( lowerCAmelCase__ , np.ones(lowerCAmelCase__ ) , bounds=lowerCAmelCase__ , constraints=[ly_contraint] ).x __magic_name__ : Union[str, Any] = l_star # calculating mean offset of separation plane to points __magic_name__ : Tuple = 0 for i in range(lowerCAmelCase__ ): for j in range(lowerCAmelCase__ ): s += classes[i] - classes[i] * self.optimum[i] * self.kernel( observations[i] , observations[j] ) __magic_name__ : Dict = s / n def __magic_name__ ( self , lowerCAmelCase__ ) -> int: __magic_name__ : Optional[int] = sum( self.optimum[n] * self.classes[n] * self.kernel(self.observations[n] , lowerCAmelCase__ ) for n in range(len(self.classes ) ) ) return 1 if s + self.offset >= 0 else -1 if __name__ == "__main__": import doctest doctest.testmod()
359
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging __magic_name__: Any = logging.get_logger(__name__) __magic_name__: Dict = { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/config.json", # See all BlenderbotSmall models at https://huggingface.co/models?filter=blenderbot_small } class snake_case__ ( _lowerCAmelCase ): lowercase__ : Any = '''blenderbot-small''' lowercase__ : Optional[int] = ['''past_key_values'''] lowercase__ : Dict = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self , lowerCAmelCase__=5_02_65 , lowerCAmelCase__=5_12 , lowerCAmelCase__=8 , lowerCAmelCase__=20_48 , lowerCAmelCase__=16 , lowerCAmelCase__=8 , lowerCAmelCase__=20_48 , lowerCAmelCase__=16 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__="gelu" , lowerCAmelCase__=5_12 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0_2 , lowerCAmelCase__=1 , lowerCAmelCase__=False , lowerCAmelCase__=0 , lowerCAmelCase__=1 , lowerCAmelCase__=2 , lowerCAmelCase__=2 , **lowerCAmelCase__ , ) -> Tuple: __magic_name__ : Tuple = vocab_size __magic_name__ : List[str] = max_position_embeddings __magic_name__ : Union[str, Any] = d_model __magic_name__ : Optional[int] = encoder_ffn_dim __magic_name__ : Union[str, Any] = encoder_layers __magic_name__ : List[str] = encoder_attention_heads __magic_name__ : List[Any] = decoder_ffn_dim __magic_name__ : str = decoder_layers __magic_name__ : List[str] = decoder_attention_heads __magic_name__ : Union[str, Any] = dropout __magic_name__ : Tuple = attention_dropout __magic_name__ : List[Any] = activation_dropout __magic_name__ : List[Any] = activation_function __magic_name__ : Optional[int] = init_std __magic_name__ : Dict = encoder_layerdrop __magic_name__ : Union[str, Any] = decoder_layerdrop __magic_name__ : Optional[int] = use_cache __magic_name__ : List[Any] = encoder_layers __magic_name__ : Dict = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , is_encoder_decoder=lowerCAmelCase__ , decoder_start_token_id=lowerCAmelCase__ , forced_eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ , ) class snake_case__ ( _lowerCAmelCase ): @property def __magic_name__ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: __magic_name__ : Optional[int] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: __magic_name__ : List[Any] = {0: """batch"""} __magic_name__ : Dict = {0: """batch""", 1: """past_decoder_sequence + sequence"""} else: __magic_name__ : List[str] = {0: """batch""", 1: """decoder_sequence"""} __magic_name__ : Any = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(lowerCAmelCase__ , direction="""inputs""" ) elif self.task == "causal-lm": # TODO: figure this case out. __magic_name__ : Union[str, Any] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: __magic_name__ ,__magic_name__ : Dict = self.num_layers for i in range(lowerCAmelCase__ ): __magic_name__ : Dict = {0: """batch""", 2: """past_sequence + sequence"""} __magic_name__ : int = {0: """batch""", 2: """past_sequence + sequence"""} else: __magic_name__ : Union[str, Any] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ("""decoder_input_ids""", {0: """batch""", 1: """decoder_sequence"""}), ("""decoder_attention_mask""", {0: """batch""", 1: """decoder_sequence"""}), ] ) return common_inputs @property def __magic_name__ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: __magic_name__ : Any = super().outputs else: __magic_name__ : int = super(lowerCAmelCase__ , self ).outputs if self.use_past: __magic_name__ ,__magic_name__ : str = self.num_layers for i in range(lowerCAmelCase__ ): __magic_name__ : Tuple = {0: """batch""", 2: """past_sequence + sequence"""} __magic_name__ : Dict = {0: """batch""", 2: """past_sequence + sequence"""} return common_outputs def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = -1 , lowerCAmelCase__ = -1 , lowerCAmelCase__ = False , lowerCAmelCase__ = None , ) -> Mapping[str, Any]: __magic_name__ : Tuple = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # Generate decoder inputs __magic_name__ : Optional[int] = seq_length if not self.use_past else 1 __magic_name__ : Optional[Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) __magic_name__ : Optional[Any] = {F'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} __magic_name__ : Dict = dict(**lowerCAmelCase__ , **lowerCAmelCase__ ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch __magic_name__ ,__magic_name__ : List[Any] = common_inputs["""input_ids"""].shape __magic_name__ : Optional[Any] = common_inputs["""decoder_input_ids"""].shape[1] __magic_name__ ,__magic_name__ : str = self.num_attention_heads __magic_name__ : Optional[Any] = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) __magic_name__ : Any = decoder_seq_length + 3 __magic_name__ : Dict = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) __magic_name__ : List[str] = torch.cat( [common_inputs["""decoder_attention_mask"""], torch.ones(lowerCAmelCase__ , lowerCAmelCase__ )] , dim=1 ) __magic_name__ : Union[str, Any] = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered __magic_name__ ,__magic_name__ : List[str] = self.num_layers __magic_name__ : Optional[Any] = min(lowerCAmelCase__ , lowerCAmelCase__ ) __magic_name__ : List[str] = max(lowerCAmelCase__ , lowerCAmelCase__ ) - min_num_layers __magic_name__ : Tuple = """encoder""" if num_encoder_layers > num_decoder_layers else """decoder""" for _ in range(lowerCAmelCase__ ): common_inputs["past_key_values"].append( ( torch.zeros(lowerCAmelCase__ ), torch.zeros(lowerCAmelCase__ ), torch.zeros(lowerCAmelCase__ ), torch.zeros(lowerCAmelCase__ ), ) ) # TODO: test this. __magic_name__ : Union[str, Any] = encoder_shape if remaining_side_name == """encoder""" else decoder_shape for _ in range(lowerCAmelCase__ , lowerCAmelCase__ ): common_inputs["past_key_values"].append((torch.zeros(lowerCAmelCase__ ), torch.zeros(lowerCAmelCase__ )) ) return common_inputs def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = -1 , lowerCAmelCase__ = -1 , lowerCAmelCase__ = False , lowerCAmelCase__ = None , ) -> Mapping[str, Any]: __magic_name__ : int = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch __magic_name__ ,__magic_name__ : Tuple = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values __magic_name__ : List[Any] = seqlen + 2 __magic_name__ ,__magic_name__ : Any = self.num_layers __magic_name__ ,__magic_name__ : int = self.num_attention_heads __magic_name__ : Optional[int] = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) __magic_name__ : Optional[int] = common_inputs["""attention_mask"""].dtype __magic_name__ : Optional[Any] = torch.cat( [common_inputs["""attention_mask"""], torch.ones(lowerCAmelCase__ , lowerCAmelCase__ , dtype=lowerCAmelCase__ )] , dim=1 ) __magic_name__ : Tuple = [ (torch.zeros(lowerCAmelCase__ ), torch.zeros(lowerCAmelCase__ )) for _ in range(lowerCAmelCase__ ) ] return common_inputs def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = -1 , lowerCAmelCase__ = -1 , lowerCAmelCase__ = False , lowerCAmelCase__ = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX __magic_name__ : Tuple = compute_effective_axis_dimension( lowerCAmelCase__ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __magic_name__ : str = tokenizer.num_special_tokens_to_add(lowerCAmelCase__ ) __magic_name__ : List[str] = compute_effective_axis_dimension( lowerCAmelCase__ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=lowerCAmelCase__ ) # Generate dummy inputs according to compute batch and sequence __magic_name__ : List[Any] = [""" """.join([tokenizer.unk_token] ) * seq_length] * batch_size __magic_name__ : List[str] = dict(tokenizer(lowerCAmelCase__ , return_tensors=lowerCAmelCase__ ) ) return common_inputs def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = -1 , lowerCAmelCase__ = -1 , lowerCAmelCase__ = False , lowerCAmelCase__ = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: __magic_name__ : Tuple = self._generate_dummy_inputs_for_default_and_seqaseq_lm( lowerCAmelCase__ , batch_size=lowerCAmelCase__ , seq_length=lowerCAmelCase__ , is_pair=lowerCAmelCase__ , framework=lowerCAmelCase__ ) elif self.task == "causal-lm": __magic_name__ : str = self._generate_dummy_inputs_for_causal_lm( lowerCAmelCase__ , batch_size=lowerCAmelCase__ , seq_length=lowerCAmelCase__ , is_pair=lowerCAmelCase__ , framework=lowerCAmelCase__ ) else: __magic_name__ : Dict = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCAmelCase__ , batch_size=lowerCAmelCase__ , seq_length=lowerCAmelCase__ , is_pair=lowerCAmelCase__ , framework=lowerCAmelCase__ ) return common_inputs def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: if self.task in ["default", "seq2seq-lm"]: __magic_name__ : List[Any] = super()._flatten_past_key_values_(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) else: __magic_name__ : Tuple = super(lowerCAmelCase__ , self )._flatten_past_key_values_( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ )
138
0
"""simple docstring""" from __future__ import annotations class _UpperCAmelCase : def __init__( self : Optional[Any] , _lowercase : int = 0 ): __UpperCAmelCase = key def a ( self : Optional[int] , _lowercase : str , _lowercase : int ): assert isinstance(_UpperCAmelCase , _UpperCAmelCase ) and isinstance(_UpperCAmelCase , _UpperCAmelCase ) __UpperCAmelCase = key or self.__key or 1 # make sure key is an appropriate size key %= 2_55 return [chr(ord(_UpperCAmelCase ) ^ key ) for ch in content] def a ( self : int , _lowercase : str , _lowercase : int ): assert isinstance(_UpperCAmelCase , _UpperCAmelCase ) and isinstance(_UpperCAmelCase , _UpperCAmelCase ) __UpperCAmelCase = key or self.__key or 1 # make sure key is an appropriate size key %= 2_55 return [chr(ord(_UpperCAmelCase ) ^ key ) for ch in content] def a ( self : str , _lowercase : str , _lowercase : int = 0 ): assert isinstance(_UpperCAmelCase , _UpperCAmelCase ) and isinstance(_UpperCAmelCase , _UpperCAmelCase ) __UpperCAmelCase = key or self.__key or 1 # make sure key can be any size while key > 2_55: key -= 2_55 # This will be returned __UpperCAmelCase = '''''' for ch in content: ans += chr(ord(_UpperCAmelCase ) ^ key ) return ans def a ( self : Dict , _lowercase : str , _lowercase : int = 0 ): assert isinstance(_UpperCAmelCase , _UpperCAmelCase ) and isinstance(_UpperCAmelCase , _UpperCAmelCase ) __UpperCAmelCase = key or self.__key or 1 # make sure key can be any size while key > 2_55: key -= 2_55 # This will be returned __UpperCAmelCase = '''''' for ch in content: ans += chr(ord(_UpperCAmelCase ) ^ key ) return ans def a ( self : List[str] , _lowercase : str , _lowercase : int = 0 ): assert isinstance(_UpperCAmelCase , _UpperCAmelCase ) and isinstance(_UpperCAmelCase , _UpperCAmelCase ) try: with open(_UpperCAmelCase ) as fin, open('''encrypt.out''' , '''w+''' ) as fout: # actual encrypt-process for line in fin: fout.write(self.encrypt_string(_UpperCAmelCase , _UpperCAmelCase ) ) except OSError: return False return True def a ( self : Optional[int] , _lowercase : str , _lowercase : int ): assert isinstance(_UpperCAmelCase , _UpperCAmelCase ) and isinstance(_UpperCAmelCase , _UpperCAmelCase ) try: with open(_UpperCAmelCase ) as fin, open('''decrypt.out''' , '''w+''' ) as fout: # actual encrypt-process for line in fin: fout.write(self.decrypt_string(_UpperCAmelCase , _UpperCAmelCase ) ) except OSError: return False return True # Tests # crypt = XORCipher() # key = 67 # # test encrypt # print(crypt.encrypt("hallo welt",key)) # # test decrypt # print(crypt.decrypt(crypt.encrypt("hallo welt",key), key)) # # test encrypt_string # print(crypt.encrypt_string("hallo welt",key)) # # test decrypt_string # print(crypt.decrypt_string(crypt.encrypt_string("hallo welt",key),key)) # if (crypt.encrypt_file("test.txt",key)): # print("encrypt successful") # else: # print("encrypt unsuccessful") # if (crypt.decrypt_file("encrypt.out",key)): # print("decrypt successful") # else: # print("decrypt unsuccessful")
332
from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_outputs import ( BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import logging from .configuration_regnet import RegNetConfig SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) # General docstring SCREAMING_SNAKE_CASE__ = """RegNetConfig""" # Base docstring SCREAMING_SNAKE_CASE__ = """facebook/regnet-y-040""" SCREAMING_SNAKE_CASE__ = [1, 1088, 7, 7] # Image classification docstring SCREAMING_SNAKE_CASE__ = """facebook/regnet-y-040""" SCREAMING_SNAKE_CASE__ = """tabby, tabby cat""" SCREAMING_SNAKE_CASE__ = [ """facebook/regnet-y-040""", # See all regnet models at https://huggingface.co/models?filter=regnet ] class A__ ( nn.Module ): def __init__( self : str , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int = 3 , _UpperCAmelCase : int = 1 , _UpperCAmelCase : int = 1 , _UpperCAmelCase : Optional[str] = "relu" , ) -> Optional[Any]: """simple docstring""" super().__init__() __lowercase = nn.Convad( _UpperCAmelCase , _UpperCAmelCase , kernel_size=_UpperCAmelCase , stride=_UpperCAmelCase , padding=kernel_size // 2 , groups=_UpperCAmelCase , bias=_UpperCAmelCase , ) __lowercase = nn.BatchNormad(_UpperCAmelCase ) __lowercase = ACTaFN[activation] if activation is not None else nn.Identity() def a__ ( self : Tuple , _UpperCAmelCase : List[str] ) -> str: """simple docstring""" __lowercase = self.convolution(_UpperCAmelCase ) __lowercase = self.normalization(_UpperCAmelCase ) __lowercase = self.activation(_UpperCAmelCase ) return hidden_state class A__ ( nn.Module ): def __init__( self : Union[str, Any] , _UpperCAmelCase : RegNetConfig ) -> Any: """simple docstring""" super().__init__() __lowercase = RegNetConvLayer( config.num_channels , config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act ) __lowercase = config.num_channels def a__ ( self : Optional[Any] , _UpperCAmelCase : Any ) -> Union[str, Any]: """simple docstring""" __lowercase = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( 'Make sure that the channel dimension of the pixel values match with the one set in the configuration.' ) __lowercase = self.embedder(_UpperCAmelCase ) return hidden_state class A__ ( nn.Module ): def __init__( self : List[str] , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int = 2 ) -> Optional[int]: """simple docstring""" super().__init__() __lowercase = nn.Convad(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 , stride=_UpperCAmelCase , bias=_UpperCAmelCase ) __lowercase = nn.BatchNormad(_UpperCAmelCase ) def a__ ( self : int , _UpperCAmelCase : Tensor ) -> Tensor: """simple docstring""" __lowercase = self.convolution(_UpperCAmelCase ) __lowercase = self.normalization(_UpperCAmelCase ) return hidden_state class A__ ( nn.Module ): def __init__( self : int , _UpperCAmelCase : int , _UpperCAmelCase : int ) -> str: """simple docstring""" super().__init__() __lowercase = nn.AdaptiveAvgPoolad((1, 1) ) __lowercase = nn.Sequential( nn.Convad(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 ) , nn.ReLU() , nn.Convad(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 ) , nn.Sigmoid() , ) def a__ ( self : str , _UpperCAmelCase : Dict ) -> str: """simple docstring""" __lowercase = self.pooler(_UpperCAmelCase ) __lowercase = self.attention(_UpperCAmelCase ) __lowercase = hidden_state * attention return hidden_state class A__ ( nn.Module ): def __init__( self : Optional[int] , _UpperCAmelCase : RegNetConfig , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int = 1 ) -> Tuple: """simple docstring""" super().__init__() __lowercase = in_channels != out_channels or stride != 1 __lowercase = max(1 , out_channels // config.groups_width ) __lowercase = ( RegNetShortCut(_UpperCAmelCase , _UpperCAmelCase , stride=_UpperCAmelCase ) if should_apply_shortcut else nn.Identity() ) __lowercase = nn.Sequential( RegNetConvLayer(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(_UpperCAmelCase , _UpperCAmelCase , stride=_UpperCAmelCase , groups=_UpperCAmelCase , activation=config.hidden_act ) , RegNetConvLayer(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 , activation=_UpperCAmelCase ) , ) __lowercase = ACTaFN[config.hidden_act] def a__ ( self : List[str] , _UpperCAmelCase : Tuple ) -> List[Any]: """simple docstring""" __lowercase = hidden_state __lowercase = self.layer(_UpperCAmelCase ) __lowercase = self.shortcut(_UpperCAmelCase ) hidden_state += residual __lowercase = self.activation(_UpperCAmelCase ) return hidden_state class A__ ( nn.Module ): def __init__( self : Union[str, Any] , _UpperCAmelCase : RegNetConfig , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int = 1 ) -> Optional[Any]: """simple docstring""" super().__init__() __lowercase = in_channels != out_channels or stride != 1 __lowercase = max(1 , out_channels // config.groups_width ) __lowercase = ( RegNetShortCut(_UpperCAmelCase , _UpperCAmelCase , stride=_UpperCAmelCase ) if should_apply_shortcut else nn.Identity() ) __lowercase = nn.Sequential( RegNetConvLayer(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(_UpperCAmelCase , _UpperCAmelCase , stride=_UpperCAmelCase , groups=_UpperCAmelCase , activation=config.hidden_act ) , RegNetSELayer(_UpperCAmelCase , reduced_channels=int(round(in_channels / 4 ) ) ) , RegNetConvLayer(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 , activation=_UpperCAmelCase ) , ) __lowercase = ACTaFN[config.hidden_act] def a__ ( self : Tuple , _UpperCAmelCase : Any ) -> List[str]: """simple docstring""" __lowercase = hidden_state __lowercase = self.layer(_UpperCAmelCase ) __lowercase = self.shortcut(_UpperCAmelCase ) hidden_state += residual __lowercase = self.activation(_UpperCAmelCase ) return hidden_state class A__ ( nn.Module ): def __init__( self : List[Any] , _UpperCAmelCase : RegNetConfig , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int = 2 , _UpperCAmelCase : int = 2 , ) -> Dict: """simple docstring""" super().__init__() __lowercase = RegNetXLayer if config.layer_type == 'x' else RegNetYLayer __lowercase = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , stride=_UpperCAmelCase , ) , *[layer(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) for _ in range(depth - 1 )] , ) def a__ ( self : Any , _UpperCAmelCase : str ) -> int: """simple docstring""" __lowercase = self.layers(_UpperCAmelCase ) return hidden_state class A__ ( nn.Module ): def __init__( self : Any , _UpperCAmelCase : RegNetConfig ) -> int: """simple docstring""" super().__init__() __lowercase = nn.ModuleList([] ) # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( RegNetStage( _UpperCAmelCase , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) __lowercase = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(_UpperCAmelCase , config.depths[1:] ): self.stages.append(RegNetStage(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , depth=_UpperCAmelCase ) ) def a__ ( self : int , _UpperCAmelCase : Tensor , _UpperCAmelCase : bool = False , _UpperCAmelCase : bool = True ) -> BaseModelOutputWithNoAttention: """simple docstring""" __lowercase = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: __lowercase = hidden_states + (hidden_state,) __lowercase = stage_module(_UpperCAmelCase ) if output_hidden_states: __lowercase = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=_UpperCAmelCase , hidden_states=_UpperCAmelCase ) class A__ ( lowerCAmelCase__ ): lowerCAmelCase__ : Optional[Any] = RegNetConfig lowerCAmelCase__ : Optional[int] = "regnet" lowerCAmelCase__ : Dict = "pixel_values" lowerCAmelCase__ : List[str] = True def a__ ( self : Any , _UpperCAmelCase : Any ) -> Dict: """simple docstring""" if isinstance(_UpperCAmelCase , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode='fan_out' , nonlinearity='relu' ) elif isinstance(_UpperCAmelCase , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def a__ ( self : Any , _UpperCAmelCase : Any , _UpperCAmelCase : Optional[Any]=False ) -> Dict: """simple docstring""" if isinstance(_UpperCAmelCase , _UpperCAmelCase ): __lowercase = value SCREAMING_SNAKE_CASE__ = r""" This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. """ SCREAMING_SNAKE_CASE__ = r""" Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConvNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~file_utils.ModelOutput`] instead of a plain tuple. """ @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , lowerCAmelCase__ , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetModel with RESNET->REGNET,ResNet->RegNet class A__ ( lowerCAmelCase__ ): def __init__( self : List[Any] , _UpperCAmelCase : Any ) -> str: """simple docstring""" super().__init__(_UpperCAmelCase ) __lowercase = config __lowercase = RegNetEmbeddings(_UpperCAmelCase ) __lowercase = RegNetEncoder(_UpperCAmelCase ) __lowercase = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_UpperCAmelCase ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=_UpperCAmelCase , config_class=_CONFIG_FOR_DOC , modality='vision' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def a__ ( self : Tuple , _UpperCAmelCase : Tensor , _UpperCAmelCase : Optional[bool] = None , _UpperCAmelCase : Optional[bool] = None ) -> BaseModelOutputWithPoolingAndNoAttention: """simple docstring""" __lowercase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __lowercase = return_dict if return_dict is not None else self.config.use_return_dict __lowercase = self.embedder(_UpperCAmelCase ) __lowercase = self.encoder( _UpperCAmelCase , output_hidden_states=_UpperCAmelCase , return_dict=_UpperCAmelCase ) __lowercase = encoder_outputs[0] __lowercase = self.pooler(_UpperCAmelCase ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=_UpperCAmelCase , pooler_output=_UpperCAmelCase , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , lowerCAmelCase__ , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetForImageClassification with RESNET->REGNET,ResNet->RegNet,resnet->regnet class A__ ( lowerCAmelCase__ ): def __init__( self : str , _UpperCAmelCase : List[Any] ) -> Tuple: """simple docstring""" super().__init__(_UpperCAmelCase ) __lowercase = config.num_labels __lowercase = RegNetModel(_UpperCAmelCase ) # classification head __lowercase = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_UpperCAmelCase ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=_UpperCAmelCase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def a__ ( self : List[Any] , _UpperCAmelCase : Optional[torch.FloatTensor] = None , _UpperCAmelCase : Optional[torch.LongTensor] = None , _UpperCAmelCase : Optional[bool] = None , _UpperCAmelCase : Optional[bool] = None , ) -> ImageClassifierOutputWithNoAttention: """simple docstring""" __lowercase = return_dict if return_dict is not None else self.config.use_return_dict __lowercase = self.regnet(_UpperCAmelCase , output_hidden_states=_UpperCAmelCase , return_dict=_UpperCAmelCase ) __lowercase = outputs.pooler_output if return_dict else outputs[1] __lowercase = self.classifier(_UpperCAmelCase ) __lowercase = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: __lowercase = 'regression' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): __lowercase = 'single_label_classification' else: __lowercase = 'multi_label_classification' if self.config.problem_type == "regression": __lowercase = MSELoss() if self.num_labels == 1: __lowercase = loss_fct(logits.squeeze() , labels.squeeze() ) else: __lowercase = loss_fct(_UpperCAmelCase , _UpperCAmelCase ) elif self.config.problem_type == "single_label_classification": __lowercase = CrossEntropyLoss() __lowercase = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": __lowercase = BCEWithLogitsLoss() __lowercase = loss_fct(_UpperCAmelCase , _UpperCAmelCase ) if not return_dict: __lowercase = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=_UpperCAmelCase , logits=_UpperCAmelCase , hidden_states=outputs.hidden_states )
325
0
import math def _SCREAMING_SNAKE_CASE ( ) -> None: __lowerCAmelCase : Union[str, Any] = input("""Enter message: """ ) __lowerCAmelCase : Any = int(input(F'''Enter key [2-{len(SCREAMING_SNAKE_CASE ) - 1}]: ''' ) ) __lowerCAmelCase : Union[str, Any] = input("""Encryption/Decryption [e/d]: """ ) if mode.lower().startswith("""e""" ): __lowerCAmelCase : List[str] = encrypt_message(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) elif mode.lower().startswith("""d""" ): __lowerCAmelCase : str = decrypt_message(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(F'''Output:\n{text + "|"}''' ) def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :int , SCREAMING_SNAKE_CASE :str ) -> str: __lowerCAmelCase : Dict = [""""""] * key for col in range(SCREAMING_SNAKE_CASE ): __lowerCAmelCase : str = col while pointer < len(SCREAMING_SNAKE_CASE ): cipher_text[col] += message[pointer] pointer += key return "".join(SCREAMING_SNAKE_CASE ) def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :int , SCREAMING_SNAKE_CASE :str ) -> str: __lowerCAmelCase : Optional[Any] = math.ceil(len(SCREAMING_SNAKE_CASE ) / key ) __lowerCAmelCase : List[Any] = key __lowerCAmelCase : Any = (num_cols * num_rows) - len(SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Tuple = [""""""] * num_cols __lowerCAmelCase : Any = 0 __lowerCAmelCase : Any = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): __lowerCAmelCase : Tuple = 0 row += 1 return "".join(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod() main()
352
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class snake_case_ ( __lowercase ): A_ = ['image_processor', 'tokenizer'] A_ = 'ChineseCLIPImageProcessor' A_ = ('BertTokenizer', 'BertTokenizerFast') def __init__( self : Optional[Any] , _snake_case : List[Any]=None , _snake_case : str=None , **_snake_case : int )->List[str]: '''simple docstring''' __lowerCAmelCase : str = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , _snake_case , ) __lowerCAmelCase : List[str] = kwargs.pop("""feature_extractor""" ) __lowerCAmelCase : Union[str, Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(_snake_case , _snake_case ) __lowerCAmelCase : Any = self.image_processor def __call__( self : Optional[Any] , _snake_case : Tuple=None , _snake_case : Tuple=None , _snake_case : List[str]=None , **_snake_case : Any )->Union[str, Any]: '''simple docstring''' if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: __lowerCAmelCase : List[Any] = self.tokenizer(_snake_case , return_tensors=_snake_case , **_snake_case ) if images is not None: __lowerCAmelCase : Tuple = self.image_processor(_snake_case , return_tensors=_snake_case , **_snake_case ) if text is not None and images is not None: __lowerCAmelCase : Optional[Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_snake_case ) , tensor_type=_snake_case ) def UpperCAmelCase__ ( self : Optional[int] , *_snake_case : Union[str, Any] , **_snake_case : int )->Optional[Any]: '''simple docstring''' return self.tokenizer.batch_decode(*_snake_case , **_snake_case ) def UpperCAmelCase__ ( self : Dict , *_snake_case : Dict , **_snake_case : Any )->Dict: '''simple docstring''' return self.tokenizer.decode(*_snake_case , **_snake_case ) @property def UpperCAmelCase__ ( self : Optional[int] )->str: '''simple docstring''' __lowerCAmelCase : Tuple = self.tokenizer.model_input_names __lowerCAmelCase : Tuple = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def UpperCAmelCase__ ( self : Dict )->Union[str, Any]: '''simple docstring''' warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" , _snake_case , ) return self.image_processor_class
232
0
'''simple docstring''' def snake_case_ ( __SCREAMING_SNAKE_CASE : int = 1000 ): """simple docstring""" lowercase_ : Optional[int] = 2**power lowercase_ : Any = str(__SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = list(__SCREAMING_SNAKE_CASE ) lowercase_ : List[Any] = 0 for i in list_num: sum_of_num += int(__SCREAMING_SNAKE_CASE ) return sum_of_num if __name__ == "__main__": _lowercase : Any = int(input("Enter the power of 2: ").strip()) print("2 ^ ", power, " = ", 2**power) _lowercase : Optional[int] = solution(power) print("Sum of the digits is: ", result)
93
import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin lowerCamelCase = random.Random() def lowerCamelCase_ ( _a , _a=1.0 , _a=None , _a=None ): """simple docstring""" if rng is None: lowerCAmelCase__ : Tuple = global_rng lowerCAmelCase__ : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class _a ( unittest.TestCase): def __init__( self : Optional[Any] , _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : str=7 , _SCREAMING_SNAKE_CASE : str=400 , _SCREAMING_SNAKE_CASE : Any=2000 , _SCREAMING_SNAKE_CASE : List[Any]=1 , _SCREAMING_SNAKE_CASE : Optional[Any]=0.0 , _SCREAMING_SNAKE_CASE : Tuple=1_6000 , _SCREAMING_SNAKE_CASE : Tuple=True , _SCREAMING_SNAKE_CASE : List[Any]=True , )-> Any: lowerCAmelCase__ : List[str] = parent lowerCAmelCase__ : str = batch_size lowerCAmelCase__ : Optional[int] = min_seq_length lowerCAmelCase__ : int = max_seq_length lowerCAmelCase__ : List[Any] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCAmelCase__ : Optional[Any] = feature_size lowerCAmelCase__ : Union[str, Any] = padding_value lowerCAmelCase__ : Tuple = sampling_rate lowerCAmelCase__ : int = return_attention_mask lowerCAmelCase__ : Optional[int] = do_normalize def UpperCAmelCase__( self : Union[str, Any] )-> Union[str, Any]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCAmelCase__( self : Optional[int] , _SCREAMING_SNAKE_CASE : Optional[Any]=False , _SCREAMING_SNAKE_CASE : str=False )-> Tuple: def _flatten(_SCREAMING_SNAKE_CASE : Union[str, Any] ): return list(itertools.chain(*_SCREAMING_SNAKE_CASE ) ) if equal_length: lowerCAmelCase__ : int = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size lowerCAmelCase__ : Tuple = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCAmelCase__ : Optional[int] = [np.asarray(_SCREAMING_SNAKE_CASE ) for x in speech_inputs] return speech_inputs class _a ( _lowercase , unittest.TestCase): _a : List[str] = WavaVecaFeatureExtractor def UpperCAmelCase__( self : Union[str, Any] )-> List[str]: lowerCAmelCase__ : List[str] = WavaVecaFeatureExtractionTester(self ) def UpperCAmelCase__( self : str , _SCREAMING_SNAKE_CASE : Optional[Any] )-> Optional[int]: self.assertTrue(np.all(np.mean(_SCREAMING_SNAKE_CASE , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(_SCREAMING_SNAKE_CASE , axis=0 ) - 1 ) < 1E-3 ) ) def UpperCAmelCase__( self : int )-> Tuple: # Tests that all call wrap to encode_plus and batch_encode_plus lowerCAmelCase__ : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCAmelCase__ : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase__ : Optional[Any] = [np.asarray(_SCREAMING_SNAKE_CASE ) for speech_input in speech_inputs] # Test not batched input lowerCAmelCase__ : Optional[Any] = feat_extract(speech_inputs[0] , return_tensors='''np''' ).input_values lowerCAmelCase__ : int = feat_extract(np_speech_inputs[0] , return_tensors='''np''' ).input_values self.assertTrue(np.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1E-3 ) ) # Test batched lowerCAmelCase__ : Any = feat_extract(_SCREAMING_SNAKE_CASE , return_tensors='''np''' ).input_values lowerCAmelCase__ : Dict = feat_extract(_SCREAMING_SNAKE_CASE , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assertTrue(np.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. lowerCAmelCase__ : str = [floats_list((1, x) )[0] for x in (800, 800, 800)] lowerCAmelCase__ : List[Any] = np.asarray(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[Any] = feat_extract(_SCREAMING_SNAKE_CASE , return_tensors='''np''' ).input_values lowerCAmelCase__ : Optional[int] = feat_extract(_SCREAMING_SNAKE_CASE , return_tensors='''np''' ).input_values for enc_seq_a, enc_seq_a in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assertTrue(np.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1E-3 ) ) def UpperCAmelCase__( self : Dict )-> Optional[Any]: lowerCAmelCase__ : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ : int = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase__ : Tuple = ['''longest''', '''max_length''', '''do_not_pad'''] lowerCAmelCase__ : str = [None, 1600, None] for max_length, padding in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): lowerCAmelCase__ : int = feat_extract(_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , return_tensors='''np''' ) lowerCAmelCase__ : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCAmelCase__( self : List[Any] )-> Union[str, Any]: lowerCAmelCase__ : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ : List[str] = range(800 , 1400 , 200 ) lowerCAmelCase__ : List[Any] = [floats_list((1, x) )[0] for x in lengths] lowerCAmelCase__ : Optional[Any] = ['''longest''', '''max_length''', '''do_not_pad'''] lowerCAmelCase__ : Optional[int] = [None, 1600, None] for max_length, padding in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): lowerCAmelCase__ : List[str] = feat_extract(_SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def UpperCAmelCase__( self : List[str] )-> int: lowerCAmelCase__ : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ : str = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase__ : str = feat_extract( _SCREAMING_SNAKE_CASE , truncation=_SCREAMING_SNAKE_CASE , max_length=1000 , padding='''max_length''' , return_tensors='''np''' ) lowerCAmelCase__ : Any = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def UpperCAmelCase__( self : Tuple )-> str: lowerCAmelCase__ : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase__ : List[str] = feat_extract( _SCREAMING_SNAKE_CASE , truncation=_SCREAMING_SNAKE_CASE , max_length=1000 , padding='''longest''' , return_tensors='''np''' ) lowerCAmelCase__ : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) lowerCAmelCase__ : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase__ : str = feat_extract( _SCREAMING_SNAKE_CASE , truncation=_SCREAMING_SNAKE_CASE , max_length=2000 , padding='''longest''' , return_tensors='''np''' ) lowerCAmelCase__ : int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def UpperCAmelCase__( self : List[Any] )-> List[str]: import torch lowerCAmelCase__ : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ : Union[str, Any] = np.random.rand(100 ).astype(np.floataa ) lowerCAmelCase__ : Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCAmelCase__ : str = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) lowerCAmelCase__ : List[Any] = feature_extractor.pad([{'''input_values''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def UpperCAmelCase__( self : Optional[int] )-> Dict: # this test makes sure that models that are using # group norm don't have their feature extractor return the # attention_mask for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: lowerCAmelCase__ : Tuple = WavaVecaConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Union[str, Any] = WavaVecaFeatureExtractor.from_pretrained(_SCREAMING_SNAKE_CASE ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == '''layer''' )
131
0
'''simple docstring''' def lowerCamelCase ( UpperCAmelCase__ : int ) -> list[int]: if num <= 0: raise ValueError("""Input must be a positive integer""" ) lowercase_ : int = [True] * (num + 1) lowercase_ : int = 2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , UpperCAmelCase__ ): lowercase_ : List[Any] = False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() _lowercase : Any = int(input("Enter a positive integer: ").strip()) print(prime_sieve_eratosthenes(user_num))
363
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline _lowercase : Union[str, Any] = logging.get_logger(__name__) # pylint: disable=invalid-name class __magic_name__ ( _UpperCAmelCase): def __init__( self : Union[str, Any] , lowercase_ : Optional[int] , lowercase_ : str ): super().__init__() self.register_modules(unet=lowercase_ , scheduler=lowercase_ ) @torch.no_grad() def __call__( self : List[str] , lowercase_ : int = 1 , lowercase_ : int = 100 , lowercase_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowercase_ : Optional[float] = None , lowercase_ : bool = True , ): if audio_length_in_s is None: lowercase_ : List[Any] = self.unet.config.sample_size / self.unet.config.sample_rate lowercase_ : Dict = audio_length_in_s * self.unet.config.sample_rate lowercase_ : Any = 2 ** len(self.unet.up_blocks ) if sample_size < 3 * down_scale_factor: raise ValueError( f'''{audio_length_in_s} is too small. Make sure it\'s bigger or equal to''' f''' {3 * down_scale_factor / self.unet.config.sample_rate}.''' ) lowercase_ : List[Any] = int(lowercase_ ) if sample_size % down_scale_factor != 0: lowercase_ : int = ( (audio_length_in_s * self.unet.config.sample_rate) // down_scale_factor + 1 ) * down_scale_factor logger.info( f'''{audio_length_in_s} is increased to {sample_size / self.unet.config.sample_rate} so that it can be handled''' f''' by the model. It will be cut to {original_sample_size / self.unet.config.sample_rate} after the denoising''' """ process.""" ) lowercase_ : Any = int(lowercase_ ) lowercase_ : List[str] = next(iter(self.unet.parameters() ) ).dtype lowercase_ : List[str] = (batch_size, self.unet.config.in_channels, sample_size) if isinstance(lowercase_ , lowercase_ ) and len(lowercase_ ) != batch_size: raise ValueError( f'''You have passed a list of generators of length {len(lowercase_ )}, but requested an effective batch''' f''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) lowercase_ : Any = randn_tensor(lowercase_ , generator=lowercase_ , device=self.device , dtype=lowercase_ ) # set step values self.scheduler.set_timesteps(lowercase_ , device=audio.device ) lowercase_ : Optional[Any] = self.scheduler.timesteps.to(lowercase_ ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output lowercase_ : Dict = self.unet(lowercase_ , lowercase_ ).sample # 2. compute previous image: x_t -> t_t-1 lowercase_ : List[str] = self.scheduler.step(lowercase_ , lowercase_ , lowercase_ ).prev_sample lowercase_ : str = audio.clamp(-1 , 1 ).float().cpu().numpy() lowercase_ : Union[str, Any] = audio[:, :, :original_sample_size] if not return_dict: return (audio,) return AudioPipelineOutput(audios=lowercase_ )
21
0
import numpy as np import datasets UpperCAmelCase__ : Dict = '\nCompute the Mahalanobis Distance\n\nMahalonobis distance is the distance between a point and a distribution.\nAnd not between two distinct points. It is effectively a multivariate equivalent of the Euclidean distance.\nIt was introduced by Prof. P. C. Mahalanobis in 1936\nand has been used in various statistical applications ever since\n[source: https://www.machinelearningplus.com/statistics/mahalanobis-distance/]\n' UpperCAmelCase__ : Any = '\\n@article{de2000mahalanobis,\n title={The mahalanobis distance},\n author={De Maesschalck, Roy and Jouan-Rimbaud, Delphine and Massart, D{\'e}sir{\'e} L},\n journal={Chemometrics and intelligent laboratory systems},\n volume={50},\n number={1},\n pages={1--18},\n year={2000},\n publisher={Elsevier}\n}\n' UpperCAmelCase__ : List[Any] = '\nArgs:\n X: List of datapoints to be compared with the `reference_distribution`.\n reference_distribution: List of datapoints from the reference distribution we want to compare to.\nReturns:\n mahalanobis: The Mahalonobis distance for each datapoint in `X`.\nExamples:\n\n >>> mahalanobis_metric = datasets.load_metric("mahalanobis")\n >>> results = mahalanobis_metric.compute(reference_distribution=[[0, 1], [1, 0]], X=[[0, 1]])\n >>> print(results)\n {\'mahalanobis\': array([0.5])}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def __magic_name__ ( self : Optional[int] ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''X''': datasets.Sequence(datasets.Value('''float''' , id='''sequence''' ) , id='''X''' ), } ) , ) def __magic_name__ ( self : Dict , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Tuple ): """simple docstring""" # convert to numpy arrays _A: Union[str, Any] = np.array(lowerCAmelCase_ ) _A: Any = np.array(lowerCAmelCase_ ) # Assert that arrays are 2D if len(X.shape ) != 2: raise ValueError('''Expected `X` to be a 2D vector''' ) if len(reference_distribution.shape ) != 2: raise ValueError('''Expected `reference_distribution` to be a 2D vector''' ) if reference_distribution.shape[0] < 2: raise ValueError( '''Expected `reference_distribution` to be a 2D vector with more than one element in the first dimension''' ) # Get mahalanobis distance for each prediction _A: List[Any] = X - np.mean(lowerCAmelCase_ ) _A: Any = np.cov(reference_distribution.T ) try: _A: str = np.linalg.inv(lowerCAmelCase_ ) except np.linalg.LinAlgError: _A: str = np.linalg.pinv(lowerCAmelCase_ ) _A: Optional[int] = np.dot(lowerCAmelCase_ , lowerCAmelCase_ ) _A: Dict = np.dot(lowerCAmelCase_ , X_minus_mu.T ).diagonal() return {"mahalanobis": mahal_dist}
121
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase__ : Any = {'configuration_swin': ['SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SwinConfig', 'SwinOnnxConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ : Union[str, Any] = [ 'SWIN_PRETRAINED_MODEL_ARCHIVE_LIST', 'SwinForImageClassification', 'SwinForMaskedImageModeling', 'SwinModel', 'SwinPreTrainedModel', 'SwinBackbone', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ : str = [ 'TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFSwinForImageClassification', 'TFSwinForMaskedImageModeling', 'TFSwinModel', 'TFSwinPreTrainedModel', ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys UpperCAmelCase__ : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
121
1
'''simple docstring''' from math import pi, sqrt def __UpperCAmelCase ( a_: float ): if num <= 0: raise ValueError("math domain error" ) if num > 171.5: raise OverflowError("math range error" ) elif num - int(a_ ) not in (0, 0.5): raise NotImplementedError("num must be an integer or a half-integer" ) elif num == 0.5: return sqrt(a_ ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def __UpperCAmelCase ( ): assert gamma(0.5 ) == sqrt(a_ ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() __a = 1.0 while num: __a = float(input('Gamma of: ')) print(f'gamma({num}) = {gamma(num)}') print('\nEnter 0 to exit...')
359
'''simple docstring''' import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class A__ ( pl.LightningModule ): """simple docstring""" def __init__( self : Any , lowerCAmelCase__ : Optional[Any] ) -> str: """simple docstring""" super().__init__() _UpperCAmelCase : List[str] = model _UpperCAmelCase : Dict = 2 _UpperCAmelCase : Tuple = nn.Linear(self.model.config.hidden_size , self.num_labels ) def _lowerCAmelCase ( self : Tuple ) -> int: """simple docstring""" pass def __UpperCAmelCase ( a_: str, a_: str, a_: str ): # load longformer model from model identifier _UpperCAmelCase : int = LongformerModel.from_pretrained(a_ ) _UpperCAmelCase : Any = LightningModel(a_ ) _UpperCAmelCase : int = torch.load(a_, map_location=torch.device("cpu" ) ) lightning_model.load_state_dict(ckpt["state_dict"] ) # init longformer question answering model _UpperCAmelCase : List[str] = LongformerForQuestionAnswering.from_pretrained(a_ ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(a_ ) print(f"""Conversion successful. Model saved under {pytorch_dump_folder_path}""" ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--longformer_model', default=None, type=str, required=True, help='model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.', ) parser.add_argument( '--longformer_question_answering_ckpt_path', default=None, type=str, required=True, help='Path the official PyTorch Lightning Checkpoint.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) __a = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
17
0
'''simple docstring''' def _SCREAMING_SNAKE_CASE (A = 1_000 ) -> int: """simple docstring""" return sum(e for e in range(3 , A ) if e % 3 == 0 or e % 5 == 0 ) if __name__ == "__main__": print(f"""{solution() = }""")
2
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging lowercase = { "cola": 2, "mnli": 3, "mrpc": 2, "sst-2": 2, "sts-b": 1, "qqp": 2, "qnli": 2, "rte": 2, "wnli": 2, } logging.set_verbosity_info() def __UpperCAmelCase ( a_ , a_ , a_ , a_=None): # Initialise PyTorch model snake_case_ = XLNetConfig.from_json_file(a_) snake_case_ = finetuning_task.lower() if finetuning_task is not None else '' if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f'''Building PyTorch XLNetForSequenceClassification model from configuration: {config}''') snake_case_ = finetuning_task snake_case_ = GLUE_TASKS_NUM_LABELS[finetuning_task] snake_case_ = XLNetForSequenceClassification(a_) elif "squad" in finetuning_task: snake_case_ = finetuning_task snake_case_ = XLNetForQuestionAnswering(a_) else: snake_case_ = XLNetLMHeadModel(a_) # Load weights from tf checkpoint load_tf_weights_in_xlnet(a_ , a_ , a_) # Save pytorch-model snake_case_ = os.path.join(a_ , a_) snake_case_ = os.path.join(a_ , a_) print(f'''Save PyTorch model to {os.path.abspath(a_)}''') torch.save(model.state_dict() , a_) print(f'''Save configuration file to {os.path.abspath(a_)}''') with open(a_ , 'w' , encoding='utf-8') as f: f.write(config.to_json_string()) if __name__ == "__main__": lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--xlnet_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained XLNet model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the folder to store the PyTorch model or dataset/vocab.", ) parser.add_argument( "--finetuning_task", default=None, type=str, help="Name of a task on which the XLNet TensorFlow model was fine-tuned", ) lowercase = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
178
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_encoder_decoder''': ['''EncoderDecoderConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['''EncoderDecoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['''TFEncoderDecoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['''FlaxEncoderDecoderModel'''] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
357
"""simple docstring""" import json import os import unittest from transformers.models.blenderbot_small.tokenization_blenderbot_small import ( VOCAB_FILES_NAMES, BlenderbotSmallTokenizer, ) from ...test_tokenization_common import TokenizerTesterMixin class __snake_case ( _lowercase , unittest.TestCase): snake_case__ : Optional[int] = BlenderbotSmallTokenizer snake_case__ : List[str] = False def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" super().setUp() _lowerCamelCase : str = ['''__start__''', '''adapt''', '''act''', '''ap@@''', '''te''', '''__end__''', '''__unk__'''] _lowerCamelCase : Any = dict(zip(__lowerCAmelCase , range(len(__lowerCAmelCase ) ) ) ) _lowerCamelCase : Any = ['''#version: 0.2''', '''a p''', '''t e</w>''', '''ap t</w>''', '''a d''', '''ad apt</w>''', '''a c''', '''ac t</w>''', ''''''] _lowerCamelCase : List[str] = {'''unk_token''': '''__unk__''', '''bos_token''': '''__start__''', '''eos_token''': '''__end__'''} _lowerCamelCase : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(__lowerCAmelCase ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(__lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Tuple , **__lowerCAmelCase : List[str] ): """simple docstring""" kwargs.update(self.special_tokens_map ) return BlenderbotSmallTokenizer.from_pretrained(self.tmpdirname , **__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[Any] , __lowerCAmelCase : List[str] ): """simple docstring""" _lowerCamelCase : int = '''adapt act apte''' _lowerCamelCase : Tuple = '''adapt act apte''' return input_text, output_text def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : Tuple = BlenderbotSmallTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) _lowerCamelCase : int = '''adapt act apte''' _lowerCamelCase : Optional[Any] = ['''adapt''', '''act''', '''ap@@''', '''te'''] _lowerCamelCase : Optional[Any] = tokenizer.tokenize(__lowerCAmelCase ) self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase : List[Any] = [tokenizer.bos_token] + tokens + [tokenizer.eos_token] _lowerCamelCase : Dict = [0, 1, 2, 3, 4, 5] self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCAmelCase ) , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" _lowerCamelCase : str = BlenderbotSmallTokenizer.from_pretrained('''facebook/blenderbot-90M''' ) assert tok('''sam''' ).input_ids == [1_3_8_4] _lowerCamelCase : List[str] = '''I am a small frog.''' _lowerCamelCase : str = tok([src_text] , padding=__lowerCAmelCase , truncation=__lowerCAmelCase )['''input_ids'''] _lowerCamelCase : Any = tok.batch_decode(__lowerCAmelCase , skip_special_tokens=__lowerCAmelCase , clean_up_tokenization_spaces=__lowerCAmelCase )[0] assert src_text != decoded # I wish it did! assert decoded == "i am a small frog ." def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase : Union[str, Any] = BlenderbotSmallTokenizer.from_pretrained('''facebook/blenderbot-90M''' ) _lowerCamelCase : Optional[Any] = '''I am a small frog .''' _lowerCamelCase : str = '''.''' _lowerCamelCase : str = tok(__lowerCAmelCase )['''input_ids'''] _lowerCamelCase : Dict = tok(__lowerCAmelCase )['''input_ids'''] assert encoded[-1] == encoded_dot[0]
175
0
import os import textwrap import pyarrow as pa import pytest from datasets import ClassLabel, Features, Image from datasets.packaged_modules.csv.csv import Csv from ..utils import require_pil @pytest.fixture def __UpperCamelCase ( lowerCAmelCase__ : Dict ): __a : Dict = tmp_path / '''file.csv''' __a : Dict = textwrap.dedent( '''\ header1,header2 1,2 10,20 ''' ) with open(lowerCAmelCase__ , '''w''' ) as f: f.write(lowerCAmelCase__ ) return str(lowerCAmelCase__ ) @pytest.fixture def __UpperCamelCase ( lowerCAmelCase__ : Union[str, Any] ): __a : List[Any] = tmp_path / '''malformed_file.csv''' __a : List[Any] = textwrap.dedent( '''\ header1,header2 1,2 10,20, ''' ) with open(lowerCAmelCase__ , '''w''' ) as f: f.write(lowerCAmelCase__ ) return str(lowerCAmelCase__ ) @pytest.fixture def __UpperCamelCase ( lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Dict ): __a : List[str] = tmp_path / '''csv_with_image.csv''' __a : Dict = textwrap.dedent( f"\\n image\n {image_file}\n " ) with open(lowerCAmelCase__ , '''w''' ) as f: f.write(lowerCAmelCase__ ) return str(lowerCAmelCase__ ) @pytest.fixture def __UpperCamelCase ( lowerCAmelCase__ : Any ): __a : Dict = tmp_path / '''csv_with_label.csv''' __a : Any = textwrap.dedent( '''\ label good bad good ''' ) with open(lowerCAmelCase__ , '''w''' ) as f: f.write(lowerCAmelCase__ ) return str(lowerCAmelCase__ ) @pytest.fixture def __UpperCamelCase ( lowerCAmelCase__ : Tuple ): __a : Union[str, Any] = tmp_path / '''csv_with_int_list.csv''' __a : str = textwrap.dedent( '''\ int_list 1 2 3 4 5 6 7 8 9 ''' ) with open(lowerCAmelCase__ , '''w''' ) as f: f.write(lowerCAmelCase__ ) return str(lowerCAmelCase__ ) def __UpperCamelCase ( lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : str , lowerCAmelCase__ : Dict ): __a : List[str] = Csv() __a : Optional[Any] = csv._generate_tables([[csv_file, malformed_csv_file]] ) with pytest.raises(lowerCAmelCase__ , match='''Error tokenizing data''' ): for _ in generator: pass assert any( record.levelname == '''ERROR''' and '''Failed to read file''' in record.message and os.path.basename(lowerCAmelCase__ ) in record.message for record in caplog.records ) @require_pil def __UpperCamelCase ( lowerCAmelCase__ : Tuple ): with open(lowerCAmelCase__ , encoding='''utf-8''' ) as f: __a : Optional[int] = f.read().splitlines()[1] __a : List[Any] = Csv(encoding='''utf-8''' , features=Features({'''image''': Image()} ) ) __a : Dict = csv._generate_tables([[csv_file_with_image]] ) __a : Optional[int] = pa.concat_tables([table for _, table in generator] ) assert pa_table.schema.field('''image''' ).type == Image()() __a : List[Any] = pa_table.to_pydict()['''image'''] assert generated_content == [{"path": image_file, "bytes": None}] def __UpperCamelCase ( lowerCAmelCase__ : Tuple ): with open(lowerCAmelCase__ , encoding='''utf-8''' ) as f: __a : Optional[int] = f.read().splitlines()[1:] __a : List[Any] = Csv(encoding='''utf-8''' , features=Features({'''label''': ClassLabel(names=['''good''', '''bad'''] )} ) ) __a : Tuple = csv._generate_tables([[csv_file_with_label]] ) __a : Tuple = pa.concat_tables([table for _, table in generator] ) assert pa_table.schema.field('''label''' ).type == ClassLabel(names=['''good''', '''bad'''] )() __a : Optional[Any] = pa_table.to_pydict()['''label'''] assert generated_content == [ClassLabel(names=['''good''', '''bad'''] ).straint(lowerCAmelCase__ ) for label in labels] def __UpperCamelCase ( lowerCAmelCase__ : int ): __a : Union[str, Any] = Csv(encoding='''utf-8''' , sep=''',''' , converters={'''int_list''': lambda lowerCAmelCase__ : [int(lowerCAmelCase__ ) for i in x.split()]} ) __a : Tuple = csv._generate_tables([[csv_file_with_int_list]] ) __a : Any = pa.concat_tables([table for _, table in generator] ) assert pa.types.is_list(pa_table.schema.field('''int_list''' ).type ) __a : List[str] = pa_table.to_pydict()['''int_list'''] assert generated_content == [[1, 2, 3], [4, 5, 6], [7, 8, 9]]
216
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowercase__ =logging.get_logger(__name__) class UpperCamelCase__ ( __lowercase ): _SCREAMING_SNAKE_CASE : Optional[Any] = ["pixel_values"] def __init__(self : str , snake_case_ : bool = True , snake_case_ : Dict[str, int] = None , snake_case_ : float = None , snake_case_ : PILImageResampling = PILImageResampling.BILINEAR , snake_case_ : bool = True , snake_case_ : Union[int, float] = 1 / 2_5_5 , snake_case_ : bool = True , snake_case_ : Optional[Union[float, List[float]]] = None , snake_case_ : Optional[Union[float, List[float]]] = None , **snake_case_ : Tuple , ): super().__init__(**snake_case_ ) __a : Optional[Any] = size if size is not None else {'''shortest_edge''': 3_8_4} __a : str = get_size_dict(snake_case_ , default_to_square=snake_case_ ) __a : Tuple = do_resize __a : List[str] = size # Default value set here for backwards compatibility where the value in config is None __a : List[str] = crop_pct if crop_pct is not None else 2_2_4 / 2_5_6 __a : Tuple = resample __a : Any = do_rescale __a : Optional[Any] = rescale_factor __a : int = do_normalize __a : List[str] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __a : Union[str, Any] = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowerCAmelCase (self : int , snake_case_ : np.ndarray , snake_case_ : Dict[str, int] , snake_case_ : float , snake_case_ : PILImageResampling = PILImageResampling.BICUBIC , snake_case_ : Optional[Union[str, ChannelDimension]] = None , **snake_case_ : str , ): __a : str = get_size_dict(snake_case_ , default_to_square=snake_case_ ) if "shortest_edge" not in size: raise ValueError(f"Size dictionary must contain 'shortest_edge' key. Got {size.keys()}" ) __a : Any = size['''shortest_edge'''] if shortest_edge < 3_8_4: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct __a : Union[str, Any] = int(shortest_edge / crop_pct ) __a : Tuple = get_resize_output_image_size(snake_case_ , size=snake_case_ , default_to_square=snake_case_ ) __a : List[str] = resize(image=snake_case_ , size=snake_case_ , resample=snake_case_ , data_format=snake_case_ , **snake_case_ ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=snake_case_ , size=(shortest_edge, shortest_edge) , data_format=snake_case_ , **snake_case_ ) else: # warping (no cropping) when evaluated at 384 or larger return resize( snake_case_ , size=(shortest_edge, shortest_edge) , resample=snake_case_ , data_format=snake_case_ , **snake_case_ ) def lowerCAmelCase (self : Optional[int] , snake_case_ : np.ndarray , snake_case_ : Union[int, float] , snake_case_ : Optional[Union[str, ChannelDimension]] = None , **snake_case_ : Union[str, Any] , ): return rescale(snake_case_ , scale=snake_case_ , data_format=snake_case_ , **snake_case_ ) def lowerCAmelCase (self : Union[str, Any] , snake_case_ : np.ndarray , snake_case_ : Union[float, List[float]] , snake_case_ : Union[float, List[float]] , snake_case_ : Optional[Union[str, ChannelDimension]] = None , **snake_case_ : Union[str, Any] , ): return normalize(snake_case_ , mean=snake_case_ , std=snake_case_ , data_format=snake_case_ , **snake_case_ ) def lowerCAmelCase (self : Dict , snake_case_ : ImageInput , snake_case_ : bool = None , snake_case_ : Dict[str, int] = None , snake_case_ : float = None , snake_case_ : PILImageResampling = None , snake_case_ : bool = None , snake_case_ : float = None , snake_case_ : bool = None , snake_case_ : Optional[Union[float, List[float]]] = None , snake_case_ : Optional[Union[float, List[float]]] = None , snake_case_ : Optional[Union[str, TensorType]] = None , snake_case_ : ChannelDimension = ChannelDimension.FIRST , **snake_case_ : Dict , ): __a : List[Any] = do_resize if do_resize is not None else self.do_resize __a : Dict = crop_pct if crop_pct is not None else self.crop_pct __a : Tuple = resample if resample is not None else self.resample __a : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale __a : Optional[int] = rescale_factor if rescale_factor is not None else self.rescale_factor __a : int = do_normalize if do_normalize is not None else self.do_normalize __a : Optional[int] = image_mean if image_mean is not None else self.image_mean __a : Optional[int] = image_std if image_std is not None else self.image_std __a : Dict = size if size is not None else self.size __a : List[str] = get_size_dict(snake_case_ , default_to_square=snake_case_ ) __a : List[str] = make_list_of_images(snake_case_ ) if not valid_images(snake_case_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_resize and size["shortest_edge"] < 3_8_4 and crop_pct is None: raise ValueError('''crop_pct must be specified if size < 384.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. __a : List[Any] = [to_numpy_array(snake_case_ ) for image in images] if do_resize: __a : List[str] = [self.resize(image=snake_case_ , size=snake_case_ , crop_pct=snake_case_ , resample=snake_case_ ) for image in images] if do_rescale: __a : List[Any] = [self.rescale(image=snake_case_ , scale=snake_case_ ) for image in images] if do_normalize: __a : Optional[int] = [self.normalize(image=snake_case_ , mean=snake_case_ , std=snake_case_ ) for image in images] __a : str = [to_channel_dimension_format(snake_case_ , snake_case_ ) for image in images] __a : Tuple = {'''pixel_values''': images} return BatchFeature(data=snake_case_ , tensor_type=snake_case_ )
216
1
def snake_case( __magic_name__ = 10_00 ) -> int: '''simple docstring''' lowercase : Dict = 3 lowercase : Tuple = 0 while a < n: if a % 3 == 0 or a % 5 == 0: result += a elif a % 15 == 0: result -= a a += 1 return result if __name__ == "__main__": print(f'''{solution() = }''')
368
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = { 'facebook/xmod-base': 'https://huggingface.co/facebook/xmod-base/resolve/main/config.json', 'facebook/xmod-large-prenorm': 'https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json', 'facebook/xmod-base-13-125k': 'https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json', 'facebook/xmod-base-30-125k': 'https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json', 'facebook/xmod-base-30-195k': 'https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json', 'facebook/xmod-base-60-125k': 'https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json', 'facebook/xmod-base-60-265k': 'https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json', 'facebook/xmod-base-75-125k': 'https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json', 'facebook/xmod-base-75-269k': 'https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json', } class _A ( _lowerCamelCase ): _UpperCamelCase : Tuple = '''xmod''' def __init__( self : Optional[Any] , _A : Union[str, Any]=30_522 , _A : List[Any]=768 , _A : Optional[Any]=12 , _A : Any=12 , _A : Tuple=3_072 , _A : Optional[int]="gelu" , _A : List[Any]=0.1 , _A : str=0.1 , _A : List[Any]=512 , _A : List[str]=2 , _A : str=0.02 , _A : Any=1E-12 , _A : Union[str, Any]=1 , _A : List[Any]=0 , _A : Dict=2 , _A : int="absolute" , _A : Dict=True , _A : int=None , _A : List[str]=False , _A : Dict=2 , _A : int=False , _A : Optional[int]=True , _A : Any=True , _A : Optional[int]=("en_XX",) , _A : Any=None , **_A : Optional[int] , ) -> Union[str, Any]: """simple docstring""" super().__init__(pad_token_id=_A , bos_token_id=_A , eos_token_id=_A , **_A ) lowercase : Optional[Any] = vocab_size lowercase : Union[str, Any] = hidden_size lowercase : Optional[Any] = num_hidden_layers lowercase : Dict = num_attention_heads lowercase : Union[str, Any] = hidden_act lowercase : Tuple = intermediate_size lowercase : List[str] = hidden_dropout_prob lowercase : Union[str, Any] = attention_probs_dropout_prob lowercase : Dict = max_position_embeddings lowercase : Any = type_vocab_size lowercase : Optional[Any] = initializer_range lowercase : str = layer_norm_eps lowercase : Tuple = position_embedding_type lowercase : Optional[Any] = use_cache lowercase : int = classifier_dropout lowercase : Optional[int] = pre_norm lowercase : Any = adapter_reduction_factor lowercase : Union[str, Any] = adapter_layer_norm lowercase : Optional[int] = adapter_reuse_layer_norm lowercase : Optional[Any] = ln_before_adapter lowercase : Union[str, Any] = list(_A ) lowercase : List[Any] = default_language class _A ( _lowerCamelCase ): @property def __a ( self : Union[str, Any] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": lowercase : Any = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowercase : Union[str, Any] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
116
0
def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase = 10, _UpperCAmelCase = 22 ) -> int: '''simple docstring''' lowerCAmelCase : str = range(1, _UpperCAmelCase ) lowerCAmelCase : Dict = range(1, _UpperCAmelCase ) return sum( 1 for power in powers for base in bases if len(str(base**power ) ) == power ) if __name__ == "__main__": print(F'{solution(10, 22) = }')
138
from ...configuration_utils import PretrainedConfig from ...utils import logging __A : List[str] = logging.get_logger(__name__) __A : Optional[Any] = { '''facebook/vit-mae-base''': '''https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json''', # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class __A ( lowerCAmelCase ): lowerCAmelCase_ : List[Any] = "vit_mae" def __init__( self : Optional[int] , UpperCAmelCase_ : Dict=768 , UpperCAmelCase_ : int=12 , UpperCAmelCase_ : Dict=12 , UpperCAmelCase_ : Optional[Any]=3072 , UpperCAmelCase_ : List[str]="gelu" , UpperCAmelCase_ : Union[str, Any]=0.0 , UpperCAmelCase_ : Optional[Any]=0.0 , UpperCAmelCase_ : List[str]=0.02 , UpperCAmelCase_ : Union[str, Any]=1E-12 , UpperCAmelCase_ : Any=224 , UpperCAmelCase_ : int=16 , UpperCAmelCase_ : Union[str, Any]=3 , UpperCAmelCase_ : Optional[Any]=True , UpperCAmelCase_ : int=16 , UpperCAmelCase_ : Union[str, Any]=512 , UpperCAmelCase_ : Any=8 , UpperCAmelCase_ : List[Any]=2048 , UpperCAmelCase_ : Tuple=0.75 , UpperCAmelCase_ : str=False , **UpperCAmelCase_ : str , ): super().__init__(**UpperCAmelCase_ ) lowerCAmelCase : Optional[Any] = hidden_size lowerCAmelCase : List[str] = num_hidden_layers lowerCAmelCase : Union[str, Any] = num_attention_heads lowerCAmelCase : Optional[int] = intermediate_size lowerCAmelCase : str = hidden_act lowerCAmelCase : Optional[int] = hidden_dropout_prob lowerCAmelCase : Optional[int] = attention_probs_dropout_prob lowerCAmelCase : List[Any] = initializer_range lowerCAmelCase : Union[str, Any] = layer_norm_eps lowerCAmelCase : Any = image_size lowerCAmelCase : Union[str, Any] = patch_size lowerCAmelCase : Union[str, Any] = num_channels lowerCAmelCase : Any = qkv_bias lowerCAmelCase : Union[str, Any] = decoder_num_attention_heads lowerCAmelCase : Tuple = decoder_hidden_size lowerCAmelCase : int = decoder_num_hidden_layers lowerCAmelCase : Optional[Any] = decoder_intermediate_size lowerCAmelCase : Union[str, Any] = mask_ratio lowerCAmelCase : Any = norm_pix_loss
138
1
"""simple docstring""" from pathlib import Path from typing import List from transformers import is_torch_available, is_vision_available from transformers.testing_utils import get_tests_dir, is_tool_test from transformers.tools.agent_types import AGENT_TYPE_MAPPING, AgentAudio, AgentImage, AgentText if is_torch_available(): import torch if is_vision_available(): from PIL import Image __lowerCamelCase = ["""text""", """image""", """audio"""] def UpperCamelCase ( __lowerCamelCase : List[str] ): snake_case : str = [] for input_type in input_types: if input_type == "text": inputs.append("Text input" ) elif input_type == "image": inputs.append( Image.open(Path(get_tests_dir("fixtures/tests_samples/COCO" ) ) / "000000039769.png" ).resize((512, 512) ) ) elif input_type == "audio": inputs.append(torch.ones(3000 ) ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): inputs.append(create_inputs(__lowerCamelCase ) ) else: raise ValueError(f"""Invalid type requested: {input_type}""" ) return inputs def UpperCamelCase ( __lowerCamelCase : List ): snake_case : List[str] = [] for output in outputs: if isinstance(__lowerCamelCase , (str, AgentText) ): output_types.append("text" ) elif isinstance(__lowerCamelCase , (Image.Image, AgentImage) ): output_types.append("image" ) elif isinstance(__lowerCamelCase , (torch.Tensor, AgentAudio) ): output_types.append("audio" ) else: raise ValueError(f"""Invalid output: {output}""" ) return output_types @is_tool_test class UpperCAmelCase : def _SCREAMING_SNAKE_CASE (self : List[str] ) -> List[str]: '''simple docstring''' self.assertTrue(hasattr(self.tool , "inputs" ) ) self.assertTrue(hasattr(self.tool , "outputs" ) ) snake_case : List[Any] = self.tool.inputs for _input in inputs: if isinstance(_input , snake_case__ ): for __input in _input: self.assertTrue(__input in authorized_types ) else: self.assertTrue(_input in authorized_types ) snake_case : str = self.tool.outputs for _output in outputs: self.assertTrue(_output in authorized_types ) def _SCREAMING_SNAKE_CASE (self : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' snake_case : List[str] = create_inputs(self.tool.inputs ) snake_case : Dict = self.tool(*snake_case__ ) # There is a single output if len(self.tool.outputs ) == 1: snake_case : List[Any] = [outputs] self.assertListEqual(output_types(snake_case__ ) , self.tool.outputs ) def _SCREAMING_SNAKE_CASE (self : List[str] ) -> List[Any]: '''simple docstring''' self.assertTrue(hasattr(self.tool , "description" ) ) self.assertTrue(hasattr(self.tool , "default_checkpoint" ) ) self.assertTrue(self.tool.description.startswith("This is a tool that" ) ) def _SCREAMING_SNAKE_CASE (self : int ) -> Union[str, Any]: '''simple docstring''' snake_case : str = create_inputs(self.tool.inputs ) snake_case : int = self.tool(*snake_case__ ) if not isinstance(snake_case__ , snake_case__ ): snake_case : Optional[Any] = [outputs] self.assertEqual(len(snake_case__ ) , len(self.tool.outputs ) ) for output, output_type in zip(snake_case__ , self.tool.outputs ): snake_case : Any = AGENT_TYPE_MAPPING[output_type] self.assertTrue(isinstance(snake_case__ , snake_case__ ) ) def _SCREAMING_SNAKE_CASE (self : Optional[Any] ) -> Optional[int]: '''simple docstring''' snake_case : List[Any] = create_inputs(self.tool.inputs ) snake_case : str = [] for _input, input_type in zip(snake_case__ , self.tool.inputs ): if isinstance(snake_case__ , snake_case__ ): _inputs.append([AGENT_TYPE_MAPPING[_input_type](_input ) for _input_type in input_type] ) else: _inputs.append(AGENT_TYPE_MAPPING[input_type](_input ) ) # Should not raise an error snake_case : Optional[int] = self.tool(*snake_case__ ) if not isinstance(snake_case__ , snake_case__ ): snake_case : List[str] = [outputs] self.assertEqual(len(snake_case__ ) , len(self.tool.outputs ) )
361
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available from ...utils import OptionalDependencyNotAvailable __lowerCamelCase = {"""configuration_dpt""": ["""DPT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """DPTConfig"""]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = ["""DPTFeatureExtractor"""] __lowerCamelCase = ["""DPTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = [ """DPT_PRETRAINED_MODEL_ARCHIVE_LIST""", """DPTForDepthEstimation""", """DPTForSemanticSegmentation""", """DPTModel""", """DPTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_dpt import DPT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPTConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_dpt import DPTFeatureExtractor from .image_processing_dpt import DPTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dpt import ( DPT_PRETRAINED_MODEL_ARCHIVE_LIST, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel, DPTPreTrainedModel, ) else: import sys __lowerCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
10
0
from ....configuration_utils import PretrainedConfig from ....utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { "CarlCochet/trajectory-transformer-halfcheetah-medium-v2": ( "https://huggingface.co/CarlCochet/trajectory-transformer-halfcheetah-medium-v2/resolve/main/config.json" ), # See all TrajectoryTransformer models at https://huggingface.co/models?filter=trajectory_transformer } class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCAmelCase_ ="trajectory_transformer" UpperCAmelCase_ =["past_key_values"] UpperCAmelCase_ ={ "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self , _A=100 , _A=5 , _A=1 , _A=1 , _A=249 , _A=6 , _A=17 , _A=25 , _A=4 , _A=4 , _A=128 , _A=0.1 , _A=0.1 , _A=0.1 , _A=0.0006 , _A=512 , _A=0.02 , _A=1E-12 , _A=1 , _A=True , _A=1 , _A=50256 , _A=50256 , **_A , ) -> int: SCREAMING_SNAKE_CASE_ = vocab_size SCREAMING_SNAKE_CASE_ = action_weight SCREAMING_SNAKE_CASE_ = reward_weight SCREAMING_SNAKE_CASE_ = value_weight SCREAMING_SNAKE_CASE_ = max_position_embeddings SCREAMING_SNAKE_CASE_ = block_size SCREAMING_SNAKE_CASE_ = action_dim SCREAMING_SNAKE_CASE_ = observation_dim SCREAMING_SNAKE_CASE_ = transition_dim SCREAMING_SNAKE_CASE_ = learning_rate SCREAMING_SNAKE_CASE_ = n_layer SCREAMING_SNAKE_CASE_ = n_head SCREAMING_SNAKE_CASE_ = n_embd SCREAMING_SNAKE_CASE_ = embd_pdrop SCREAMING_SNAKE_CASE_ = attn_pdrop SCREAMING_SNAKE_CASE_ = resid_pdrop SCREAMING_SNAKE_CASE_ = initializer_range SCREAMING_SNAKE_CASE_ = layer_norm_eps SCREAMING_SNAKE_CASE_ = kaiming_initializer_range SCREAMING_SNAKE_CASE_ = use_cache super().__init__(pad_token_id=_A , bos_token_id=_A , eos_token_id=_A , **_A )
299
from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput __UpperCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): """simple docstring""" @register_to_config def __init__( self , _A , _A = None , _A = None ) -> Optional[Any]: super().__init__() SCREAMING_SNAKE_CASE_ = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" SCREAMING_SNAKE_CASE_ = torch.zeros(_A , _A ) else: SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = torch.nn.Parameter(_A ) class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCAmelCase_ =42 UpperCAmelCase_ =42 UpperCAmelCase_ =42 UpperCAmelCase_ =42 UpperCAmelCase_ =42 UpperCAmelCase_ =42 def __init__( self , _A , _A , _A , _A , _A , _A , ) -> Any: super().__init__() self.register_modules( vqvae=_A , transformer=_A , text_encoder=_A , tokenizer=_A , scheduler=_A , learned_classifier_free_sampling_embeddings=_A , ) def _UpperCamelCase ( self , _A , _A , _A ) -> Optional[int]: SCREAMING_SNAKE_CASE_ = len(_A ) if isinstance(_A , _A ) else 1 # get prompt text embeddings SCREAMING_SNAKE_CASE_ = self.tokenizer( _A , padding='''max_length''' , max_length=self.tokenizer.model_max_length , return_tensors='''pt''' , ) SCREAMING_SNAKE_CASE_ = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: SCREAMING_SNAKE_CASE_ = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( '''The following part of your input was truncated because CLIP can only handle sequences up to''' F''' {self.tokenizer.model_max_length} tokens: {removed_text}''' ) SCREAMING_SNAKE_CASE_ = text_input_ids[:, : self.tokenizer.model_max_length] SCREAMING_SNAKE_CASE_ = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 SCREAMING_SNAKE_CASE_ = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=_A ) # duplicate text embeddings for each generation per prompt SCREAMING_SNAKE_CASE_ = prompt_embeds.repeat_interleave(_A , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: SCREAMING_SNAKE_CASE_ = self.learned_classifier_free_sampling_embeddings.embeddings SCREAMING_SNAKE_CASE_ = negative_prompt_embeds.unsqueeze(0 ).repeat(_A , 1 , 1 ) else: SCREAMING_SNAKE_CASE_ = [''''''] * batch_size SCREAMING_SNAKE_CASE_ = text_input_ids.shape[-1] SCREAMING_SNAKE_CASE_ = self.tokenizer( _A , padding='''max_length''' , max_length=_A , truncation=_A , return_tensors='''pt''' , ) SCREAMING_SNAKE_CASE_ = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings SCREAMING_SNAKE_CASE_ = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=_A ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method SCREAMING_SNAKE_CASE_ = negative_prompt_embeds.shape[1] SCREAMING_SNAKE_CASE_ = negative_prompt_embeds.repeat(1 , _A , 1 ) SCREAMING_SNAKE_CASE_ = negative_prompt_embeds.view(batch_size * num_images_per_prompt , _A , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes SCREAMING_SNAKE_CASE_ = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self , _A , _A = 100 , _A = 5.0 , _A = 1.0 , _A = 1 , _A = None , _A = None , _A = "pil" , _A = True , _A = None , _A = 1 , ) -> Union[ImagePipelineOutput, Tuple]: if isinstance(_A , _A ): SCREAMING_SNAKE_CASE_ = 1 elif isinstance(_A , _A ): SCREAMING_SNAKE_CASE_ = len(_A ) else: raise ValueError(F'''`prompt` has to be of type `str` or `list` but is {type(_A )}''' ) SCREAMING_SNAKE_CASE_ = batch_size * num_images_per_prompt SCREAMING_SNAKE_CASE_ = guidance_scale > 1.0 SCREAMING_SNAKE_CASE_ = self._encode_prompt(_A , _A , _A ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(_A , _A ) or callback_steps <= 0) ): raise ValueError( F'''`callback_steps` has to be a positive integer but is {callback_steps} of type''' F''' {type(_A )}.''' ) # get the initial completely masked latents unless the user supplied it SCREAMING_SNAKE_CASE_ = (batch_size, self.transformer.num_latent_pixels) if latents is None: SCREAMING_SNAKE_CASE_ = self.transformer.num_vector_embeds - 1 SCREAMING_SNAKE_CASE_ = torch.full(_A , _A ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( '''Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0,''' F''' {self.transformer.num_vector_embeds - 1} (inclusive).''' ) SCREAMING_SNAKE_CASE_ = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(_A , device=self.device ) SCREAMING_SNAKE_CASE_ = self.scheduler.timesteps.to(self.device ) SCREAMING_SNAKE_CASE_ = latents for i, t in enumerate(self.progress_bar(_A ) ): # expand the sample if we are doing classifier free guidance SCREAMING_SNAKE_CASE_ = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` SCREAMING_SNAKE_CASE_ = self.transformer(_A , encoder_hidden_states=_A , timestep=_A ).sample if do_classifier_free_guidance: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = model_output.chunk(2 ) SCREAMING_SNAKE_CASE_ = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(_A , dim=1 , keepdim=_A ) SCREAMING_SNAKE_CASE_ = self.truncate(_A , _A ) # remove `log(0)`'s (`-inf`s) SCREAMING_SNAKE_CASE_ = model_output.clamp(-70 ) # compute the previous noisy sample x_t -> x_t-1 SCREAMING_SNAKE_CASE_ = self.scheduler.step(_A , timestep=_A , sample=_A , generator=_A ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(_A , _A , _A ) SCREAMING_SNAKE_CASE_ = self.vqvae.config.vq_embed_dim SCREAMING_SNAKE_CASE_ = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) SCREAMING_SNAKE_CASE_ = self.vqvae.quantize.get_codebook_entry(_A , shape=_A ) SCREAMING_SNAKE_CASE_ = self.vqvae.decode(_A , force_not_quantize=_A ).sample SCREAMING_SNAKE_CASE_ = (image / 2 + 0.5).clamp(0 , 1 ) SCREAMING_SNAKE_CASE_ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": SCREAMING_SNAKE_CASE_ = self.numpy_to_pil(_A ) if not return_dict: return (image,) return ImagePipelineOutput(images=_A ) def _UpperCamelCase ( self , _A , _A ) -> torch.FloatTensor: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = torch.sort(_A , 1 , descending=_A ) SCREAMING_SNAKE_CASE_ = torch.exp(_A ) SCREAMING_SNAKE_CASE_ = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out SCREAMING_SNAKE_CASE_ = torch.full_like(keep_mask[:, 0:1, :] , _A ) SCREAMING_SNAKE_CASE_ = torch.cat((all_true, keep_mask) , dim=1 ) SCREAMING_SNAKE_CASE_ = keep_mask[:, :-1, :] SCREAMING_SNAKE_CASE_ = keep_mask.gather(1 , indices.argsort(1 ) ) SCREAMING_SNAKE_CASE_ = log_p_x_0.clone() SCREAMING_SNAKE_CASE_ = -torch.inf # -inf = log(0) return rv
299
1
def UpperCamelCase ( __lowerCamelCase : list[list[int]] , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : set ): snake_case , snake_case : Tuple = len(__lowerCamelCase ), len(grid[0] ) if ( min(__lowerCamelCase , __lowerCamelCase ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) snake_case : str = 0 count += depth_first_search(__lowerCamelCase , row + 1 , __lowerCamelCase , __lowerCamelCase ) count += depth_first_search(__lowerCamelCase , row - 1 , __lowerCamelCase , __lowerCamelCase ) count += depth_first_search(__lowerCamelCase , __lowerCamelCase , col + 1 , __lowerCamelCase ) count += depth_first_search(__lowerCamelCase , __lowerCamelCase , col - 1 , __lowerCamelCase ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
10
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging __lowerCamelCase = logging.get_logger(__name__) class UpperCAmelCase ( A_ ): A__ : int = ["pixel_values"] def __init__(self : Tuple , snake_case__ : bool = True , snake_case__ : Union[int, float] = 1 / 2_55 , snake_case__ : bool = True , snake_case__ : int = 8 , **snake_case__ : Dict , ) -> None: '''simple docstring''' super().__init__(**snake_case__ ) snake_case : int = do_rescale snake_case : List[str] = rescale_factor snake_case : Optional[Any] = do_pad snake_case : Dict = pad_size def _SCREAMING_SNAKE_CASE (self : Dict , snake_case__ : np.ndarray , snake_case__ : float , snake_case__ : Optional[Union[str, ChannelDimension]] = None , **snake_case__ : List[str] ) -> np.ndarray: '''simple docstring''' return rescale(snake_case__ , scale=snake_case__ , data_format=snake_case__ , **snake_case__ ) def _SCREAMING_SNAKE_CASE (self : List[Any] , snake_case__ : np.ndarray , snake_case__ : int , snake_case__ : Optional[Union[str, ChannelDimension]] = None ) -> Dict: '''simple docstring''' snake_case , snake_case : Union[str, Any] = get_image_size(snake_case__ ) snake_case : str = (old_height // size + 1) * size - old_height snake_case : List[str] = (old_width // size + 1) * size - old_width return pad(snake_case__ , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=snake_case__ ) def _SCREAMING_SNAKE_CASE (self : str , snake_case__ : ImageInput , snake_case__ : Optional[bool] = None , snake_case__ : Optional[float] = None , snake_case__ : Optional[bool] = None , snake_case__ : Optional[int] = None , snake_case__ : Optional[Union[str, TensorType]] = None , snake_case__ : Union[str, ChannelDimension] = ChannelDimension.FIRST , **snake_case__ : List[Any] , ) -> Tuple: '''simple docstring''' snake_case : str = do_rescale if do_rescale is not None else self.do_rescale snake_case : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case : Optional[Any] = do_pad if do_pad is not None else self.do_pad snake_case : Dict = pad_size if pad_size is not None else self.pad_size snake_case : Union[str, Any] = make_list_of_images(snake_case__ ) if not valid_images(snake_case__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. snake_case : str = [to_numpy_array(snake_case__ ) for image in images] if do_rescale: snake_case : str = [self.rescale(image=snake_case__ , scale=snake_case__ ) for image in images] if do_pad: snake_case : List[Any] = [self.pad(snake_case__ , size=snake_case__ ) for image in images] snake_case : Union[str, Any] = [to_channel_dimension_format(snake_case__ , snake_case__ ) for image in images] snake_case : Optional[Any] = {"pixel_values": images} return BatchFeature(data=snake_case__ , tensor_type=snake_case__ )
10
1
'''simple docstring''' import json import os import unittest from transformers.models.blenderbot_small.tokenization_blenderbot_small import ( VOCAB_FILES_NAMES, BlenderbotSmallTokenizer, ) from ...test_tokenization_common import TokenizerTesterMixin class _lowerCAmelCase ( __UpperCAmelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE : int = BlenderbotSmallTokenizer __SCREAMING_SNAKE_CASE : int = False def _a (self ): super().setUp() A_ : Optional[Any] = ["""__start__""", """adapt""", """act""", """ap@@""", """te""", """__end__""", """__unk__"""] A_ : Union[str, Any] = dict(zip(lowercase , range(len(lowercase ) ) ) ) A_ : Optional[Any] = ["""#version: 0.2""", """a p""", """t e</w>""", """ap t</w>""", """a d""", """ad apt</w>""", """a c""", """ac t</w>""", """"""] A_ : List[Any] = {"""unk_token""": """__unk__""", """bos_token""": """__start__""", """eos_token""": """__end__"""} A_ : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) A_ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(lowercase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(lowercase ) ) def _a (self , **lowercase ): kwargs.update(self.special_tokens_map ) return BlenderbotSmallTokenizer.from_pretrained(self.tmpdirname , **lowercase ) def _a (self , lowercase ): A_ : Tuple = """adapt act apte""" A_ : List[str] = """adapt act apte""" return input_text, output_text def _a (self ): A_ : int = BlenderbotSmallTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) A_ : Optional[Any] = """adapt act apte""" A_ : Union[str, Any] = ["""adapt""", """act""", """ap@@""", """te"""] A_ : List[Any] = tokenizer.tokenize(lowercase ) self.assertListEqual(lowercase , lowercase ) A_ : Union[str, Any] = [tokenizer.bos_token] + tokens + [tokenizer.eos_token] A_ : Optional[int] = [0, 1, 2, 3, 4, 5] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase ) , lowercase ) def _a (self ): A_ : Dict = BlenderbotSmallTokenizer.from_pretrained("""facebook/blenderbot-90M""" ) assert tok("""sam""" ).input_ids == [1384] A_ : Tuple = """I am a small frog.""" A_ : Tuple = tok([src_text] , padding=lowercase , truncation=lowercase )["""input_ids"""] A_ : Tuple = tok.batch_decode(lowercase , skip_special_tokens=lowercase , clean_up_tokenization_spaces=lowercase )[0] assert src_text != decoded # I wish it did! assert decoded == "i am a small frog ." def _a (self ): A_ : Any = BlenderbotSmallTokenizer.from_pretrained("""facebook/blenderbot-90M""" ) A_ : List[Any] = """I am a small frog .""" A_ : Any = """.""" A_ : Dict = tok(lowercase )["""input_ids"""] A_ : int = tok(lowercase )["""input_ids"""] assert encoded[-1] == encoded_dot[0]
206
'''simple docstring''' import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class _lowerCAmelCase : def __init__(self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=True , lowercase=False , lowercase=True , lowercase=99 , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=512 , lowercase=16 , lowercase=2 , lowercase=0.02 , lowercase=3 , lowercase=4 , lowercase=None , ): A_ : str = parent A_ : int = batch_size A_ : Dict = seq_length A_ : Any = is_training A_ : List[str] = use_input_mask A_ : Any = use_token_type_ids A_ : int = use_labels A_ : str = vocab_size A_ : Any = hidden_size A_ : Optional[Any] = num_hidden_layers A_ : Dict = num_attention_heads A_ : Dict = intermediate_size A_ : Optional[int] = hidden_act A_ : int = hidden_dropout_prob A_ : Optional[int] = attention_probs_dropout_prob A_ : Any = max_position_embeddings A_ : List[Any] = type_vocab_size A_ : Any = type_sequence_label_size A_ : Tuple = initializer_range A_ : int = num_labels A_ : Optional[int] = num_choices A_ : Optional[int] = scope def _a (self ): A_ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ : List[str] = None if self.use_input_mask: A_ : Any = random_attention_mask([self.batch_size, self.seq_length] ) A_ : int = None if self.use_token_type_ids: A_ : int = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A_ : Tuple = None A_ : List[str] = None A_ : List[str] = None if self.use_labels: A_ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A_ : str = ids_tensor([self.batch_size] , self.num_choices ) A_ : Optional[Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _a (self ): return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase , initializer_range=self.initializer_range , ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : List[Any] = BioGptModel(config=lowercase ) model.to(lowercase ) model.eval() A_ : List[Any] = model(lowercase , attention_mask=lowercase ) A_ : str = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ): A_ : List[Any] = BioGptForCausalLM(config=lowercase ) model.to(lowercase ) model.eval() A_ : Dict = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ): A_ : Optional[Any] = BioGptModel(config=lowercase ) model.to(lowercase ) model.eval() # create attention mask A_ : str = torch.ones(input_ids.shape , dtype=torch.long , device=lowercase ) A_ : Optional[Any] = self.seq_length // 2 A_ : List[Any] = 0 # first forward pass A_, A_ : List[str] = model(lowercase , attention_mask=lowercase ).to_tuple() # create hypothetical next token and extent to next_input_ids A_ : Optional[Any] = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids A_ : Union[str, Any] = ids_tensor((1,) , lowercase ).item() + 1 A_ : List[str] = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) A_ : Optional[int] = random_other_next_tokens # append to next input_ids and attn_mask A_ : Union[str, Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) A_ : Any = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=lowercase )] , dim=1 , ) # get two different outputs A_ : List[Any] = model(lowercase , attention_mask=lowercase )["""last_hidden_state"""] A_ : Optional[int] = model(lowercase , past_key_values=lowercase , attention_mask=lowercase )["""last_hidden_state"""] # select random slice A_ : List[str] = ids_tensor((1,) , output_from_past.shape[-1] ).item() A_ : int = output_from_no_past[:, -1, random_slice_idx].detach() A_ : List[Any] = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase , lowercase , atol=1E-3 ) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ): A_ : Optional[int] = BioGptModel(config=lowercase ).to(lowercase ).eval() A_ : Union[str, Any] = torch.ones(input_ids.shape , dtype=torch.long , device=lowercase ) # first forward pass A_ : int = model(lowercase , attention_mask=lowercase , use_cache=lowercase ) A_, A_ : int = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids A_ : Optional[int] = ids_tensor((self.batch_size, 3) , config.vocab_size ) A_ : str = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and A_ : Union[str, Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) A_ : int = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) A_ : List[str] = model(lowercase , attention_mask=lowercase )["""last_hidden_state"""] A_ : Tuple = model(lowercase , attention_mask=lowercase , past_key_values=lowercase )[ """last_hidden_state""" ] # select random slice A_ : List[str] = ids_tensor((1,) , output_from_past.shape[-1] ).item() A_ : List[str] = output_from_no_past[:, -3:, random_slice_idx].detach() A_ : Any = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase , lowercase , atol=1E-3 ) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase , lowercase=False ): A_ : Union[str, Any] = BioGptForCausalLM(lowercase ) model.to(lowercase ) if gradient_checkpointing: model.gradient_checkpointing_enable() A_ : Dict = model(lowercase , labels=lowercase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def _a (self , lowercase , *lowercase ): A_ : Union[str, Any] = BioGptModel(lowercase ) A_ : str = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.0_01 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.01 ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ): A_ : Union[str, Any] = self.num_labels A_ : Optional[int] = BioGptForTokenClassification(lowercase ) model.to(lowercase ) model.eval() A_ : Dict = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a (self ): A_ : List[Any] = self.prepare_config_and_inputs() ( ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ) : Dict = config_and_inputs A_ : Tuple = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class _lowerCAmelCase ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE : str = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE : int = (BioGptForCausalLM,) if is_torch_available() else () __SCREAMING_SNAKE_CASE : Optional[Any] = ( { 'feature-extraction': BioGptModel, 'text-classification': BioGptForSequenceClassification, 'text-generation': BioGptForCausalLM, 'token-classification': BioGptForTokenClassification, 'zero-shot': BioGptForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE : List[str] = False def _a (self ): A_ : Tuple = BioGptModelTester(self ) A_ : List[Any] = ConfigTester(self , config_class=lowercase , hidden_size=37 ) def _a (self ): self.config_tester.run_common_tests() def _a (self ): A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def _a (self ): A_ : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A_ : int = type self.model_tester.create_and_check_model(*lowercase ) def _a (self ): A_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*lowercase ) def _a (self ): A_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*lowercase , gradient_checkpointing=lowercase ) def _a (self ): A_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*lowercase ) def _a (self ): A_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*lowercase ) def _a (self ): A_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*lowercase ) @slow def _a (self ): A_ : str = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(lowercase ) A_ : List[str] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A_ : List[str] = """left""" # Define PAD Token = EOS Token = 50256 A_ : Any = tokenizer.eos_token A_ : Dict = model.config.eos_token_id # use different length sentences to test batching A_ : List[Any] = [ """Hello, my dog is a little""", """Today, I""", ] A_ : List[str] = tokenizer(lowercase , return_tensors="""pt""" , padding=lowercase ) A_ : List[str] = inputs["""input_ids"""].to(lowercase ) A_ : List[Any] = model.generate( input_ids=lowercase , attention_mask=inputs["""attention_mask"""].to(lowercase ) , ) A_ : List[str] = tokenizer(sentences[0] , return_tensors="""pt""" ).input_ids.to(lowercase ) A_ : List[Any] = model.generate(input_ids=lowercase ) A_ : List[str] = inputs_non_padded.shape[-1] - inputs["""attention_mask"""][-1].long().sum().cpu().item() A_ : Union[str, Any] = tokenizer(sentences[1] , return_tensors="""pt""" ).input_ids.to(lowercase ) A_ : Any = model.generate(input_ids=lowercase , max_length=model.config.max_length - num_paddings ) A_ : List[str] = tokenizer.batch_decode(lowercase , skip_special_tokens=lowercase ) A_ : str = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowercase ) A_ : List[str] = tokenizer.decode(output_padded[0] , skip_special_tokens=lowercase ) A_ : Union[str, Any] = [ """Hello, my dog is a little bit bigger than a little bit.""", """Today, I have a good idea of how to use the information""", ] self.assertListEqual(lowercase , lowercase ) self.assertListEqual(lowercase , [non_padded_sentence, padded_sentence] ) @slow def _a (self ): for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : int = BioGptModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) def _a (self ): A_, A_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() A_ : List[Any] = 3 A_ : Union[str, Any] = input_dict["""input_ids"""] A_ : List[Any] = input_ids.ne(1 ).to(lowercase ) A_ : str = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) A_ : Union[str, Any] = BioGptForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() A_ : Tuple = model(lowercase , attention_mask=lowercase , labels=lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _a (self ): A_, A_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() A_ : Tuple = 3 A_ : Dict = """multi_label_classification""" A_ : List[Any] = input_dict["""input_ids"""] A_ : Tuple = input_ids.ne(1 ).to(lowercase ) A_ : List[str] = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) A_ : Dict = BioGptForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() A_ : int = model(lowercase , attention_mask=lowercase , labels=lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class _lowerCAmelCase ( unittest.TestCase ): @slow def _a (self ): A_ : Dict = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) A_ : Optional[Any] = torch.tensor([[2, 4805, 9, 656, 21]] ) A_ : Dict = model(lowercase )[0] A_ : Any = 42384 A_ : Union[str, Any] = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , lowercase ) A_ : Union[str, Any] = torch.tensor( [[[-9.52_36, -9.89_18, 10.45_57], [-11.04_69, -9.64_23, 8.10_22], [-8.86_64, -7.88_26, 5.53_25]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase , atol=1E-4 ) ) @slow def _a (self ): A_ : Union[str, Any] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A_ : Dict = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(lowercase ) torch.manual_seed(0 ) A_ : Dict = tokenizer("""COVID-19 is""" , return_tensors="""pt""" ).to(lowercase ) A_ : Optional[int] = model.generate( **lowercase , min_length=100 , max_length=1024 , num_beams=5 , early_stopping=lowercase , ) A_ : Union[str, Any] = tokenizer.decode(output_ids[0] , skip_special_tokens=lowercase ) A_ : Any = ( """COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the""" """ causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and""" """ territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),""" """ and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and""" """ more than 800,000 deaths.""" ) self.assertEqual(lowercase , lowercase )
206
1
from __future__ import annotations def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_,snake_case_ ): _A : List[Any] = [] _A : List[str] = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) _A : Dict = result + left + right return input_list def lowerCAmelCase_ ( snake_case_ ): if len(snake_case_ ) <= 1: return input_list _A : Tuple = list(snake_case_ ) # iteration for two-way merging _A : int = 2 while p <= len(snake_case_ ): # getting low, high and middle value for merge-sort of single list for i in range(0,len(snake_case_ ),snake_case_ ): _A : List[str] = i _A : Tuple = i + p - 1 _A : Dict = (low + high + 1) // 2 _A : Tuple = merge(snake_case_,snake_case_,snake_case_,snake_case_ ) # final merge of last two parts if p * 2 >= len(snake_case_ ): _A : Any = i _A : List[Any] = merge(snake_case_,0,snake_case_,len(snake_case_ ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": _snake_case = input("Enter numbers separated by a comma:\n").strip() if user_input == "": _snake_case = [] else: _snake_case = [int(item.strip()) for item in user_input.split(",")] print(iter_merge_sort(unsorted))
356
import inspect import unittest from transformers import ViTMSNConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMSNForImageClassification, ViTMSNModel from transformers.models.vit_msn.modeling_vit_msn import VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowercase : def __init__( self , _a , _a=13 , _a=30 , _a=2 , _a=3 , _a=True , _a=True , _a=32 , _a=5 , _a=4 , _a=37 , _a="gelu" , _a=0.1 , _a=0.1 , _a=10 , _a=0.02 , _a=None , ) -> Union[str, Any]: _A : Optional[int] = parent _A : Dict = batch_size _A : Any = image_size _A : Optional[int] = patch_size _A : Optional[int] = num_channels _A : List[Any] = is_training _A : Optional[Any] = use_labels _A : Any = hidden_size _A : Any = num_hidden_layers _A : List[Any] = num_attention_heads _A : int = intermediate_size _A : Dict = hidden_act _A : Optional[int] = hidden_dropout_prob _A : str = attention_probs_dropout_prob _A : Any = type_sequence_label_size _A : str = initializer_range _A : Tuple = scope # in ViT MSN, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) _A : List[Any] = (image_size // patch_size) ** 2 _A : str = num_patches + 1 def a__ ( self ) -> Dict: _A : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _A : List[str] = None if self.use_labels: _A : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _A : List[Any] = self.get_config() return config, pixel_values, labels def a__ ( self ) -> Union[str, Any]: return ViTMSNConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def a__ ( self , _a , _a , _a ) -> Dict: _A : List[str] = ViTMSNModel(config=_a ) model.to(_a ) model.eval() _A : List[str] = model(_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self , _a , _a , _a ) -> List[str]: _A : Union[str, Any] = self.type_sequence_label_size _A : Tuple = ViTMSNForImageClassification(_a ) model.to(_a ) model.eval() _A : Optional[int] = model(_a , labels=_a ) print("""Pixel and labels shape: {pixel_values.shape}, {labels.shape}""" ) print("""Labels: {labels}""" ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images _A : Dict = 1 _A : str = ViTMSNForImageClassification(_a ) model.to(_a ) model.eval() _A : int = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _A : int = model(_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a__ ( self ) -> Any: _A : Optional[int] = self.prepare_config_and_inputs() _A , _A , _A : Dict = config_and_inputs _A : List[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class lowercase ( UpperCamelCase__,UpperCamelCase__,unittest.TestCase ): _a = (ViTMSNModel, ViTMSNForImageClassification) if is_torch_available() else () _a = ( {"feature-extraction": ViTMSNModel, "image-classification": ViTMSNForImageClassification} if is_torch_available() else {} ) _a = False _a = False _a = False _a = False def a__ ( self ) -> Tuple: _A : Tuple = ViTMSNModelTester(self ) _A : List[Any] = ConfigTester(self , config_class=_a , has_text_modality=_a , hidden_size=37 ) def a__ ( self ) -> Optional[int]: self.config_tester.run_common_tests() @unittest.skip(reason="""ViTMSN does not use inputs_embeds""" ) def a__ ( self ) -> int: pass def a__ ( self ) -> Any: _A , _A : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A : Tuple = model_class(_a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _A : str = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_a , nn.Linear ) ) def a__ ( self ) -> str: _A , _A : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A : int = model_class(_a ) _A : Optional[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _A : str = [*signature.parameters.keys()] _A : Optional[int] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _a ) def a__ ( self ) -> List[Any]: _A : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def a__ ( self ) -> Any: _A : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_a ) @slow def a__ ( self ) -> int: for model_name in VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _A : int = ViTMSNModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def lowerCAmelCase_ ( ): _A : Dict = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class lowercase ( unittest.TestCase ): @cached_property def a__ ( self ) -> int: return ViTImageProcessor.from_pretrained("""facebook/vit-msn-small""" ) if is_vision_available() else None @slow def a__ ( self ) -> Optional[int]: torch.manual_seed(2 ) _A : Tuple = ViTMSNForImageClassification.from_pretrained("""facebook/vit-msn-small""" ).to(_a ) _A : Tuple = self.default_image_processor _A : Dict = prepare_img() _A : Optional[Any] = image_processor(images=_a , return_tensors="""pt""" ).to(_a ) # forward pass with torch.no_grad(): _A : int = model(**_a ) # verify the logits _A : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _a ) _A : Optional[int] = torch.tensor([-0.0803, -0.4454, -0.2375] ).to(_a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _a , atol=1e-4 ) )
343
0
from torch import nn class SCREAMING_SNAKE_CASE_ ( nn.Module ): def __init__( self : Any , lowerCamelCase_ : Tuple , lowerCamelCase_ : Tuple ): """simple docstring""" super().__init__() UpperCamelCase = class_size UpperCamelCase = embed_size # self.mlp1 = nn.Linear(embed_size, embed_size) # self.mlp2 = (nn.Linear(embed_size, class_size)) UpperCamelCase = nn.Linear(lowerCamelCase_ , lowerCamelCase_ ) def lowerCamelCase_ ( self : Optional[Any] , lowerCamelCase_ : int ): """simple docstring""" UpperCamelCase = self.mlp(lowerCamelCase_ ) return logits
343
from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def UpperCamelCase_( lowerCamelCase_ ) -> bool: _lowercase : int = int(number**0.5 ) return number == sq * sq def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> tuple[int, int]: _lowercase : int = x_num * y_den * z_den + y_num * x_den * z_den + z_num * x_den * y_den _lowercase : int = x_den * y_den * z_den _lowercase : int = gcd(lowerCamelCase_ , lowerCamelCase_ ) top //= hcf bottom //= hcf return top, bottom def UpperCamelCase_( lowerCamelCase_ = 35 ) -> int: _lowercase : set = set() _lowercase : int _lowercase : Fraction = Fraction(0 ) _lowercase : tuple[int, int] for x_num in range(1 , order + 1 ): for x_den in range(x_num + 1 , order + 1 ): for y_num in range(1 , order + 1 ): for y_den in range(y_num + 1 , order + 1 ): # n=1 _lowercase : int = x_num * y_den + x_den * y_num _lowercase : int = x_den * y_den _lowercase : str = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: _lowercase : List[Any] = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) # n=2 _lowercase : Dict = ( x_num * x_num * y_den * y_den + x_den * x_den * y_num * y_num ) _lowercase : List[Any] = x_den * x_den * y_den * y_den if is_sq(lowerCamelCase_ ) and is_sq(lowerCamelCase_ ): _lowercase : Tuple = int(sqrt(lowerCamelCase_ ) ) _lowercase : int = int(sqrt(lowerCamelCase_ ) ) _lowercase : Any = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: _lowercase : Optional[int] = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) # n=-1 _lowercase : Any = x_num * y_num _lowercase : str = x_den * y_num + x_num * y_den _lowercase : Any = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: _lowercase : int = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) # n=2 _lowercase : str = x_num * x_num * y_num * y_num _lowercase : Optional[Any] = ( x_den * x_den * y_num * y_num + x_num * x_num * y_den * y_den ) if is_sq(lowerCamelCase_ ) and is_sq(lowerCamelCase_ ): _lowercase : Tuple = int(sqrt(lowerCamelCase_ ) ) _lowercase : List[str] = int(sqrt(lowerCamelCase_ ) ) _lowercase : Union[str, Any] = gcd(lowerCamelCase_ , lowerCamelCase_ ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: _lowercase : Tuple = add_three( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) unique_s.add(lowerCamelCase_ ) for num, den in unique_s: total += Fraction(lowerCamelCase_ , lowerCamelCase_ ) return total.denominator + total.numerator if __name__ == "__main__": print(F"{solution() = }")
21
0
import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class __a ( nn.Module ): def __init__( self : Tuple ): super().__init__() lowerCAmelCase_ : str = nn.Linear(3 , 4 ) lowerCAmelCase_ : str = nn.BatchNormad(4 ) lowerCAmelCase_ : List[Any] = nn.Linear(4 , 5 ) def A ( self : Union[str, Any] , UpperCAmelCase : Union[str, Any] ): return self.lineara(self.batchnorm(self.lineara(UpperCAmelCase ) ) ) class __a ( unittest.TestCase ): def A ( self : Optional[Any] ): lowerCAmelCase_ : int = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCAmelCase , model.state_dict() ) lowerCAmelCase_ : Any = os.path.join(UpperCAmelCase , """index.json""" ) self.assertTrue(os.path.isfile(UpperCAmelCase ) ) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: lowerCAmelCase_ : Optional[int] = os.path.join(UpperCAmelCase , F'{key}.dat' ) self.assertTrue(os.path.isfile(UpperCAmelCase ) ) # TODO: add tests on the fact weights are properly loaded def A ( self : List[Any] ): lowerCAmelCase_ : Union[str, Any] = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: lowerCAmelCase_ : List[str] = torch.randn(2 , 3 , dtype=UpperCAmelCase ) with TemporaryDirectory() as tmp_dir: lowerCAmelCase_ : List[str] = offload_weight(UpperCAmelCase , """weight""" , UpperCAmelCase , {} ) lowerCAmelCase_ : Optional[Any] = os.path.join(UpperCAmelCase , """weight.dat""" ) self.assertTrue(os.path.isfile(UpperCAmelCase ) ) self.assertDictEqual(UpperCAmelCase , {"""weight""": {"""shape""": [2, 3], """dtype""": str(UpperCAmelCase ).split(""".""" )[1]}} ) lowerCAmelCase_ : Optional[Any] = load_offloaded_weight(UpperCAmelCase , index["""weight"""] ) self.assertTrue(torch.equal(UpperCAmelCase , UpperCAmelCase ) ) def A ( self : int ): lowerCAmelCase_ : Optional[Any] = ModelForTest() lowerCAmelCase_ : Optional[int] = model.state_dict() lowerCAmelCase_ : List[Any] = {k: v for k, v in state_dict.items() if """linear2""" not in k} lowerCAmelCase_ : Any = {k: v for k, v in state_dict.items() if """linear2""" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCAmelCase , UpperCAmelCase ) lowerCAmelCase_ : Union[str, Any] = OffloadedWeightsLoader(state_dict=UpperCAmelCase , save_folder=UpperCAmelCase ) # Every key is there with the right value self.assertEqual(sorted(UpperCAmelCase ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCAmelCase , weight_map[key] ) ) lowerCAmelCase_ : Dict = {k: v for k, v in state_dict.items() if """weight""" in k} lowerCAmelCase_ : Dict = {k: v for k, v in state_dict.items() if """weight""" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCAmelCase , UpperCAmelCase ) lowerCAmelCase_ : List[Any] = OffloadedWeightsLoader(state_dict=UpperCAmelCase , save_folder=UpperCAmelCase ) # Every key is there with the right value self.assertEqual(sorted(UpperCAmelCase ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCAmelCase , weight_map[key] ) ) with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCAmelCase , UpperCAmelCase ) # Duplicates are removed lowerCAmelCase_ : List[str] = OffloadedWeightsLoader(state_dict=UpperCAmelCase , save_folder=UpperCAmelCase ) # Every key is there with the right value self.assertEqual(sorted(UpperCAmelCase ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCAmelCase , weight_map[key] ) ) def A ( self : str ): lowerCAmelCase_ : Optional[Any] = {"""a.1""": 0, """a.10""": 1, """a.2""": 2} lowerCAmelCase_ : Dict = extract_submodules_state_dict(UpperCAmelCase , ["""a.1""", """a.2"""] ) self.assertDictEqual(UpperCAmelCase , {"""a.1""": 0, """a.2""": 2} ) lowerCAmelCase_ : Union[str, Any] = {"""a.1.a""": 0, """a.10.a""": 1, """a.2.a""": 2} lowerCAmelCase_ : Optional[Any] = extract_submodules_state_dict(UpperCAmelCase , ["""a.1""", """a.2"""] ) self.assertDictEqual(UpperCAmelCase , {"""a.1.a""": 0, """a.2.a""": 2} )
28
from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'EleutherAI/gpt-j-6B': 'https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json', # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class __a ( __UpperCamelCase ): __snake_case : Union[str, Any] = """gptj""" __snake_case : int = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : List[str] , UpperCAmelCase : Optional[int]=5_04_00 , UpperCAmelCase : Optional[int]=20_48 , UpperCAmelCase : str=40_96 , UpperCAmelCase : Any=28 , UpperCAmelCase : Dict=16 , UpperCAmelCase : List[str]=64 , UpperCAmelCase : int=None , UpperCAmelCase : Union[str, Any]="gelu_new" , UpperCAmelCase : Tuple=0.0 , UpperCAmelCase : Dict=0.0 , UpperCAmelCase : str=0.0 , UpperCAmelCase : Optional[Any]=1e-5 , UpperCAmelCase : List[Any]=0.02 , UpperCAmelCase : Optional[int]=True , UpperCAmelCase : Dict=5_02_56 , UpperCAmelCase : int=5_02_56 , UpperCAmelCase : Tuple=False , **UpperCAmelCase : Any , ): lowerCAmelCase_ : Tuple = vocab_size lowerCAmelCase_ : Union[str, Any] = n_positions lowerCAmelCase_ : Union[str, Any] = n_embd lowerCAmelCase_ : List[Any] = n_layer lowerCAmelCase_ : List[Any] = n_head lowerCAmelCase_ : Tuple = n_inner lowerCAmelCase_ : Optional[Any] = rotary_dim lowerCAmelCase_ : str = activation_function lowerCAmelCase_ : str = resid_pdrop lowerCAmelCase_ : List[Any] = embd_pdrop lowerCAmelCase_ : Dict = attn_pdrop lowerCAmelCase_ : Any = layer_norm_epsilon lowerCAmelCase_ : Optional[int] = initializer_range lowerCAmelCase_ : Optional[int] = use_cache lowerCAmelCase_ : Optional[int] = bos_token_id lowerCAmelCase_ : Any = eos_token_id super().__init__( bos_token_id=UpperCAmelCase , eos_token_id=UpperCAmelCase , tie_word_embeddings=UpperCAmelCase , **UpperCAmelCase ) class __a ( __UpperCamelCase ): def __init__( self : Any , UpperCAmelCase : PretrainedConfig , UpperCAmelCase : str = "default" , UpperCAmelCase : List[PatchingSpec] = None , UpperCAmelCase : bool = False , ): super().__init__(UpperCAmelCase , task=UpperCAmelCase , patching_specs=UpperCAmelCase , use_past=UpperCAmelCase ) if not getattr(self._config , """pad_token_id""" , UpperCAmelCase ): # TODO: how to do that better? lowerCAmelCase_ : List[Any] = 0 @property def A ( self : List[Any] ): lowerCAmelCase_ : Optional[int] = OrderedDict({"""input_ids""": {0: """batch""", 1: """sequence"""}} ) if self.use_past: self.fill_with_past_key_values_(UpperCAmelCase , direction="""inputs""" ) lowerCAmelCase_ : Any = {0: """batch""", 1: """past_sequence + sequence"""} else: lowerCAmelCase_ : List[Any] = {0: """batch""", 1: """sequence"""} return common_inputs @property def A ( self : Union[str, Any] ): return self._config.n_layer @property def A ( self : Optional[Any] ): return self._config.n_head def A ( self : Optional[Any] , UpperCAmelCase : PreTrainedTokenizer , UpperCAmelCase : int = -1 , UpperCAmelCase : int = -1 , UpperCAmelCase : bool = False , UpperCAmelCase : Optional[TensorType] = None , ): lowerCAmelCase_ : Optional[Any] = super(UpperCAmelCase , self ).generate_dummy_inputs( UpperCAmelCase , batch_size=UpperCAmelCase , seq_length=UpperCAmelCase , is_pair=UpperCAmelCase , framework=UpperCAmelCase ) # We need to order the input in the way they appears in the forward() lowerCAmelCase_ : List[Any] = OrderedDict({"""input_ids""": common_inputs["""input_ids"""]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch lowerCAmelCase_ , lowerCAmelCase_ : int = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values lowerCAmelCase_ : Optional[Any] = seqlen + 2 lowerCAmelCase_ : Optional[int] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) lowerCAmelCase_ : Optional[int] = [ (torch.zeros(UpperCAmelCase ), torch.zeros(UpperCAmelCase )) for _ in range(self.num_layers ) ] lowerCAmelCase_ : Dict = common_inputs["""attention_mask"""] if self.use_past: lowerCAmelCase_ : Union[str, Any] = ordered_inputs["""attention_mask"""].dtype lowerCAmelCase_ : str = torch.cat( [ordered_inputs["""attention_mask"""], torch.ones(UpperCAmelCase , UpperCAmelCase , dtype=UpperCAmelCase )] , dim=1 ) return ordered_inputs @property def A ( self : Optional[int] ): return 13
28
1
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
221
"""simple docstring""" import random import unittest import numpy as np import torch from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionUpscalePipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class _lowerCAmelCase ( lowercase ,unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = "ssube/stable-diffusion-x4-upscaler-onnx" def _lowercase ( self : Union[str, Any], UpperCAmelCase__ : List[str]=0 ): __lowercase = floats_tensor((1, 3, 1_2_8, 1_2_8), rng=random.Random(UpperCAmelCase__ ) ) __lowercase = torch.manual_seed(UpperCAmelCase__ ) __lowercase = { "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def _lowercase ( self : Any ): __lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider="CPUExecutionProvider" ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) __lowercase = self.get_dummy_inputs() __lowercase = pipe(**UpperCAmelCase__ ).images __lowercase = image[0, -3:, -3:, -1].flatten() # started as 128, should now be 512 assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowercase = np.array( [0.6_974_782, 0.68_902_093, 0.70_135_885, 0.7_583_618, 0.7_804_545, 0.7_854_912, 0.78_667_426, 0.78_743_863, 0.78_070_223] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def _lowercase ( self : Optional[Any] ): __lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider="CPUExecutionProvider" ) __lowercase = PNDMScheduler.from_config(pipe.scheduler.config, skip_prk_steps=UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) __lowercase = self.get_dummy_inputs() __lowercase = pipe(**UpperCAmelCase__ ).images __lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowercase = np.array( [0.6_898_892, 0.59_240_556, 0.52_499_527, 0.58_866_215, 0.52_258_235, 0.52_572_715, 0.62_414_473, 0.6_174_387, 0.6_214_964] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def _lowercase ( self : int ): __lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider="CPUExecutionProvider" ) __lowercase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) __lowercase = self.get_dummy_inputs() __lowercase = pipe(**UpperCAmelCase__ ).images __lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowercase = np.array( [0.7_659_278, 0.76_437_664, 0.75_579_107, 0.7_691_116, 0.77_666_986, 0.7_727_672, 0.7_758_664, 0.7_812_226, 0.76_942_515] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def _lowercase ( self : str ): __lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider="CPUExecutionProvider" ) __lowercase = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) __lowercase = self.get_dummy_inputs() __lowercase = pipe(**UpperCAmelCase__ ).images __lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowercase = np.array( [0.6_974_782, 0.68_902_093, 0.70_135_885, 0.7_583_618, 0.7_804_545, 0.7_854_912, 0.78_667_426, 0.78_743_863, 0.78_070_223] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def _lowercase ( self : Any ): __lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider="CPUExecutionProvider" ) __lowercase = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) __lowercase = self.get_dummy_inputs() __lowercase = pipe(**UpperCAmelCase__ ).images __lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowercase = np.array( [0.77_424_496, 0.773_601, 0.7_645_288, 0.7_769_598, 0.7_772_739, 0.7_738_688, 0.78_187_233, 0.77_879_584, 0.767_043] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @property def _lowercase ( self : Tuple ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def _lowercase ( self : Dict ): __lowercase = ort.SessionOptions() __lowercase = False return options def _lowercase ( self : Dict ): __lowercase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) __lowercase = init_image.resize((1_2_8, 1_2_8) ) # using the PNDM scheduler by default __lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained( "ssube/stable-diffusion-x4-upscaler-onnx", provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) __lowercase = "A fantasy landscape, trending on artstation" __lowercase = torch.manual_seed(0 ) __lowercase = pipe( prompt=UpperCAmelCase__, image=UpperCAmelCase__, guidance_scale=7.5, num_inference_steps=1_0, generator=UpperCAmelCase__, output_type="np", ) __lowercase = output.images __lowercase = images[0, 2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert images.shape == (1, 5_1_2, 5_1_2, 3) __lowercase = np.array([0.4_883, 0.4_947, 0.4_980, 0.4_975, 0.4_982, 0.4_980, 0.5_000, 0.5_006, 0.4_972] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def _lowercase ( self : str ): __lowercase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) __lowercase = init_image.resize((1_2_8, 1_2_8) ) __lowercase = LMSDiscreteScheduler.from_pretrained( "ssube/stable-diffusion-x4-upscaler-onnx", subfolder="scheduler" ) __lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained( "ssube/stable-diffusion-x4-upscaler-onnx", scheduler=UpperCAmelCase__, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) __lowercase = "A fantasy landscape, trending on artstation" __lowercase = torch.manual_seed(0 ) __lowercase = pipe( prompt=UpperCAmelCase__, image=UpperCAmelCase__, guidance_scale=7.5, num_inference_steps=2_0, generator=UpperCAmelCase__, output_type="np", ) __lowercase = output.images __lowercase = images[0, 2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert images.shape == (1, 5_1_2, 5_1_2, 3) __lowercase = np.array( [0.50_173_753, 0.50_223_356, 0.502_039, 0.50_233_036, 0.5_023_725, 0.5_022_601, 0.5_018_758, 0.50_234_085, 0.50_241_566] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
17
0
"""simple docstring""" import jax.numpy as jnp from ...utils import logging from ..ta.modeling_flax_ta import FlaxTaEncoderModel, FlaxTaForConditionalGeneration, FlaxTaModel from .configuration_mta import MTaConfig __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = '''T5Config''' def lowercase (SCREAMING_SNAKE_CASE_ : jnp.array , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int ) -> jnp.ndarray: SCREAMING_SNAKE_CASE = jnp.zeros_like(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = shifted_input_ids.at[:, 1:].set(input_ids[:, :-1] ) SCREAMING_SNAKE_CASE = shifted_input_ids.at[:, 0].set(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = jnp.where(shifted_input_ids == -1_00 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return shifted_input_ids class lowerCAmelCase ( lowerCamelCase_ ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[Any] = """mt5""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = MTaConfig class lowerCAmelCase ( lowerCamelCase_ ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : int = """mt5""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = MTaConfig class lowerCAmelCase ( lowerCamelCase_ ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = """mt5""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = MTaConfig
351
"""simple docstring""" from math import sqrt def lowercase (SCREAMING_SNAKE_CASE_ : int ) -> int: SCREAMING_SNAKE_CASE = 0 for i in range(1 , int(sqrt(SCREAMING_SNAKE_CASE_ ) + 1 ) ): if n % i == 0 and i != sqrt(SCREAMING_SNAKE_CASE_ ): total += i + n // i elif i == sqrt(SCREAMING_SNAKE_CASE_ ): total += i return total - n def lowercase (SCREAMING_SNAKE_CASE_ : int = 1_00_00 ) -> int: SCREAMING_SNAKE_CASE = sum( i for i in range(1 , SCREAMING_SNAKE_CASE_ ) if sum_of_divisors(sum_of_divisors(SCREAMING_SNAKE_CASE_ ) ) == i and sum_of_divisors(SCREAMING_SNAKE_CASE_ ) != i ) return total if __name__ == "__main__": print(solution(int(str(input()).strip())))
38
0
import os import textwrap import pyarrow as pa import pytest from datasets import ClassLabel, Features, Image from datasets.packaged_modules.csv.csv import Csv from ..utils import require_pil @pytest.fixture def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> str: '''simple docstring''' UpperCAmelCase = tmp_path / '''file.csv''' UpperCAmelCase = textwrap.dedent( '''\ header1,header2 1,2 10,20 ''' ) with open(UpperCamelCase__ , '''w''' ) as f: f.write(UpperCamelCase__ ) return str(UpperCamelCase__ ) @pytest.fixture def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> int: '''simple docstring''' UpperCAmelCase = tmp_path / '''malformed_file.csv''' UpperCAmelCase = textwrap.dedent( '''\ header1,header2 1,2 10,20, ''' ) with open(UpperCamelCase__ , '''w''' ) as f: f.write(UpperCamelCase__ ) return str(UpperCamelCase__ ) @pytest.fixture def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ ) -> List[str]: '''simple docstring''' UpperCAmelCase = tmp_path / '''csv_with_image.csv''' UpperCAmelCase = textwrap.dedent( F"""\ image {image_file} """ ) with open(UpperCamelCase__ , '''w''' ) as f: f.write(UpperCamelCase__ ) return str(UpperCamelCase__ ) @pytest.fixture def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase = tmp_path / '''csv_with_label.csv''' UpperCAmelCase = textwrap.dedent( '''\ label good bad good ''' ) with open(UpperCamelCase__ , '''w''' ) as f: f.write(UpperCamelCase__ ) return str(UpperCamelCase__ ) @pytest.fixture def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> Dict: '''simple docstring''' UpperCAmelCase = tmp_path / '''csv_with_int_list.csv''' UpperCAmelCase = textwrap.dedent( '''\ int_list 1 2 3 4 5 6 7 8 9 ''' ) with open(UpperCamelCase__ , '''w''' ) as f: f.write(UpperCamelCase__ ) return str(UpperCamelCase__ ) def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase = Csv() UpperCAmelCase = csv._generate_tables([[csv_file, malformed_csv_file]] ) with pytest.raises(UpperCamelCase__ , match='''Error tokenizing data''' ): for _ in generator: pass assert any( record.levelname == '''ERROR''' and '''Failed to read file''' in record.message and os.path.basename(UpperCamelCase__ ) in record.message for record in caplog.records ) @require_pil def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> Dict: '''simple docstring''' with open(UpperCamelCase__ , encoding='''utf-8''' ) as f: UpperCAmelCase = f.read().splitlines()[1] UpperCAmelCase = Csv(encoding='''utf-8''' , features=Features({'''image''': Image()} ) ) UpperCAmelCase = csv._generate_tables([[csv_file_with_image]] ) UpperCAmelCase = pa.concat_tables([table for _, table in generator] ) assert pa_table.schema.field('''image''' ).type == Image()() UpperCAmelCase = pa_table.to_pydict()['''image'''] assert generated_content == [{"path": image_file, "bytes": None}] def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> Optional[Any]: '''simple docstring''' with open(UpperCamelCase__ , encoding='''utf-8''' ) as f: UpperCAmelCase = f.read().splitlines()[1:] UpperCAmelCase = Csv(encoding='''utf-8''' , features=Features({'''label''': ClassLabel(names=['''good''', '''bad'''] )} ) ) UpperCAmelCase = csv._generate_tables([[csv_file_with_label]] ) UpperCAmelCase = pa.concat_tables([table for _, table in generator] ) assert pa_table.schema.field('''label''' ).type == ClassLabel(names=['''good''', '''bad'''] )() UpperCAmelCase = pa_table.to_pydict()['''label'''] assert generated_content == [ClassLabel(names=['''good''', '''bad'''] ).straint(UpperCamelCase__ ) for label in labels] def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> Dict: '''simple docstring''' UpperCAmelCase = Csv(encoding='''utf-8''' , sep=''',''' , converters={'''int_list''': lambda UpperCamelCase__ : [int(UpperCamelCase__ ) for i in x.split()]} ) UpperCAmelCase = csv._generate_tables([[csv_file_with_int_list]] ) UpperCAmelCase = pa.concat_tables([table for _, table in generator] ) assert pa.types.is_list(pa_table.schema.field('''int_list''' ).type ) UpperCAmelCase = pa_table.to_pydict()['''int_list'''] assert generated_content == [[1, 2, 3], [4, 5, 6], [7, 8, 9]]
273
from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class A_ : UpperCAmelCase__ = MBartConfig UpperCAmelCase__ = {} UpperCAmelCase__ = '''gelu''' def __init__( self , _A , _A=1_3 , _A=7 , _A=True , _A=False , _A=9_9 , _A=3_2 , _A=2 , _A=4 , _A=3_7 , _A=0.1 , _A=0.1 , _A=2_0 , _A=2 , _A=1 , _A=0 , ): '''simple docstring''' UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = eos_token_id UpperCAmelCase = pad_token_id UpperCAmelCase = bos_token_id def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) UpperCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) UpperCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) UpperCAmelCase = prepare_mbart_inputs_dict(_A , _A , _A ) return config, inputs_dict def _lowercase ( self , _A , _A ): '''simple docstring''' UpperCAmelCase = TFMBartModel(config=_A ).get_decoder() UpperCAmelCase = inputs_dict['''input_ids'''] UpperCAmelCase = input_ids[:1, :] UpperCAmelCase = inputs_dict['''attention_mask'''][:1, :] UpperCAmelCase = inputs_dict['''head_mask'''] UpperCAmelCase = 1 # first forward pass UpperCAmelCase = model(_A , attention_mask=_A , head_mask=_A , use_cache=_A ) UpperCAmelCase , UpperCAmelCase = outputs.to_tuple() UpperCAmelCase = past_key_values[1] def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , ) -> List[str]: '''simple docstring''' if attention_mask is None: UpperCAmelCase = tf.cast(tf.math.not_equal(UpperCamelCase__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: UpperCAmelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: UpperCAmelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: UpperCAmelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: UpperCAmelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class A_ (a_ , a_ , unittest.TestCase ): UpperCAmelCase__ = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () UpperCAmelCase__ = (TFMBartForConditionalGeneration,) if is_tf_available() else () UpperCAmelCase__ = ( { '''conversational''': TFMBartForConditionalGeneration, '''feature-extraction''': TFMBartModel, '''summarization''': TFMBartForConditionalGeneration, '''text2text-generation''': TFMBartForConditionalGeneration, '''translation''': TFMBartForConditionalGeneration, } if is_tf_available() else {} ) UpperCAmelCase__ = True UpperCAmelCase__ = False UpperCAmelCase__ = False def _lowercase ( self , _A , _A , _A , _A , _A ): '''simple docstring''' if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = TFMBartModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=_A ) def _lowercase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_A ) @require_sentencepiece @require_tokenizers @require_tf class A_ (unittest.TestCase ): UpperCAmelCase__ = [ ''' UN Chief Says There Is No Military Solution in Syria''', ] UpperCAmelCase__ = [ '''Şeful ONU declară că nu există o soluţie militară în Siria''', ] UpperCAmelCase__ = '''facebook/mbart-large-en-ro''' @cached_property def _lowercase ( self ): '''simple docstring''' return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def _lowercase ( self , **_A ): '''simple docstring''' UpperCAmelCase = self.translate_src_text(**_A ) self.assertListEqual(self.expected_text , _A ) def _lowercase ( self , **_A ): '''simple docstring''' UpperCAmelCase = self.tokenizer(self.src_text , **_A , return_tensors='''tf''' ) UpperCAmelCase = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 ) UpperCAmelCase = self.tokenizer.batch_decode(_A , skip_special_tokens=_A ) return generated_words @slow def _lowercase ( self ): '''simple docstring''' self._assert_generated_batch_equal_expected()
273
1
"""simple docstring""" import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel a : str = HfApi() a : List[str] = {} # fmt: off a : Optional[int] = torch.tensor([ -0.75_15, -1.68_83, 0.24_20, 0.03_00, 0.63_47, 1.34_33, -1.17_43, -3.74_67, 1.23_42, -2.24_85, 0.46_36, 0.80_76, -0.79_91, 0.39_69, 0.84_98, 0.91_89, -1.88_87, -3.35_22, 0.76_39, 0.20_40, 0.62_71, -2.71_48, -1.63_16, 3.08_39, 0.31_86, 0.27_21, -0.97_59, -1.24_61, 2.62_57, 1.35_57 ]) a : Union[str, Any] = torch.tensor([ -2.36_39, -2.53_44, 0.00_54, -0.66_74, 1.59_90, 1.01_58, 0.31_24, -2.14_36, 1.87_95, -2.54_29, -0.15_66, -0.39_73, 1.24_90, 2.64_47, 1.22_83, -0.52_08, -2.81_54, -3.51_19, 2.38_38, 1.20_33, 1.72_01, -2.12_56, -1.45_76, 2.79_48, 2.42_04, -0.97_52, -1.25_46, 0.80_27, 3.27_58, 3.13_65 ]) a : Optional[Any] = torch.tensor([ -0.65_31, -0.68_91, -0.31_72, -0.53_75, -0.91_40, -0.53_67, -0.11_75, -0.78_69, -0.38_08, -0.45_13, -0.20_98, -0.00_83, 0.31_83, 0.51_40, 0.22_47, -0.13_04, -0.13_02, -0.28_02, -0.20_84, -0.20_25, -0.49_67, -0.48_73, -0.08_61, 0.69_25, 0.02_50, 0.12_90, -0.15_43, 0.63_16, 1.04_60, 1.49_43 ]) a : Optional[int] = torch.tensor([ 0.09_11, 0.11_07, 0.01_82, 0.04_35, -0.08_05, -0.06_08, 0.03_81, 0.21_72, -0.02_80, 0.13_27, -0.02_99, -0.02_55, -0.00_50, -0.11_70, -0.10_46, 0.03_09, 0.13_67, 0.17_28, -0.05_33, -0.07_48, -0.05_34, 0.16_24, 0.03_84, -0.18_05, -0.07_07, 0.06_42, 0.02_20, -0.01_34, -0.13_33, -0.15_05 ]) a : Optional[Any] = torch.tensor([ 0.13_21, 0.13_37, 0.04_40, 0.06_22, -0.05_91, -0.03_70, 0.05_03, 0.21_33, -0.01_77, 0.14_15, -0.01_16, -0.01_12, 0.00_44, -0.09_80, -0.07_89, 0.03_95, 0.15_02, 0.17_85, -0.04_88, -0.05_14, -0.04_04, 0.15_39, 0.04_54, -0.15_59, -0.06_65, 0.06_59, 0.03_83, -0.00_05, -0.12_66, -0.13_86 ]) a : int = torch.tensor([ 0.11_54, 0.12_18, 0.03_07, 0.05_26, -0.07_11, -0.05_41, 0.03_66, 0.20_78, -0.02_67, 0.13_17, -0.02_26, -0.01_93, -0.00_14, -0.10_55, -0.09_02, 0.03_30, 0.13_91, 0.17_09, -0.05_62, -0.06_93, -0.05_60, 0.14_82, 0.03_81, -0.16_83, -0.06_81, 0.06_61, 0.03_31, -0.00_46, -0.12_68, -0.14_31 ]) a : Tuple = torch.tensor([ 0.11_92, 0.12_40, 0.04_14, 0.06_06, -0.05_57, -0.04_12, 0.04_30, 0.20_42, -0.02_00, 0.13_85, -0.01_15, -0.01_32, 0.00_17, -0.09_65, -0.08_02, 0.03_98, 0.14_33, 0.17_47, -0.04_58, -0.05_33, -0.04_07, 0.15_45, 0.04_19, -0.15_74, -0.06_45, 0.06_26, 0.03_41, -0.00_10, -0.11_99, -0.13_90 ]) a : str = torch.tensor([ 0.10_75, 0.10_74, 0.02_05, 0.04_31, -0.07_74, -0.06_07, 0.02_98, 0.20_42, -0.03_20, 0.12_67, -0.02_81, -0.02_50, -0.00_64, -0.10_91, -0.09_46, 0.02_90, 0.13_28, 0.16_50, -0.05_80, -0.07_38, -0.05_86, 0.14_40, 0.03_37, -0.17_46, -0.07_12, 0.06_05, 0.02_50, -0.00_99, -0.13_16, -0.14_73 ]) a : int = torch.tensor([ -1.45_72, -2.04_81, -0.04_14, -0.60_05, 1.41_36, 0.58_48, 0.40_28, -2.73_30, 1.22_12, -2.12_28, 0.21_55, 0.40_39, 0.76_62, 2.05_35, 0.74_77, -0.32_43, -2.17_58, -2.76_48, 1.69_47, 0.70_26, 1.23_38, -1.60_78, -0.86_82, 2.28_10, 1.85_74, -0.57_18, -0.55_86, -0.01_86, 2.34_15, 2.12_51]) a : int = torch.tensor([ -1.36_90, -1.97_20, -0.40_90, -0.69_66, 1.46_60, 0.99_38, -0.13_85, -2.73_24, 0.77_36, -1.89_17, 0.29_23, 0.42_93, 0.16_93, 1.41_12, 1.18_87, -0.31_81, -2.21_60, -2.63_81, 1.31_70, 0.81_63, 0.92_40, -1.65_44, -0.60_99, 2.52_59, 1.64_30, -0.90_90, -0.93_92, -0.01_26, 2.42_68, 2.32_66 ]) a : List[Any] = torch.tensor([ -1.35_25, -1.96_28, -0.39_56, -0.68_60, 1.46_64, 1.00_14, -0.12_59, -2.72_12, 0.77_72, -1.88_11, 0.29_96, 0.43_88, 0.17_04, 1.40_29, 1.17_01, -0.30_27, -2.20_53, -2.62_87, 1.33_50, 0.81_31, 0.92_74, -1.62_92, -0.60_98, 2.51_31, 1.65_05, -0.89_58, -0.92_98, -0.01_51, 2.42_57, 2.33_55 ]) a : Optional[Any] = torch.tensor([ -2.05_85, -2.78_97, -0.28_50, -0.89_40, 1.90_52, 0.57_02, 0.63_45, -3.89_59, 1.59_32, -3.23_19, 0.19_74, 0.02_87, 1.75_66, 2.65_43, 0.83_87, -0.53_51, -3.27_36, -4.33_75, 2.90_29, 1.63_90, 1.46_40, -2.17_01, -1.90_13, 2.93_41, 3.49_81, -0.62_55, -1.16_44, -0.15_91, 3.70_97, 3.20_66 ]) a : List[str] = torch.tensor([ -2.31_39, -2.55_94, -0.01_97, -0.67_85, 1.70_01, 1.16_06, 0.30_75, -2.17_40, 1.80_71, -2.56_30, -0.09_26, -0.38_11, 1.21_16, 2.62_46, 1.27_31, -0.53_98, -2.81_53, -3.61_40, 2.38_93, 1.32_62, 1.62_58, -2.18_56, -1.32_67, 2.83_95, 2.37_79, -1.06_23, -1.24_68, 0.89_59, 3.33_67, 3.22_43 ]) a : Tuple = torch.tensor([ -2.06_28, -2.76_67, -0.20_89, -0.82_63, 2.05_39, 0.59_92, 0.64_95, -3.83_36, 1.60_25, -3.28_17, 0.17_21, -0.06_33, 1.75_16, 2.70_39, 0.81_00, -0.59_08, -3.21_13, -4.43_43, 2.92_57, 1.36_32, 1.55_62, -2.14_89, -1.98_94, 3.05_60, 3.33_96, -0.73_28, -1.04_17, 0.03_83, 3.70_93, 3.23_43 ]) a : List[Any] = torch.tensor([ -1.45_74, -2.05_69, -0.04_73, -0.61_17, 1.40_18, 0.57_69, 0.41_29, -2.73_44, 1.22_41, -2.13_97, 0.20_00, 0.39_37, 0.76_16, 2.04_53, 0.73_24, -0.33_91, -2.17_46, -2.77_44, 1.69_63, 0.69_21, 1.21_87, -1.61_72, -0.88_77, 2.24_39, 1.84_71, -0.58_39, -0.56_05, -0.04_64, 2.32_50, 2.12_19 ]) # fmt: on a : Dict = api.list_models(filter="""diffusers""") for mod in models: if "google" in mod.author or mod.modelId == "CompVis/ldm-celebahq-256": a : Tuple = """/home/patrick/google_checkpoints/""" + mod.modelId.split("""/""")[-1] print(F"""Started running {mod.modelId}!!!""") if mod.modelId.startswith("""CompVis"""): a : Any = UNetaDModel.from_pretrained(local_checkpoint, subfolder="""unet""") else: a : Dict = UNetaDModel.from_pretrained(local_checkpoint) torch.manual_seed(0) random.seed(0) a : List[Any] = torch.randn(1, model.config.in_channels, model.config.sample_size, model.config.sample_size) a : Dict = torch.tensor([10] * noise.shape[0]) with torch.no_grad(): a : Any = model(noise, time_step).sample assert torch.allclose( logits[0, 0, 0, :30], results["""_""".join("""_""".join(mod.modelId.split("""/""")).split("""-"""))], atol=1e-3 ) print(F"""{mod.modelId} has passed successfully!!!""")
150
"""simple docstring""" import inspect import jax import jax.lax as lax import jax.numpy as jnp from ..utils import add_start_docstrings from ..utils.logging import get_logger a : Any = get_logger(__name__) a : Any = r""" Args: input_ids (`jnp.ndarray` of shape `(batch_size, sequence_length)`): Indices of input sequence tokens in the vocabulary. Indices can be obtained using [`PreTrainedTokenizer`]. See [`PreTrainedTokenizer.encode`] and [`PreTrainedTokenizer.__call__`] for details. [What are input IDs?](../glossary#input-ids) scores (`jnp.ndarray` of shape `(batch_size, config.vocab_size)`): Prediction scores of a language modeling head. These can be logits for each vocabulary when not using beam search or log softmax for each vocabulary token when using beam search kwargs (`Dict[str, Any]`, *optional*): Additional logits processor specific kwargs. Return: `jnp.ndarray` of shape `(batch_size, config.vocab_size)`: The processed prediction scores. """ class __UpperCAmelCase: """simple docstring""" @add_start_docstrings(snake_case__ ) def __call__( self , snake_case__ , snake_case__ ): '''simple docstring''' raise NotImplementedError( F'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) class __UpperCAmelCase: """simple docstring""" @add_start_docstrings(snake_case__ ) def __call__( self , snake_case__ , snake_case__ ): '''simple docstring''' raise NotImplementedError( F'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) class __UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): """simple docstring""" @add_start_docstrings(snake_case__ ) def __call__( self , snake_case__ , snake_case__ , snake_case__ , **snake_case__ ): '''simple docstring''' for processor in self: lowercase__ : Optional[Any]= inspect.signature(processor.__call__ ).parameters if len(snake_case__ ) > 3: if not all(arg in kwargs for arg in list(function_args.keys() )[2:] ): raise ValueError( F'''Make sure that all the required parameters: {list(function_args.keys() )} for ''' F'''{processor.__class__} are passed to the logits processor.''' ) lowercase__ : Union[str, Any]= processor(snake_case__ , snake_case__ , snake_case__ , **snake_case__ ) else: lowercase__ : Dict= processor(snake_case__ , snake_case__ , snake_case__ ) return scores class __UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , snake_case__ ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ) or not (temperature > 0): raise ValueError(F'''`temperature` has to be a strictly positive float, but is {temperature}''' ) lowercase__ : Any= temperature def __call__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' lowercase__ : int= scores / self.temperature return scores class __UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = -float("Inf" ) , snake_case__ = 1 ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ) or (top_p < 0 or top_p > 1.0): raise ValueError(F'''`top_p` has to be a float > 0 and < 1, but is {top_p}''' ) if not isinstance(snake_case__ , snake_case__ ) or (min_tokens_to_keep < 1): raise ValueError(F'''`min_tokens_to_keep` has to be a positive integer, but is {min_tokens_to_keep}''' ) lowercase__ : int= top_p lowercase__ : Optional[int]= filter_value lowercase__ : Tuple= min_tokens_to_keep def __call__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' lowercase__, lowercase__ : Dict= lax.top_k(snake_case__ , scores.shape[-1] ) lowercase__ : Optional[int]= jnp.full_like(snake_case__ , self.filter_value ) lowercase__ : Union[str, Any]= jax.nn.softmax(snake_case__ , axis=-1 ).cumsum(axis=-1 ) lowercase__ : str= cumulative_probs < self.top_p # include the token that is higher than top_p as well lowercase__ : str= jnp.roll(snake_case__ , 1 ) score_mask |= score_mask.at[:, 0].set(snake_case__ ) # min tokens to keep lowercase__ : Optional[int]= score_mask.at[:, : self.min_tokens_to_keep].set(snake_case__ ) lowercase__ : str= jnp.where(snake_case__ , snake_case__ , snake_case__ ) lowercase__ : str= jax.lax.sort_key_val(snake_case__ , snake_case__ )[-1] return next_scores class __UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ = -float("Inf" ) , snake_case__ = 1 ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ) or top_k <= 0: raise ValueError(F'''`top_k` has to be a strictly positive integer, but is {top_k}''' ) lowercase__ : List[Any]= max(snake_case__ , snake_case__ ) lowercase__ : Dict= filter_value def __call__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' lowercase__, lowercase__ : Optional[Any]= scores.shape lowercase__ : int= jnp.full(batch_size * vocab_size , self.filter_value ) lowercase__ : Dict= min(self.top_k , scores.shape[-1] ) # Safety check lowercase__, lowercase__ : List[Any]= lax.top_k(snake_case__ , snake_case__ ) lowercase__ : Optional[int]= jnp.broadcast_to((jnp.arange(snake_case__ ) * vocab_size)[:, None] , (batch_size, topk) ).flatten() lowercase__ : str= topk_scores.flatten() lowercase__ : Any= topk_indices.flatten() + shift lowercase__ : Optional[Any]= next_scores_flat.at[topk_indices_flat].set(snake_case__ ) lowercase__ : str= next_scores_flat.reshape(snake_case__ , snake_case__ ) return next_scores class __UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , snake_case__ ): '''simple docstring''' lowercase__ : Any= bos_token_id def __call__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' lowercase__ : Any= jnp.full(scores.shape , -float("inf" ) ) lowercase__ : int= 1 - jnp.bool_(cur_len - 1 ) lowercase__ : int= jnp.where(snake_case__ , new_scores.at[:, self.bos_token_id].set(0 ) , snake_case__ ) return scores class __UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ ): '''simple docstring''' lowercase__ : Tuple= max_length lowercase__ : str= eos_token_id def __call__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' lowercase__ : List[Any]= jnp.full(scores.shape , -float("inf" ) ) lowercase__ : Any= 1 - jnp.bool_(cur_len - self.max_length + 1 ) lowercase__ : Optional[int]= jnp.where(snake_case__ , new_scores.at[:, self.eos_token_id].set(0 ) , snake_case__ ) return scores class __UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ) or min_length < 0: raise ValueError(F'''`min_length` has to be a positive integer, but is {min_length}''' ) if not isinstance(snake_case__ , snake_case__ ) or eos_token_id < 0: raise ValueError(F'''`eos_token_id` has to be a positive integer, but is {eos_token_id}''' ) lowercase__ : List[str]= min_length lowercase__ : Dict= eos_token_id def __call__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' # create boolean flag to decide if min length penalty should be applied lowercase__ : Tuple= 1 - jnp.clip(cur_len - self.min_length , 0 , 1 ) lowercase__ : Dict= jnp.where(snake_case__ , scores.at[:, self.eos_token_id].set(-float("inf" ) ) , snake_case__ ) return scores class __UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ ): '''simple docstring''' lowercase__ : Optional[Any]= list(snake_case__ ) lowercase__ : List[Any]= begin_index def __call__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' lowercase__ : str= 1 - jnp.bool_(cur_len - self.begin_index ) lowercase__ : str= jnp.where(snake_case__ , scores.at[:, self.begin_suppress_tokens].set(-float("inf" ) ) , snake_case__ ) return scores class __UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , snake_case__ ): '''simple docstring''' lowercase__ : List[Any]= list(snake_case__ ) def __call__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' lowercase__ : Any= scores.at[..., self.suppress_tokens].set(-float("inf" ) ) return scores class __UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , snake_case__ ): '''simple docstring''' lowercase__ : int= dict(snake_case__ ) # Converts the dictionary of format {index: token} containing the tokens to be forced to an array, where the # index of the array corresponds to the index of the token to be forced, for XLA compatibility. # Indexes without forced tokens will have a negative value. lowercase__ : List[Any]= jnp.ones((max(force_token_map.keys() ) + 1) , dtype=jnp.intaa ) * -1 for index, token in force_token_map.items(): if token is not None: lowercase__ : List[Any]= force_token_array.at[index].set(snake_case__ ) lowercase__ : int= jnp.intaa(snake_case__ ) def __call__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' def _force_token(snake_case__ ): lowercase__ : Dict= scores.shape[0] lowercase__ : Any= self.force_token_array[generation_idx] lowercase__ : List[Any]= jnp.ones_like(snake_case__ , dtype=scores.dtype ) * -float("inf" ) lowercase__ : List[Any]= jnp.zeros((batch_size, 1) , dtype=scores.dtype ) lowercase__ : List[str]= lax.dynamic_update_slice(snake_case__ , snake_case__ , (0, current_token) ) return new_scores lowercase__ : Dict= lax.cond( cur_len >= self.force_token_array.shape[0] , lambda: scores , lambda: lax.cond( self.force_token_array[cur_len] >= 0 , lambda: _force_token(snake_case__ ) , lambda: scores , ) , ) return scores class __UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' lowercase__ : str= generate_config.eos_token_id lowercase__ : Optional[int]= generate_config.no_timestamps_token_id lowercase__ : Dict= generate_config.no_timestamps_token_id + 1 lowercase__ : List[Any]= decoder_input_length + 1 if generate_config.is_multilingual: # room for language token and task token self.begin_index += 2 if hasattr(snake_case__ , "max_initial_timestamp_index" ): lowercase__ : int= generate_config.max_initial_timestamp_index else: lowercase__ : Dict= model_config.vocab_size if self.max_initial_timestamp_index is None: lowercase__ : str= model_config.vocab_size def __call__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' # suppress <|notimestamps|> which is handled by without_timestamps lowercase__ : int= scores.at[:, self.no_timestamps_token_id].set(-float("inf" ) ) def handle_pairs(snake_case__ , snake_case__ ): lowercase__ : Union[str, Any]= jnp.where((cur_len - self.begin_index) >= 1 , snake_case__ , snake_case__ ) lowercase__ : Tuple= jnp.where( input_ids_k[cur_len - 1] >= self.timestamp_begin , True and last_was_timestamp , snake_case__ , ) lowercase__ : int= jnp.where((cur_len - self.begin_index) < 2 , snake_case__ , snake_case__ ) lowercase__ : Optional[int]= jnp.where( input_ids_k[cur_len - 2] >= self.timestamp_begin , snake_case__ , snake_case__ , ) return jnp.where( snake_case__ , jnp.where( penultimate_was_timestamp > 0 , scores_k.at[self.timestamp_begin :].set(-float("inf" ) ) , scores_k.at[: self.eos_token_id].set(-float("inf" ) ) , ) , snake_case__ , ) lowercase__ : List[str]= jax.vmap(snake_case__ )(snake_case__ , snake_case__ ) lowercase__ : str= jnp.where(cur_len == self.begin_index , snake_case__ , snake_case__ ) lowercase__ : List[Any]= jnp.where( self.max_initial_timestamp_index is not None , True and apply_max_initial_timestamp , snake_case__ , ) lowercase__ : Any= self.timestamp_begin + self.max_initial_timestamp_index lowercase__ : str= jnp.where( snake_case__ , scores.at[:, last_allowed + 1 :].set(-float("inf" ) ) , snake_case__ , ) # if sum of probability over timestamps is above any other token, sample timestamp lowercase__ : str= jax.nn.log_softmax(snake_case__ , axis=-1 ) def handle_cumulative_probs(snake_case__ , snake_case__ ): lowercase__ : Dict= jax.nn.logsumexp(logprobs_k[self.timestamp_begin :] , axis=-1 ) lowercase__ : Union[str, Any]= jnp.max(logprobs_k[: self.timestamp_begin] ) return jnp.where( timestamp_logprob > max_text_token_logprob , scores_k.at[: self.timestamp_begin].set(-float("inf" ) ) , snake_case__ , ) lowercase__ : Optional[int]= jax.vmap(snake_case__ )(snake_case__ , snake_case__ ) return scores
150
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) A_ : Tuple = {"configuration_vit_mae": ["VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP", "ViTMAEConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Any = [ "VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST", "ViTMAEForPreTraining", "ViTMAELayer", "ViTMAEModel", "ViTMAEPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Dict = [ "TFViTMAEForPreTraining", "TFViTMAEModel", "TFViTMAEPreTrainedModel", ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys A_ : Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
165
"""simple docstring""" from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class _a ( lowerCAmelCase): """simple docstring""" def lowercase__ ( self : List[Any] , __UpperCamelCase : float )->float: return 0.0 def lowercase ( _SCREAMING_SNAKE_CASE : np.ndarray , _SCREAMING_SNAKE_CASE : int ): '''simple docstring''' _UpperCAmelCase = min([-20, np.min(fft_results[1 : samplerate // 2 - 1] )] ) _UpperCAmelCase = max([20, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def lowercase ( _SCREAMING_SNAKE_CASE : FilterType , _SCREAMING_SNAKE_CASE : int ): '''simple docstring''' _UpperCAmelCase = 512 _UpperCAmelCase = [1] + [0] * (size - 1) _UpperCAmelCase = [filter_type.process(_SCREAMING_SNAKE_CASE ) for item in inputs] _UpperCAmelCase = [0] * (samplerate - size) # zero-padding outputs += filler _UpperCAmelCase = np.abs(np.fft.fft(_SCREAMING_SNAKE_CASE ) ) _UpperCAmelCase = 20 * np.logaa(_SCREAMING_SNAKE_CASE ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('''Frequency (Hz)''' ) plt.xscale('''log''' ) # Display within reasonable bounds _UpperCAmelCase = get_bounds(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) plt.ylim(max([-80, bounds[0]] ) , min([80, bounds[1]] ) ) plt.ylabel('''Gain (dB)''' ) plt.plot(_SCREAMING_SNAKE_CASE ) plt.show() def lowercase ( _SCREAMING_SNAKE_CASE : FilterType , _SCREAMING_SNAKE_CASE : int ): '''simple docstring''' _UpperCAmelCase = 512 _UpperCAmelCase = [1] + [0] * (size - 1) _UpperCAmelCase = [filter_type.process(_SCREAMING_SNAKE_CASE ) for item in inputs] _UpperCAmelCase = [0] * (samplerate - size) # zero-padding outputs += filler _UpperCAmelCase = np.angle(np.fft.fft(_SCREAMING_SNAKE_CASE ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('''Frequency (Hz)''' ) plt.xscale('''log''' ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel('''Phase shift (Radians)''' ) plt.plot(np.unwrap(_SCREAMING_SNAKE_CASE , -2 * pi ) ) plt.show()
260
0
from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging __snake_case : List[Any] = logging.get_logger(__name__) def _UpperCamelCase ( UpperCamelCase_ : Union[tf.Tensor, np.ndarray] ) -> List[int]: """simple docstring""" if isinstance(UpperCamelCase_ , np.ndarray ): return list(tensor.shape ) lowerCAmelCase__ = tf.shape(UpperCamelCase_ ) if tensor.shape == tf.TensorShape(UpperCamelCase_ ): return dynamic lowerCAmelCase__ = tensor.shape.as_list() return [dynamic[i] if s is None else s for i, s in enumerate(UpperCamelCase_ )] def _UpperCamelCase ( UpperCamelCase_ : tf.Tensor , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[str] = None ) -> tf.Tensor: """simple docstring""" return tf.nn.softmax(logits=logits + 1e-9 , axis=UpperCamelCase_ , name=UpperCamelCase_ ) def _UpperCamelCase ( UpperCamelCase_ : Tuple , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Tuple=1e-5 , UpperCamelCase_ : Union[str, Any]=-1 ) -> Union[str, Any]: """simple docstring""" if weight.shape.rank != 1 or bias.shape.rank != 1 or not isinstance(UpperCamelCase_ , UpperCamelCase_ ): raise NotImplementedError('Only 1D weight and bias tensors are supported for now, with only a single axis.' ) # Get mean and variance on the axis to be normalized lowerCAmelCase__ , lowerCAmelCase__ = tf.nn.moments(UpperCamelCase_ , axes=[axis] , keepdims=UpperCamelCase_ ) if axis != -1: # Reshape scale and weight to have the same rank as inputs, but with 1 dimensions # on every dimension except axis lowerCAmelCase__ = [1] * inputs.shape.rank lowerCAmelCase__ = shape_list(UpperCamelCase_ )[axis] lowerCAmelCase__ = tf.reshape(UpperCamelCase_ , UpperCamelCase_ ) lowerCAmelCase__ = tf.reshape(UpperCamelCase_ , UpperCamelCase_ ) # Compute layer normalization using the batch_normalization # function. lowerCAmelCase__ = tf.nn.batch_normalization( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , offset=UpperCamelCase_ , scale=UpperCamelCase_ , variance_epsilon=UpperCamelCase_ , ) return outputs def _UpperCamelCase ( UpperCamelCase_ : Any , UpperCamelCase_ : Optional[int]=0 , UpperCamelCase_ : List[str]=-1 ) -> List[str]: """simple docstring""" if end_dim < 0: end_dim += input.shape.rank if start_dim < 0: start_dim += input.shape.rank if start_dim == end_dim: return input lowerCAmelCase__ = tf.shape(UpperCamelCase_ ) lowerCAmelCase__ = tf.math.reduce_prod(in_shape[start_dim : end_dim + 1] ) lowerCAmelCase__ = tf.concat([in_shape[:start_dim], [flattened_dim], in_shape[end_dim + 1 :]] , axis=0 ) return tf.reshape(UpperCamelCase_ , UpperCamelCase_ ) def _UpperCamelCase ( UpperCamelCase_ : tf.Tensor ) -> tf.Tensor: """simple docstring""" if not isinstance(UpperCamelCase_ , tf.Tensor ): lowerCAmelCase__ = tf.convert_to_tensor(UpperCamelCase_ ) # Catches stray NumPy inputs if encoder_attention_mask.shape.rank == 3: lowerCAmelCase__ = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.shape.rank == 2: lowerCAmelCase__ = encoder_attention_mask[:, None, None, :] # T5 has a mask that can compare sequence ids, we can simulate this here with this transposition # Cf. https://github.com/tensorflow/mesh/blob/8d2465e9bc93129b913b5ccc6a59aa97abd96ec6/mesh_tensorflow # /transformer/transformer_layers.py#L270 # encoder_extended_attention_mask = (encoder_extended_attention_mask == # encoder_extended_attention_mask.transpose(-1, -2)) lowerCAmelCase__ = ( tf.cast(1 , encoder_attention_mask.dtype ) - encoder_extended_attention_mask ) * encoder_extended_attention_mask.dtype.min return encoder_extended_attention_mask def _UpperCamelCase ( UpperCamelCase_ : tf.Tensor , UpperCamelCase_ : int , UpperCamelCase_ : str = "input_ids" ) -> None: """simple docstring""" tf.debugging.assert_less( UpperCamelCase_ , tf.cast(UpperCamelCase_ , dtype=tensor.dtype ) , message=( F"The maximum value of {tensor_name} ({tf.math.reduce_max(UpperCamelCase_ )}) must be smaller than the embedding " F"layer's input dimension ({embed_dim}). The likely cause is some problem at tokenization time." ) , ) def _UpperCamelCase ( UpperCamelCase_ : List[Any] , UpperCamelCase_ : List[str] , UpperCamelCase_ : str ) -> List[str]: """simple docstring""" lowerCAmelCase__ = 6_4512 # Check that no item in `data` is larger than `HDF5_OBJECT_HEADER_LIMIT` # because in that case even chunking the array would not make the saving # possible. lowerCAmelCase__ = [x for x in data if len(UpperCamelCase_ ) > HDF5_OBJECT_HEADER_LIMIT] # Expecting this to never be true. if bad_attributes: raise RuntimeError( 'The following attributes cannot be saved to HDF5 file because ' F"they are larger than {HDF5_OBJECT_HEADER_LIMIT} " F"bytes: {bad_attributes}" ) lowerCAmelCase__ = np.asarray(UpperCamelCase_ ) lowerCAmelCase__ = 1 lowerCAmelCase__ = np.array_split(UpperCamelCase_ , UpperCamelCase_ ) # This will never loop forever thanks to the test above. while any(x.nbytes > HDF5_OBJECT_HEADER_LIMIT for x in chunked_data ): num_chunks += 1 lowerCAmelCase__ = np.array_split(UpperCamelCase_ , UpperCamelCase_ ) if num_chunks > 1: for chunk_id, chunk_data in enumerate(UpperCamelCase_ ): lowerCAmelCase__ = chunk_data else: lowerCAmelCase__ = data def _UpperCamelCase ( UpperCamelCase_ : List[Any] , UpperCamelCase_ : int ) -> Any: """simple docstring""" if name in group.attrs: lowerCAmelCase__ = [n.decode('utf8' ) if hasattr(UpperCamelCase_ , 'decode' ) else n for n in group.attrs[name]] else: lowerCAmelCase__ = [] lowerCAmelCase__ = 0 while "%s%d" % (name, chunk_id) in group.attrs: data.extend( [n.decode('utf8' ) if hasattr(UpperCamelCase_ , 'decode' ) else n for n in group.attrs['%s%d' % (name, chunk_id)]] ) chunk_id += 1 return data def _UpperCamelCase ( UpperCamelCase_ : Union[str, Any] ) -> str: """simple docstring""" def _expand_single_ad_tensor(UpperCamelCase_ : Union[str, Any] ): if isinstance(UpperCamelCase_ , tf.Tensor ) and t.shape.rank == 1: return tf.expand_dims(UpperCamelCase_ , axis=-1 ) return t return tf.nest.map_structure(_expand_single_ad_tensor , UpperCamelCase_ )
350
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __snake_case : Any = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( __lowercase): _SCREAMING_SNAKE_CASE : List[Any] = ['''pixel_values'''] def __init__( self , _UpperCamelCase = True , _UpperCamelCase = None , _UpperCamelCase = PILImageResampling.BILINEAR , _UpperCamelCase = True , _UpperCamelCase = None , _UpperCamelCase = True , _UpperCamelCase = 1 / 2_55 , _UpperCamelCase = True , _UpperCamelCase = None , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" super().__init__(**_UpperCamelCase ) lowerCAmelCase__ = size if size is not None else {'shortest_edge': 2_56} lowerCAmelCase__ = get_size_dict(_UpperCamelCase , default_to_square=_UpperCamelCase ) lowerCAmelCase__ = crop_size if crop_size is not None else {'height': 2_24, 'width': 2_24} lowerCAmelCase__ = get_size_dict(_UpperCamelCase ) lowerCAmelCase__ = do_resize lowerCAmelCase__ = size lowerCAmelCase__ = resample lowerCAmelCase__ = do_center_crop lowerCAmelCase__ = crop_size lowerCAmelCase__ = do_rescale lowerCAmelCase__ = rescale_factor lowerCAmelCase__ = do_normalize lowerCAmelCase__ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCAmelCase__ = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = PILImageResampling.BICUBIC , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" lowerCAmelCase__ = get_size_dict(_UpperCamelCase , default_to_square=_UpperCamelCase ) if "shortest_edge" not in size: raise ValueError(F"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) lowerCAmelCase__ = get_resize_output_image_size(_UpperCamelCase , size=size['shortest_edge'] , default_to_square=_UpperCamelCase ) return resize(_UpperCamelCase , size=_UpperCamelCase , resample=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" lowerCAmelCase__ = get_size_dict(_UpperCamelCase ) return center_crop(_UpperCamelCase , size=(size['height'], size['width']) , data_format=_UpperCamelCase , **_UpperCamelCase ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = None , **_UpperCamelCase ): """simple docstring""" return rescale(_UpperCamelCase , scale=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" return normalize(_UpperCamelCase , mean=_UpperCamelCase , std=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = ChannelDimension.FIRST , **_UpperCamelCase , ): """simple docstring""" lowerCAmelCase__ = do_resize if do_resize is not None else self.do_resize lowerCAmelCase__ = size if size is not None else self.size lowerCAmelCase__ = get_size_dict(_UpperCamelCase , default_to_square=_UpperCamelCase ) lowerCAmelCase__ = resample if resample is not None else self.resample lowerCAmelCase__ = do_center_crop if do_center_crop is not None else self.do_center_crop lowerCAmelCase__ = crop_size if crop_size is not None else self.crop_size lowerCAmelCase__ = get_size_dict(_UpperCamelCase ) lowerCAmelCase__ = do_rescale if do_rescale is not None else self.do_rescale lowerCAmelCase__ = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCAmelCase__ = do_normalize if do_normalize is not None else self.do_normalize lowerCAmelCase__ = image_mean if image_mean is not None else self.image_mean lowerCAmelCase__ = image_std if image_std is not None else self.image_std lowerCAmelCase__ = make_list_of_images(_UpperCamelCase ) if not valid_images(_UpperCamelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. lowerCAmelCase__ = [to_numpy_array(_UpperCamelCase ) for image in images] if do_resize: lowerCAmelCase__ = [self.resize(image=_UpperCamelCase , size=_UpperCamelCase , resample=_UpperCamelCase ) for image in images] if do_center_crop: lowerCAmelCase__ = [self.center_crop(image=_UpperCamelCase , size=_UpperCamelCase ) for image in images] if do_rescale: lowerCAmelCase__ = [self.rescale(image=_UpperCamelCase , scale=_UpperCamelCase ) for image in images] if do_normalize: lowerCAmelCase__ = [self.normalize(image=_UpperCamelCase , mean=_UpperCamelCase , std=_UpperCamelCase ) for image in images] lowerCAmelCase__ = [to_channel_dimension_format(_UpperCamelCase , _UpperCamelCase ) for image in images] lowerCAmelCase__ = {'pixel_values': images} return BatchFeature(data=_UpperCamelCase , tensor_type=_UpperCamelCase )
122
0
'''simple docstring''' from __future__ import annotations def __A ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = list(range(len(__a ) ) ) SCREAMING_SNAKE_CASE : Optional[Any] = [v / w for v, w in zip(__a , __a )] index.sort(key=lambda lowerCamelCase_ : ratio[i] , reverse=__a ) SCREAMING_SNAKE_CASE : float = 0 SCREAMING_SNAKE_CASE : list[float] = [0] * len(__a ) for i in index: if weight[i] <= capacity: SCREAMING_SNAKE_CASE : Optional[int] = 1 max_value += value[i] capacity -= weight[i] else: SCREAMING_SNAKE_CASE : Optional[int] = capacity / weight[i] max_value += value[i] * capacity / weight[i] break return max_value, fractions if __name__ == "__main__": import doctest doctest.testmod()
323
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __A = {"configuration_ibert": ["IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "IBertConfig", "IBertOnnxConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "IBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "IBertForMaskedLM", "IBertForMultipleChoice", "IBertForQuestionAnswering", "IBertForSequenceClassification", "IBertForTokenClassification", "IBertModel", "IBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
10
0
import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def __A ( _lowercase ): '''simple docstring''' _A = checkpoints.load_tax_checkpoint(_lowercase ) _A = flatten_dict(_lowercase ) return flax_params def __A ( _lowercase ): '''simple docstring''' _A = {} _A = { '''token_embedder''': '''embeddings''', '''encoder_norm''': '''layernorm''', '''kernel''': '''weight''', '''.out''': '''.output''', '''scale''': '''weight''', '''embedders_0.pos_embedding''': '''row_embedder.weight''', '''embedders_1.pos_embedding''': '''column_embedder.weight''', } _A = { '''query''': '''attention.query''', '''key''': '''attention.key''', '''value''': '''attention.value''', '''output.dense''': '''output''', '''encoder_decoder_attention.o''': '''encoder_decoder_attention.attention.o''', '''pre_self_attention_layer_norm''': '''self_attention.layer_norm''', '''pre_cross_attention_layer_norm''': '''encoder_decoder_attention.layer_norm''', '''mlp.''': '''mlp.DenseReluDense.''', '''pre_mlp_layer_norm''': '''mlp.layer_norm''', '''self_attention.o''': '''self_attention.attention.o''', '''decoder.embeddings.embedding''': '''decoder.embed_tokens.weight''', '''decoder.relpos_bias.rel_embedding''': '''decoder.layer.0.self_attention.attention.relative_attention_bias.weight''', '''decoder.decoder_norm.weight''': '''decoder.final_layer_norm.weight''', '''decoder.logits_dense.weight''': '''decoder.lm_head.weight''', } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key _A = '''.'''.join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): _A = new_key.replace(_lowercase , _lowercase ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): _A = new_key.replace(_lowercase , _lowercase ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number _A = re.sub(R'''layers_(\d+)''' , R'''layer.\1''' , _lowercase ) _A = new_key.replace('''encoder''' , '''encoder.encoder''' ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number _A = re.sub(R'''layers_(\d+)''' , R'''layer.\1''' , _lowercase ) _A = flax_dict[key] _A = {} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): _A = torch.from_numpy(converted_dict[key].T ) else: _A = torch.from_numpy(converted_dict[key] ) return converted_torch_dict def __A ( _lowercase , _lowercase , _lowercase=False , _lowercase=False ): '''simple docstring''' _A = get_flax_param(_lowercase ) if not use_large: _A = PixaStructVisionConfig() _A = PixaStructTextConfig() else: _A = PixaStructVisionConfig( hidden_size=15_36 , d_ff=39_68 , num_attention_heads=24 , num_hidden_layers=18 ) _A = PixaStructTextConfig(hidden_size=15_36 , d_ff=39_68 , num_heads=24 , num_layers=18 ) _A = PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=_lowercase ) _A = PixaStructForConditionalGeneration(_lowercase ) _A = rename_and_convert_flax_params(_lowercase ) model.load_state_dict(_lowercase ) _A = AutoTokenizer.from_pretrained('''ybelkada/test-pix2struct-tokenizer''' ) _A = PixaStructImageProcessor() _A = PixaStructProcessor(image_processor=_lowercase , tokenizer=_lowercase ) if use_large: _A = 40_96 _A = True # mkdir if needed os.makedirs(_lowercase , exist_ok=_lowercase ) model.save_pretrained(_lowercase ) processor.save_pretrained(_lowercase ) print('''Model saved in {}'''.format(_lowercase ) ) if __name__ == "__main__": __A = argparse.ArgumentParser() parser.add_argument('--t5x_checkpoint_path', default=None, type=str, help='Path to the original T5x checkpoint.') parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--use_large', action='store_true', help='Use large model.') parser.add_argument('--is_vqa', action='store_true', help='Use large model.') __A = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
75
from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class SCREAMING_SNAKE_CASE ( snake_case ): """simple docstring""" A_ = 42 A_ = 42 A_ = None class SCREAMING_SNAKE_CASE ( snake_case , snake_case ): """simple docstring""" A_ = 2 @register_to_config def __init__( self: List[str] , __A: float = 0.02 , __A: float = 1_00 , __A: float = 1.007 , __A: float = 80 , __A: float = 0.05 , __A: float = 50 , ) -> Optional[int]: # standard deviation of the initial noise distribution _A = sigma_max # setable values _A = None _A = None _A = None # sigma(t_i) def __A ( self: Any , __A: torch.FloatTensor , __A: Optional[int] = None ) -> torch.FloatTensor: return sample def __A ( self: Union[str, Any] , __A: int , __A: Union[str, torch.device] = None ) -> List[Any]: _A = num_inference_steps _A = np.arange(0 , self.num_inference_steps )[::-1].copy() _A = torch.from_numpy(__A ).to(__A ) _A = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] _A = torch.tensor(__A , dtype=torch.floataa , device=__A ) def __A ( self: List[Any] , __A: torch.FloatTensor , __A: float , __A: Optional[torch.Generator] = None ) -> Tuple[torch.FloatTensor, float]: if self.config.s_min <= sigma <= self.config.s_max: _A = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: _A = 0 # sample eps ~ N(0, S_noise^2 * I) _A = self.config.s_noise * randn_tensor(sample.shape , generator=__A ).to(sample.device ) _A = sigma + gamma * sigma _A = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __A ( self: Optional[Any] , __A: torch.FloatTensor , __A: float , __A: float , __A: torch.FloatTensor , __A: bool = True , ) -> Union[KarrasVeOutput, Tuple]: _A = sample_hat + sigma_hat * model_output _A = (sample_hat - pred_original_sample) / sigma_hat _A = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=__A , derivative=__A , pred_original_sample=__A ) def __A ( self: Dict , __A: torch.FloatTensor , __A: float , __A: float , __A: torch.FloatTensor , __A: torch.FloatTensor , __A: torch.FloatTensor , __A: bool = True , ) -> Union[KarrasVeOutput, Tuple]: _A = sample_prev + sigma_prev * model_output _A = (sample_prev - pred_original_sample) / sigma_prev _A = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=__A , derivative=__A , pred_original_sample=__A ) def __A ( self: Any , __A: List[Any] , __A: int , __A: int ) -> int: raise NotImplementedError()
75
1
def lowerCAmelCase_ ( __a , __a , __a , __a ) -> int: """simple docstring""" lowerCamelCase__ , lowerCamelCase__: int =len(__a ), len(grid[0] ) if ( min(__a , __a ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) lowerCamelCase__: Dict =0 count += depth_first_search(__a , row + 1 , __a , __a ) count += depth_first_search(__a , row - 1 , __a , __a ) count += depth_first_search(__a , __a , col + 1 , __a ) count += depth_first_search(__a , __a , col - 1 , __a ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
10
import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin __A = get_tests_dir("fixtures/test_sentencepiece_bpe.model") class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' lowercase_ = BartphoTokenizer lowercase_ = False lowercase_ = True def SCREAMING_SNAKE_CASE_ (self : List[Any]) ->Tuple: '''simple docstring''' super().setUp() lowerCamelCase__: int =["▁This", "▁is", "▁a", "▁t", "est"] lowerCamelCase__: Tuple =dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_)))) lowerCamelCase__: List[Any] ={"unk_token": "<unk>"} lowerCamelCase__: Dict =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["monolingual_vocab_file"]) with open(self.monolingual_vocab_file , "w" , encoding="utf-8") as fp: for token in vocab_tokens: fp.write(F"""{token} {vocab_tokens[token]}\n""") lowerCamelCase__: Dict =BartphoTokenizer(UpperCAmelCase_ , self.monolingual_vocab_file , **self.special_tokens_map) tokenizer.save_pretrained(self.tmpdirname) def SCREAMING_SNAKE_CASE_ (self : Optional[int] , **UpperCAmelCase_ : Optional[Any]) ->str: '''simple docstring''' kwargs.update(self.special_tokens_map) return BartphoTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : Union[str, Any] , UpperCAmelCase_ : Optional[Any]) ->List[Any]: '''simple docstring''' lowerCamelCase__: Optional[int] ="This is a là test" lowerCamelCase__: Optional[Any] ="This is a<unk><unk> test" return input_text, output_text def SCREAMING_SNAKE_CASE_ (self : List[Any]) ->Optional[Any]: '''simple docstring''' lowerCamelCase__: str =BartphoTokenizer(UpperCAmelCase_ , self.monolingual_vocab_file , **self.special_tokens_map) lowerCamelCase__: List[Any] ="This is a là test" lowerCamelCase__: Optional[int] ="▁This ▁is ▁a ▁l à ▁t est".split() lowerCamelCase__: Optional[int] =tokenizer.tokenize(UpperCAmelCase_) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_) lowerCamelCase__: Tuple =tokens + [tokenizer.unk_token] lowerCamelCase__: List[Any] =[4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_) , UpperCAmelCase_)
10
1
import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all feature extractors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...feature_extraction_utils import FeatureExtractionMixin from ...utils import CONFIG_NAME, FEATURE_EXTRACTOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) _snake_case = logging.get_logger(__name__) _snake_case = OrderedDict( [ ("audio-spectrogram-transformer", "ASTFeatureExtractor"), ("beit", "BeitFeatureExtractor"), ("chinese_clip", "ChineseCLIPFeatureExtractor"), ("clap", "ClapFeatureExtractor"), ("clip", "CLIPFeatureExtractor"), ("clipseg", "ViTFeatureExtractor"), ("conditional_detr", "ConditionalDetrFeatureExtractor"), ("convnext", "ConvNextFeatureExtractor"), ("cvt", "ConvNextFeatureExtractor"), ("data2vec-audio", "Wav2Vec2FeatureExtractor"), ("data2vec-vision", "BeitFeatureExtractor"), ("deformable_detr", "DeformableDetrFeatureExtractor"), ("deit", "DeiTFeatureExtractor"), ("detr", "DetrFeatureExtractor"), ("dinat", "ViTFeatureExtractor"), ("donut-swin", "DonutFeatureExtractor"), ("dpt", "DPTFeatureExtractor"), ("encodec", "EncodecFeatureExtractor"), ("flava", "FlavaFeatureExtractor"), ("glpn", "GLPNFeatureExtractor"), ("groupvit", "CLIPFeatureExtractor"), ("hubert", "Wav2Vec2FeatureExtractor"), ("imagegpt", "ImageGPTFeatureExtractor"), ("layoutlmv2", "LayoutLMv2FeatureExtractor"), ("layoutlmv3", "LayoutLMv3FeatureExtractor"), ("levit", "LevitFeatureExtractor"), ("maskformer", "MaskFormerFeatureExtractor"), ("mctct", "MCTCTFeatureExtractor"), ("mobilenet_v1", "MobileNetV1FeatureExtractor"), ("mobilenet_v2", "MobileNetV2FeatureExtractor"), ("mobilevit", "MobileViTFeatureExtractor"), ("nat", "ViTFeatureExtractor"), ("owlvit", "OwlViTFeatureExtractor"), ("perceiver", "PerceiverFeatureExtractor"), ("poolformer", "PoolFormerFeatureExtractor"), ("regnet", "ConvNextFeatureExtractor"), ("resnet", "ConvNextFeatureExtractor"), ("segformer", "SegformerFeatureExtractor"), ("sew", "Wav2Vec2FeatureExtractor"), ("sew-d", "Wav2Vec2FeatureExtractor"), ("speech_to_text", "Speech2TextFeatureExtractor"), ("speecht5", "SpeechT5FeatureExtractor"), ("swiftformer", "ViTFeatureExtractor"), ("swin", "ViTFeatureExtractor"), ("swinv2", "ViTFeatureExtractor"), ("table-transformer", "DetrFeatureExtractor"), ("timesformer", "VideoMAEFeatureExtractor"), ("tvlt", "TvltFeatureExtractor"), ("unispeech", "Wav2Vec2FeatureExtractor"), ("unispeech-sat", "Wav2Vec2FeatureExtractor"), ("van", "ConvNextFeatureExtractor"), ("videomae", "VideoMAEFeatureExtractor"), ("vilt", "ViltFeatureExtractor"), ("vit", "ViTFeatureExtractor"), ("vit_mae", "ViTFeatureExtractor"), ("vit_msn", "ViTFeatureExtractor"), ("wav2vec2", "Wav2Vec2FeatureExtractor"), ("wav2vec2-conformer", "Wav2Vec2FeatureExtractor"), ("wavlm", "Wav2Vec2FeatureExtractor"), ("whisper", "WhisperFeatureExtractor"), ("xclip", "CLIPFeatureExtractor"), ("yolos", "YolosFeatureExtractor"), ] ) _snake_case = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FEATURE_EXTRACTOR_MAPPING_NAMES) def lowerCAmelCase_ ( snake_case_ ): for module_name, extractors in FEATURE_EXTRACTOR_MAPPING_NAMES.items(): if class_name in extractors: _A : int = model_type_to_module_name(snake_case_ ) _A : Tuple = importlib.import_module(f'''.{module_name}''',"""transformers.models""" ) try: return getattr(snake_case_,snake_case_ ) except AttributeError: continue for _, extractor in FEATURE_EXTRACTOR_MAPPING._extra_content.items(): if getattr(snake_case_,"""__name__""",snake_case_ ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. _A : Dict = importlib.import_module("""transformers""" ) if hasattr(snake_case_,snake_case_ ): return getattr(snake_case_,snake_case_ ) return None def lowerCAmelCase_ ( snake_case_,snake_case_ = None,snake_case_ = False,snake_case_ = False,snake_case_ = None,snake_case_ = None,snake_case_ = None,snake_case_ = False,**snake_case_,): _A : Optional[int] = get_file_from_repo( snake_case_,snake_case_,cache_dir=snake_case_,force_download=snake_case_,resume_download=snake_case_,proxies=snake_case_,use_auth_token=snake_case_,revision=snake_case_,local_files_only=snake_case_,) if resolved_config_file is None: logger.info( """Could not locate the feature extractor configuration file, will try to use the model config instead.""" ) return {} with open(snake_case_,encoding="""utf-8""" ) as reader: return json.load(snake_case_ ) class lowercase : def __init__( self ) -> Any: raise EnvironmentError( """AutoFeatureExtractor is designed to be instantiated """ """using the `AutoFeatureExtractor.from_pretrained(pretrained_model_name_or_path)` method.""" ) @classmethod @replace_list_option_in_docstrings(_a ) def a__ ( cls , _a , **_a ) -> str: _A : Any = kwargs.pop("""config""" , _a ) _A : Tuple = kwargs.pop("""trust_remote_code""" , _a ) _A : Dict = True _A , _A : List[str] = FeatureExtractionMixin.get_feature_extractor_dict(_a , **_a ) _A : str = config_dict.get("""feature_extractor_type""" , _a ) _A : Optional[Any] = None if "AutoFeatureExtractor" in config_dict.get("""auto_map""" , {} ): _A : Union[str, Any] = config_dict["""auto_map"""]["""AutoFeatureExtractor"""] # If we don't find the feature extractor class in the feature extractor config, let's try the model config. if feature_extractor_class is None and feature_extractor_auto_map is None: if not isinstance(_a , _a ): _A : str = AutoConfig.from_pretrained(_a , **_a ) # It could be in `config.feature_extractor_type`` _A : Dict = getattr(_a , """feature_extractor_type""" , _a ) if hasattr(_a , """auto_map""" ) and "AutoFeatureExtractor" in config.auto_map: _A : Optional[Any] = config.auto_map["""AutoFeatureExtractor"""] if feature_extractor_class is not None: _A : int = feature_extractor_class_from_name(_a ) _A : List[str] = feature_extractor_auto_map is not None _A : Dict = feature_extractor_class is not None or type(_a ) in FEATURE_EXTRACTOR_MAPPING _A : Optional[int] = resolve_trust_remote_code( _a , _a , _a , _a ) if has_remote_code and trust_remote_code: _A : str = get_class_from_dynamic_module( _a , _a , **_a ) _A : Dict = kwargs.pop("""code_revision""" , _a ) if os.path.isdir(_a ): feature_extractor_class.register_for_auto_class() return feature_extractor_class.from_dict(_a , **_a ) elif feature_extractor_class is not None: return feature_extractor_class.from_dict(_a , **_a ) # Last try: we use the FEATURE_EXTRACTOR_MAPPING. elif type(_a ) in FEATURE_EXTRACTOR_MAPPING: _A : Optional[int] = FEATURE_EXTRACTOR_MAPPING[type(_a )] return feature_extractor_class.from_dict(_a , **_a ) raise ValueError( F'''Unrecognized feature extractor in {pretrained_model_name_or_path}. Should have a ''' F'''`feature_extractor_type` key in its {FEATURE_EXTRACTOR_NAME} of {CONFIG_NAME}, or one of the following ''' F'''`model_type` keys in its {CONFIG_NAME}: {", ".join(c for c in FEATURE_EXTRACTOR_MAPPING_NAMES.keys() )}''' ) @staticmethod def a__ ( _a , _a ) -> Any: FEATURE_EXTRACTOR_MAPPING.register(_a , _a )
343
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def lowerCAmelCase_ ( snake_case_ ): _A : Tuple = filter(lambda snake_case_ : p.requires_grad,model.parameters() ) _A : str = sum([np.prod(p.size() ) for p in model_parameters] ) return params _snake_case = logging.getLogger(__name__) def lowerCAmelCase_ ( snake_case_,snake_case_ ): if metric == "rouge2": _A : Optional[int] = """{val_avg_rouge2:.4f}-{step_count}""" elif metric == "bleu": _A : Dict = """{val_avg_bleu:.4f}-{step_count}""" elif metric == "em": _A : List[str] = """{val_avg_em:.4f}-{step_count}""" else: raise NotImplementedError( f'''seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this''' """ function.""" ) _A : Optional[int] = ModelCheckpoint( dirpath=snake_case_,filename=snake_case_,monitor=f'''val_{metric}''',mode="""max""",save_top_k=3,every_n_epochs=1,) return checkpoint_callback def lowerCAmelCase_ ( snake_case_,snake_case_ ): return EarlyStopping( monitor=f'''val_{metric}''',mode="""min""" if """loss""" in metric else """max""",patience=snake_case_,verbose=snake_case_,) class lowercase ( pl.Callback ): def a__ ( self , _a , _a ) -> Optional[Any]: _A : List[Any] = {F'''lr_group_{i}''': param["""lr"""] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(_a ) @rank_zero_only def a__ ( self , _a , _a , _a , _a=True ) -> None: logger.info(F'''***** {type_path} results at step {trainer.global_step:05d} *****''' ) _A : int = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ["""log""", """progress_bar""", """preds"""]} ) # Log results _A : Dict = Path(pl_module.hparams.output_dir ) if type_path == "test": _A : List[Any] = od / """test_results.txt""" _A : List[Any] = od / """test_generations.txt""" else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. _A : Optional[int] = od / F'''{type_path}_results/{trainer.global_step:05d}.txt''' _A : int = od / F'''{type_path}_generations/{trainer.global_step:05d}.txt''' results_file.parent.mkdir(exist_ok=_a ) generations_file.parent.mkdir(exist_ok=_a ) with open(_a , """a+""" ) as writer: for key in sorted(_a ): if key in ["log", "progress_bar", "preds"]: continue _A : List[Any] = metrics[key] if isinstance(_a , torch.Tensor ): _A : str = val.item() _A : str = F'''{key}: {val:.6f}\n''' writer.write(_a ) if not save_generations: return if "preds" in metrics: _A : List[Any] = """\n""".join(metrics["""preds"""] ) generations_file.open("""w+""" ).write(_a ) @rank_zero_only def a__ ( self , _a , _a ) -> str: try: _A : int = pl_module.model.model.num_parameters() except AttributeError: _A : str = pl_module.model.num_parameters() _A : Optional[int] = count_trainable_parameters(_a ) # mp stands for million parameters trainer.logger.log_metrics({"""n_params""": npars, """mp""": npars / 1e6, """grad_mp""": n_trainable_pars / 1e6} ) @rank_zero_only def a__ ( self , _a , _a ) -> Optional[int]: save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(_a , _a , """test""" ) @rank_zero_only def a__ ( self , _a , _a ) -> Tuple: save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
343
1
"""simple docstring""" import numpy as np # Importing the Keras libraries and packages import tensorflow as tf from tensorflow.keras import layers, models if __name__ == "__main__": # Initialising the CNN # (Sequential- Building the model layer by layer) _a : List[Any]= models.Sequential() # Step 1 - Convolution # Here 64,64 is the length & breadth of dataset images and 3 is for the RGB channel # (3,3) is the kernel size (filter matrix) classifier.add( layers.ConvaD(32, (3, 3), input_shape=(64, 64, 3), activation="relu") ) # Step 2 - Pooling classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Adding a second convolutional layer classifier.add(layers.ConvaD(32, (3, 3), activation="relu")) classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Step 3 - Flattening classifier.add(layers.Flatten()) # Step 4 - Full connection classifier.add(layers.Dense(units=128, activation="relu")) classifier.add(layers.Dense(units=1, activation="sigmoid")) # Compiling the CNN classifier.compile( optimizer="adam", loss="binary_crossentropy", metrics=["accuracy"] ) # Part 2 - Fitting the CNN to the images # Load Trained model weights # from keras.models import load_model # regressor=load_model('cnn.h5') _a : Any= tf.keras.preprocessing.image.ImageDataGenerator( rescale=1.0 / 255, shear_range=0.2, zoom_range=0.2, horizontal_flip=True ) _a : List[Any]= tf.keras.preprocessing.image.ImageDataGenerator(rescale=1.0 / 255) _a : Tuple= train_datagen.flow_from_directory( "dataset/training_set", target_size=(64, 64), batch_size=32, class_mode="binary" ) _a : Optional[int]= test_datagen.flow_from_directory( "dataset/test_set", target_size=(64, 64), batch_size=32, class_mode="binary" ) classifier.fit_generator( training_set, steps_per_epoch=5, epochs=30, validation_data=test_set ) classifier.save("cnn.h5") # Part 3 - Making new predictions _a : Any= tf.keras.preprocessing.image.load_img( "dataset/single_prediction/image.png", target_size=(64, 64) ) _a : Union[str, Any]= tf.keras.preprocessing.image.img_to_array(test_image) _a : List[str]= np.expand_dims(test_image, axis=0) _a : Dict= classifier.predict(test_image) # training_set.class_indices if result[0][0] == 0: _a : Union[str, Any]= "Normal" if result[0][0] == 1: _a : Union[str, Any]= "Abnormality detected"
172
from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class SCREAMING_SNAKE_CASE_ : def __init__( self : Tuple , lowerCamelCase_ : Any , lowerCamelCase_ : Dict=13 , lowerCamelCase_ : str=30 , lowerCamelCase_ : List[str]=2 , lowerCamelCase_ : Union[str, Any]=3 , lowerCamelCase_ : Any=True , lowerCamelCase_ : int=True , lowerCamelCase_ : Tuple=32 , lowerCamelCase_ : Optional[Any]=2 , lowerCamelCase_ : int=4 , lowerCamelCase_ : str=37 , lowerCamelCase_ : Optional[Any]="gelu" , lowerCamelCase_ : Optional[int]=0.1 , lowerCamelCase_ : List[Any]=0.1 , lowerCamelCase_ : List[Any]=10 , lowerCamelCase_ : List[Any]=0.0_2 , lowerCamelCase_ : Optional[int]=3 , lowerCamelCase_ : List[Any]=0.6 , lowerCamelCase_ : Optional[Any]=None , ): """simple docstring""" UpperCamelCase = parent UpperCamelCase = batch_size UpperCamelCase = image_size UpperCamelCase = patch_size UpperCamelCase = num_channels UpperCamelCase = is_training UpperCamelCase = use_labels UpperCamelCase = hidden_size UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = intermediate_size UpperCamelCase = hidden_act UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = type_sequence_label_size UpperCamelCase = initializer_range UpperCamelCase = mask_ratio UpperCamelCase = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) UpperCamelCase = (image_size // patch_size) ** 2 UpperCamelCase = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase = None if self.use_labels: UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase = self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase_ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def lowerCamelCase_ ( self : Any , lowerCamelCase_ : Union[str, Any] , lowerCamelCase_ : Any , lowerCamelCase_ : Union[str, Any] ): """simple docstring""" UpperCamelCase = TFViTMAEModel(config=lowerCamelCase_ ) UpperCamelCase = model(lowerCamelCase_ , training=lowerCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self : List[str] , lowerCamelCase_ : Union[str, Any] , lowerCamelCase_ : List[str] , lowerCamelCase_ : str ): """simple docstring""" UpperCamelCase = TFViTMAEForPreTraining(lowerCamelCase_ ) UpperCamelCase = model(lowerCamelCase_ , training=lowerCamelCase_ ) # expected sequence length = num_patches UpperCamelCase = (self.image_size // self.patch_size) ** 2 UpperCamelCase = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images UpperCamelCase = 1 UpperCamelCase = TFViTMAEForPreTraining(lowerCamelCase_ ) UpperCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCamelCase = model(lowerCamelCase_ , training=lowerCamelCase_ ) UpperCamelCase = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def lowerCamelCase_ ( self : Dict ): """simple docstring""" UpperCamelCase = self.prepare_config_and_inputs() ((UpperCamelCase) , (UpperCamelCase) , (UpperCamelCase)) = config_and_inputs UpperCamelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): __lowerCAmelCase = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () __lowerCAmelCase = {"""feature-extraction""": TFViTMAEModel} if is_tf_available() else {} __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" UpperCamelCase = TFViTMAEModelTester(self ) UpperCamelCase = ConfigTester(self , config_class=lowerCamelCase_ , has_text_modality=lowerCamelCase_ , hidden_size=37 ) def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""ViTMAE does not use inputs_embeds""" ) def lowerCamelCase_ ( self : str ): """simple docstring""" pass def lowerCamelCase_ ( self : Tuple ): """simple docstring""" UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase_ , tf.keras.layers.Layer ) ) def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase = [*signature.parameters.keys()] UpperCamelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowerCamelCase_ ) def lowerCamelCase_ ( self : str ): """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase_ ) def lowerCamelCase_ ( self : int ): """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase_ ) def lowerCamelCase_ ( self : Dict ): """simple docstring""" np.random.seed(2 ) UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = int((config.image_size // config.patch_size) ** 2 ) UpperCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) UpperCamelCase = model(lowerCamelCase_ , noise=lowerCamelCase_ ) UpperCamelCase = copy.deepcopy(self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) ) UpperCamelCase = model(**lowerCamelCase_ , noise=lowerCamelCase_ ) UpperCamelCase = outputs_dict[0].numpy() UpperCamelCase = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1E-6 ) def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" np.random.seed(2 ) UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = int((config.image_size // config.patch_size) ** 2 ) UpperCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(lowerCamelCase_ : List[Any] ): UpperCamelCase = {} for k, v in inputs_dict.items(): if tf.is_tensor(lowerCamelCase_ ): UpperCamelCase = v.numpy() else: UpperCamelCase = np.array(lowerCamelCase_ ) return inputs_np_dict for model_class in self.all_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) UpperCamelCase = prepare_numpy_arrays(lowerCamelCase_ ) UpperCamelCase = model(lowerCamelCase_ , noise=lowerCamelCase_ ) UpperCamelCase = model(**lowerCamelCase_ , noise=lowerCamelCase_ ) self.assert_outputs_same(lowerCamelCase_ , lowerCamelCase_ ) def lowerCamelCase_ ( self : Any , lowerCamelCase_ : Optional[int] , lowerCamelCase_ : Optional[int] , lowerCamelCase_ : Optional[int] ): """simple docstring""" np.random.seed(2 ) UpperCamelCase = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) UpperCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) UpperCamelCase = tf.constant(lowerCamelCase_ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument UpperCamelCase = tf_noise super().check_pt_tf_models(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) def lowerCamelCase_ ( self : Dict ): """simple docstring""" np.random.seed(2 ) UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(lowerCamelCase_ ) if module_member_name.endswith("""MainLayer""" ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len("""MainLayer""" )] == model_class.__name__[: -len("""Model""" )] for module_member in (getattr(lowerCamelCase_ , lowerCamelCase_ ),) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(lowerCamelCase_ , """_keras_serializable""" , lowerCamelCase_ ) } UpperCamelCase = int((config.image_size // config.patch_size) ** 2 ) UpperCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) UpperCamelCase = tf.convert_to_tensor(lowerCamelCase_ ) inputs_dict.update({"""noise""": noise} ) for main_layer_class in tf_main_layer_classes: UpperCamelCase = main_layer_class(lowerCamelCase_ ) UpperCamelCase = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } UpperCamelCase = tf.keras.Model(lowerCamelCase_ , outputs=main_layer(lowerCamelCase_ ) ) UpperCamelCase = model(lowerCamelCase_ ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCamelCase = os.path.join(lowerCamelCase_ , """keras_model.h5""" ) model.save(lowerCamelCase_ ) UpperCamelCase = tf.keras.models.load_model( lowerCamelCase_ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(lowerCamelCase_ , tf.keras.Model ) UpperCamelCase = model(lowerCamelCase_ ) self.assert_outputs_same(lowerCamelCase_ , lowerCamelCase_ ) @slow def lowerCamelCase_ ( self : Dict ): """simple docstring""" np.random.seed(2 ) UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = int((config.image_size // config.patch_size) ** 2 ) UpperCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) UpperCamelCase = model(lowerCamelCase_ , noise=lowerCamelCase_ ) if model_class.__name__ == "TFViTMAEModel": UpperCamelCase = outputs.last_hidden_state.numpy() UpperCamelCase = 0 else: UpperCamelCase = outputs.logits.numpy() UpperCamelCase = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase_ , saved_model=lowerCamelCase_ ) UpperCamelCase = model_class.from_pretrained(lowerCamelCase_ ) UpperCamelCase = model(lowerCamelCase_ , noise=lowerCamelCase_ ) if model_class.__name__ == "TFViTMAEModel": UpperCamelCase = after_outputs["""last_hidden_state"""].numpy() UpperCamelCase = 0 else: UpperCamelCase = after_outputs["""logits"""].numpy() UpperCamelCase = 0 UpperCamelCase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCamelCase_ , 1E-5 ) def lowerCamelCase_ ( self : List[str] ): """simple docstring""" np.random.seed(2 ) UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = int((config.image_size // config.patch_size) ** 2 ) UpperCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) UpperCamelCase = model(lowerCamelCase_ , noise=lowerCamelCase_ ) UpperCamelCase = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(lowerCamelCase_ ) UpperCamelCase = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config UpperCamelCase = model_class.from_config(model.config ) UpperCamelCase = new_model(lowerCamelCase_ ) # Build model new_model.set_weights(model.get_weights() ) UpperCamelCase = new_model(lowerCamelCase_ , noise=lowerCamelCase_ ) self.assert_outputs_same(lowerCamelCase_ , lowerCamelCase_ ) @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.""" ) def lowerCamelCase_ ( self : int ): """simple docstring""" pass @unittest.skip(reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load""" ) def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" pass @slow def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" UpperCamelCase = TFViTMAEModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(lowerCamelCase_ ) def lowercase( ) -> int: '''simple docstring''' UpperCamelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self : Dict ): """simple docstring""" return ViTImageProcessor.from_pretrained("""facebook/vit-mae-base""" ) if is_vision_available() else None @slow def lowerCamelCase_ ( self : List[str] ): """simple docstring""" np.random.seed(2 ) UpperCamelCase = TFViTMAEForPreTraining.from_pretrained("""facebook/vit-mae-base""" ) UpperCamelCase = self.default_image_processor UpperCamelCase = prepare_img() UpperCamelCase = image_processor(images=lowerCamelCase_ , return_tensors="""tf""" ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) UpperCamelCase = ViTMAEConfig() UpperCamelCase = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) UpperCamelCase = np.random.uniform(size=(1, num_patches) ) # forward pass UpperCamelCase = model(**lowerCamelCase_ , noise=lowerCamelCase_ ) # verify the logits UpperCamelCase = tf.convert_to_tensor([1, 196, 768] ) self.assertEqual(outputs.logits.shape , lowerCamelCase_ ) UpperCamelCase = tf.convert_to_tensor( [[-0.0_5_4_8, -1.7_0_2_3, -0.9_3_2_5], [0.3_7_2_1, -0.5_6_7_0, -0.2_2_3_3], [0.8_2_3_5, -1.3_8_7_8, -0.3_5_2_4]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , lowerCamelCase_ , atol=1E-4 )
343
0
'''simple docstring''' import pytest __lowercase : Any = '__dummy_dataset1__' __lowercase : int = '\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/"\nURLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n "tokens": datasets.Sequence(datasets.Value("string")),\n "ner_tags": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n "O",\n "B-PER",\n "I-PER",\n "B-ORG",\n "I-ORG",\n "B-LOC",\n "I-LOC",\n ]\n )\n ),\n "langs": datasets.Sequence(datasets.Value("string")),\n "spans": datasets.Sequence(datasets.Value("string")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, "r", encoding="utf-8") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n' @pytest.fixture def lowerCamelCase (): return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def lowerCamelCase (): return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def lowerCamelCase (_SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : str ): __a : int = dataset_loading_script_name __a : Optional[Any] = tmp_path / "datasets" / script_name script_dir.mkdir(parents=a_ ) __a : int = script_dir / F"""{script_name}.py""" with open(a_ , 'w' ) as f: f.write(a_ ) return str(a_ )
367
'''simple docstring''' import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () __lowercase : int = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). __lowercase : Any = [0, 25, 50] __lowercase : int = [25, 50, 75] __lowercase : List[str] = fuzz.membership.trimf(X, abca) __lowercase : Any = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. __lowercase : List[Any] = np.ones(75) __lowercase : Any = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) __lowercase : int = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) __lowercase : int = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) __lowercase : str = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) __lowercase : List[Any] = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] __lowercase : Optional[Any] = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) __lowercase : str = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] __lowercase : Optional[Any] = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] __lowercase : Union[str, Any] = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
294
0
'''simple docstring''' import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEmbeddings, BertLayer, BertPooler, BertPreTrainedModel, ) def __lowerCamelCase ( A__ ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = torch.exp(A__ ) UpperCamelCase = torch.sum(A__ , dim=1 ) # sum of exp(x_i) UpperCamelCase = torch.sum(x * exp_x , dim=1 ) # sum of x_i * exp(x_i) return torch.log(A__ ) - B / A class SCREAMING_SNAKE_CASE ( nn.Module ): """simple docstring""" def __init__( self : Any , UpperCamelCase__ : Any ): """simple docstring""" super().__init__() UpperCamelCase = config.output_attentions UpperCamelCase = config.output_hidden_states UpperCamelCase = nn.ModuleList([BertLayer(UpperCamelCase__ ) for _ in range(config.num_hidden_layers )] ) UpperCamelCase = nn.ModuleList([BertHighway(UpperCamelCase__ ) for _ in range(config.num_hidden_layers )] ) UpperCamelCase = [-1 for _ in range(config.num_hidden_layers )] def A ( self : Dict , UpperCamelCase__ : str ): """simple docstring""" if (type(UpperCamelCase__ ) is float) or (type(UpperCamelCase__ ) is int): for i in range(len(self.early_exit_entropy ) ): UpperCamelCase = x else: UpperCamelCase = x def A ( self : Union[str, Any] , UpperCamelCase__ : str ): """simple docstring""" UpperCamelCase = pooler.state_dict() for highway in self.highway: for name, param in highway.pooler.state_dict().items(): param.copy_(loaded_model[name] ) def A ( self : List[Any] , UpperCamelCase__ : Tuple , UpperCamelCase__ : str=None , UpperCamelCase__ : Union[str, Any]=None , UpperCamelCase__ : Any=None , UpperCamelCase__ : List[Any]=None , ): """simple docstring""" UpperCamelCase = () UpperCamelCase = () UpperCamelCase = () for i, layer_module in enumerate(self.layer ): if self.output_hidden_states: UpperCamelCase = all_hidden_states + (hidden_states,) UpperCamelCase = layer_module( UpperCamelCase__ , UpperCamelCase__ , head_mask[i] , UpperCamelCase__ , UpperCamelCase__ ) UpperCamelCase = layer_outputs[0] if self.output_attentions: UpperCamelCase = all_attentions + (layer_outputs[1],) UpperCamelCase = (hidden_states,) if self.output_hidden_states: UpperCamelCase = current_outputs + (all_hidden_states,) if self.output_attentions: UpperCamelCase = current_outputs + (all_attentions,) UpperCamelCase = self.highway[i](UpperCamelCase__ ) # logits, pooled_output if not self.training: UpperCamelCase = highway_exit[0] UpperCamelCase = entropy(UpperCamelCase__ ) UpperCamelCase = highway_exit + (highway_entropy,) # logits, hidden_states(?), entropy UpperCamelCase = all_highway_exits + (highway_exit,) if highway_entropy < self.early_exit_entropy[i]: UpperCamelCase = (highway_logits,) + current_outputs[1:] + (all_highway_exits,) raise HighwayException(UpperCamelCase__ , i + 1 ) else: UpperCamelCase = all_highway_exits + (highway_exit,) # Add last layer if self.output_hidden_states: UpperCamelCase = all_hidden_states + (hidden_states,) UpperCamelCase = (hidden_states,) if self.output_hidden_states: UpperCamelCase = outputs + (all_hidden_states,) if self.output_attentions: UpperCamelCase = outputs + (all_attentions,) UpperCamelCase = outputs + (all_highway_exits,) return outputs # last-layer hidden state, (all hidden states), (all attentions), all highway exits @add_start_docstrings( """The Bert Model transformer with early exiting (DeeBERT). """ , _a , ) class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" def __init__( self : Any , UpperCamelCase__ : Any ): """simple docstring""" super().__init__(UpperCamelCase__ ) UpperCamelCase = config UpperCamelCase = BertEmbeddings(UpperCamelCase__ ) UpperCamelCase = DeeBertEncoder(UpperCamelCase__ ) UpperCamelCase = BertPooler(UpperCamelCase__ ) self.init_weights() def A ( self : int ): """simple docstring""" self.encoder.init_highway_pooler(self.pooler ) def A ( self : List[Any] ): """simple docstring""" return self.embeddings.word_embeddings def A ( self : List[str] , UpperCamelCase__ : Optional[Any] ): """simple docstring""" UpperCamelCase = value def A ( self : Any , UpperCamelCase__ : List[str] ): """simple docstring""" for layer, heads in heads_to_prune.items(): self.encoder.layer[layer].attention.prune_heads(UpperCamelCase__ ) @add_start_docstrings_to_model_forward(UpperCamelCase__ ) def A ( self : Optional[Any] , UpperCamelCase__ : Union[str, Any]=None , UpperCamelCase__ : List[Any]=None , UpperCamelCase__ : Optional[int]=None , UpperCamelCase__ : int=None , UpperCamelCase__ : Any=None , UpperCamelCase__ : int=None , UpperCamelCase__ : Tuple=None , UpperCamelCase__ : Optional[int]=None , ): """simple docstring""" if input_ids is not None and inputs_embeds is not None: raise ValueError('You cannot specify both input_ids and inputs_embeds at the same time' ) elif input_ids is not None: UpperCamelCase = input_ids.size() elif inputs_embeds is not None: UpperCamelCase = inputs_embeds.size()[:-1] else: raise ValueError('You have to specify either input_ids or inputs_embeds' ) UpperCamelCase = input_ids.device if input_ids is not None else inputs_embeds.device if attention_mask is None: UpperCamelCase = torch.ones(UpperCamelCase__ , device=UpperCamelCase__ ) if encoder_attention_mask is None: UpperCamelCase = torch.ones(UpperCamelCase__ , device=UpperCamelCase__ ) if token_type_ids is None: UpperCamelCase = torch.zeros(UpperCamelCase__ , dtype=torch.long , device=UpperCamelCase__ ) # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] # ourselves in which case we just need to make it broadcastable to all heads. UpperCamelCase = self.get_extended_attention_mask(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # If a 2D ou 3D attention mask is provided for the cross-attention # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] if encoder_attention_mask.dim() == 3: UpperCamelCase = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.dim() == 2: UpperCamelCase = encoder_attention_mask[:, None, None, :] UpperCamelCase = encoder_extended_attention_mask.to( dtype=next(self.parameters() ).dtype ) # fp16 compatibility UpperCamelCase = (1.0 - encoder_extended_attention_mask) * -1_0_0_0_0.0 # Prepare head mask if needed # 1.0 in head_mask indicate we keep the head # attention_probs has shape bsz x n_heads x N x N # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] UpperCamelCase = self.get_head_mask(UpperCamelCase__ , self.config.num_hidden_layers ) UpperCamelCase = self.embeddings( input_ids=UpperCamelCase__ , position_ids=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , inputs_embeds=UpperCamelCase__ ) UpperCamelCase = self.encoder( UpperCamelCase__ , attention_mask=UpperCamelCase__ , head_mask=UpperCamelCase__ , encoder_hidden_states=UpperCamelCase__ , encoder_attention_mask=UpperCamelCase__ , ) UpperCamelCase = encoder_outputs[0] UpperCamelCase = self.pooler(UpperCamelCase__ ) UpperCamelCase = ( sequence_output, pooled_output, ) + encoder_outputs[ 1: ] # add hidden_states and attentions if they are here return outputs # sequence_output, pooled_output, (hidden_states), (attentions), highway exits class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" def __init__( self : Union[str, Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : List[Any] ): """simple docstring""" UpperCamelCase = message UpperCamelCase = exit_layer # start from 1! class SCREAMING_SNAKE_CASE ( nn.Module ): """simple docstring""" def __init__( self : Optional[int] , UpperCamelCase__ : Union[str, Any] ): """simple docstring""" super().__init__() UpperCamelCase = BertPooler(UpperCamelCase__ ) UpperCamelCase = nn.Dropout(config.hidden_dropout_prob ) UpperCamelCase = nn.Linear(config.hidden_size , config.num_labels ) def A ( self : Dict , UpperCamelCase__ : Tuple ): """simple docstring""" UpperCamelCase = encoder_outputs[0] UpperCamelCase = self.pooler(UpperCamelCase__ ) # "return" pooler_output # BertModel UpperCamelCase = (pooler_input, pooler_output) + encoder_outputs[1:] # "return" bmodel_output # Dropout and classification UpperCamelCase = bmodel_output[1] UpperCamelCase = self.dropout(UpperCamelCase__ ) UpperCamelCase = self.classifier(UpperCamelCase__ ) return logits, pooled_output @add_start_docstrings( """Bert Model (with early exiting - DeeBERT) with a classifier on top, also takes care of multi-layer training. """ , _a , ) class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" def __init__( self : Optional[int] , UpperCamelCase__ : Optional[Any] ): """simple docstring""" super().__init__(UpperCamelCase__ ) UpperCamelCase = config.num_labels UpperCamelCase = config.num_hidden_layers UpperCamelCase = DeeBertModel(UpperCamelCase__ ) UpperCamelCase = nn.Dropout(config.hidden_dropout_prob ) UpperCamelCase = nn.Linear(config.hidden_size , self.config.num_labels ) self.init_weights() @add_start_docstrings_to_model_forward(UpperCamelCase__ ) def A ( self : str , UpperCamelCase__ : int=None , UpperCamelCase__ : List[Any]=None , UpperCamelCase__ : Union[str, Any]=None , UpperCamelCase__ : int=None , UpperCamelCase__ : str=None , UpperCamelCase__ : Optional[int]=None , UpperCamelCase__ : int=None , UpperCamelCase__ : str=-1 , UpperCamelCase__ : Union[str, Any]=False , ): """simple docstring""" UpperCamelCase = self.num_layers try: UpperCamelCase = self.bert( UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , position_ids=UpperCamelCase__ , head_mask=UpperCamelCase__ , inputs_embeds=UpperCamelCase__ , ) # sequence_output, pooled_output, (hidden_states), (attentions), highway exits UpperCamelCase = outputs[1] UpperCamelCase = self.dropout(UpperCamelCase__ ) UpperCamelCase = self.classifier(UpperCamelCase__ ) UpperCamelCase = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: UpperCamelCase = e.message UpperCamelCase = e.exit_layer UpperCamelCase = outputs[0] if not self.training: UpperCamelCase = entropy(UpperCamelCase__ ) UpperCamelCase = [] UpperCamelCase = [] if labels is not None: if self.num_labels == 1: # We are doing regression UpperCamelCase = MSELoss() UpperCamelCase = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: UpperCamelCase = CrossEntropyLoss() UpperCamelCase = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits UpperCamelCase = [] for highway_exit in outputs[-1]: UpperCamelCase = highway_exit[0] if not self.training: highway_logits_all.append(UpperCamelCase__ ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression UpperCamelCase = MSELoss() UpperCamelCase = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: UpperCamelCase = CrossEntropyLoss() UpperCamelCase = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(UpperCamelCase__ ) if train_highway: UpperCamelCase = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: UpperCamelCase = (loss,) + outputs if not self.training: UpperCamelCase = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: UpperCamelCase = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), (highway_exits)
28
'''simple docstring''' import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging _lowerCamelCase : Union[str, Any] = "\\n\n" _lowerCamelCase : List[str] = "\nPerplexity (PPL) is one of the most common metrics for evaluating language models.\nIt is defined as the exponentiated average negative log-likelihood of a sequence.\n\nFor more information, see https://huggingface.co/docs/transformers/perplexity\n" _lowerCamelCase : Dict = "\nArgs:\n model_id (str): model used for calculating Perplexity\n NOTE: Perplexity can only be calculated for causal language models.\n This includes models such as gpt2, causal variations of bert,\n causal versions of t5, and more (the full list can be found\n in the AutoModelForCausalLM documentation here:\n https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM )\n\n input_texts (list of str): input text, each separate text snippet\n is one list entry.\n batch_size (int): the batch size to run texts through the model. Defaults to 16.\n add_start_token (bool): whether to add the start token to the texts,\n so the perplexity can include the probability of the first word. Defaults to True.\n device (str): device to run on, defaults to 'cuda' when available\nReturns:\n perplexity: dictionary containing the perplexity scores for the texts\n in the input list, as well as the mean perplexity. If one of the input texts is\n longer than the max input length of the model, then it is truncated to the\n max length for the perplexity computation.\nExamples:\n Example 1:\n >>> perplexity = datasets.load_metric(\"perplexity\")\n >>> input_texts = [\"lorem ipsum\", \"Happy Birthday!\", \"Bienvenue\"]\n >>> results = perplexity.compute(model_id='gpt2',\n ... add_start_token=False,\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n ['perplexities', 'mean_perplexity']\n >>> print(round(results[\"mean_perplexity\"], 2))\n 78.22\n >>> print(round(results[\"perplexities\"][0], 2))\n 11.11\n\n Example 2:\n >>> perplexity = datasets.load_metric(\"perplexity\")\n >>> input_texts = datasets.load_dataset(\"wikitext\",\n ... \"wikitext-2-raw-v1\",\n ... split=\"test\")[\"text\"][:50] # doctest:+ELLIPSIS\n [...]\n >>> input_texts = [s for s in input_texts if s!='']\n >>> results = perplexity.compute(model_id='gpt2',\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n ['perplexities', 'mean_perplexity']\n >>> print(round(results[\"mean_perplexity\"], 2))\n 60.35\n >>> print(round(results[\"perplexities\"][0], 2))\n 81.12\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE ( datasets.Metric ): """simple docstring""" def A ( self : Tuple ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'input_texts': datasets.Value('string' ), } ) , reference_urls=['https://huggingface.co/docs/transformers/perplexity'] , ) def A ( self : Optional[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int = 1_6 , UpperCamelCase__ : bool = True , UpperCamelCase__ : List[Any]=None ): """simple docstring""" if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": UpperCamelCase = 'cuda' else: UpperCamelCase = 'cuda' if torch.cuda.is_available() else 'cpu' UpperCamelCase = AutoModelForCausalLM.from_pretrained(UpperCamelCase__ ) UpperCamelCase = model.to(UpperCamelCase__ ) UpperCamelCase = AutoTokenizer.from_pretrained(UpperCamelCase__ ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: UpperCamelCase = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(UpperCamelCase__ ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({'pad_token': existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" UpperCamelCase = model.config.max_length - 1 else: UpperCamelCase = model.config.max_length UpperCamelCase = tokenizer( UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , padding=UpperCamelCase__ , truncation=UpperCamelCase__ , max_length=UpperCamelCase__ , return_tensors='pt' , return_attention_mask=UpperCamelCase__ , ).to(UpperCamelCase__ ) UpperCamelCase = encodings['input_ids'] UpperCamelCase = encodings['attention_mask'] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." UpperCamelCase = [] UpperCamelCase = CrossEntropyLoss(reduction='none' ) for start_index in logging.tqdm(range(0 , len(UpperCamelCase__ ) , UpperCamelCase__ ) ): UpperCamelCase = min(start_index + batch_size , len(UpperCamelCase__ ) ) UpperCamelCase = encoded_texts[start_index:end_index] UpperCamelCase = attn_masks[start_index:end_index] if add_start_token: UpperCamelCase = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(UpperCamelCase__ ) UpperCamelCase = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) UpperCamelCase = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(UpperCamelCase__ ), attn_mask] , dim=1 ) UpperCamelCase = encoded_batch with torch.no_grad(): UpperCamelCase = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ ).logits UpperCamelCase = out_logits[..., :-1, :].contiguous() UpperCamelCase = labels[..., 1:].contiguous() UpperCamelCase = attn_mask[..., 1:].contiguous() UpperCamelCase = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , UpperCamelCase__ ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(UpperCamelCase__ )}
28
1
from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy __UpperCAmelCase = logging.get_logger(__name__) class __a ( _a ): def __init__( self : Union[str, Any] , UpperCAmelCase : int , UpperCAmelCase : int , UpperCAmelCase : float , **UpperCAmelCase : str ): lowerCAmelCase_ : List[str] = feature_size lowerCAmelCase_ : str = sampling_rate lowerCAmelCase_ : Dict = padding_value lowerCAmelCase_ : List[str] = kwargs.pop("""padding_side""" , """right""" ) lowerCAmelCase_ : List[Any] = kwargs.pop("""return_attention_mask""" , __lowerCamelCase ) super().__init__(**__lowerCamelCase ) def A ( self : int , UpperCAmelCase : Union[ BatchFeature, List[BatchFeature], Dict[str, BatchFeature], Dict[str, List[BatchFeature]], List[Dict[str, BatchFeature]], ] , UpperCAmelCase : Union[bool, str, PaddingStrategy] = True , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : bool = False , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : Optional[bool] = None , UpperCAmelCase : Optional[Union[str, TensorType]] = None , ): # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(__lowerCamelCase , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): lowerCAmelCase_ : Optional[Any] = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( """You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`""" F' to this method that includes {self.model_input_names[0]}, but you provided' F' {list(processed_features.keys() )}' ) lowerCAmelCase_ : Optional[int] = processed_features[self.model_input_names[0]] lowerCAmelCase_ : int = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(__lowerCamelCase ) == 0: if return_attention_mask: lowerCAmelCase_ : Any = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch lowerCAmelCase_ : List[str] = required_input[0] if isinstance(__lowerCamelCase , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. lowerCAmelCase_ : Optional[Any] = 0 while len(required_input[index] ) == 0: index += 1 if index < len(__lowerCamelCase ): lowerCAmelCase_ : str = required_input[index][0] if return_tensors is None: if is_tf_tensor(__lowerCamelCase ): lowerCAmelCase_ : int = """tf""" elif is_torch_tensor(__lowerCamelCase ): lowerCAmelCase_ : Optional[Any] = """pt""" elif isinstance(__lowerCamelCase , (int, float, list, tuple, np.ndarray) ): lowerCAmelCase_ : List[str] = """np""" else: raise ValueError( F'type of {first_element} unknown: {type(__lowerCamelCase )}. ' """Should be one of a python, numpy, pytorch or tensorflow object.""" ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): lowerCAmelCase_ : str = to_numpy(__lowerCamelCase ) else: lowerCAmelCase_ : int = [to_numpy(__lowerCamelCase ) for v in value] # Convert padding_strategy in PaddingStrategy lowerCAmelCase_ : Dict = self._get_padding_strategies(padding=__lowerCamelCase , max_length=__lowerCamelCase ) lowerCAmelCase_ : str = processed_features[self.model_input_names[0]] lowerCAmelCase_ : int = len(__lowerCamelCase ) if not all(len(__lowerCamelCase ) == batch_size for v in processed_features.values() ): raise ValueError("""Some items in the output dictionary have a different batch size than others.""" ) lowerCAmelCase_ : Optional[Any] = [] for i in range(__lowerCamelCase ): lowerCAmelCase_ : Union[str, Any] = {k: v[i] for k, v in processed_features.items()} # truncation lowerCAmelCase_ : List[str] = self._truncate( __lowerCamelCase , max_length=__lowerCamelCase , pad_to_multiple_of=__lowerCamelCase , truncation=__lowerCamelCase , ) truncated_inputs.append(__lowerCamelCase ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length lowerCAmelCase_ : str = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) lowerCAmelCase_ : Union[str, Any] = PaddingStrategy.MAX_LENGTH lowerCAmelCase_ : Dict = {} for i in range(__lowerCamelCase ): # padding lowerCAmelCase_ : Optional[int] = self._pad( truncated_inputs[i] , max_length=__lowerCamelCase , padding_strategy=__lowerCamelCase , pad_to_multiple_of=__lowerCamelCase , return_attention_mask=__lowerCamelCase , ) for key, value in outputs.items(): if key not in batch_outputs: lowerCAmelCase_ : List[str] = [] if value.dtype is np.dtype(np.floataa ): lowerCAmelCase_ : List[str] = value.astype(np.floataa ) batch_outputs[key].append(__lowerCamelCase ) return BatchFeature(__lowerCamelCase , tensor_type=__lowerCamelCase ) def A ( self : str , UpperCAmelCase : Union[Dict[str, np.ndarray], BatchFeature] , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : PaddingStrategy = PaddingStrategy.DO_NOT_PAD , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : Optional[bool] = None , ): lowerCAmelCase_ : Optional[Any] = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: lowerCAmelCase_ : List[Any] = len(__lowerCamelCase ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): lowerCAmelCase_ : Tuple = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of lowerCAmelCase_ : Union[str, Any] = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(__lowerCamelCase ) < max_length if return_attention_mask and "attention_mask" not in processed_features: lowerCAmelCase_ : Union[str, Any] = np.ones(len(__lowerCamelCase ) , dtype=np.intaa ) if needs_to_be_padded: lowerCAmelCase_ : Tuple = max_length - len(__lowerCamelCase ) if self.padding_side == "right": if return_attention_mask: lowerCAmelCase_ : Tuple = np.pad( processed_features["""attention_mask"""] , (0, difference) ) lowerCAmelCase_ : int = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) lowerCAmelCase_ : Tuple = np.pad( __lowerCamelCase , __lowerCamelCase , """constant""" , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: lowerCAmelCase_ : Optional[Any] = np.pad( processed_features["""attention_mask"""] , (difference, 0) ) lowerCAmelCase_ : Tuple = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) lowerCAmelCase_ : Tuple = np.pad( __lowerCamelCase , __lowerCamelCase , """constant""" , constant_values=self.padding_value ) else: raise ValueError("""Invalid padding strategy:""" + str(self.padding_side ) ) return processed_features def A ( self : Any , UpperCAmelCase : Union[Dict[str, np.ndarray], BatchFeature] , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : Optional[int] = None , UpperCAmelCase : Optional[bool] = None , ): if not truncation: return processed_features elif truncation and max_length is None: raise ValueError("""When setting ``truncation=True``, make sure that ``max_length`` is defined.""" ) lowerCAmelCase_ : Optional[Any] = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): lowerCAmelCase_ : Optional[Any] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of lowerCAmelCase_ : Any = len(__lowerCamelCase ) > max_length if needs_to_be_truncated: lowerCAmelCase_ : str = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: lowerCAmelCase_ : List[Any] = processed_features["""attention_mask"""][:max_length] return processed_features def A ( self : str , UpperCAmelCase : Tuple=False , UpperCAmelCase : Optional[int]=None ): # Get padding strategy if padding is not False: if padding is True: lowerCAmelCase_ : List[Any] = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(__lowerCamelCase , __lowerCamelCase ): lowerCAmelCase_ : Optional[int] = PaddingStrategy(__lowerCamelCase ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): lowerCAmelCase_ : Any = padding else: lowerCAmelCase_ : Dict = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F'When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined' ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( """Asking to pad but the feature_extractor does not have a padding value. Please select a value to use""" """ as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.""" ) return padding_strategy
367
from datetime import datetime as dt import os from github import Github __UpperCAmelCase = [ 'good first issue', 'good second issue', 'good difficult issue', 'feature request', 'new model', 'wip', ] def __UpperCamelCase ( ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : Dict = Github(os.environ["""GITHUB_TOKEN"""] ) lowerCAmelCase_ : Tuple = g.get_repo("""huggingface/transformers""" ) lowerCAmelCase_ : Any = repo.get_issues(state="""open""" ) for issue in open_issues: lowerCAmelCase_ : Union[str, Any] = sorted([comment for comment in issue.get_comments()] , key=lambda lowercase__ : i.created_at , reverse=lowercase__ ) lowerCAmelCase_ : str = comments[0] if len(lowercase__ ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state="""closed""" ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( """This issue has been automatically marked as stale because it has not had """ """recent activity. If you think this still needs to be addressed """ """please comment on this thread.\n\nPlease note that issues that do not follow the """ """[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) """ """are likely to be ignored.""" ) if __name__ == "__main__": main()
28
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class __lowerCamelCase : '''simple docstring''' a_ : List[str] = XGLMConfig a_ : Optional[Any] = {} a_ : Optional[int] = """gelu""" def __init__( self : Union[str, Any] , a_ : str , a_ : Optional[Any]=14 , a_ : Union[str, Any]=7 , a_ : List[Any]=True , a_ : Any=True , a_ : int=True , a_ : Any=99 , a_ : Union[str, Any]=32 , a_ : List[Any]=2 , a_ : Optional[Any]=4 , a_ : Optional[Any]=37 , a_ : int="gelu" , a_ : int=0.1 , a_ : Any=0.1 , a_ : List[str]=5_12 , a_ : List[Any]=0.02 , ): lowerCAmelCase_ : Tuple = parent lowerCAmelCase_ : Optional[Any] = batch_size lowerCAmelCase_ : List[Any] = seq_length lowerCAmelCase_ : Union[str, Any] = is_training lowerCAmelCase_ : Tuple = use_input_mask lowerCAmelCase_ : List[str] = use_labels lowerCAmelCase_ : Union[str, Any] = vocab_size lowerCAmelCase_ : Tuple = d_model lowerCAmelCase_ : Any = num_hidden_layers lowerCAmelCase_ : Union[str, Any] = num_attention_heads lowerCAmelCase_ : Optional[Any] = ffn_dim lowerCAmelCase_ : List[Any] = activation_function lowerCAmelCase_ : str = activation_dropout lowerCAmelCase_ : List[str] = attention_dropout lowerCAmelCase_ : List[str] = max_position_embeddings lowerCAmelCase_ : Dict = initializer_range lowerCAmelCase_ : List[Any] = None lowerCAmelCase_ : str = 0 lowerCAmelCase_ : Any = 2 lowerCAmelCase_ : Any = 1 def lowerCamelCase ( self : int ): return XGLMConfig.from_pretrained("facebook/xglm-564M" ) def lowerCamelCase ( self : List[str] ): lowerCAmelCase_ : Union[str, Any] = tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) lowerCAmelCase_ : List[str] = None if self.use_input_mask: lowerCAmelCase_ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase_ : str = self.get_config() lowerCAmelCase_ : List[Any] = floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def lowerCamelCase ( self : str ): return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=__lowerCamelCase , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=__lowerCamelCase , ) def lowerCamelCase ( self : Union[str, Any] ): lowerCAmelCase_ : Optional[Any] = self.prepare_config_and_inputs() ( lowerCAmelCase_ ) : Optional[Any] = config_and_inputs lowerCAmelCase_ : str = { """input_ids""": input_ids, """head_mask""": head_mask, } return config, inputs_dict @require_tf class __lowerCamelCase ( _a , _a , unittest.TestCase ): '''simple docstring''' a_ : int = (TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () a_ : Dict = (TFXGLMForCausalLM,) if is_tf_available() else () a_ : Optional[Any] = ( {"""feature-extraction""": TFXGLMModel, """text-generation""": TFXGLMForCausalLM} if is_tf_available() else {} ) a_ : List[str] = False a_ : Optional[int] = False a_ : List[str] = False def lowerCamelCase ( self : List[Any] ): lowerCAmelCase_ : Optional[Any] = TFXGLMModelTester(self ) lowerCAmelCase_ : List[Any] = ConfigTester(self , config_class=__lowerCamelCase , n_embd=37 ) def lowerCamelCase ( self : int ): self.config_tester.run_common_tests() @slow def lowerCamelCase ( self : int ): for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : str = TFXGLMModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) @unittest.skip(reason="Currently, model embeddings are going to undergo a major refactor." ) def lowerCamelCase ( self : Dict ): super().test_resize_token_embeddings() @require_tf class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def lowerCamelCase ( self : Optional[Any] , a_ : Optional[int]=True ): lowerCAmelCase_ : Dict = TFXGLMForCausalLM.from_pretrained("facebook/xglm-564M" ) lowerCAmelCase_ : str = tf.convert_to_tensor([[2, 2_68, 98_65]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off lowerCAmelCase_ : int = [2, 2_68, 98_65, 67, 11, 19_88, 5_72_52, 98_65, 5, 9_84, 67, 19_88, 21_38_38, 16_58, 53, 7_04_46, 33, 66_57, 2_78, 15_81] # fmt: on lowerCAmelCase_ : str = model.generate(__lowerCamelCase , do_sample=__lowerCamelCase , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , __lowerCamelCase ) @slow def lowerCamelCase ( self : Tuple ): lowerCAmelCase_ : Tuple = XGLMTokenizer.from_pretrained("facebook/xglm-564M" ) lowerCAmelCase_ : Tuple = TFXGLMForCausalLM.from_pretrained("facebook/xglm-564M" ) tf.random.set_seed(0 ) lowerCAmelCase_ : Optional[int] = tokenizer("Today is a nice day and" , return_tensors="tf" ) lowerCAmelCase_ : str = tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(":/CPU:0" ): lowerCAmelCase_ : Any = model.generate(__lowerCamelCase , do_sample=__lowerCamelCase , seed=[7, 0] ) lowerCAmelCase_ : str = tokenizer.decode(output_ids[0] , skip_special_tokens=__lowerCamelCase ) lowerCAmelCase_ : Union[str, Any] = ( """Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due""" ) self.assertEqual(__lowerCamelCase , __lowerCamelCase ) @slow def lowerCamelCase ( self : int ): lowerCAmelCase_ : List[str] = TFXGLMForCausalLM.from_pretrained("facebook/xglm-564M" ) lowerCAmelCase_ : Tuple = XGLMTokenizer.from_pretrained("facebook/xglm-564M" ) lowerCAmelCase_ : str = """left""" # use different length sentences to test batching lowerCAmelCase_ : Union[str, Any] = [ """This is an extremelly long sentence that only exists to test the ability of the model to cope with """ """left-padding, such as in batched generation. The output for the sequence below should be the same """ """regardless of whether left padding is applied or not. When""", """Hello, my dog is a little""", ] lowerCAmelCase_ : str = tokenizer(__lowerCamelCase , return_tensors="tf" , padding=__lowerCamelCase ) lowerCAmelCase_ : Optional[int] = inputs["""input_ids"""] lowerCAmelCase_ : Dict = model.generate(input_ids=__lowerCamelCase , attention_mask=inputs["attention_mask"] , max_new_tokens=12 ) lowerCAmelCase_ : List[str] = tokenizer(sentences[0] , return_tensors="tf" ).input_ids lowerCAmelCase_ : List[Any] = model.generate(input_ids=__lowerCamelCase , max_new_tokens=12 ) lowerCAmelCase_ : str = tokenizer(sentences[1] , return_tensors="tf" ).input_ids lowerCAmelCase_ : str = model.generate(input_ids=__lowerCamelCase , max_new_tokens=12 ) lowerCAmelCase_ : Any = tokenizer.batch_decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) lowerCAmelCase_ : Union[str, Any] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__lowerCamelCase ) lowerCAmelCase_ : List[Any] = tokenizer.decode(output_padded[0] , skip_special_tokens=__lowerCamelCase ) lowerCAmelCase_ : List[Any] = [ """This is an extremelly long sentence that only exists to test the ability of the model to cope with """ """left-padding, such as in batched generation. The output for the sequence below should be the same """ """regardless of whether left padding is applied or not. When left padding is applied, the sequence will be """ """a single""", """Hello, my dog is a little bit of a shy one, but he is very friendly""", ] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) self.assertListEqual(__lowerCamelCase , [non_padded_sentence, padded_sentence] )
241
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _SCREAMING_SNAKE_CASE ( _a , unittest.TestCase ): snake_case__ : Optional[Any] = TransfoXLTokenizer snake_case__ : List[Any] = False snake_case__ : Tuple = False def _A ( self : str ): super().setUp() UpperCamelCase :Dict = [ """<unk>""", """[CLS]""", """[SEP]""", """want""", """unwanted""", """wa""", """un""", """running""", """,""", """low""", """l""", ] UpperCamelCase :str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def _A ( self : List[str] , **__lowerCamelCase : Any ): UpperCamelCase :Any = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _A ( self : Any , __lowerCamelCase : int ): UpperCamelCase :List[Any] = """<unk> UNwanted , running""" UpperCamelCase :int = """<unk> unwanted, running""" return input_text, output_text def _A ( self : Tuple ): UpperCamelCase :List[str] = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=__lowerCamelCase ) UpperCamelCase :Any = tokenizer.tokenize("""<unk> UNwanted , running""" ) self.assertListEqual(__lowerCamelCase , ["""<unk>""", """unwanted""", """,""", """running"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [0, 4, 8, 7] ) def _A ( self : Optional[Any] ): UpperCamelCase :List[Any] = TransfoXLTokenizer(lower_case=__lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo ! how \n Are yoU ? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) def _A ( self : Union[str, Any] ): UpperCamelCase :int = TransfoXLTokenizer(lower_case=__lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo ! how \n Are yoU ? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _A ( self : Tuple ): UpperCamelCase :Any = TransfoXLTokenizer(lower_case=__lowerCamelCase ) UpperCamelCase :Optional[int] = """Hello (bracket) and side-scrolled [and] Henry's $5,000 with 3.34 m. What's up!?""" UpperCamelCase :Optional[int] = [ """Hello""", """(""", """bracket""", """)""", """and""", """side""", """@-@""", """scrolled""", """[""", """and""", """]""", """Henry""", """'s""", """$""", """5""", """@,@""", """000""", """with""", """3""", """@.@""", """34""", """m""", """.""", """What""", """'s""", """up""", """!""", """?""", ] self.assertListEqual(tokenizer.tokenize(__lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(tokenizer.convert_tokens_to_string(__lowerCamelCase ) , __lowerCamelCase ) def _A ( self : List[Any] ): UpperCamelCase :Any = self.get_tokenizer() UpperCamelCase :List[str] = len(__lowerCamelCase ) tokenizer.add_tokens(["""new1""", """new2"""] ) tokenizer.move_added_token("""new1""" , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(__lowerCamelCase ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode("""new1""" ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , """new1""" )
38
0
import argparse import json import pickle from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import MaskFormerConfig, MaskFormerForInstanceSegmentation, MaskFormerImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() __lowercase = logging.get_logger(__name__) def lowerCamelCase ( SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = SwinConfig.from_pretrained( '''microsoft/swin-tiny-patch4-window7-224''' , out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] ) __UpperCamelCase :Optional[Any] = MaskFormerConfig(backbone_config=SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = '''huggingface/label-files''' if "ade20k-full" in model_name: # this should be ok __UpperCamelCase :List[Any] = 847 __UpperCamelCase :str = '''maskformer-ade20k-full-id2label.json''' elif "ade" in model_name: # this should be ok __UpperCamelCase :List[str] = 150 __UpperCamelCase :int = '''ade20k-id2label.json''' elif "coco-stuff" in model_name: # this should be ok __UpperCamelCase :Any = 171 __UpperCamelCase :int = '''maskformer-coco-stuff-id2label.json''' elif "coco" in model_name: # TODO __UpperCamelCase :List[str] = 133 __UpperCamelCase :List[str] = '''coco-panoptic-id2label.json''' elif "cityscapes" in model_name: # this should be ok __UpperCamelCase :Optional[Any] = 19 __UpperCamelCase :Optional[int] = '''cityscapes-id2label.json''' elif "vistas" in model_name: # this should be ok __UpperCamelCase :Dict = 65 __UpperCamelCase :Optional[Any] = '''mapillary-vistas-id2label.json''' __UpperCamelCase :List[Any] = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , repo_type='''dataset''' ) , '''r''' ) ) __UpperCamelCase :List[Any] = {int(SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} return config def lowerCamelCase ( SCREAMING_SNAKE_CASE : Tuple ): '''simple docstring''' __UpperCamelCase :int = [] # stem # fmt: off rename_keys.append(('''backbone.patch_embed.proj.weight''', '''model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.weight''') ) rename_keys.append(('''backbone.patch_embed.proj.bias''', '''model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.bias''') ) rename_keys.append(('''backbone.patch_embed.norm.weight''', '''model.pixel_level_module.encoder.model.embeddings.norm.weight''') ) rename_keys.append(('''backbone.patch_embed.norm.bias''', '''model.pixel_level_module.encoder.model.embeddings.norm.bias''') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.norm1.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.norm1.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.attn.relative_position_bias_table""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.attn.relative_position_index""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.attn.proj.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.attn.proj.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.norm2.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.norm2.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.mlp.fc1.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.mlp.fc1.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.mlp.fc2.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.mlp.fc2.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.bias""") ) if i < 3: rename_keys.append((f"""backbone.layers.{i}.downsample.reduction.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.reduction.weight""") ) rename_keys.append((f"""backbone.layers.{i}.downsample.norm.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.weight""") ) rename_keys.append((f"""backbone.layers.{i}.downsample.norm.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.bias""") ) rename_keys.append((f"""backbone.norm{i}.weight""", f"""model.pixel_level_module.encoder.hidden_states_norms.{i}.weight""") ) rename_keys.append((f"""backbone.norm{i}.bias""", f"""model.pixel_level_module.encoder.hidden_states_norms.{i}.bias""") ) # FPN rename_keys.append(('''sem_seg_head.layer_4.weight''', '''model.pixel_level_module.decoder.fpn.stem.0.weight''') ) rename_keys.append(('''sem_seg_head.layer_4.norm.weight''', '''model.pixel_level_module.decoder.fpn.stem.1.weight''') ) rename_keys.append(('''sem_seg_head.layer_4.norm.bias''', '''model.pixel_level_module.decoder.fpn.stem.1.bias''') ) for source_index, target_index in zip(range(3 , 0 , -1 ) , range(0 , 3 ) ): rename_keys.append((f"""sem_seg_head.adapter_{source_index}.weight""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.0.weight""") ) rename_keys.append((f"""sem_seg_head.adapter_{source_index}.norm.weight""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.weight""") ) rename_keys.append((f"""sem_seg_head.adapter_{source_index}.norm.bias""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.bias""") ) rename_keys.append((f"""sem_seg_head.layer_{source_index}.weight""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.block.0.weight""") ) rename_keys.append((f"""sem_seg_head.layer_{source_index}.norm.weight""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.weight""") ) rename_keys.append((f"""sem_seg_head.layer_{source_index}.norm.bias""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.bias""") ) rename_keys.append(('''sem_seg_head.mask_features.weight''', '''model.pixel_level_module.decoder.mask_projection.weight''') ) rename_keys.append(('''sem_seg_head.mask_features.bias''', '''model.pixel_level_module.decoder.mask_projection.bias''') ) # Transformer decoder for idx in range(config.decoder_config.decoder_layers ): # self-attention out projection rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.weight""", f"""model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.bias""", f"""model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.bias""") ) # cross-attention out projection rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.weight""", f"""model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.bias""", f"""model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.bias""") ) # MLP 1 rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.weight""", f"""model.transformer_module.decoder.layers.{idx}.fc1.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.bias""", f"""model.transformer_module.decoder.layers.{idx}.fc1.bias""") ) # MLP 2 rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.weight""", f"""model.transformer_module.decoder.layers.{idx}.fc2.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.bias""", f"""model.transformer_module.decoder.layers.{idx}.fc2.bias""") ) # layernorm 1 (self-attention layernorm) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.weight""", f"""model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.bias""", f"""model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.bias""") ) # layernorm 2 (cross-attention layernorm) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.weight""", f"""model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.bias""", f"""model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.bias""") ) # layernorm 3 (final layernorm) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.weight""", f"""model.transformer_module.decoder.layers.{idx}.final_layer_norm.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.bias""", f"""model.transformer_module.decoder.layers.{idx}.final_layer_norm.bias""") ) rename_keys.append(('''sem_seg_head.predictor.transformer.decoder.norm.weight''', '''model.transformer_module.decoder.layernorm.weight''') ) rename_keys.append(('''sem_seg_head.predictor.transformer.decoder.norm.bias''', '''model.transformer_module.decoder.layernorm.bias''') ) # heads on top rename_keys.append(('''sem_seg_head.predictor.query_embed.weight''', '''model.transformer_module.queries_embedder.weight''') ) rename_keys.append(('''sem_seg_head.predictor.input_proj.weight''', '''model.transformer_module.input_projection.weight''') ) rename_keys.append(('''sem_seg_head.predictor.input_proj.bias''', '''model.transformer_module.input_projection.bias''') ) rename_keys.append(('''sem_seg_head.predictor.class_embed.weight''', '''class_predictor.weight''') ) rename_keys.append(('''sem_seg_head.predictor.class_embed.bias''', '''class_predictor.bias''') ) for i in range(3 ): rename_keys.append((f"""sem_seg_head.predictor.mask_embed.layers.{i}.weight""", f"""mask_embedder.{i}.0.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.mask_embed.layers.{i}.bias""", f"""mask_embedder.{i}.0.bias""") ) # fmt: on return rename_keys def lowerCamelCase ( SCREAMING_SNAKE_CASE : Any , SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' __UpperCamelCase :Tuple = dct.pop(SCREAMING_SNAKE_CASE ) __UpperCamelCase :str = val def lowerCamelCase ( SCREAMING_SNAKE_CASE : List[str] , SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' __UpperCamelCase :List[Any] = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): __UpperCamelCase :Any = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) __UpperCamelCase :str = state_dict.pop(f"""backbone.layers.{i}.blocks.{j}.attn.qkv.weight""" ) __UpperCamelCase :Any = state_dict.pop(f"""backbone.layers.{i}.blocks.{j}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict __UpperCamelCase :Union[str, Any] = in_proj_weight[:dim, :] __UpperCamelCase :Optional[Any] = in_proj_bias[: dim] __UpperCamelCase :Union[str, Any] = in_proj_weight[ dim : dim * 2, : ] __UpperCamelCase :Any = in_proj_bias[ dim : dim * 2 ] __UpperCamelCase :Union[str, Any] = in_proj_weight[ -dim :, : ] __UpperCamelCase :Union[str, Any] = in_proj_bias[-dim :] # fmt: on def lowerCamelCase ( SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' __UpperCamelCase :Optional[int] = config.decoder_config.hidden_size for idx in range(config.decoder_config.decoder_layers ): # read in weights + bias of self-attention input projection layer (in the original implementation, this is a single matrix + bias) __UpperCamelCase :List[str] = state_dict.pop(f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_weight""" ) __UpperCamelCase :Optional[Any] = state_dict.pop(f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict __UpperCamelCase :int = in_proj_weight[: hidden_size, :] __UpperCamelCase :List[str] = in_proj_bias[:config.hidden_size] __UpperCamelCase :Optional[Any] = in_proj_weight[hidden_size : hidden_size * 2, :] __UpperCamelCase :Optional[Any] = in_proj_bias[hidden_size : hidden_size * 2] __UpperCamelCase :Tuple = in_proj_weight[-hidden_size :, :] __UpperCamelCase :List[Any] = in_proj_bias[-hidden_size :] # read in weights + bias of cross-attention input projection layer (in the original implementation, this is a single matrix + bias) __UpperCamelCase :int = state_dict.pop(f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_weight""" ) __UpperCamelCase :Any = state_dict.pop(f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict __UpperCamelCase :List[str] = in_proj_weight[: hidden_size, :] __UpperCamelCase :List[Any] = in_proj_bias[:config.hidden_size] __UpperCamelCase :Optional[Any] = in_proj_weight[hidden_size : hidden_size * 2, :] __UpperCamelCase :Tuple = in_proj_bias[hidden_size : hidden_size * 2] __UpperCamelCase :str = in_proj_weight[-hidden_size :, :] __UpperCamelCase :List[str] = in_proj_bias[-hidden_size :] # fmt: on def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __UpperCamelCase :Any = Image.open(requests.get(SCREAMING_SNAKE_CASE , stream=SCREAMING_SNAKE_CASE ).raw ) return im @torch.no_grad() def lowerCamelCase ( SCREAMING_SNAKE_CASE : Any , SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : List[Any] = False ): '''simple docstring''' __UpperCamelCase :Optional[int] = get_maskformer_config(SCREAMING_SNAKE_CASE ) # load original state_dict with open(SCREAMING_SNAKE_CASE , '''rb''' ) as f: __UpperCamelCase :Optional[int] = pickle.load(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = data['''model'''] # for name, param in state_dict.items(): # print(name, param.shape) # rename keys __UpperCamelCase :Optional[int] = create_rename_keys(SCREAMING_SNAKE_CASE ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) read_in_swin_q_k_v(SCREAMING_SNAKE_CASE , config.backbone_config ) read_in_decoder_q_k_v(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # update to torch tensors for key, value in state_dict.items(): __UpperCamelCase :int = torch.from_numpy(SCREAMING_SNAKE_CASE ) # load 🤗 model __UpperCamelCase :Union[str, Any] = MaskFormerForInstanceSegmentation(SCREAMING_SNAKE_CASE ) model.eval() for name, param in model.named_parameters(): print(SCREAMING_SNAKE_CASE , param.shape ) __UpperCamelCase :List[str] = model.load_state_dict(SCREAMING_SNAKE_CASE , strict=SCREAMING_SNAKE_CASE ) assert missing_keys == [ "model.pixel_level_module.encoder.model.layernorm.weight", "model.pixel_level_module.encoder.model.layernorm.bias", ] assert len(SCREAMING_SNAKE_CASE ) == 0, f"""Unexpected keys: {unexpected_keys}""" # verify results __UpperCamelCase :Tuple = prepare_img() if "vistas" in model_name: __UpperCamelCase :Dict = 65 elif "cityscapes" in model_name: __UpperCamelCase :Dict = 65_535 else: __UpperCamelCase :Tuple = 255 __UpperCamelCase :str = True if '''ade''' in model_name else False __UpperCamelCase :Dict = MaskFormerImageProcessor(ignore_index=SCREAMING_SNAKE_CASE , reduce_labels=SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = image_processor(SCREAMING_SNAKE_CASE , return_tensors='''pt''' ) __UpperCamelCase :List[Any] = model(**SCREAMING_SNAKE_CASE ) print('''Logits:''' , outputs.class_queries_logits[0, :3, :3] ) if model_name == "maskformer-swin-tiny-ade": __UpperCamelCase :List[str] = torch.tensor( [[3.6_353, -4.4_770, -2.6_065], [0.5_081, -4.2_394, -3.5_343], [2.1_909, -5.0_353, -1.9_323]] ) assert torch.allclose(outputs.class_queries_logits[0, :3, :3] , SCREAMING_SNAKE_CASE , atol=1e-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(f"""Saving model and image processor to {pytorch_dump_folder_path}""" ) Path(SCREAMING_SNAKE_CASE ).mkdir(exist_ok=SCREAMING_SNAKE_CASE ) model.save_pretrained(SCREAMING_SNAKE_CASE ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE ) if push_to_hub: print('''Pushing model and image processor to the hub...''' ) model.push_to_hub(f"""nielsr/{model_name}""" ) image_processor.push_to_hub(f"""nielsr/{model_name}""" ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''maskformer-swin-tiny-ade''', type=str, help=('''Name of the MaskFormer model you\'d like to convert''',), ) parser.add_argument( '''--checkpoint_path''', default='''/Users/nielsrogge/Documents/MaskFormer_checkpoints/MaskFormer-Swin-tiny-ADE20k/model.pkl''', type=str, help='''Path to the original state dict (.pth file).''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) __lowercase = parser.parse_args() convert_maskformer_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
359
import os import pytest from transformers.dynamic_module_utils import get_imports __lowercase = ''' import os ''' __lowercase = ''' def foo(): import os return False ''' __lowercase = ''' def foo(): def bar(): if True: import os return False return bar() ''' __lowercase = ''' import os try: import bar except ImportError: raise ValueError() ''' __lowercase = ''' import os def foo(): try: import bar except ImportError: raise ValueError() ''' __lowercase = ''' import os try: import bar except (ImportError, AttributeError): raise ValueError() ''' __lowercase = ''' import os try: import bar except ImportError as e: raise ValueError() ''' __lowercase = ''' import os try: import bar except: raise ValueError() ''' __lowercase = ''' import os try: import bar import baz except ImportError: raise ValueError() ''' __lowercase = ''' import os try: import bar import baz except ImportError: x = 1 raise ValueError() ''' __lowercase = [ TOP_LEVEL_IMPORT, IMPORT_IN_FUNCTION, DEEPLY_NESTED_IMPORT, TOP_LEVEL_TRY_IMPORT, GENERIC_EXCEPT_IMPORT, MULTILINE_TRY_IMPORT, MULTILINE_BOTH_IMPORT, MULTIPLE_EXCEPTS_IMPORT, EXCEPT_AS_IMPORT, TRY_IMPORT_IN_FUNCTION, ] @pytest.mark.parametrize('''case''' , SCREAMING_SNAKE_CASE ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = os.path.join(SCREAMING_SNAKE_CASE , '''test_file.py''' ) with open(SCREAMING_SNAKE_CASE , '''w''' ) as _tmp_file: _tmp_file.write(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Dict = get_imports(SCREAMING_SNAKE_CASE ) assert parsed_imports == ["os"]
105
0