code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class a_ ( a , a , unittest.TestCase ): A__ : Union[str, Any] = CycleDiffusionPipeline A__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { 'negative_prompt', 'height', 'width', 'negative_prompt_embeds', } A__ : Union[str, Any] = PipelineTesterMixin.required_optional_params - {'latents'} A__ : Dict = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'source_prompt'} ) A__ : str = IMAGE_TO_IMAGE_IMAGE_PARAMS A__ : List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowerCAmelCase( self : Dict ): """simple docstring""" torch.manual_seed(0 ) snake_case : Optional[int] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) snake_case : int = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , num_train_timesteps=1_000 , clip_sample=UpperCAmelCase__ , set_alpha_to_one=UpperCAmelCase__ , ) torch.manual_seed(0 ) snake_case : Tuple = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) torch.manual_seed(0 ) snake_case : Any = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) snake_case : List[str] = CLIPTextModel(UpperCAmelCase__ ) snake_case : Union[str, Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) snake_case : str = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Dict=0 ): """simple docstring""" snake_case : List[Any] = floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCAmelCase__ ) ).to(UpperCAmelCase__ ) snake_case : str = image / 2 + 0.5 if str(UpperCAmelCase__ ).startswith('''mps''' ): snake_case : Any = torch.manual_seed(UpperCAmelCase__ ) else: snake_case : Dict = torch.Generator(device=UpperCAmelCase__ ).manual_seed(UpperCAmelCase__ ) snake_case : Tuple = { '''prompt''': '''An astronaut riding an elephant''', '''source_prompt''': '''An astronaut riding a horse''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''eta''': 0.1, '''strength''': 0.8, '''guidance_scale''': 3, '''source_guidance_scale''': 1, '''output_type''': '''numpy''', } return inputs def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator snake_case : Dict = self.get_dummy_components() snake_case : List[str] = CycleDiffusionPipeline(**UpperCAmelCase__ ) snake_case : Dict = pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) snake_case : int = self.get_dummy_inputs(UpperCAmelCase__ ) snake_case : Optional[int] = pipe(**UpperCAmelCase__ ) snake_case : str = output.images snake_case : str = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) snake_case : Optional[int] = np.array([0.4459, 0.4943, 0.4544, 0.6643, 0.5474, 0.4327, 0.5701, 0.5959, 0.5179] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Optional[int] = self.get_dummy_components() for name, module in components.items(): if hasattr(UpperCAmelCase__ , '''half''' ): snake_case : List[str] = module.half() snake_case : List[str] = CycleDiffusionPipeline(**UpperCAmelCase__ ) snake_case : Tuple = pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) snake_case : Tuple = self.get_dummy_inputs(UpperCAmelCase__ ) snake_case : Optional[int] = pipe(**UpperCAmelCase__ ) snake_case : Optional[int] = output.images snake_case : List[Any] = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) snake_case : str = np.array([0.3506, 0.4543, 0.446, 0.4575, 0.5195, 0.4155, 0.5273, 0.518, 0.4116] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def lowerCAmelCase( self : Tuple ): """simple docstring""" return super().test_save_load_local() @unittest.skip('''non-deterministic pipeline''' ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" return super().test_inference_batch_single_identical() @skip_mps def lowerCAmelCase( self : Dict ): """simple docstring""" return super().test_dict_tuple_outputs_equivalent() @skip_mps def lowerCAmelCase( self : int ): """simple docstring""" return super().test_save_load_optional_components() @skip_mps def lowerCAmelCase( self : int ): """simple docstring""" return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class a_ ( unittest.TestCase ): def lowerCAmelCase( self : int ): """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Any = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/cycle-diffusion/black_colored_car.png''' ) snake_case : Tuple = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy''' ) snake_case : List[Any] = init_image.resize((512, 512) ) snake_case : Dict = '''CompVis/stable-diffusion-v1-4''' snake_case : Any = DDIMScheduler.from_pretrained(UpperCAmelCase__ , subfolder='''scheduler''' ) snake_case : List[Any] = CycleDiffusionPipeline.from_pretrained( UpperCAmelCase__ , scheduler=UpperCAmelCase__ , safety_checker=UpperCAmelCase__ , torch_dtype=torch.floataa , revision='''fp16''' ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) pipe.enable_attention_slicing() snake_case : Tuple = '''A black colored car''' snake_case : str = '''A blue colored car''' snake_case : Union[str, Any] = torch.manual_seed(0 ) snake_case : List[Any] = pipe( prompt=UpperCAmelCase__ , source_prompt=UpperCAmelCase__ , image=UpperCAmelCase__ , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=UpperCAmelCase__ , output_type='''np''' , ) snake_case : Any = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5e-1 def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Optional[int] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/cycle-diffusion/black_colored_car.png''' ) snake_case : Optional[int] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy''' ) snake_case : str = init_image.resize((512, 512) ) snake_case : Optional[Any] = '''CompVis/stable-diffusion-v1-4''' snake_case : Union[str, Any] = DDIMScheduler.from_pretrained(UpperCAmelCase__ , subfolder='''scheduler''' ) snake_case : Tuple = CycleDiffusionPipeline.from_pretrained(UpperCAmelCase__ , scheduler=UpperCAmelCase__ , safety_checker=UpperCAmelCase__ ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) pipe.enable_attention_slicing() snake_case : List[Any] = '''A black colored car''' snake_case : Tuple = '''A blue colored car''' snake_case : Optional[int] = torch.manual_seed(0 ) snake_case : Tuple = pipe( prompt=UpperCAmelCase__ , source_prompt=UpperCAmelCase__ , image=UpperCAmelCase__ , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=UpperCAmelCase__ , output_type='''np''' , ) snake_case : str = output.images assert np.abs(image - expected_image ).max() < 2e-2
84
import torch from diffusers import DiffusionPipeline class a_ ( a ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[Any] ): """simple docstring""" super().__init__() self.register_modules(unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ ) def __call__( self : Optional[int] ): """simple docstring""" snake_case : Any = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) snake_case : Dict = 1 snake_case : Optional[Any] = self.unet(UpperCAmelCase__ , UpperCAmelCase__ ).sample snake_case : List[Any] = self.scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample snake_case : List[Any] = scheduler_output - scheduler_output + torch.ones_like(UpperCAmelCase__ ) return result
84
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _a : List[str] = { 'configuration_owlvit': [ 'OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'OwlViTConfig', 'OwlViTOnnxConfig', 'OwlViTTextConfig', 'OwlViTVisionConfig', ], 'processing_owlvit': ['OwlViTProcessor'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Union[str, Any] = ['OwlViTFeatureExtractor'] _a : int = ['OwlViTImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[Any] = [ 'OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'OwlViTModel', 'OwlViTPreTrainedModel', 'OwlViTTextModel', 'OwlViTVisionModel', 'OwlViTForObjectDetection', ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _a : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
84
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class a_ ( a ): A__ : List[str] = ['image_processor', 'tokenizer'] A__ : Any = 'CLIPImageProcessor' A__ : Optional[int] = ('CLIPTokenizer', 'CLIPTokenizerFast') def __init__( self : Union[str, Any] , UpperCAmelCase__ : str=None , UpperCAmelCase__ : Union[str, Any]=None , **UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : Optional[int] = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , UpperCAmelCase__ , ) snake_case : List[Any] = kwargs.pop('''feature_extractor''' ) snake_case : Optional[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(UpperCAmelCase__ , UpperCAmelCase__ ) def __call__( self : Any , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : int=None , **UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: snake_case : int = self.tokenizer(UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , **UpperCAmelCase__ ) if images is not None: snake_case : Dict = self.image_processor(UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , **UpperCAmelCase__ ) if text is not None and images is not None: snake_case : Tuple = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**UpperCAmelCase__ ) , tensor_type=UpperCAmelCase__ ) def lowerCAmelCase( self : List[str] , *UpperCAmelCase__ : Dict , **UpperCAmelCase__ : int ): """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : str ): """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : int = self.tokenizer.model_input_names snake_case : Tuple = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , UpperCAmelCase__ , ) return self.image_processor_class @property def lowerCAmelCase( self : Any ): """simple docstring""" warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , UpperCAmelCase__ , ) return self.image_processor
84
1
import logging import os from typing import Dict, List, Optional, Union import torch import torch.nn as nn from accelerate.utils.imports import ( is_abit_bnb_available, is_abit_bnb_available, is_bnb_available, ) from ..big_modeling import dispatch_model, init_empty_weights from .dataclasses import BnbQuantizationConfig from .modeling import ( find_tied_parameters, get_balanced_memory, infer_auto_device_map, load_checkpoint_in_model, offload_weight, set_module_tensor_to_device, ) if is_bnb_available(): import bitsandbytes as bnb from copy import deepcopy _a : List[Any] = logging.getLogger(__name__) def a_ ( __magic_name__ , __magic_name__ , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = False , ) -> List[Any]: """simple docstring""" snake_case : str = bnb_quantization_config.load_in_abit snake_case : Tuple = bnb_quantization_config.load_in_abit if load_in_abit and not is_abit_bnb_available(): raise ImportError( '''You have a version of `bitsandbytes` that is not compatible with 8bit quantization,''' ''' make sure you have the latest version of `bitsandbytes` installed.''' ) if load_in_abit and not is_abit_bnb_available(): raise ValueError( '''You have a version of `bitsandbytes` that is not compatible with 4bit quantization,''' '''make sure you have the latest version of `bitsandbytes` installed.''' ) snake_case : Union[str, Any] = [] # custom device map if isinstance(__magic_name__ , __magic_name__ ) and len(device_map.keys() ) > 1: snake_case : Any = [key for key, value in device_map.items() if value in ['''disk''', '''cpu''']] # We keep some modules such as the lm_head in their original dtype for numerical stability reasons if bnb_quantization_config.skip_modules is None: snake_case : Union[str, Any] = get_keys_to_not_convert(__magic_name__ ) # add cpu modules to skip modules only for 4-bit modules if load_in_abit: bnb_quantization_config.skip_modules.extend(__magic_name__ ) snake_case : Union[str, Any] = bnb_quantization_config.skip_modules # We add the modules we want to keep in full precision if bnb_quantization_config.keep_in_fpaa_modules is None: snake_case : str = [] snake_case : List[Any] = bnb_quantization_config.keep_in_fpaa_modules modules_to_not_convert.extend(__magic_name__ ) # compatibility with peft snake_case : Tuple = load_in_abit snake_case : Dict = load_in_abit snake_case : Optional[Any] = get_parameter_device(__magic_name__ ) if model_device.type != "meta": # quantization of an already loaded model logger.warning( '''It is not recommended to quantize a loaded model. ''' '''The model should be instantiated under the `init_empty_weights` context manager.''' ) snake_case : List[Any] = replace_with_bnb_layers(__magic_name__ , __magic_name__ , modules_to_not_convert=__magic_name__ ) # convert param to the right dtype snake_case : Tuple = bnb_quantization_config.torch_dtype for name, param in model.state_dict().items(): if any(module_to_keep_in_fpaa in name for module_to_keep_in_fpaa in keep_in_fpaa_modules ): param.to(torch.floataa ) if param.dtype != torch.floataa: snake_case : Tuple = name.replace('''.weight''' , '''''' ).replace('''.bias''' , '''''' ) snake_case : Tuple = getattr(__magic_name__ , __magic_name__ , __magic_name__ ) if param is not None: param.to(torch.floataa ) elif torch.is_floating_point(__magic_name__ ): param.to(__magic_name__ ) if model_device.type == "cuda": # move everything to cpu in the first place because we can't do quantization if the weights are already on cuda model.cuda(torch.cuda.current_device() ) torch.cuda.empty_cache() elif torch.cuda.is_available(): model.to(torch.cuda.current_device() ) else: raise RuntimeError('''No GPU found. A GPU is needed for quantization.''' ) logger.info( F"The model device type is {model_device.type}. However, cuda is needed for quantization." '''We move the model to cuda.''' ) return model elif weights_location is None: raise RuntimeError( F"`weights_location` needs to be the folder path containing the weights of the model, but we found {weights_location} " ) else: with init_empty_weights(): snake_case : Optional[int] = replace_with_bnb_layers( __magic_name__ , __magic_name__ , modules_to_not_convert=__magic_name__ ) snake_case : Any = get_quantized_model_device_map( __magic_name__ , __magic_name__ , __magic_name__ , max_memory=__magic_name__ , no_split_module_classes=__magic_name__ , ) if offload_state_dict is None and device_map is not None and "disk" in device_map.values(): snake_case : Any = True snake_case : int = any(x in list(device_map.values() ) for x in ['''cpu''', '''disk'''] ) load_checkpoint_in_model( __magic_name__ , __magic_name__ , __magic_name__ , dtype=bnb_quantization_config.torch_dtype , offload_folder=__magic_name__ , offload_state_dict=__magic_name__ , keep_in_fpaa_modules=bnb_quantization_config.keep_in_fpaa_modules , offload_abit_bnb=load_in_abit and offload , ) return dispatch_model(__magic_name__ , device_map=__magic_name__ , offload_dir=__magic_name__ ) def a_ ( __magic_name__ , __magic_name__ , __magic_name__=None , __magic_name__=None , __magic_name__=None ) -> Union[str, Any]: """simple docstring""" if device_map is None: if torch.cuda.is_available(): snake_case : Optional[int] = {'''''': torch.cuda.current_device()} else: raise RuntimeError('''No GPU found. A GPU is needed for quantization.''' ) logger.info('''The device_map was not initialized.''' '''Setting device_map to `{\'\':torch.cuda.current_device()}`.''' ) if isinstance(__magic_name__ , __magic_name__ ): if device_map not in ["auto", "balanced", "balanced_low_0", "sequential"]: raise ValueError( '''If passing a string for `device_map`, please choose \'auto\', \'balanced\', \'balanced_low_0\' or ''' '''\'sequential\'.''' ) snake_case : Optional[Any] = {} special_dtypes.update( { name: bnb_quantization_config.torch_dtype for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.skip_modules ) } ) special_dtypes.update( { name: torch.floataa for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.keep_in_fpaa_modules ) } ) snake_case : List[Any] = {} snake_case : Dict = special_dtypes snake_case : Optional[Any] = no_split_module_classes snake_case : Tuple = bnb_quantization_config.target_dtype # get max_memory for each device. if device_map != "sequential": snake_case : Optional[int] = get_balanced_memory( __magic_name__ , low_zero=(device_map == '''balanced_low_0''') , max_memory=__magic_name__ , **__magic_name__ , ) snake_case : List[Any] = max_memory snake_case : Tuple = infer_auto_device_map(__magic_name__ , **__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ): # check if don't have any quantized module on the cpu snake_case : Union[str, Any] = bnb_quantization_config.skip_modules + bnb_quantization_config.keep_in_fpaa_modules snake_case : Union[str, Any] = { key: device_map[key] for key in device_map.keys() if key not in modules_not_to_convert } for device in ["cpu", "disk"]: if device in device_map_without_some_modules.values(): if bnb_quantization_config.load_in_abit: raise ValueError( ''' Some modules are dispatched on the CPU or the disk. Make sure you have enough GPU RAM to fit the quantized model. If you want to dispatch the model on the CPU or the disk while keeping these modules in `torch_dtype`, you need to pass a custom `device_map` to `load_and_quantize_model`. Check https://huggingface.co/docs/accelerate/main/en/usage_guides/quantization#offload-modules-to-cpu-and-disk for more details. ''' ) else: logger.info( '''Some modules are are offloaded to the CPU or the disk. Note that these modules will be converted to 8-bit''' ) del device_map_without_some_modules return device_map def a_ ( __magic_name__ , __magic_name__ , __magic_name__=None , __magic_name__=None ) -> Union[str, Any]: """simple docstring""" if modules_to_not_convert is None: snake_case : Optional[int] = [] snake_case , snake_case : List[Any] = _replace_with_bnb_layers( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) if not has_been_replaced: logger.warning( '''You are loading your model in 8bit or 4bit but no linear modules were found in your model.''' ''' this can happen for some architectures such as gpt2 that uses Conv1D instead of Linear layers.''' ''' Please double check your model architecture, or submit an issue on github if you think this is''' ''' a bug.''' ) return model def a_ ( __magic_name__ , __magic_name__ , __magic_name__=None , __magic_name__=None , ) -> Optional[Any]: """simple docstring""" snake_case : Optional[Any] = False for name, module in model.named_children(): if current_key_name is None: snake_case : List[str] = [] current_key_name.append(__magic_name__ ) if isinstance(__magic_name__ , nn.Linear ) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` snake_case : Dict = '''.'''.join(__magic_name__ ) snake_case : List[Any] = True for key in modules_to_not_convert: if ( (key in current_key_name_str) and (key + "." in current_key_name_str) ) or key == current_key_name_str: snake_case : Union[str, Any] = False break if proceed: # Load bnb module with empty weight and replace ``nn.Linear` module if bnb_quantization_config.load_in_abit: snake_case : int = bnb.nn.LinearabitLt( module.in_features , module.out_features , module.bias is not None , has_fpaa_weights=__magic_name__ , threshold=bnb_quantization_config.llm_inta_threshold , ) elif bnb_quantization_config.load_in_abit: snake_case : int = bnb.nn.Linearabit( module.in_features , module.out_features , module.bias is not None , bnb_quantization_config.bnb_abit_compute_dtype , compress_statistics=bnb_quantization_config.bnb_abit_use_double_quant , quant_type=bnb_quantization_config.bnb_abit_quant_type , ) else: raise ValueError('''load_in_8bit and load_in_4bit can\'t be both False''' ) snake_case : Union[str, Any] = module.weight.data if module.bias is not None: snake_case : Dict = module.bias.data bnb_module.requires_grad_(__magic_name__ ) setattr(__magic_name__ , __magic_name__ , __magic_name__ ) snake_case : Union[str, Any] = True if len(list(module.children() ) ) > 0: snake_case , snake_case : Optional[Any] = _replace_with_bnb_layers( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) snake_case : Tuple = has_been_replaced | _has_been_replaced # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def a_ ( __magic_name__ ) -> Union[str, Any]: """simple docstring""" with init_empty_weights(): snake_case : int = deepcopy(__magic_name__ ) # this has 0 cost since it is done inside `init_empty_weights` context manager` snake_case : int = find_tied_parameters(__magic_name__ ) # For compatibility with Accelerate < 0.18 if isinstance(__magic_name__ , __magic_name__ ): snake_case : Union[str, Any] = sum(list(tied_params.values() ) , [] ) + list(tied_params.keys() ) else: snake_case : Any = sum(__magic_name__ , [] ) snake_case : Optional[int] = len(__magic_name__ ) > 0 # Check if it is a base model snake_case : Any = False if hasattr(__magic_name__ , '''base_model_prefix''' ): snake_case : List[str] = not hasattr(__magic_name__ , model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head snake_case : Optional[Any] = list(model.named_children() ) snake_case : Any = [list_modules[-1][0]] # add last module together with tied weights snake_case : Any = set(__magic_name__ ) - set(__magic_name__ ) snake_case : Union[str, Any] = list(set(__magic_name__ ) ) + list(__magic_name__ ) # remove ".weight" from the keys snake_case : List[str] = ['''.weight''', '''.bias'''] snake_case : Dict = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: snake_case : int = name.replace(__magic_name__ , '''''' ) filtered_module_names.append(__magic_name__ ) return filtered_module_names def a_ ( __magic_name__ ) -> int: """simple docstring""" for m in model.modules(): if isinstance(__magic_name__ , bnb.nn.Linearabit ): return True return False def a_ ( __magic_name__ ) -> Any: """simple docstring""" return next(parameter.parameters() ).device def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" if fpaa_statistics is None: set_module_tensor_to_device(__magic_name__ , __magic_name__ , 0 , dtype=__magic_name__ , value=__magic_name__ ) snake_case : Optional[Any] = param_name snake_case : int = model if "." in tensor_name: snake_case : List[Any] = tensor_name.split('''.''' ) for split in splits[:-1]: snake_case : str = getattr(__magic_name__ , __magic_name__ ) if new_module is None: raise ValueError(F"{module} has no attribute {split}." ) snake_case : Dict = new_module snake_case : Optional[Any] = splits[-1] # offload weights snake_case : Any = False offload_weight(module._parameters[tensor_name] , __magic_name__ , __magic_name__ , index=__magic_name__ ) if hasattr(module._parameters[tensor_name] , '''SCB''' ): offload_weight( module._parameters[tensor_name].SCB , param_name.replace('''weight''' , '''SCB''' ) , __magic_name__ , index=__magic_name__ , ) else: offload_weight(__magic_name__ , __magic_name__ , __magic_name__ , index=__magic_name__ ) offload_weight(__magic_name__ , param_name.replace('''weight''' , '''SCB''' ) , __magic_name__ , index=__magic_name__ ) set_module_tensor_to_device(__magic_name__ , __magic_name__ , '''meta''' , dtype=__magic_name__ , value=torch.empty(*param.size() ) )
84
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets _a : Optional[Any] = '\\n@inproceedings{pillutla-etal:mauve:neurips2021,\n title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers},\n author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid},\n booktitle = {NeurIPS},\n year = {2021}\n}\n\n' _a : str = '\\nMAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure.\n\nMAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences.\n\nFor details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021).\n\nThis metrics is a wrapper around the official implementation of MAUVE:\nhttps://github.com/krishnap25/mauve\n' _a : List[Any] = '\nCalculates MAUVE scores between two lists of generated text and reference text.\nArgs:\n predictions: list of generated text to score. Each predictions\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\nOptional Args:\n num_buckets: the size of the histogram to quantize P and Q. Options: \'auto\' (default) or an integer\n pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1\n kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9\n kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5\n kmeans_max_iter: maximum number of k-means iterations. Default 500\n featurize_model_name: name of the model from which features are obtained. Default \'gpt2-large\' Use one of [\'gpt2\', \'gpt2-medium\', \'gpt2-large\', \'gpt2-xl\'].\n device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU\n max_text_length: maximum number of tokens to consider. Default 1024\n divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25\n mauve_scaling_factor: "c" from the paper. Default 5.\n verbose: If True (default), print running time updates\n seed: random seed to initialize k-means cluster assignments.\nReturns:\n mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer,\n frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer,\n divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve,\n p_hist: a discrete distribution, which is a quantized version of the text distribution p_text,\n q_hist: same as above, but with q_text.\nExamples:\n\n >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest\n >>> import datasets\n >>> mauve = datasets.load_metric(\'mauve\')\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP\n >>> print(out.mauve) # doctest: +SKIP\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): def lowerCAmelCase( self : Any ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://github.com/krishnap25/mauve''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/krishnap25/mauve'''] , reference_urls=[ '''https://arxiv.org/abs/2102.01454''', '''https://github.com/krishnap25/mauve''', ] , ) def lowerCAmelCase( self : int , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : List[str]="auto" , UpperCAmelCase__ : Tuple=-1 , UpperCAmelCase__ : Optional[int]=0.9 , UpperCAmelCase__ : List[Any]=5 , UpperCAmelCase__ : List[Any]=500 , UpperCAmelCase__ : Union[str, Any]="gpt2-large" , UpperCAmelCase__ : Optional[Any]=-1 , UpperCAmelCase__ : int=1_024 , UpperCAmelCase__ : List[Any]=25 , UpperCAmelCase__ : Union[str, Any]=5 , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : List[Any]=25 , ): """simple docstring""" snake_case : List[str] = compute_mauve( p_text=UpperCAmelCase__ , q_text=UpperCAmelCase__ , p_features=UpperCAmelCase__ , q_features=UpperCAmelCase__ , p_tokens=UpperCAmelCase__ , q_tokens=UpperCAmelCase__ , num_buckets=UpperCAmelCase__ , pca_max_data=UpperCAmelCase__ , kmeans_explained_var=UpperCAmelCase__ , kmeans_num_redo=UpperCAmelCase__ , kmeans_max_iter=UpperCAmelCase__ , featurize_model_name=UpperCAmelCase__ , device_id=UpperCAmelCase__ , max_text_length=UpperCAmelCase__ , divergence_curve_discretization_size=UpperCAmelCase__ , mauve_scaling_factor=UpperCAmelCase__ , verbose=UpperCAmelCase__ , seed=UpperCAmelCase__ , ) return out
84
1
def a_ ( __magic_name__ ) -> int: """simple docstring""" if divisor % 5 == 0 or divisor % 2 == 0: return 0 snake_case : Optional[Any] = 1 snake_case : int = 1 while repunit: snake_case : Dict = (10 * repunit + 1) % divisor repunit_index += 1 return repunit_index def a_ ( __magic_name__ = 1_000_000 ) -> int: """simple docstring""" snake_case : str = limit - 1 if divisor % 2 == 0: divisor += 1 while least_divisible_repunit(__magic_name__ ) <= limit: divisor += 2 return divisor if __name__ == "__main__": print(f"{solution() = }")
84
import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" if "cls_token" in name: snake_case : Tuple = name.replace('''cls_token''' , '''vit.embeddings.cls_token''' ) if "mask_token" in name: snake_case : Optional[int] = name.replace('''mask_token''' , '''decoder.mask_token''' ) if "decoder_pos_embed" in name: snake_case : List[str] = name.replace('''decoder_pos_embed''' , '''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: snake_case : List[str] = name.replace('''pos_embed''' , '''vit.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: snake_case : List[Any] = name.replace('''patch_embed.proj''' , '''vit.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: snake_case : int = name.replace('''patch_embed.norm''' , '''vit.embeddings.norm''' ) if "decoder_blocks" in name: snake_case : int = name.replace('''decoder_blocks''' , '''decoder.decoder_layers''' ) if "blocks" in name: snake_case : Optional[Any] = name.replace('''blocks''' , '''vit.encoder.layer''' ) if "attn.proj" in name: snake_case : str = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: snake_case : Any = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: snake_case : List[str] = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: snake_case : Tuple = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: snake_case : Tuple = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: snake_case : Tuple = name.replace('''mlp.fc2''' , '''output.dense''' ) if "decoder_embed" in name: snake_case : Dict = name.replace('''decoder_embed''' , '''decoder.decoder_embed''' ) if "decoder_norm" in name: snake_case : Dict = name.replace('''decoder_norm''' , '''decoder.decoder_norm''' ) if "decoder_pred" in name: snake_case : Dict = name.replace('''decoder_pred''' , '''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name: snake_case : Optional[int] = name.replace('''norm.weight''' , '''vit.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name: snake_case : List[str] = name.replace('''norm.bias''' , '''vit.layernorm.bias''' ) return name def a_ ( __magic_name__ , __magic_name__ ) -> str: """simple docstring""" for key in orig_state_dict.copy().keys(): snake_case : Union[str, Any] = orig_state_dict.pop(__magic_name__ ) if "qkv" in key: snake_case : Optional[int] = key.split('''.''' ) snake_case : int = int(key_split[1] ) if "decoder_blocks" in key: snake_case : List[str] = config.decoder_hidden_size snake_case : List[Any] = '''decoder.decoder_layers.''' if "weight" in key: snake_case : str = val[:dim, :] snake_case : Optional[Any] = val[dim : dim * 2, :] snake_case : Any = val[-dim:, :] elif "bias" in key: snake_case : Optional[Any] = val[:dim] snake_case : List[Any] = val[dim : dim * 2] snake_case : List[Any] = val[-dim:] else: snake_case : Optional[int] = config.hidden_size snake_case : Tuple = '''vit.encoder.layer.''' if "weight" in key: snake_case : Optional[Any] = val[:dim, :] snake_case : str = val[dim : dim * 2, :] snake_case : Union[str, Any] = val[-dim:, :] elif "bias" in key: snake_case : Tuple = val[:dim] snake_case : int = val[dim : dim * 2] snake_case : Optional[Any] = val[-dim:] else: snake_case : Optional[Any] = val return orig_state_dict def a_ ( __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" snake_case : List[str] = ViTMAEConfig() if "large" in checkpoint_url: snake_case : str = 1_024 snake_case : Tuple = 4_096 snake_case : Optional[Any] = 24 snake_case : List[Any] = 16 elif "huge" in checkpoint_url: snake_case : Tuple = 14 snake_case : int = 1_280 snake_case : Dict = 5_120 snake_case : Tuple = 32 snake_case : Optional[Any] = 16 snake_case : Optional[Any] = ViTMAEForPreTraining(__magic_name__ ) snake_case : Optional[Any] = torch.hub.load_state_dict_from_url(__magic_name__ , map_location='''cpu''' )['''model'''] snake_case : int = ViTMAEImageProcessor(size=config.image_size ) snake_case : Dict = convert_state_dict(__magic_name__ , __magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() snake_case : Tuple = '''https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg''' snake_case : List[Any] = Image.open(requests.get(__magic_name__ , stream=__magic_name__ ).raw ) snake_case : Dict = ViTMAEImageProcessor(size=config.image_size ) snake_case : str = image_processor(images=__magic_name__ , return_tensors='''pt''' ) # forward pass torch.manual_seed(2 ) snake_case : Union[str, Any] = model(**__magic_name__ ) snake_case : Optional[Any] = outputs.logits if "large" in checkpoint_url: snake_case : Any = torch.tensor( [[-0.7309, -0.7128, -1.0169], [-1.0161, -0.9058, -1.1878], [-1.0478, -0.9411, -1.1911]] ) elif "huge" in checkpoint_url: snake_case : List[Any] = torch.tensor( [[-1.1599, -0.9199, -1.2221], [-1.1952, -0.9269, -1.2307], [-1.2143, -0.9337, -1.2262]] ) else: snake_case : Dict = torch.tensor( [[-0.9192, -0.8481, -1.1259], [-1.1349, -1.0034, -1.2599], [-1.1757, -1.0429, -1.2726]] ) # verify logits assert torch.allclose(logits[0, :3, :3] , __magic_name__ , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(__magic_name__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__magic_name__ ) if __name__ == "__main__": _a : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth', type=str, help='URL of the checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _a : str = parser.parse_args() convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
84
1
import argparse from pathlib import Path from transformers import AutoConfig, AutoTokenizer, RagConfig, RagSequenceForGeneration, RagTokenForGeneration def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , ) -> Union[str, Any]: """simple docstring""" if config_name_or_path is None: snake_case : Dict = '''facebook/rag-token-base''' if model_type == '''rag_token''' else '''facebook/rag-sequence-base''' if generator_tokenizer_name_or_path is None: snake_case : Tuple = generator_name_or_path if question_encoder_tokenizer_name_or_path is None: snake_case : Optional[int] = question_encoder_name_or_path snake_case : List[Any] = RagTokenForGeneration if model_type == '''rag_token''' else RagSequenceForGeneration # Save model. snake_case : str = RagConfig.from_pretrained(__magic_name__ ) snake_case : Any = AutoConfig.from_pretrained(__magic_name__ ) snake_case : Optional[int] = AutoConfig.from_pretrained(__magic_name__ ) snake_case : Optional[Any] = gen_config snake_case : Optional[int] = question_encoder_config snake_case : int = model_class.from_pretrained_question_encoder_generator( __magic_name__ , __magic_name__ , config=__magic_name__ ) rag_model.save_pretrained(__magic_name__ ) # Sanity check. model_class.from_pretrained(__magic_name__ ) # Save tokenizers. snake_case : Optional[Any] = AutoTokenizer.from_pretrained(__magic_name__ ) gen_tokenizer.save_pretrained(dest_dir / '''generator_tokenizer/''' ) snake_case : Optional[Any] = AutoTokenizer.from_pretrained(__magic_name__ ) question_encoder_tokenizer.save_pretrained(dest_dir / '''question_encoder_tokenizer/''' ) if __name__ == "__main__": _a : Dict = argparse.ArgumentParser() parser.add_argument( '--model_type', choices=['rag_sequence', 'rag_token'], required=True, type=str, help='RAG model type: rag_sequence, rag_token', ) parser.add_argument('--dest', type=str, required=True, help='Path to the output checkpoint directory.') parser.add_argument('--generator_name_or_path', type=str, required=True, help='Generator model identifier') parser.add_argument( '--question_encoder_name_or_path', type=str, required=True, help='Question encoder model identifier' ) parser.add_argument( '--generator_tokenizer_name_or_path', type=str, help='Generator tokenizer identifier, if not specified, resolves to ``generator_name_or_path``', ) parser.add_argument( '--question_encoder_tokenizer_name_or_path', type=str, help='Question encoder tokenizer identifier, if not specified, resolves to ``question_encoder_name_or_path``', ) parser.add_argument( '--config_name_or_path', type=str, help=( 'Identifier of the model config to use, if not provided, resolves to a base config for a given' ' ``model_type``' ), ) _a : int = parser.parse_args() _a : List[str] = Path(args.dest) dest_dir.mkdir(exist_ok=True) consolidate( args.model_type, args.generator_name_or_path, args.question_encoder_name_or_path, dest_dir, args.config_name_or_path, args.generator_tokenizer_name_or_path, args.question_encoder_tokenizer_name_or_path, )
84
import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _a : Optional[Any] = 16 _a : Union[str, Any] = 32 def a_ ( __magic_name__ , __magic_name__ = 16 ) -> Dict: """simple docstring""" snake_case : Tuple = AutoTokenizer.from_pretrained('''bert-base-cased''' ) snake_case : Any = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(__magic_name__ ): # max_length=None => use the model max length (it's actually the default) snake_case : Union[str, Any] = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=__magic_name__ , max_length=__magic_name__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): snake_case : Union[str, Any] = datasets.map( __magic_name__ , batched=__magic_name__ , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library snake_case : Optional[Any] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(__magic_name__ ): # On TPU it's best to pad everything to the same length or training will be very slow. snake_case : str = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": snake_case : Tuple = 16 elif accelerator.mixed_precision != "no": snake_case : Dict = 8 else: snake_case : Union[str, Any] = None return tokenizer.pad( __magic_name__ , padding='''longest''' , max_length=__magic_name__ , pad_to_multiple_of=__magic_name__ , return_tensors='''pt''' , ) # Instantiate dataloaders. snake_case : str = DataLoader( tokenized_datasets['''train'''] , shuffle=__magic_name__ , collate_fn=__magic_name__ , batch_size=__magic_name__ ) snake_case : List[str] = DataLoader( tokenized_datasets['''validation'''] , shuffle=__magic_name__ , collate_fn=__magic_name__ , batch_size=__magic_name__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders _a : Optional[int] = mocked_dataloaders # noqa: F811 def a_ ( __magic_name__ , __magic_name__ ) -> Optional[Any]: """simple docstring""" if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , __magic_name__ ) == "1": snake_case : Optional[int] = 2 # Initialize accelerator snake_case : int = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs snake_case : Dict = config['''lr'''] snake_case : Any = int(config['''num_epochs'''] ) snake_case : List[str] = int(config['''seed'''] ) snake_case : List[Any] = int(config['''batch_size'''] ) snake_case : Tuple = evaluate.load('''glue''' , '''mrpc''' ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=__magic_name__ ) def inner_training_loop(__magic_name__ ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(__magic_name__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) snake_case : str = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=__magic_name__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). snake_case : Optional[int] = model.to(accelerator.device ) # Instantiate optimizer snake_case : Optional[int] = AdamW(params=model.parameters() , lr=__magic_name__ ) snake_case , snake_case : List[Any] = get_dataloaders(__magic_name__ , __magic_name__ ) # Instantiate scheduler snake_case : int = get_linear_schedule_with_warmup( optimizer=__magic_name__ , num_warmup_steps=100 , num_training_steps=(len(__magic_name__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. snake_case , snake_case , snake_case , snake_case , snake_case : Tuple = accelerator.prepare( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # Now we train the model for epoch in range(__magic_name__ ): model.train() for step, batch in enumerate(__magic_name__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) snake_case : int = model(**__magic_name__ ) snake_case : Optional[int] = outputs.loss accelerator.backward(__magic_name__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__magic_name__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): snake_case : List[str] = model(**__magic_name__ ) snake_case : List[Any] = outputs.logits.argmax(dim=-1 ) snake_case , snake_case : Dict = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=__magic_name__ , references=__magic_name__ , ) snake_case : Tuple = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"epoch {epoch}:" , __magic_name__ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def a_ ( ) -> Union[str, Any]: """simple docstring""" snake_case : int = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=__magic_name__ , default=__magic_name__ , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) snake_case : Optional[Any] = parser.parse_args() snake_case : Optional[int] = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(__magic_name__ , __magic_name__ ) if __name__ == "__main__": main()
84
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer _a : Optional[int] = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast _a : Any = TaTokenizerFast _a : Union[str, Any] = {'configuration_mt5': ['MT5Config', 'MT5OnnxConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[int] = [ 'MT5EncoderModel', 'MT5ForConditionalGeneration', 'MT5ForQuestionAnswering', 'MT5Model', 'MT5PreTrainedModel', 'MT5Stack', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[int] = ['TFMT5EncoderModel', 'TFMT5ForConditionalGeneration', 'TFMT5Model'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Tuple = ['FlaxMT5EncoderModel', 'FlaxMT5ForConditionalGeneration', 'FlaxMT5Model'] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys _a : List[Any] = _LazyModule( __name__, globals()['__file__'], _import_structure, extra_objects={'MT5Tokenizer': MTaTokenizer, 'MT5TokenizerFast': MTaTokenizerFast}, module_spec=__spec__, )
84
from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract _a : Dict = logging.get_logger(__name__) def a_ ( __magic_name__ , __magic_name__ , __magic_name__ ) -> List[Any]: """simple docstring""" return [ int(1_000 * (box[0] / width) ), int(1_000 * (box[1] / height) ), int(1_000 * (box[2] / width) ), int(1_000 * (box[3] / height) ), ] def a_ ( __magic_name__ , __magic_name__ , __magic_name__ = None ) -> str: """simple docstring""" snake_case : Any = tesseract_config if tesseract_config is not None else '''''' # apply OCR snake_case : str = to_pil_image(__magic_name__ ) snake_case , snake_case : Union[str, Any] = pil_image.size snake_case : List[Any] = pytesseract.image_to_data(__magic_name__ , lang=__magic_name__ , output_type='''dict''' , config=__magic_name__ ) snake_case , snake_case , snake_case , snake_case , snake_case : Optional[Any] = data['''text'''], data['''left'''], data['''top'''], data['''width'''], data['''height'''] # filter empty words and corresponding coordinates snake_case : Union[str, Any] = [idx for idx, word in enumerate(__magic_name__ ) if not word.strip()] snake_case : Union[str, Any] = [word for idx, word in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : Optional[Any] = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format snake_case : List[Any] = [] for x, y, w, h in zip(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): snake_case : Optional[int] = [x, y, x + w, y + h] actual_boxes.append(__magic_name__ ) # finally, normalize the bounding boxes snake_case : List[Any] = [] for box in actual_boxes: normalized_boxes.append(normalize_box(__magic_name__ , __magic_name__ , __magic_name__ ) ) assert len(__magic_name__ ) == len(__magic_name__ ), "Not as many words as there are bounding boxes" return words, normalized_boxes class a_ ( a ): A__ : int = ['pixel_values'] def __init__( self : Optional[int] , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : PILImageResampling = PILImageResampling.BILINEAR , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[str] = "" , **UpperCAmelCase__ : int , ): """simple docstring""" super().__init__(**UpperCAmelCase__ ) snake_case : Any = size if size is not None else {'''height''': 224, '''width''': 224} snake_case : Tuple = get_size_dict(UpperCAmelCase__ ) snake_case : Dict = do_resize snake_case : str = size snake_case : Optional[int] = resample snake_case : Union[str, Any] = apply_ocr snake_case : int = ocr_lang snake_case : Union[str, Any] = tesseract_config def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : np.ndarray , UpperCAmelCase__ : Dict[str, int] , UpperCAmelCase__ : PILImageResampling = PILImageResampling.BILINEAR , UpperCAmelCase__ : Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase__ : Any , ): """simple docstring""" snake_case : Dict = get_size_dict(UpperCAmelCase__ ) if "height" not in size or "width" not in size: raise ValueError(F"The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}" ) snake_case : Tuple = (size['''height'''], size['''width''']) return resize(UpperCAmelCase__ , size=UpperCAmelCase__ , resample=UpperCAmelCase__ , data_format=UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : ImageInput , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : PILImageResampling = None , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[Union[str, TensorType]] = None , UpperCAmelCase__ : ChannelDimension = ChannelDimension.FIRST , **UpperCAmelCase__ : List[Any] , ): """simple docstring""" snake_case : Tuple = do_resize if do_resize is not None else self.do_resize snake_case : List[Any] = size if size is not None else self.size snake_case : Tuple = get_size_dict(UpperCAmelCase__ ) snake_case : str = resample if resample is not None else self.resample snake_case : Optional[int] = apply_ocr if apply_ocr is not None else self.apply_ocr snake_case : Any = ocr_lang if ocr_lang is not None else self.ocr_lang snake_case : Optional[int] = tesseract_config if tesseract_config is not None else self.tesseract_config snake_case : List[str] = make_list_of_images(UpperCAmelCase__ ) if not valid_images(UpperCAmelCase__ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) # All transformations expect numpy arrays. snake_case : Any = [to_numpy_array(UpperCAmelCase__ ) for image in images] if apply_ocr: requires_backends(self , '''pytesseract''' ) snake_case : Optional[int] = [] snake_case : Union[str, Any] = [] for image in images: snake_case , snake_case : List[Any] = apply_tesseract(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) words_batch.append(UpperCAmelCase__ ) boxes_batch.append(UpperCAmelCase__ ) if do_resize: snake_case : Any = [self.resize(image=UpperCAmelCase__ , size=UpperCAmelCase__ , resample=UpperCAmelCase__ ) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) snake_case : int = [flip_channel_order(UpperCAmelCase__ ) for image in images] snake_case : Optional[Any] = [to_channel_dimension_format(UpperCAmelCase__ , UpperCAmelCase__ ) for image in images] snake_case : List[Any] = BatchFeature(data={'''pixel_values''': images} , tensor_type=UpperCAmelCase__ ) if apply_ocr: snake_case : Dict = words_batch snake_case : Dict = boxes_batch return data
84
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _a : Union[str, Any] = {'configuration_vit_msn': ['VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ViTMSNConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[int] = [ 'VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST', 'ViTMSNModel', 'ViTMSNForImageClassification', 'ViTMSNPreTrainedModel', ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys _a : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
84
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class a_ : def __init__( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Union[str, Any]=13 , UpperCAmelCase__ : Union[str, Any]=2 , UpperCAmelCase__ : List[Any]=24 , UpperCAmelCase__ : Union[str, Any]=16 , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : str=True , UpperCAmelCase__ : int=32 , UpperCAmelCase__ : Tuple=5 , UpperCAmelCase__ : Dict=4 , UpperCAmelCase__ : Optional[int]=37 , UpperCAmelCase__ : Optional[int]="gelu" , UpperCAmelCase__ : Dict=0.1 , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : Optional[int]=10 , UpperCAmelCase__ : Any=0.02 , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : List[Any]=2 , UpperCAmelCase__ : Optional[Any]=2 , ): """simple docstring""" snake_case : Tuple = parent snake_case : Dict = batch_size snake_case : str = patch_size snake_case : Union[str, Any] = max_length snake_case : str = num_mel_bins snake_case : Any = is_training snake_case : Union[str, Any] = use_labels snake_case : Tuple = hidden_size snake_case : Dict = num_hidden_layers snake_case : Any = num_attention_heads snake_case : Any = intermediate_size snake_case : List[Any] = hidden_act snake_case : str = hidden_dropout_prob snake_case : str = attention_probs_dropout_prob snake_case : str = type_sequence_label_size snake_case : Optional[int] = initializer_range snake_case : str = scope snake_case : int = frequency_stride snake_case : Union[str, Any] = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) snake_case : Optional[int] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 snake_case : Any = (self.max_length - self.patch_size) // self.time_stride + 1 snake_case : Union[str, Any] = frequency_out_dimension * time_out_dimension snake_case : Union[str, Any] = num_patches + 2 def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Optional[int] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) snake_case : str = None if self.use_labels: snake_case : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case : List[str] = self.get_config() return config, input_values, labels def lowerCAmelCase( self : Any ): """simple docstring""" return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCAmelCase__ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : str = ASTModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : Any = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : int = self.prepare_config_and_inputs() ( ( snake_case ) , ( snake_case ) , ( snake_case ) , ) : int = config_and_inputs snake_case : Tuple = {'''input_values''': input_values} return config, inputs_dict @require_torch class a_ ( a , a , unittest.TestCase ): A__ : List[Any] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) A__ : int = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) A__ : Optional[Any] = False A__ : Dict = False A__ : int = False A__ : Optional[int] = False def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : int ): """simple docstring""" if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[int] = ASTModelTester(self ) snake_case : Optional[int] = ConfigTester(self , config_class=UpperCAmelCase__ , has_text_modality=UpperCAmelCase__ , hidden_size=37 ) def lowerCAmelCase( self : List[str] ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def lowerCAmelCase( self : Tuple ): """simple docstring""" pass def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Optional[Any] = model_class(UpperCAmelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase__ , nn.Linear ) ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Any = model_class(UpperCAmelCase__ ) snake_case : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case : str = [*signature.parameters.keys()] snake_case : List[str] = ['''input_values'''] self.assertListEqual(arg_names[:1] , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case : List[str] = ASTModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) def a_ ( ) -> Dict: """simple docstring""" snake_case : Dict = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) snake_case , snake_case : int = torchaudio.load(__magic_name__ ) return audio, sampling_rate @require_torch @require_torchaudio class a_ ( unittest.TestCase ): @cached_property def lowerCAmelCase( self : Any ): """simple docstring""" return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : List[str] = self.default_feature_extractor snake_case : str = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(UpperCAmelCase__ ) snake_case : str = self.default_feature_extractor snake_case , snake_case : int = prepare_audio() snake_case : Optional[int] = audio.squeeze().numpy() snake_case : Optional[Any] = feature_extractor(UpperCAmelCase__ , sampling_rate=UpperCAmelCase__ , return_tensors='''pt''' ).to(UpperCAmelCase__ ) # forward pass with torch.no_grad(): snake_case : Union[str, Any] = model(**UpperCAmelCase__ ) # verify the logits snake_case : Any = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase__ ) snake_case : str = torch.tensor([-0.8760, -7.0042, -8.6602] ).to(UpperCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase__ , atol=1e-4 ) )
84
1
import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class a_ ( a ): A__ : BigBirdConfig A__ : jnp.dtype = jnp.floataa A__ : bool = True def lowerCAmelCase( self : int ): """simple docstring""" super().setup() snake_case : Tuple = nn.Dense(5 , dtype=self.dtype ) def __call__( self : Union[str, Any] , *UpperCAmelCase__ : int , **UpperCAmelCase__ : Dict ): """simple docstring""" snake_case : Optional[int] = super().__call__(*UpperCAmelCase__ , **UpperCAmelCase__ ) snake_case : Union[str, Any] = self.cls(outputs[2] ) return outputs[:2] + (cls_out,) class a_ ( a ): A__ : Optional[Any] = FlaxBigBirdForNaturalQuestionsModule def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Union[str, Any]: """simple docstring""" def cross_entropy(__magic_name__ , __magic_name__ , __magic_name__=None ): snake_case : Dict = logits.shape[-1] snake_case : List[str] = (labels[..., None] == jnp.arange(__magic_name__ )[None]).astype('''f4''' ) snake_case : List[str] = jax.nn.log_softmax(__magic_name__ , axis=-1 ) snake_case : Tuple = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: snake_case : Tuple = reduction(__magic_name__ ) return loss snake_case : List[Any] = partial(__magic_name__ , reduction=jnp.mean ) snake_case : int = cross_entropy(__magic_name__ , __magic_name__ ) snake_case : List[str] = cross_entropy(__magic_name__ , __magic_name__ ) snake_case : Optional[Any] = cross_entropy(__magic_name__ , __magic_name__ ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class a_ : A__ : str = "google/bigbird-roberta-base" A__ : int = 3000 A__ : int = 1_0500 A__ : int = 128 A__ : int = 3 A__ : int = 1 A__ : int = 5 # tx_args A__ : float = 3e-5 A__ : float = 0.0 A__ : int = 2_0000 A__ : float = 0.0095 A__ : str = "bigbird-roberta-natural-questions" A__ : str = "training-expt" A__ : str = "data/nq-training.jsonl" A__ : str = "data/nq-validation.jsonl" def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" os.makedirs(self.base_dir , exist_ok=UpperCAmelCase__ ) snake_case : str = os.path.join(self.base_dir , self.save_dir ) snake_case : Any = self.batch_size_per_device * jax.device_count() @dataclass class a_ : A__ : int A__ : int = 4096 # no dynamic padding on TPUs def __call__( self : Tuple , UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : Optional[Any] = self.collate_fn(UpperCAmelCase__ ) snake_case : List[Any] = jax.tree_util.tree_map(UpperCAmelCase__ , UpperCAmelCase__ ) return batch def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : Dict ): """simple docstring""" snake_case , snake_case : str = self.fetch_inputs(features['''input_ids'''] ) snake_case : List[str] = { '''input_ids''': jnp.array(UpperCAmelCase__ , dtype=jnp.intaa ), '''attention_mask''': jnp.array(UpperCAmelCase__ , dtype=jnp.intaa ), '''start_labels''': jnp.array(features['''start_token'''] , dtype=jnp.intaa ), '''end_labels''': jnp.array(features['''end_token'''] , dtype=jnp.intaa ), '''pooled_labels''': jnp.array(features['''category'''] , dtype=jnp.intaa ), } return batch def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : list ): """simple docstring""" snake_case : str = [self._fetch_inputs(UpperCAmelCase__ ) for ids in input_ids] return zip(*UpperCAmelCase__ ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : list ): """simple docstring""" snake_case : List[Any] = [1 for _ in range(len(UpperCAmelCase__ ) )] while len(UpperCAmelCase__ ) < self.max_length: input_ids.append(self.pad_id ) attention_mask.append(0 ) return input_ids, attention_mask def a_ ( __magic_name__ , __magic_name__ , __magic_name__=None ) -> List[Any]: """simple docstring""" if seed is not None: snake_case : List[Any] = dataset.shuffle(seed=__magic_name__ ) for i in range(len(__magic_name__ ) // batch_size ): snake_case : List[Any] = dataset[i * batch_size : (i + 1) * batch_size] yield dict(__magic_name__ ) @partial(jax.pmap , axis_name='''batch''' ) def a_ ( __magic_name__ , __magic_name__ , **__magic_name__ ) -> List[str]: """simple docstring""" def loss_fn(__magic_name__ ): snake_case : int = model_inputs.pop('''start_labels''' ) snake_case : Any = model_inputs.pop('''end_labels''' ) snake_case : str = model_inputs.pop('''pooled_labels''' ) snake_case : Dict = state.apply_fn(**__magic_name__ , params=__magic_name__ , dropout_rng=__magic_name__ , train=__magic_name__ ) snake_case , snake_case , snake_case : List[Any] = outputs return state.loss_fn( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , ) snake_case , snake_case : Any = jax.random.split(__magic_name__ ) snake_case : Tuple = jax.value_and_grad(__magic_name__ ) snake_case , snake_case : List[Any] = grad_fn(state.params ) snake_case : str = jax.lax.pmean({'''loss''': loss} , axis_name='''batch''' ) snake_case : Optional[Any] = jax.lax.pmean(__magic_name__ , '''batch''' ) snake_case : int = state.apply_gradients(grads=__magic_name__ ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name='''batch''' ) def a_ ( __magic_name__ , **__magic_name__ ) -> Optional[int]: """simple docstring""" snake_case : List[str] = model_inputs.pop('''start_labels''' ) snake_case : int = model_inputs.pop('''end_labels''' ) snake_case : Optional[int] = model_inputs.pop('''pooled_labels''' ) snake_case : List[Any] = state.apply_fn(**__magic_name__ , params=state.params , train=__magic_name__ ) snake_case , snake_case , snake_case : str = outputs snake_case : Any = state.loss_fn(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) snake_case : Any = jax.lax.pmean({'''loss''': loss} , axis_name='''batch''' ) return metrics class a_ ( train_state.TrainState ): A__ : Callable = struct.field(pytree_node=a ) @dataclass class a_ : A__ : Args A__ : Callable A__ : Callable A__ : Callable A__ : Callable A__ : wandb A__ : Callable = None def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[str]=None ): """simple docstring""" snake_case : str = model.params snake_case : Union[str, Any] = TrainState.create( apply_fn=model.__call__ , params=UpperCAmelCase__ , tx=UpperCAmelCase__ , loss_fn=UpperCAmelCase__ , ) if ckpt_dir is not None: snake_case , snake_case , snake_case , snake_case , snake_case : Any = restore_checkpoint(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : Optional[Any] = { '''lr''': args.lr, '''init_lr''': args.init_lr, '''warmup_steps''': args.warmup_steps, '''num_train_steps''': num_train_steps, '''weight_decay''': args.weight_decay, } snake_case , snake_case : int = build_tx(**UpperCAmelCase__ ) snake_case : str = train_state.TrainState( step=UpperCAmelCase__ , apply_fn=model.__call__ , params=UpperCAmelCase__ , tx=UpperCAmelCase__ , opt_state=UpperCAmelCase__ , ) snake_case : List[Any] = args snake_case : List[Any] = data_collator snake_case : Optional[Any] = lr snake_case : Union[str, Any] = params snake_case : Union[str, Any] = jax_utils.replicate(UpperCAmelCase__ ) return state def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Any ): """simple docstring""" snake_case : int = self.args snake_case : List[Any] = len(UpperCAmelCase__ ) // args.batch_size snake_case : str = jax.random.PRNGKey(0 ) snake_case : Any = jax.random.split(UpperCAmelCase__ , jax.device_count() ) for epoch in range(args.max_epochs ): snake_case : Union[str, Any] = jnp.array(0 , dtype=jnp.floataa ) snake_case : Union[str, Any] = get_batched_dataset(UpperCAmelCase__ , args.batch_size , seed=UpperCAmelCase__ ) snake_case : Any = 0 for batch in tqdm(UpperCAmelCase__ , total=UpperCAmelCase__ , desc=F"Running EPOCH-{epoch}" ): snake_case : Any = self.data_collator(UpperCAmelCase__ ) snake_case , snake_case , snake_case : Optional[int] = self.train_step_fn(UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ) running_loss += jax_utils.unreplicate(metrics['''loss'''] ) i += 1 if i % args.logging_steps == 0: snake_case : List[Any] = jax_utils.unreplicate(state.step ) snake_case : str = running_loss.item() / i snake_case : Optional[int] = self.scheduler_fn(state_step - 1 ) snake_case : List[str] = self.evaluate(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : Optional[int] = { '''step''': state_step.item(), '''eval_loss''': eval_loss.item(), '''tr_loss''': tr_loss, '''lr''': lr.item(), } tqdm.write(str(UpperCAmelCase__ ) ) self.logger.log(UpperCAmelCase__ , commit=UpperCAmelCase__ ) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + F"-e{epoch}-s{i}" , state=UpperCAmelCase__ ) def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Tuple ): """simple docstring""" snake_case : Optional[int] = get_batched_dataset(UpperCAmelCase__ , self.args.batch_size ) snake_case : List[Any] = len(UpperCAmelCase__ ) // self.args.batch_size snake_case : str = jnp.array(0 , dtype=jnp.floataa ) snake_case : Union[str, Any] = 0 for batch in tqdm(UpperCAmelCase__ , total=UpperCAmelCase__ , desc='''Evaluating ... ''' ): snake_case : int = self.data_collator(UpperCAmelCase__ ) snake_case : Any = self.val_step_fn(UpperCAmelCase__ , **UpperCAmelCase__ ) running_loss += jax_utils.unreplicate(metrics['''loss'''] ) i += 1 return running_loss / i def lowerCAmelCase( self : int , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Dict ): """simple docstring""" snake_case : Union[str, Any] = jax_utils.unreplicate(UpperCAmelCase__ ) print(F"SAVING CHECKPOINT IN {save_dir}" , end=''' ... ''' ) self.model_save_fn(UpperCAmelCase__ , params=state.params ) with open(os.path.join(UpperCAmelCase__ , '''opt_state.msgpack''' ) , '''wb''' ) as f: f.write(to_bytes(state.opt_state ) ) joblib.dump(self.args , os.path.join(UpperCAmelCase__ , '''args.joblib''' ) ) joblib.dump(self.data_collator , os.path.join(UpperCAmelCase__ , '''data_collator.joblib''' ) ) with open(os.path.join(UpperCAmelCase__ , '''training_state.json''' ) , '''w''' ) as f: json.dump({'''step''': state.step.item()} , UpperCAmelCase__ ) print('''DONE''' ) def a_ ( __magic_name__ , __magic_name__ ) -> Optional[int]: """simple docstring""" print(F"RESTORING CHECKPOINT FROM {save_dir}" , end=''' ... ''' ) with open(os.path.join(__magic_name__ , '''flax_model.msgpack''' ) , '''rb''' ) as f: snake_case : Any = from_bytes(state.params , f.read() ) with open(os.path.join(__magic_name__ , '''opt_state.msgpack''' ) , '''rb''' ) as f: snake_case : Optional[Any] = from_bytes(state.opt_state , f.read() ) snake_case : Tuple = joblib.load(os.path.join(__magic_name__ , '''args.joblib''' ) ) snake_case : Optional[int] = joblib.load(os.path.join(__magic_name__ , '''data_collator.joblib''' ) ) with open(os.path.join(__magic_name__ , '''training_state.json''' ) , '''r''' ) as f: snake_case : Optional[Any] = json.load(__magic_name__ ) snake_case : List[str] = training_state['''step'''] print('''DONE''' ) return params, opt_state, step, args, data_collator def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Tuple: """simple docstring""" snake_case : Tuple = num_train_steps - warmup_steps snake_case : List[str] = optax.linear_schedule(init_value=__magic_name__ , end_value=__magic_name__ , transition_steps=__magic_name__ ) snake_case : Optional[Any] = optax.linear_schedule(init_value=__magic_name__ , end_value=1e-7 , transition_steps=__magic_name__ ) snake_case : str = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" def weight_decay_mask(__magic_name__ ): snake_case : Optional[int] = traverse_util.flatten_dict(__magic_name__ ) snake_case : Dict = {k: (v[-1] != '''bias''' and v[-2:] != ('''LayerNorm''', '''scale''')) for k, v in params.items()} return traverse_util.unflatten_dict(__magic_name__ ) snake_case : int = scheduler_fn(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) snake_case : str = optax.adamw(learning_rate=__magic_name__ , weight_decay=__magic_name__ , mask=__magic_name__ ) return tx, lr
84
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _a : Union[str, Any] = logging.getLogger(__name__) def a_ ( __magic_name__ , __magic_name__ ) -> Optional[int]: """simple docstring""" return (preds == labels).mean() @dataclass class a_ : A__ : str = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) @dataclass class a_ : A__ : str = field(metadata={'help': 'The name of the task to train on: ' + ', '.join(processors.keys() )} ) A__ : str = field(metadata={'help': 'Should contain the data files for the task.'} ) A__ : int = field( default=128 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) A__ : bool = field( default=a , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) def a_ ( ) -> Dict: """simple docstring""" snake_case : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) snake_case , snake_case , snake_case : Tuple = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"Output directory ({training_args.output_dir}) already exists and is not empty. Use" ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , __magic_name__ ) # Set seed set_seed(training_args.seed ) try: snake_case : int = processors[data_args.task_name]() snake_case : List[str] = processor.get_labels() snake_case : str = len(__magic_name__ ) except KeyError: raise ValueError('''Task not found: %s''' % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=__magic_name__ , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) snake_case : str = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) snake_case : Any = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=__magic_name__ , cache_dir=model_args.cache_dir , ) # Get datasets snake_case : Optional[int] = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=__magic_name__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) snake_case : Any = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=__magic_name__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(__magic_name__ ) -> Dict: snake_case : str = np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(__magic_name__ , p.label_ids )} # Data collator snake_case : Dict = DataCollatorWithPadding(__magic_name__ , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer snake_case : List[Any] = Trainer( model=__magic_name__ , args=__magic_name__ , train_dataset=__magic_name__ , eval_dataset=__magic_name__ , compute_metrics=__magic_name__ , data_collator=__magic_name__ , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation snake_case : Optional[int] = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) snake_case : Optional[Any] = trainer.evaluate() snake_case : Union[str, Any] = os.path.join(training_args.output_dir , '''eval_results.txt''' ) if trainer.is_world_master(): with open(__magic_name__ , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(''' %s = %s''' , __magic_name__ , __magic_name__ ) writer.write('''%s = %s\n''' % (key, value) ) results.update(__magic_name__ ) return results def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" main() if __name__ == "__main__": main()
84
1
import argparse import json from collections import OrderedDict from functools import partial from pathlib import Path import timm import torch from huggingface_hub import hf_hub_download from transformers import LevitConfig, LevitForImageClassificationWithTeacher, LevitImageProcessor from transformers.utils import logging logging.set_verbosity_info() _a : Optional[Any] = logging.get_logger() def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = True ) -> Any: """simple docstring""" print(F"Converting {name}..." ) with torch.no_grad(): if hidden_sizes == 128: if name[-1] == "S": snake_case : int = timm.create_model('''levit_128s''' , pretrained=__magic_name__ ) else: snake_case : List[Any] = timm.create_model('''levit_128''' , pretrained=__magic_name__ ) if hidden_sizes == 192: snake_case : Union[str, Any] = timm.create_model('''levit_192''' , pretrained=__magic_name__ ) if hidden_sizes == 256: snake_case : Optional[int] = timm.create_model('''levit_256''' , pretrained=__magic_name__ ) if hidden_sizes == 384: snake_case : int = timm.create_model('''levit_384''' , pretrained=__magic_name__ ) from_model.eval() snake_case : List[Any] = LevitForImageClassificationWithTeacher(__magic_name__ ).eval() snake_case : Optional[Any] = OrderedDict() snake_case : Any = from_model.state_dict() snake_case : int = list(from_model.state_dict().keys() ) snake_case : Union[str, Any] = list(our_model.state_dict().keys() ) print(len(__magic_name__ ) , len(__magic_name__ ) ) for i in range(len(__magic_name__ ) ): snake_case : Tuple = weights[og_keys[i]] our_model.load_state_dict(__magic_name__ ) snake_case : Union[str, Any] = torch.randn((2, 3, 224, 224) ) snake_case : Tuple = from_model(__magic_name__ ) snake_case : Any = our_model(__magic_name__ ).logits assert torch.allclose(__magic_name__ , __magic_name__ ), "The model logits don't match the original one." snake_case : Optional[Any] = name print(__magic_name__ ) if push_to_hub: our_model.save_pretrained(save_directory / checkpoint_name ) snake_case : str = LevitImageProcessor() image_processor.save_pretrained(save_directory / checkpoint_name ) print(F"Pushed {checkpoint_name}" ) def a_ ( __magic_name__ , __magic_name__ = None , __magic_name__ = True ) -> Union[str, Any]: """simple docstring""" snake_case : Union[str, Any] = '''imagenet-1k-id2label.json''' snake_case : List[Any] = 1_000 snake_case : Dict = (1, num_labels) snake_case : Dict = '''huggingface/label-files''' snake_case : Optional[Any] = num_labels snake_case : int = json.load(open(hf_hub_download(__magic_name__ , __magic_name__ , repo_type='''dataset''' ) , '''r''' ) ) snake_case : Union[str, Any] = {int(__magic_name__ ): v for k, v in idalabel.items()} snake_case : Dict = idalabel snake_case : Any = {v: k for k, v in idalabel.items()} snake_case : Optional[int] = partial(__magic_name__ , num_labels=__magic_name__ , idalabel=__magic_name__ , labelaid=__magic_name__ ) snake_case : Union[str, Any] = { '''levit-128S''': 128, '''levit-128''': 128, '''levit-192''': 192, '''levit-256''': 256, '''levit-384''': 384, } snake_case : List[str] = { '''levit-128S''': ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 6, 8] , depths=[2, 3, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), '''levit-128''': ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 8, 12] , depths=[4, 4, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), '''levit-192''': ImageNetPreTrainedConfig( hidden_sizes=[192, 288, 384] , num_attention_heads=[3, 5, 6] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), '''levit-256''': ImageNetPreTrainedConfig( hidden_sizes=[256, 384, 512] , num_attention_heads=[4, 6, 8] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), '''levit-384''': ImageNetPreTrainedConfig( hidden_sizes=[384, 512, 768] , num_attention_heads=[6, 9, 12] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0.1 , ), } if model_name: convert_weight_and_push( names_to_hidden_sizes[model_name] , __magic_name__ , names_to_config[model_name] , __magic_name__ , __magic_name__ ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(names_to_hidden_sizes[model_name] , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) return config, expected_shape if __name__ == "__main__": _a : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default=None, type=str, help='The name of the model you wish to convert, it must be one of the supported Levit* architecture,', ) parser.add_argument( '--pytorch_dump_folder_path', default='levit-dump-folder/', type=Path, required=False, help='Path to the output PyTorch model directory.', ) parser.add_argument('--push_to_hub', action='store_true', help='Push model and image processor to the hub') parser.add_argument( '--no-push_to_hub', dest='push_to_hub', action='store_false', help='Do not push model and image processor to the hub', ) _a : Dict = parser.parse_args() _a : Path = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
84
import re def a_ ( __magic_name__ ) -> bool: """simple docstring""" snake_case : List[str] = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(__magic_name__ , __magic_name__ ) ) if __name__ == "__main__": _a : Any = '0094702343221' print(is_sri_lankan_phone_number(phone))
84
1
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _a : List[Any] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class a_ ( a , unittest.TestCase ): A__ : Dict = ReformerTokenizer A__ : Optional[int] = ReformerTokenizerFast A__ : str = True A__ : Tuple = False A__ : str = True def lowerCAmelCase( self : List[Any] ): """simple docstring""" super().setUp() snake_case : str = ReformerTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : int = '''<s>''' snake_case : List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase__ ) , UpperCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase__ ) , UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Any = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(UpperCAmelCase__ ) , 1_000 ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def lowerCAmelCase( self : Dict ): """simple docstring""" if not self.test_rust_tokenizer: return snake_case : Any = self.get_tokenizer() snake_case : str = self.get_rust_tokenizer() snake_case : Tuple = '''I was born in 92000, and this is falsé.''' snake_case : str = tokenizer.tokenize(UpperCAmelCase__ ) snake_case : int = rust_tokenizer.tokenize(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : Union[str, Any] = tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) snake_case : List[str] = rust_tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : List[str] = self.get_rust_tokenizer() snake_case : Optional[int] = tokenizer.encode(UpperCAmelCase__ ) snake_case : Optional[Any] = rust_tokenizer.encode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : List[Any]=15 ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): snake_case : str = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) # Simple input snake_case : Union[str, Any] = '''This is a simple input''' snake_case : List[str] = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case : int = ('''This is a simple input''', '''This is a pair''') snake_case : int = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Simple input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Simple input self.assertRaises( UpperCAmelCase__ , tokenizer_r.batch_encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' , ) # Pair input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Pair input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Pair input self.assertRaises( UpperCAmelCase__ , tokenizer_r.batch_encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' , ) def lowerCAmelCase( self : str ): """simple docstring""" pass def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Union[str, Any] = ReformerTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) snake_case : List[str] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(UpperCAmelCase__ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , [285, 46, 10, 170, 382] , ) snake_case : int = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) snake_case : int = tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) snake_case : List[Any] = tokenizer.convert_ids_to_tokens(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def lowerCAmelCase( self : Tuple ): """simple docstring""" return ReformerTokenizer.from_pretrained('''google/reformer-crime-and-punishment''' ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : Any = '''Hello World!''' snake_case : Optional[Any] = [126, 32, 262, 152, 38, 72, 287] self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @slow def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[Any] = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) snake_case : Dict = [ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @require_torch @slow def lowerCAmelCase( self : List[Any] ): """simple docstring""" import torch from transformers import ReformerConfig, ReformerModel # Build sequence snake_case : Any = list(self.big_tokenizer.get_vocab().keys() )[:10] snake_case : Union[str, Any] = ''' '''.join(UpperCAmelCase__ ) snake_case : Optional[int] = self.big_tokenizer.encode_plus(UpperCAmelCase__ , return_tensors='''pt''' ) snake_case : List[str] = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors='''pt''' ) snake_case : Optional[Any] = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) snake_case : Tuple = encoded_sequence['''input_ids'''].shape snake_case : List[Any] = ReformerModel(UpperCAmelCase__ ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**UpperCAmelCase__ ) model(**UpperCAmelCase__ ) @slow def lowerCAmelCase( self : Optional[int] ): """simple docstring""" # fmt: off snake_case : Tuple = {'''input_ids''': [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 snake_case : Tuple = [ '''This is a very simple sentence.''', '''The quick brown fox jumps over the lazy dog.''', ] self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase__ , model_name='''google/reformer-crime-and-punishment''' , revision='''0e6c3decb8211d49bf881013425dc8b0448b3f5a''' , padding=UpperCAmelCase__ , sequences=UpperCAmelCase__ , )
84
import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class a_ ( unittest.TestCase ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[int]=7 , UpperCAmelCase__ : Union[str, Any]=3 , UpperCAmelCase__ : int=18 , UpperCAmelCase__ : Optional[int]=30 , UpperCAmelCase__ : Optional[int]=400 , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : int=True , ): """simple docstring""" snake_case : int = size if size is not None else {'''height''': 18, '''width''': 18} snake_case : Optional[Any] = parent snake_case : Any = batch_size snake_case : Any = num_channels snake_case : Union[str, Any] = image_size snake_case : Dict = min_resolution snake_case : Dict = max_resolution snake_case : int = do_resize snake_case : List[str] = size snake_case : List[Any] = apply_ocr def lowerCAmelCase( self : int ): """simple docstring""" return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class a_ ( a , unittest.TestCase ): A__ : List[Any] = LayoutLMvaImageProcessor if is_pytesseract_available() else None def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : Optional[Any] = LayoutLMvaImageProcessingTester(self ) @property def lowerCAmelCase( self : Dict ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[str] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase__ , '''do_resize''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''size''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''apply_ocr''' ) ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 18, '''width''': 18} ) snake_case : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) def lowerCAmelCase( self : str ): """simple docstring""" pass def lowerCAmelCase( self : Dict ): """simple docstring""" # Initialize image_processing snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , Image.Image ) # Test not batched input snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) self.assertIsInstance(encoding.words , UpperCAmelCase__ ) self.assertIsInstance(encoding.boxes , UpperCAmelCase__ ) # Test batched snake_case : Dict = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" # Initialize image_processing snake_case : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , numpify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , np.ndarray ) # Test not batched input snake_case : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched snake_case : List[Any] = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" # Initialize image_processing snake_case : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , torchify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , torch.Tensor ) # Test not batched input snake_case : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched snake_case : Tuple = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" # with apply_OCR = True snake_case : int = LayoutLMvaImageProcessor() from datasets import load_dataset snake_case : List[Any] = load_dataset('''hf-internal-testing/fixtures_docvqa''' , split='''test''' ) snake_case : List[Any] = Image.open(ds[0]['''file'''] ).convert('''RGB''' ) snake_case : Any = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 snake_case : Optional[Any] = [['''11:14''', '''to''', '''11:39''', '''a.m''', '''11:39''', '''to''', '''11:44''', '''a.m.''', '''11:44''', '''a.m.''', '''to''', '''12:25''', '''p.m.''', '''12:25''', '''to''', '''12:58''', '''p.m.''', '''12:58''', '''to''', '''4:00''', '''p.m.''', '''2:00''', '''to''', '''5:00''', '''p.m.''', '''Coffee''', '''Break''', '''Coffee''', '''will''', '''be''', '''served''', '''for''', '''men''', '''and''', '''women''', '''in''', '''the''', '''lobby''', '''adjacent''', '''to''', '''exhibit''', '''area.''', '''Please''', '''move''', '''into''', '''exhibit''', '''area.''', '''(Exhibits''', '''Open)''', '''TRRF''', '''GENERAL''', '''SESSION''', '''(PART''', '''|)''', '''Presiding:''', '''Lee''', '''A.''', '''Waller''', '''TRRF''', '''Vice''', '''President''', '''“Introductory''', '''Remarks”''', '''Lee''', '''A.''', '''Waller,''', '''TRRF''', '''Vice''', '''Presi-''', '''dent''', '''Individual''', '''Interviews''', '''with''', '''TRRF''', '''Public''', '''Board''', '''Members''', '''and''', '''Sci-''', '''entific''', '''Advisory''', '''Council''', '''Mem-''', '''bers''', '''Conducted''', '''by''', '''TRRF''', '''Treasurer''', '''Philip''', '''G.''', '''Kuehn''', '''to''', '''get''', '''answers''', '''which''', '''the''', '''public''', '''refrigerated''', '''warehousing''', '''industry''', '''is''', '''looking''', '''for.''', '''Plus''', '''questions''', '''from''', '''the''', '''floor.''', '''Dr.''', '''Emil''', '''M.''', '''Mrak,''', '''University''', '''of''', '''Cal-''', '''ifornia,''', '''Chairman,''', '''TRRF''', '''Board;''', '''Sam''', '''R.''', '''Cecil,''', '''University''', '''of''', '''Georgia''', '''College''', '''of''', '''Agriculture;''', '''Dr.''', '''Stanley''', '''Charm,''', '''Tufts''', '''University''', '''School''', '''of''', '''Medicine;''', '''Dr.''', '''Robert''', '''H.''', '''Cotton,''', '''ITT''', '''Continental''', '''Baking''', '''Company;''', '''Dr.''', '''Owen''', '''Fennema,''', '''University''', '''of''', '''Wis-''', '''consin;''', '''Dr.''', '''Robert''', '''E.''', '''Hardenburg,''', '''USDA.''', '''Questions''', '''and''', '''Answers''', '''Exhibits''', '''Open''', '''Capt.''', '''Jack''', '''Stoney''', '''Room''', '''TRRF''', '''Scientific''', '''Advisory''', '''Council''', '''Meeting''', '''Ballroom''', '''Foyer''']] # noqa: E231 snake_case : Union[str, Any] = [[[141, 57, 214, 69], [228, 58, 252, 69], [141, 75, 216, 88], [230, 79, 280, 88], [142, 260, 218, 273], [230, 261, 255, 273], [143, 279, 218, 290], [231, 282, 290, 291], [143, 342, 218, 354], [231, 345, 289, 355], [202, 362, 227, 373], [143, 379, 220, 392], [231, 382, 291, 394], [144, 714, 220, 726], [231, 715, 256, 726], [144, 732, 220, 745], [232, 736, 291, 747], [144, 769, 218, 782], [231, 770, 256, 782], [141, 788, 202, 801], [215, 791, 274, 804], [143, 826, 204, 838], [215, 826, 240, 838], [142, 844, 202, 857], [215, 847, 274, 859], [334, 57, 427, 69], [440, 57, 522, 69], [369, 75, 461, 88], [469, 75, 516, 88], [528, 76, 562, 88], [570, 76, 667, 88], [675, 75, 711, 87], [721, 79, 778, 88], [789, 75, 840, 88], [369, 97, 470, 107], [484, 94, 507, 106], [518, 94, 562, 107], [576, 94, 655, 110], [668, 94, 792, 109], [804, 95, 829, 107], [369, 113, 465, 125], [477, 116, 547, 125], [562, 113, 658, 125], [671, 116, 748, 125], [761, 113, 811, 125], [369, 131, 465, 143], [477, 133, 548, 143], [563, 130, 698, 145], [710, 130, 802, 146], [336, 171, 412, 183], [423, 171, 572, 183], [582, 170, 716, 184], [728, 171, 817, 187], [829, 171, 844, 186], [338, 197, 482, 212], [507, 196, 557, 209], [569, 196, 595, 208], [610, 196, 702, 209], [505, 214, 583, 226], [595, 214, 656, 227], [670, 215, 807, 227], [335, 259, 543, 274], [556, 259, 708, 272], [372, 279, 422, 291], [435, 279, 460, 291], [474, 279, 574, 292], [587, 278, 664, 291], [676, 278, 738, 291], [751, 279, 834, 291], [372, 298, 434, 310], [335, 341, 483, 354], [497, 341, 655, 354], [667, 341, 728, 354], [740, 341, 825, 354], [335, 360, 430, 372], [442, 360, 534, 372], [545, 359, 687, 372], [697, 360, 754, 372], [765, 360, 823, 373], [334, 378, 428, 391], [440, 378, 577, 394], [590, 378, 705, 391], [720, 378, 801, 391], [334, 397, 400, 409], [370, 416, 529, 429], [544, 416, 576, 432], [587, 416, 665, 428], [677, 416, 814, 429], [372, 435, 452, 450], [465, 434, 495, 447], [511, 434, 600, 447], [611, 436, 637, 447], [649, 436, 694, 451], [705, 438, 824, 447], [369, 453, 452, 466], [464, 454, 509, 466], [522, 453, 611, 469], [625, 453, 792, 469], [370, 472, 556, 488], [570, 472, 684, 487], [697, 472, 718, 485], [732, 472, 835, 488], [369, 490, 411, 503], [425, 490, 484, 503], [496, 490, 635, 506], [645, 490, 707, 503], [718, 491, 761, 503], [771, 490, 840, 503], [336, 510, 374, 521], [388, 510, 447, 522], [460, 510, 489, 521], [503, 510, 580, 522], [592, 509, 736, 525], [745, 509, 770, 522], [781, 509, 840, 522], [338, 528, 434, 541], [448, 528, 596, 541], [609, 527, 687, 540], [700, 528, 792, 541], [336, 546, 397, 559], [407, 546, 431, 559], [443, 546, 525, 560], [537, 546, 680, 562], [688, 546, 714, 559], [722, 546, 837, 562], [336, 565, 449, 581], [461, 565, 485, 577], [497, 565, 665, 581], [681, 565, 718, 577], [732, 565, 837, 580], [337, 584, 438, 597], [452, 583, 521, 596], [535, 584, 677, 599], [690, 583, 787, 596], [801, 583, 825, 596], [338, 602, 478, 615], [492, 602, 530, 614], [543, 602, 638, 615], [650, 602, 676, 614], [688, 602, 788, 615], [802, 602, 843, 614], [337, 621, 502, 633], [516, 621, 615, 637], [629, 621, 774, 636], [789, 621, 827, 633], [337, 639, 418, 652], [432, 640, 571, 653], [587, 639, 731, 655], [743, 639, 769, 652], [780, 639, 841, 652], [338, 658, 440, 673], [455, 658, 491, 670], [508, 658, 602, 671], [616, 658, 638, 670], [654, 658, 835, 674], [337, 677, 429, 689], [337, 714, 482, 726], [495, 714, 548, 726], [561, 714, 683, 726], [338, 770, 461, 782], [474, 769, 554, 785], [489, 788, 562, 803], [576, 788, 643, 801], [656, 787, 751, 804], [764, 788, 844, 801], [334, 825, 421, 838], [430, 824, 574, 838], [584, 824, 723, 841], [335, 844, 450, 857], [464, 843, 583, 860], [628, 862, 755, 875], [769, 861, 848, 878]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , UpperCAmelCase__ ) self.assertListEqual(encoding.boxes , UpperCAmelCase__ ) # with apply_OCR = False snake_case : str = LayoutLMvaImageProcessor(apply_ocr=UpperCAmelCase__ ) snake_case : Optional[Any] = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) )
84
1
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.local_sgd import LocalSGD ######################################################################## # This is a fully working simple example to use Accelerate # with LocalSGD, which is a method to synchronize model # parameters every K batches. It is different, but complementary # to gradient accumulation. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _a : Optional[int] = 16 _a : Optional[int] = 32 def a_ ( __magic_name__ , __magic_name__ = 16 ) -> Optional[Any]: """simple docstring""" snake_case : Tuple = AutoTokenizer.from_pretrained('''bert-base-cased''' ) snake_case : Any = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(__magic_name__ ): # max_length=None => use the model max length (it's actually the default) snake_case : Union[str, Any] = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=__magic_name__ , max_length=__magic_name__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): snake_case : Dict = datasets.map( __magic_name__ , batched=__magic_name__ , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library snake_case : Any = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(__magic_name__ ): # On TPU it's best to pad everything to the same length or training will be very slow. snake_case : Any = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": snake_case : Optional[Any] = 16 elif accelerator.mixed_precision != "no": snake_case : str = 8 else: snake_case : Optional[int] = None return tokenizer.pad( __magic_name__ , padding='''longest''' , max_length=__magic_name__ , pad_to_multiple_of=__magic_name__ , return_tensors='''pt''' , ) # Instantiate dataloaders. snake_case : Any = DataLoader( tokenized_datasets['''train'''] , shuffle=__magic_name__ , collate_fn=__magic_name__ , batch_size=__magic_name__ ) snake_case : Optional[int] = DataLoader( tokenized_datasets['''validation'''] , shuffle=__magic_name__ , collate_fn=__magic_name__ , batch_size=__magic_name__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders _a : Optional[int] = mocked_dataloaders # noqa: F811 def a_ ( __magic_name__ , __magic_name__ ) -> Optional[int]: """simple docstring""" if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , __magic_name__ ) == "1": snake_case : Dict = 2 # New Code # snake_case : Dict = int(args.gradient_accumulation_steps ) snake_case : Tuple = int(args.local_sgd_steps ) # Initialize accelerator snake_case : str = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , gradient_accumulation_steps=__magic_name__ ) if accelerator.distributed_type not in [DistributedType.NO, DistributedType.MULTI_CPU, DistributedType.MULTI_GPU]: raise NotImplementedError('''LocalSGD is supported only for CPUs and GPUs (no DeepSpeed or MegatronLM)''' ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs snake_case : Any = config['''lr'''] snake_case : Optional[Any] = int(config['''num_epochs'''] ) snake_case : Dict = int(config['''seed'''] ) snake_case : Tuple = int(config['''batch_size'''] ) snake_case : List[str] = evaluate.load('''glue''' , '''mrpc''' ) set_seed(__magic_name__ ) snake_case , snake_case : Tuple = get_dataloaders(__magic_name__ , __magic_name__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) snake_case : Any = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=__magic_name__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). snake_case : Union[str, Any] = model.to(accelerator.device ) # Instantiate optimizer snake_case : Dict = AdamW(params=model.parameters() , lr=__magic_name__ ) # Instantiate scheduler snake_case : Optional[int] = get_linear_schedule_with_warmup( optimizer=__magic_name__ , num_warmup_steps=100 , num_training_steps=(len(__magic_name__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. snake_case , snake_case , snake_case , snake_case , snake_case : Dict = accelerator.prepare( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # Now we train the model for epoch in range(__magic_name__ ): model.train() with LocalSGD( accelerator=__magic_name__ , model=__magic_name__ , local_sgd_steps=__magic_name__ , enabled=local_sgd_steps is not None ) as local_sgd: for step, batch in enumerate(__magic_name__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(__magic_name__ ): snake_case : List[str] = model(**__magic_name__ ) snake_case : str = output.loss accelerator.backward(__magic_name__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() # LocalSGD-specific line local_sgd.step() model.eval() for step, batch in enumerate(__magic_name__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): snake_case : List[Any] = model(**__magic_name__ ) snake_case : Tuple = outputs.logits.argmax(dim=-1 ) snake_case , snake_case : int = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=__magic_name__ , references=__magic_name__ , ) snake_case : List[str] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"epoch {epoch}:" , __magic_name__ ) def a_ ( ) -> int: """simple docstring""" snake_case : Optional[Any] = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=__magic_name__ , default=__magic_name__ , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) # New Code # parser.add_argument( '''--gradient_accumulation_steps''' , type=__magic_name__ , default=1 , help='''The number of minibatches to be ran before gradients are accumulated.''' , ) parser.add_argument( '''--local_sgd_steps''' , type=__magic_name__ , default=8 , help='''Number of local SGD steps or None to disable local SGD''' ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) snake_case : Dict = parser.parse_args() snake_case : Tuple = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(__magic_name__ , __magic_name__ ) if __name__ == "__main__": main()
84
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _a : Tuple = logging.get_logger(__name__) # pylint: disable=invalid-name _a : Dict = '\n Examples:\n ```py\n >>> import torch\n >>> import numpy as np\n\n >>> from diffusers import KandinskyV22PriorPipeline, KandinskyV22ControlnetPipeline\n >>> from transformers import pipeline\n >>> from diffusers.utils import load_image\n\n\n >>> def make_hint(image, depth_estimator):\n ... image = depth_estimator(image)["depth"]\n ... image = np.array(image)\n ... image = image[:, :, None]\n ... image = np.concatenate([image, image, image], axis=2)\n ... detected_map = torch.from_numpy(image).float() / 255.0\n ... hint = detected_map.permute(2, 0, 1)\n ... return hint\n\n\n >>> depth_estimator = pipeline("depth-estimation")\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16\n ... )\n >>> pipe_prior = pipe_prior.to("cuda")\n\n >>> pipe = KandinskyV22ControlnetPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-controlnet-depth", torch_dtype=torch.float16\n ... )\n >>> pipe = pipe.to("cuda")\n\n\n >>> img = load_image(\n ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main"\n ... "/kandinsky/cat.png"\n ... ).resize((768, 768))\n\n >>> hint = make_hint(img, depth_estimator).unsqueeze(0).half().to("cuda")\n\n >>> prompt = "A robot, 4k photo"\n >>> negative_prior_prompt = "lowres, text, error, cropped, worst quality, low quality, jpeg artifacts, ugly, duplicate, morbid, mutilated, out of frame, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, mutation, deformed, blurry, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, username, watermark, signature"\n\n >>> generator = torch.Generator(device="cuda").manual_seed(43)\n\n >>> image_emb, zero_image_emb = pipe_prior(\n ... prompt=prompt, negative_prompt=negative_prior_prompt, generator=generator\n ... ).to_tuple()\n\n >>> images = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... hint=hint,\n ... num_inference_steps=50,\n ... generator=generator,\n ... height=768,\n ... width=768,\n ... ).images\n\n >>> images[0].save("robot_cat.png")\n ```\n' def a_ ( __magic_name__ , __magic_name__ , __magic_name__=8 ) -> str: """simple docstring""" snake_case : List[Any] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 snake_case : Tuple = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class a_ ( a ): def __init__( self : Optional[int] , UpperCAmelCase__ : UNetaDConditionModel , UpperCAmelCase__ : DDPMScheduler , UpperCAmelCase__ : VQModel , ): """simple docstring""" super().__init__() self.register_modules( unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ , movq=UpperCAmelCase__ , ) snake_case : List[Any] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def lowerCAmelCase( self : int , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Any ): """simple docstring""" if latents is None: snake_case : int = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=UpperCAmelCase__ , dtype=UpperCAmelCase__ ) else: if latents.shape != shape: raise ValueError(F"Unexpected latents shape, got {latents.shape}, expected {shape}" ) snake_case : Optional[Any] = latents.to(UpperCAmelCase__ ) snake_case : List[Any] = latents * scheduler.init_noise_sigma return latents def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Optional[int]=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) snake_case : Union[str, Any] = torch.device(F"cuda:{gpu_id}" ) snake_case : Dict = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : Any=0 ): """simple docstring""" if is_accelerate_available() and is_accelerate_version('''>=''' , '''0.17.0.dev0''' ): from accelerate import cpu_offload_with_hook else: raise ImportError('''`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.''' ) snake_case : Optional[int] = torch.device(F"cuda:{gpu_id}" ) if self.device.type != "cpu": self.to('''cpu''' , silence_dtype_warnings=UpperCAmelCase__ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) snake_case : List[str] = None for cpu_offloaded_model in [self.unet, self.movq]: snake_case , snake_case : Optional[int] = cpu_offload_with_hook(UpperCAmelCase__ , UpperCAmelCase__ , prev_module_hook=UpperCAmelCase__ ) # We'll offload the last model manually. snake_case : Tuple = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" if not hasattr(self.unet , '''_hf_hook''' ): return self.device for module in self.unet.modules(): if ( hasattr(UpperCAmelCase__ , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(UpperCAmelCase__ ) def __call__( self : List[str] , UpperCAmelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase__ : torch.FloatTensor , UpperCAmelCase__ : int = 512 , UpperCAmelCase__ : int = 512 , UpperCAmelCase__ : int = 100 , UpperCAmelCase__ : float = 4.0 , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase__ : Optional[torch.FloatTensor] = None , UpperCAmelCase__ : Optional[str] = "pil" , UpperCAmelCase__ : bool = True , ): """simple docstring""" snake_case : Optional[int] = self._execution_device snake_case : Union[str, Any] = guidance_scale > 1.0 if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Any = torch.cat(UpperCAmelCase__ , dim=0 ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Union[str, Any] = torch.cat(UpperCAmelCase__ , dim=0 ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : int = torch.cat(UpperCAmelCase__ , dim=0 ) snake_case : List[Any] = image_embeds.shape[0] * num_images_per_prompt if do_classifier_free_guidance: snake_case : Dict = image_embeds.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Optional[Any] = negative_image_embeds.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Tuple = hint.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Optional[Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=UpperCAmelCase__ ) snake_case : List[str] = torch.cat([hint, hint] , dim=0 ).to(dtype=self.unet.dtype , device=UpperCAmelCase__ ) self.scheduler.set_timesteps(UpperCAmelCase__ , device=UpperCAmelCase__ ) snake_case : str = self.scheduler.timesteps snake_case : Optional[Any] = self.movq.config.latent_channels snake_case , snake_case : Optional[Any] = downscale_height_and_width(UpperCAmelCase__ , UpperCAmelCase__ , self.movq_scale_factor ) # create initial latent snake_case : Dict = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , self.scheduler , ) for i, t in enumerate(self.progress_bar(UpperCAmelCase__ ) ): # expand the latents if we are doing classifier free guidance snake_case : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents snake_case : Optional[int] = {'''image_embeds''': image_embeds, '''hint''': hint} snake_case : Any = self.unet( sample=UpperCAmelCase__ , timestep=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , added_cond_kwargs=UpperCAmelCase__ , return_dict=UpperCAmelCase__ , )[0] if do_classifier_free_guidance: snake_case , snake_case : Dict = noise_pred.split(latents.shape[1] , dim=1 ) snake_case , snake_case : Any = noise_pred.chunk(2 ) snake_case , snake_case : Dict = variance_pred.chunk(2 ) snake_case : str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) snake_case : List[str] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , '''variance_type''' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): snake_case , snake_case : Tuple = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 snake_case : List[Any] = self.scheduler.step( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , generator=UpperCAmelCase__ , )[0] # post-processing snake_case : List[Any] = self.movq.decode(UpperCAmelCase__ , force_not_quantize=UpperCAmelCase__ )['''sample'''] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}" ) if output_type in ["np", "pil"]: snake_case : Optional[Any] = image * 0.5 + 0.5 snake_case : int = image.clamp(0 , 1 ) snake_case : List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": snake_case : str = self.numpy_to_pil(UpperCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase__ )
84
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _a : List[Any] = logging.get_logger(__name__) _a : Optional[int] = { 'google/switch-base-8': 'https://huggingface.co/google/switch-base-8/blob/main/config.json', } class a_ ( a ): A__ : Dict = 'switch_transformers' A__ : Optional[int] = ['past_key_values'] A__ : Tuple = {'hidden_size': 'd_model', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers'} def __init__( self : Any , UpperCAmelCase__ : Optional[Any]=32_128 , UpperCAmelCase__ : List[str]=768 , UpperCAmelCase__ : List[Any]=64 , UpperCAmelCase__ : List[str]=2_048 , UpperCAmelCase__ : Any=64 , UpperCAmelCase__ : Optional[Any]=12 , UpperCAmelCase__ : List[str]=3 , UpperCAmelCase__ : Any=12 , UpperCAmelCase__ : Tuple=3 , UpperCAmelCase__ : str=12 , UpperCAmelCase__ : Optional[Any]=8 , UpperCAmelCase__ : Dict=False , UpperCAmelCase__ : Dict=0.01 , UpperCAmelCase__ : str="float32" , UpperCAmelCase__ : Tuple=False , UpperCAmelCase__ : int=32 , UpperCAmelCase__ : Tuple=128 , UpperCAmelCase__ : int=0.1 , UpperCAmelCase__ : Optional[int]=1e-6 , UpperCAmelCase__ : Optional[Any]=0.001 , UpperCAmelCase__ : int=0.001 , UpperCAmelCase__ : Dict=1.0 , UpperCAmelCase__ : Any="relu" , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : Union[str, Any]=0 , UpperCAmelCase__ : List[Any]=1 , **UpperCAmelCase__ : Dict , ): """simple docstring""" snake_case : int = vocab_size snake_case : Any = d_model snake_case : Optional[Any] = d_kv snake_case : List[str] = d_ff snake_case : Optional[int] = num_sparse_encoder_layers snake_case : Dict = num_layers snake_case : List[str] = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry snake_case : Optional[Any] = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: snake_case : Dict = self.num_layers // self.num_sparse_encoder_layers else: snake_case : List[str] = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: snake_case : Any = self.num_decoder_layers // self.num_sparse_decoder_layers else: snake_case : List[str] = self.num_decoder_layers # HACK: this will create 0 sparse layers snake_case : Tuple = num_heads snake_case : int = num_experts snake_case : Optional[int] = expert_capacity snake_case : Optional[int] = router_bias snake_case : Union[str, Any] = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}" ) snake_case : int = router_dtype snake_case : Tuple = router_ignore_padding_tokens snake_case : List[str] = relative_attention_num_buckets snake_case : Tuple = relative_attention_max_distance snake_case : List[str] = dropout_rate snake_case : List[Any] = layer_norm_epsilon snake_case : str = initializer_factor snake_case : Dict = feed_forward_proj snake_case : str = use_cache snake_case : List[str] = add_router_probs snake_case : Dict = router_z_loss_coef snake_case : Union[str, Any] = router_aux_loss_coef snake_case : Tuple = self.feed_forward_proj.split('''-''' ) snake_case : int = act_info[-1] snake_case : Union[str, Any] = act_info[0] == '''gated''' if len(UpperCAmelCase__ ) > 1 and act_info[0] != "gated" or len(UpperCAmelCase__ ) > 2: raise ValueError( F"`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer." '''Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ''' '''\'gated-gelu\' or \'relu\'''' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": snake_case : Tuple = '''gelu_new''' super().__init__( pad_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , is_encoder_decoder=UpperCAmelCase__ , **UpperCAmelCase__ , )
84
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _a : List[Any] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class a_ ( a , unittest.TestCase ): A__ : Dict = ReformerTokenizer A__ : Optional[int] = ReformerTokenizerFast A__ : str = True A__ : Tuple = False A__ : str = True def lowerCAmelCase( self : List[Any] ): """simple docstring""" super().setUp() snake_case : str = ReformerTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : int = '''<s>''' snake_case : List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase__ ) , UpperCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase__ ) , UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Any = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(UpperCAmelCase__ ) , 1_000 ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def lowerCAmelCase( self : Dict ): """simple docstring""" if not self.test_rust_tokenizer: return snake_case : Any = self.get_tokenizer() snake_case : str = self.get_rust_tokenizer() snake_case : Tuple = '''I was born in 92000, and this is falsé.''' snake_case : str = tokenizer.tokenize(UpperCAmelCase__ ) snake_case : int = rust_tokenizer.tokenize(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : Union[str, Any] = tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) snake_case : List[str] = rust_tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : List[str] = self.get_rust_tokenizer() snake_case : Optional[int] = tokenizer.encode(UpperCAmelCase__ ) snake_case : Optional[Any] = rust_tokenizer.encode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : List[Any]=15 ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): snake_case : str = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) # Simple input snake_case : Union[str, Any] = '''This is a simple input''' snake_case : List[str] = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case : int = ('''This is a simple input''', '''This is a pair''') snake_case : int = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Simple input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Simple input self.assertRaises( UpperCAmelCase__ , tokenizer_r.batch_encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' , ) # Pair input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Pair input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Pair input self.assertRaises( UpperCAmelCase__ , tokenizer_r.batch_encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' , ) def lowerCAmelCase( self : str ): """simple docstring""" pass def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Union[str, Any] = ReformerTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) snake_case : List[str] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(UpperCAmelCase__ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , [285, 46, 10, 170, 382] , ) snake_case : int = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) snake_case : int = tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) snake_case : List[Any] = tokenizer.convert_ids_to_tokens(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def lowerCAmelCase( self : Tuple ): """simple docstring""" return ReformerTokenizer.from_pretrained('''google/reformer-crime-and-punishment''' ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : Any = '''Hello World!''' snake_case : Optional[Any] = [126, 32, 262, 152, 38, 72, 287] self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @slow def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[Any] = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) snake_case : Dict = [ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @require_torch @slow def lowerCAmelCase( self : List[Any] ): """simple docstring""" import torch from transformers import ReformerConfig, ReformerModel # Build sequence snake_case : Any = list(self.big_tokenizer.get_vocab().keys() )[:10] snake_case : Union[str, Any] = ''' '''.join(UpperCAmelCase__ ) snake_case : Optional[int] = self.big_tokenizer.encode_plus(UpperCAmelCase__ , return_tensors='''pt''' ) snake_case : List[str] = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors='''pt''' ) snake_case : Optional[Any] = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) snake_case : Tuple = encoded_sequence['''input_ids'''].shape snake_case : List[Any] = ReformerModel(UpperCAmelCase__ ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**UpperCAmelCase__ ) model(**UpperCAmelCase__ ) @slow def lowerCAmelCase( self : Optional[int] ): """simple docstring""" # fmt: off snake_case : Tuple = {'''input_ids''': [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 snake_case : Tuple = [ '''This is a very simple sentence.''', '''The quick brown fox jumps over the lazy dog.''', ] self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase__ , model_name='''google/reformer-crime-and-punishment''' , revision='''0e6c3decb8211d49bf881013425dc8b0448b3f5a''' , padding=UpperCAmelCase__ , sequences=UpperCAmelCase__ , )
84
1
from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
84
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def a_ ( __magic_name__ ) -> Tuple: """simple docstring""" snake_case , snake_case : Any = image.size snake_case , snake_case : List[str] = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 snake_case : List[str] = image.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) snake_case : Dict = np.array(__magic_name__ ).astype(np.floataa ) / 255.0 snake_case : Optional[Any] = image[None].transpose(0 , 3 , 1 , 2 ) snake_case : Tuple = torch.from_numpy(__magic_name__ ) return 2.0 * image - 1.0 class a_ ( a ): def __init__( self : Optional[Any] , UpperCAmelCase__ : VQModel , UpperCAmelCase__ : UNetaDModel , UpperCAmelCase__ : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): """simple docstring""" super().__init__() self.register_modules(vqvae=UpperCAmelCase__ , unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ ) @torch.no_grad() def __call__( self : Any , UpperCAmelCase__ : Union[torch.Tensor, PIL.Image.Image] = None , UpperCAmelCase__ : Optional[int] = 1 , UpperCAmelCase__ : Optional[int] = 100 , UpperCAmelCase__ : Optional[float] = 0.0 , UpperCAmelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase__ : Optional[str] = "pil" , UpperCAmelCase__ : bool = True , ): """simple docstring""" if isinstance(UpperCAmelCase__ , PIL.Image.Image ): snake_case : Optional[int] = 1 elif isinstance(UpperCAmelCase__ , torch.Tensor ): snake_case : Any = image.shape[0] else: raise ValueError(F"`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(UpperCAmelCase__ )}" ) if isinstance(UpperCAmelCase__ , PIL.Image.Image ): snake_case : Optional[Any] = preprocess(UpperCAmelCase__ ) snake_case , snake_case : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image snake_case : List[Any] = (batch_size, self.unet.config.in_channels // 2, height, width) snake_case : str = next(self.unet.parameters() ).dtype snake_case : Dict = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=self.device , dtype=UpperCAmelCase__ ) snake_case : Any = image.to(device=self.device , dtype=UpperCAmelCase__ ) # set timesteps and move to the correct device self.scheduler.set_timesteps(UpperCAmelCase__ , device=self.device ) snake_case : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler snake_case : Union[str, Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] snake_case : Any = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) snake_case : Optional[Any] = {} if accepts_eta: snake_case : Dict = eta for t in self.progress_bar(UpperCAmelCase__ ): # concat latents and low resolution image in the channel dimension. snake_case : Optional[int] = torch.cat([latents, image] , dim=1 ) snake_case : str = self.scheduler.scale_model_input(UpperCAmelCase__ , UpperCAmelCase__ ) # predict the noise residual snake_case : int = self.unet(UpperCAmelCase__ , UpperCAmelCase__ ).sample # compute the previous noisy sample x_t -> x_t-1 snake_case : Any = self.scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample # decode the image latents with the VQVAE snake_case : Optional[int] = self.vqvae.decode(UpperCAmelCase__ ).sample snake_case : int = torch.clamp(UpperCAmelCase__ , -1.0 , 1.0 ) snake_case : Dict = image / 2 + 0.5 snake_case : int = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case : Any = self.numpy_to_pil(UpperCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase__ )
84
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _a : Union[str, Any] = {'configuration_mbart': ['MBART_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MBartConfig', 'MBartOnnxConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = ['MBartTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[str] = ['MBartTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[int] = [ 'MBART_PRETRAINED_MODEL_ARCHIVE_LIST', 'MBartForCausalLM', 'MBartForConditionalGeneration', 'MBartForQuestionAnswering', 'MBartForSequenceClassification', 'MBartModel', 'MBartPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = [ 'TFMBartForConditionalGeneration', 'TFMBartModel', 'TFMBartPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[int] = [ 'FlaxMBartForConditionalGeneration', 'FlaxMBartForQuestionAnswering', 'FlaxMBartForSequenceClassification', 'FlaxMBartModel', 'FlaxMBartPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mbart import MBART_PRETRAINED_CONFIG_ARCHIVE_MAP, MBartConfig, MBartOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart import MBartTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart_fast import MBartTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mbart import ( MBART_PRETRAINED_MODEL_ARCHIVE_LIST, MBartForCausalLM, MBartForConditionalGeneration, MBartForQuestionAnswering, MBartForSequenceClassification, MBartModel, MBartPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mbart import TFMBartForConditionalGeneration, TFMBartModel, TFMBartPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mbart import ( FlaxMBartForConditionalGeneration, FlaxMBartForQuestionAnswering, FlaxMBartForSequenceClassification, FlaxMBartModel, FlaxMBartPreTrainedModel, ) else: import sys _a : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
84
import os import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers.models.realm.configuration_realm import RealmConfig from transformers.models.realm.retrieval_realm import _REALM_BLOCK_RECORDS_FILENAME, RealmRetriever from transformers.models.realm.tokenization_realm import VOCAB_FILES_NAMES, RealmTokenizer class a_ ( a ): def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[Any] = tempfile.mkdtemp() snake_case : Dict = 5 # Realm tok snake_case : str = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''test''', '''question''', '''this''', '''is''', '''the''', '''first''', '''second''', '''third''', '''fourth''', '''fifth''', '''record''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] snake_case : Tuple = os.path.join(self.tmpdirname , '''realm_tokenizer''' ) os.makedirs(UpperCAmelCase__ , exist_ok=UpperCAmelCase__ ) snake_case : Any = os.path.join(UpperCAmelCase__ , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) snake_case : Tuple = os.path.join(self.tmpdirname , '''realm_block_records''' ) os.makedirs(UpperCAmelCase__ , exist_ok=UpperCAmelCase__ ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" return RealmTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''realm_tokenizer''' ) ) def lowerCAmelCase( self : Any ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Any = RealmConfig(num_block_records=self.num_block_records ) return config def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Optional[int] = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''question''': ['''foo''', '''bar'''], '''answers''': [['''Foo''', '''Bar'''], ['''Bar''']], } ) return dataset def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Dict = np.array( [ b'''This is the first record''', b'''This is the second record''', b'''This is the third record''', b'''This is the fourth record''', b'''This is the fifth record''', b'''This is a longer longer longer record''', ] , dtype=UpperCAmelCase__ , ) return block_records def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Tuple = RealmRetriever( block_records=self.get_dummy_block_records() , tokenizer=self.get_tokenizer() , ) return retriever def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[str] = self.get_config() snake_case : Optional[Any] = self.get_dummy_retriever() snake_case : Optional[int] = retriever.tokenizer snake_case : Dict = np.array([0, 3] , dtype='''long''' ) snake_case : Optional[int] = tokenizer(['''Test question'''] ).input_ids snake_case : Union[str, Any] = tokenizer( ['''the fourth'''] , add_special_tokens=UpperCAmelCase__ , return_token_type_ids=UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , ).input_ids snake_case : Optional[Any] = config.reader_seq_len snake_case , snake_case , snake_case , snake_case : List[str] = retriever( UpperCAmelCase__ , UpperCAmelCase__ , answer_ids=UpperCAmelCase__ , max_length=UpperCAmelCase__ , return_tensors='''np''' ) self.assertEqual(len(UpperCAmelCase__ ) , 2 ) self.assertEqual(len(UpperCAmelCase__ ) , 2 ) self.assertEqual(len(UpperCAmelCase__ ) , 2 ) self.assertEqual(concat_inputs.input_ids.shape , (2, 10) ) self.assertEqual(concat_inputs.attention_mask.shape , (2, 10) ) self.assertEqual(concat_inputs.token_type_ids.shape , (2, 10) ) self.assertEqual(concat_inputs.special_tokens_mask.shape , (2, 10) ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[0] ) , ['''[CLS]''', '''test''', '''question''', '''[SEP]''', '''this''', '''is''', '''the''', '''first''', '''record''', '''[SEP]'''] , ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[1] ) , ['''[CLS]''', '''test''', '''question''', '''[SEP]''', '''this''', '''is''', '''the''', '''fourth''', '''record''', '''[SEP]'''] , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[Any] = self.get_config() snake_case : Optional[int] = self.get_dummy_retriever() snake_case : List[str] = retriever.tokenizer snake_case : Optional[Any] = np.array([0, 3, 5] , dtype='''long''' ) snake_case : Optional[int] = tokenizer(['''Test question'''] ).input_ids snake_case : Any = tokenizer( ['''the fourth''', '''longer longer'''] , add_special_tokens=UpperCAmelCase__ , return_token_type_ids=UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , ).input_ids snake_case : List[Any] = config.reader_seq_len snake_case , snake_case , snake_case , snake_case : Union[str, Any] = retriever( UpperCAmelCase__ , UpperCAmelCase__ , answer_ids=UpperCAmelCase__ , max_length=UpperCAmelCase__ , return_tensors='''np''' ) self.assertEqual([False, True, True] , UpperCAmelCase__ ) self.assertEqual([[-1, -1, -1], [6, -1, -1], [6, 7, 8]] , UpperCAmelCase__ ) self.assertEqual([[-1, -1, -1], [7, -1, -1], [7, 8, 9]] , UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : int = self.get_dummy_retriever() retriever.save_pretrained(os.path.join(self.tmpdirname , '''realm_block_records''' ) ) # Test local path snake_case : Optional[Any] = retriever.from_pretrained(os.path.join(self.tmpdirname , '''realm_block_records''' ) ) self.assertEqual(retriever.block_records[0] , b'''This is the first record''' ) # Test mocked remote path with patch('''transformers.models.realm.retrieval_realm.hf_hub_download''' ) as mock_hf_hub_download: snake_case : Any = os.path.join( os.path.join(self.tmpdirname , '''realm_block_records''' ) , _REALM_BLOCK_RECORDS_FILENAME ) snake_case : Any = RealmRetriever.from_pretrained('''google/realm-cc-news-pretrained-openqa''' ) self.assertEqual(retriever.block_records[0] , b'''This is the first record''' )
84
1
import math def a_ ( __magic_name__ ) -> list: """simple docstring""" snake_case : str = [True] * n snake_case : int = False snake_case : Any = False snake_case : Any = True for i in range(3 , int(n**0.5 + 1 ) , 2 ): snake_case : Union[str, Any] = i * 2 while index < n: snake_case : Dict = False snake_case : Dict = index + i snake_case : Tuple = [2] for i in range(3 , __magic_name__ , 2 ): if is_prime[i]: primes.append(__magic_name__ ) return primes def a_ ( __magic_name__ = 999_966_663_333 ) -> int: """simple docstring""" snake_case : Dict = math.floor(math.sqrt(__magic_name__ ) ) + 100 snake_case : Union[str, Any] = prime_sieve(__magic_name__ ) snake_case : Tuple = 0 snake_case : Dict = 0 snake_case : str = primes[prime_index] while (last_prime**2) <= limit: snake_case : Dict = primes[prime_index + 1] snake_case : Optional[Any] = last_prime**2 snake_case : Any = next_prime**2 # Get numbers divisible by lps(current) snake_case : List[str] = lower_bound + last_prime while upper_bound > current <= limit: matches_sum += current current += last_prime # Reset the upper_bound while (upper_bound - next_prime) > limit: upper_bound -= next_prime # Add the numbers divisible by ups(current) snake_case : List[Any] = upper_bound - next_prime while current > lower_bound: matches_sum += current current -= next_prime # Remove the numbers divisible by both ups and lps snake_case : Optional[Any] = 0 while upper_bound > current <= limit: if current <= lower_bound: # Increment the current number current += last_prime * next_prime continue if current > limit: break # Remove twice since it was added by both ups and lps matches_sum -= current * 2 # Increment the current number current += last_prime * next_prime # Setup for next pair snake_case : Tuple = next_prime prime_index += 1 return matches_sum if __name__ == "__main__": print(solution())
84
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _a : str = {'configuration_encoder_decoder': ['EncoderDecoderConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = ['EncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Dict = ['TFEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = ['FlaxEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys _a : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
84
1
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging _a : Any = logging.get_logger(__name__) _a : Any = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class a_ ( a ): A__ : List[str] = 'mvp' A__ : Optional[int] = ['past_key_values'] A__ : Optional[int] = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self : str , UpperCAmelCase__ : Optional[Any]=50_267 , UpperCAmelCase__ : Tuple=1_024 , UpperCAmelCase__ : Dict=12 , UpperCAmelCase__ : List[str]=4_096 , UpperCAmelCase__ : Optional[int]=16 , UpperCAmelCase__ : Optional[int]=12 , UpperCAmelCase__ : Union[str, Any]=4_096 , UpperCAmelCase__ : Optional[Any]=16 , UpperCAmelCase__ : Dict=0.0 , UpperCAmelCase__ : str=0.0 , UpperCAmelCase__ : List[Any]="gelu" , UpperCAmelCase__ : Optional[int]=1_024 , UpperCAmelCase__ : Optional[int]=0.1 , UpperCAmelCase__ : Optional[Any]=0.0 , UpperCAmelCase__ : Any=0.0 , UpperCAmelCase__ : Dict=0.02 , UpperCAmelCase__ : List[Any]=0.0 , UpperCAmelCase__ : str=False , UpperCAmelCase__ : Any=True , UpperCAmelCase__ : Optional[int]=1 , UpperCAmelCase__ : Optional[int]=0 , UpperCAmelCase__ : Tuple=2 , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : int=2 , UpperCAmelCase__ : List[Any]=2 , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : int=100 , UpperCAmelCase__ : List[str]=800 , **UpperCAmelCase__ : Optional[int] , ): """simple docstring""" snake_case : Any = vocab_size snake_case : Tuple = max_position_embeddings snake_case : List[Any] = d_model snake_case : Optional[Any] = encoder_ffn_dim snake_case : Union[str, Any] = encoder_layers snake_case : Union[str, Any] = encoder_attention_heads snake_case : Any = decoder_ffn_dim snake_case : str = decoder_layers snake_case : List[str] = decoder_attention_heads snake_case : int = dropout snake_case : Union[str, Any] = attention_dropout snake_case : str = activation_dropout snake_case : Any = activation_function snake_case : int = init_std snake_case : Union[str, Any] = encoder_layerdrop snake_case : Dict = decoder_layerdrop snake_case : Dict = classifier_dropout snake_case : Tuple = use_cache snake_case : int = encoder_layers snake_case : Union[str, Any] = scale_embedding # scale factor will be sqrt(d_model) if True snake_case : int = use_prompt snake_case : str = prompt_length snake_case : Tuple = prompt_mid_dim super().__init__( pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , is_encoder_decoder=UpperCAmelCase__ , decoder_start_token_id=UpperCAmelCase__ , forced_eos_token_id=UpperCAmelCase__ , **UpperCAmelCase__ , ) if self.forced_bos_token_id is None and kwargs.get('''force_bos_token_to_be_generated''' , UpperCAmelCase__ ): snake_case : List[str] = self.bos_token_id warnings.warn( F"Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. " '''The config can simply be saved and uploaded again to be fixed.''' )
84
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer _a : str = logging.get_logger(__name__) _a : Optional[int] = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} _a : Optional[Any] = { 'vocab_file': { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json' ), }, } _a : Union[str, Any] = { 'yjernite/retribert-base-uncased': 512, } _a : Tuple = { 'yjernite/retribert-base-uncased': {'do_lower_case': True}, } class a_ ( a ): A__ : List[str] = VOCAB_FILES_NAMES A__ : Any = PRETRAINED_VOCAB_FILES_MAP A__ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Any = PRETRAINED_INIT_CONFIGURATION A__ : Optional[Any] = RetriBertTokenizer A__ : Any = ['input_ids', 'attention_mask'] def __init__( self : Optional[int] , UpperCAmelCase__ : Any=None , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Dict="[UNK]" , UpperCAmelCase__ : str="[SEP]" , UpperCAmelCase__ : Union[str, Any]="[PAD]" , UpperCAmelCase__ : Dict="[CLS]" , UpperCAmelCase__ : Optional[Any]="[MASK]" , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Optional[int]=None , **UpperCAmelCase__ : Dict , ): """simple docstring""" super().__init__( UpperCAmelCase__ , tokenizer_file=UpperCAmelCase__ , do_lower_case=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , tokenize_chinese_chars=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ , **UpperCAmelCase__ , ) snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , UpperCAmelCase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' , UpperCAmelCase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , UpperCAmelCase__ ) != tokenize_chinese_chars ): snake_case : int = getattr(UpperCAmelCase__ , normalizer_state.pop('''type''' ) ) snake_case : List[Any] = do_lower_case snake_case : Union[str, Any] = strip_accents snake_case : int = tokenize_chinese_chars snake_case : int = normalizer_class(**UpperCAmelCase__ ) snake_case : Union[str, Any] = do_lower_case def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Union[str, Any]=None ): """simple docstring""" snake_case : str = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ): """simple docstring""" snake_case : List[Any] = [self.sep_token_id] snake_case : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCAmelCase( self : str , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ): """simple docstring""" snake_case : Tuple = self._tokenizer.model.save(UpperCAmelCase__ , name=UpperCAmelCase__ ) return tuple(UpperCAmelCase__ )
84
1
def a_ ( __magic_name__ ) -> int: """simple docstring""" snake_case : list[list[int]] = [[0 for _ in range(__magic_name__ )] for _ in range(m + 1 )] for i in range(m + 1 ): snake_case : Optional[int] = 1 for n in range(m + 1 ): for k in range(1 , __magic_name__ ): memo[n][k] += memo[n][k - 1] if n - k > 0: memo[n][k] += memo[n - k - 1][k] return memo[m][m - 1] if __name__ == "__main__": import sys if len(sys.argv) == 1: try: _a : List[str] = int(input('Enter a number: ').strip()) print(partition(n)) except ValueError: print('Please enter a number.') else: try: _a : Dict = int(sys.argv[1]) print(partition(n)) except ValueError: print('Please pass a number.')
84
import string import numpy def a_ ( __magic_name__ , __magic_name__ ) -> int: """simple docstring""" return b if a == 0 else greatest_common_divisor(b % a , __magic_name__ ) class a_ : A__ : List[Any] = string.ascii_uppercase + string.digits # This cipher takes alphanumerics into account # i.e. a total of 36 characters # take x and return x % len(key_string) A__ : List[str] = numpy.vectorize(lambda a : x % 36 ) A__ : Dict = numpy.vectorize(a ) def __init__( self : List[str] , UpperCAmelCase__ : numpy.ndarray ): """simple docstring""" snake_case : int = self.modulus(UpperCAmelCase__ ) # mod36 calc's on the encrypt key self.check_determinant() # validate the determinant of the encryption key snake_case : List[str] = encrypt_key.shape[0] def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : str ): """simple docstring""" return self.key_string.index(UpperCAmelCase__ ) def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : int ): """simple docstring""" return self.key_string[round(UpperCAmelCase__ )] def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[Any] = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: snake_case : Tuple = det % len(self.key_string ) snake_case : Tuple = len(self.key_string ) if greatest_common_divisor(UpperCAmelCase__ , len(self.key_string ) ) != 1: snake_case : List[Any] = ( F"determinant modular {req_l} of encryption key({det}) " F"is not co prime w.r.t {req_l}.\nTry another key." ) raise ValueError(UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Optional[int] = [char for char in text.upper() if char in self.key_string] snake_case : Optional[int] = chars[-1] while len(UpperCAmelCase__ ) % self.break_key != 0: chars.append(UpperCAmelCase__ ) return "".join(UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Optional[int] = self.process_text(text.upper() ) snake_case : Optional[int] = '''''' for i in range(0 , len(UpperCAmelCase__ ) - self.break_key + 1 , self.break_key ): snake_case : int = text[i : i + self.break_key] snake_case : int = [self.replace_letters(UpperCAmelCase__ ) for char in batch] snake_case : Tuple = numpy.array([vec] ).T snake_case : Optional[Any] = self.modulus(self.encrypt_key.dot(UpperCAmelCase__ ) ).T.tolist()[ 0 ] snake_case : Dict = ''''''.join( self.replace_digits(UpperCAmelCase__ ) for num in batch_encrypted ) encrypted += encrypted_batch return encrypted def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Optional[int] = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: snake_case : int = det % len(self.key_string ) snake_case : Dict = None for i in range(len(self.key_string ) ): if (det * i) % len(self.key_string ) == 1: snake_case : Any = i break snake_case : Any = ( det_inv * numpy.linalg.det(self.encrypt_key ) * numpy.linalg.inv(self.encrypt_key ) ) return self.to_int(self.modulus(UpperCAmelCase__ ) ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Any = self.make_decrypt_key() snake_case : Optional[Any] = self.process_text(text.upper() ) snake_case : int = '''''' for i in range(0 , len(UpperCAmelCase__ ) - self.break_key + 1 , self.break_key ): snake_case : Any = text[i : i + self.break_key] snake_case : int = [self.replace_letters(UpperCAmelCase__ ) for char in batch] snake_case : List[str] = numpy.array([vec] ).T snake_case : Optional[Any] = self.modulus(decrypt_key.dot(UpperCAmelCase__ ) ).T.tolist()[0] snake_case : int = ''''''.join( self.replace_digits(UpperCAmelCase__ ) for num in batch_decrypted ) decrypted += decrypted_batch return decrypted def a_ ( ) -> None: """simple docstring""" snake_case : Any = int(input('''Enter the order of the encryption key: ''' ) ) snake_case : List[Any] = [] print('''Enter each row of the encryption key with space separated integers''' ) for _ in range(__magic_name__ ): snake_case : Optional[Any] = [int(__magic_name__ ) for x in input().split()] hill_matrix.append(__magic_name__ ) snake_case : List[str] = HillCipher(numpy.array(__magic_name__ ) ) print('''Would you like to encrypt or decrypt some text? (1 or 2)''' ) snake_case : int = input('''\n1. Encrypt\n2. Decrypt\n''' ) if option == "1": snake_case : List[Any] = input('''What text would you like to encrypt?: ''' ) print('''Your encrypted text is:''' ) print(hc.encrypt(__magic_name__ ) ) elif option == "2": snake_case : int = input('''What text would you like to decrypt?: ''' ) print('''Your decrypted text is:''' ) print(hc.decrypt(__magic_name__ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
84
1
import warnings from typing import Dict, List, Optional, Tuple from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _a : List[str] = logging.get_logger(__name__) class a_ ( a ): A__ : int = ['input_ids', 'attention_mask'] def __init__( self : List[Any] , UpperCAmelCase__ : Any="</s>" , UpperCAmelCase__ : Any="<unk>" , UpperCAmelCase__ : Optional[int]="<pad>" , UpperCAmelCase__ : int=125 , UpperCAmelCase__ : Dict=None , **UpperCAmelCase__ : Dict , ): """simple docstring""" # Add extra_ids to the special token list if extra_ids > 0 and additional_special_tokens is None: snake_case : Optional[Any] = [F"<extra_id_{i}>" for i in range(UpperCAmelCase__ )] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens snake_case : List[str] = len(set(filter(lambda UpperCAmelCase__ : bool('''extra_id''' in str(UpperCAmelCase__ ) ) , UpperCAmelCase__ ) ) ) if extra_tokens != extra_ids: raise ValueError( F"Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are" ''' provided to ByT5Tokenizer. In this case the additional_special_tokens must include the''' ''' extra_ids tokens''' ) snake_case : Optional[int] = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else pad_token snake_case : str = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else eos_token snake_case : Optional[int] = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else unk_token super().__init__( eos_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , extra_ids=UpperCAmelCase__ , additional_special_tokens=UpperCAmelCase__ , **UpperCAmelCase__ , ) snake_case : Optional[Any] = extra_ids snake_case : Optional[int] = 2**8 # utf is 8 bits # define special tokens dict snake_case : Dict[int, str] = { self.pad_token: 0, self.eos_token: 1, self.unk_token: 2, } snake_case : List[str] = len(self.special_tokens_encoder ) snake_case : Tuple = len(UpperCAmelCase__ ) for i, token in enumerate(UpperCAmelCase__ ): snake_case : str = self.vocab_size + i - n snake_case : Dict[str, int] = {v: k for k, v in self.special_tokens_encoder.items()} @property def lowerCAmelCase( self : List[Any] ): """simple docstring""" return self._utf_vocab_size + self._num_special_tokens + self._extra_ids def lowerCAmelCase( self : Dict , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None , UpperCAmelCase__ : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase__ , token_ids_a=UpperCAmelCase__ , already_has_special_tokens=UpperCAmelCase__ ) # normal case: some special tokens if token_ids_a is None: return ([0] * len(UpperCAmelCase__ )) + [1] return ([0] * len(UpperCAmelCase__ )) + [1] + ([0] * len(UpperCAmelCase__ )) + [1] def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : List[int] ): """simple docstring""" if len(UpperCAmelCase__ ) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( F"This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated" ''' eos tokens being added.''' ) return token_ids else: return token_ids + [self.eos_token_id] def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ): """simple docstring""" snake_case : Any = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos ) * [0] return len(token_ids_a + eos + token_ids_a + eos ) * [0] def lowerCAmelCase( self : str , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ): """simple docstring""" snake_case : Any = self._add_eos_if_not_present(UpperCAmelCase__ ) if token_ids_a is None: return token_ids_a else: snake_case : Tuple = self._add_eos_if_not_present(UpperCAmelCase__ ) return token_ids_a + token_ids_a def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Optional[int] = [chr(UpperCAmelCase__ ) for i in text.encode('''utf-8''' )] return tokens def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : Tuple ): """simple docstring""" if token in self.special_tokens_encoder: snake_case : List[Any] = self.special_tokens_encoder[token] elif token in self.added_tokens_encoder: snake_case : Union[str, Any] = self.added_tokens_encoder[token] elif len(UpperCAmelCase__ ) != 1: snake_case : Optional[Any] = self.unk_token_id else: snake_case : Dict = ord(UpperCAmelCase__ ) + self._num_special_tokens return token_id def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : Optional[int] ): """simple docstring""" if index in self.special_tokens_decoder: snake_case : Optional[int] = self.special_tokens_decoder[index] else: snake_case : List[Any] = chr(index - self._num_special_tokens ) return token def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" snake_case : List[str] = b'''''' for token in tokens: if token in self.special_tokens_decoder: snake_case : Dict = self.special_tokens_decoder[token].encode('''utf-8''' ) elif token in self.added_tokens_decoder: snake_case : Any = self.special_tokens_decoder[token].encode('''utf-8''' ) elif token in self.special_tokens_encoder: snake_case : Optional[int] = token.encode('''utf-8''' ) elif token in self.added_tokens_encoder: snake_case : int = token.encode('''utf-8''' ) else: snake_case : str = bytes([ord(UpperCAmelCase__ )] ) bstring += tok_string snake_case : Union[str, Any] = bstring.decode('''utf-8''' , errors='''ignore''' ) return string def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ): """simple docstring""" return ()
84
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class a_ ( a ): A__ : List[Any] = 'Salesforce/blip-image-captioning-base' A__ : Dict = ( 'This is a tool that generates a description of an image. It takes an input named `image` which should be the ' 'image to caption, and returns a text that contains the description in English.' ) A__ : str = 'image_captioner' A__ : Dict = AutoModelForVisionaSeq A__ : Optional[Any] = ['image'] A__ : List[str] = ['text'] def __init__( self : List[str] , *UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" requires_backends(self , ['''vision'''] ) super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : "Image" ): """simple docstring""" return self.pre_processor(images=UpperCAmelCase__ , return_tensors='''pt''' ) def lowerCAmelCase( self : Any , UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" return self.model.generate(**UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : List[Any] ): """simple docstring""" return self.pre_processor.batch_decode(UpperCAmelCase__ , skip_special_tokens=UpperCAmelCase__ )[0].strip()
84
1
from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class a_ ( nn.Module ): def __init__( self : Tuple , UpperCAmelCase__ : int = 16 , UpperCAmelCase__ : int = 88 , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : float = 0.0 , UpperCAmelCase__ : int = 32 , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : str = "geglu" , UpperCAmelCase__ : Optional[int] = None , ): """simple docstring""" super().__init__() snake_case : Tuple = nn.ModuleList( [ TransformeraDModel( num_attention_heads=UpperCAmelCase__ , attention_head_dim=UpperCAmelCase__ , in_channels=UpperCAmelCase__ , num_layers=UpperCAmelCase__ , dropout=UpperCAmelCase__ , norm_num_groups=UpperCAmelCase__ , cross_attention_dim=UpperCAmelCase__ , attention_bias=UpperCAmelCase__ , sample_size=UpperCAmelCase__ , num_vector_embeds=UpperCAmelCase__ , activation_fn=UpperCAmelCase__ , num_embeds_ada_norm=UpperCAmelCase__ , ) for _ in range(2 ) ] ) # Variables that can be set by a pipeline: # The ratio of transformer1 to transformer2's output states to be combined during inference snake_case : Tuple = 0.5 # The shape of `encoder_hidden_states` is expected to be # `(batch_size, condition_lengths[0]+condition_lengths[1], num_features)` snake_case : Optional[Any] = [77, 257] # Which transformer to use to encode which condition. # E.g. `(1, 0)` means that we'll use `transformers[1](conditions[0])` and `transformers[0](conditions[1])` snake_case : int = [1, 0] def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : Optional[Any]=None , UpperCAmelCase__ : bool = True , ): """simple docstring""" snake_case : Tuple = hidden_states snake_case : Optional[Any] = [] snake_case : Tuple = 0 # attention_mask is not used yet for i in range(2 ): # for each of the two transformers, pass the corresponding condition tokens snake_case : Union[str, Any] = encoder_hidden_states[:, tokens_start : tokens_start + self.condition_lengths[i]] snake_case : Dict = self.transformer_index_for_condition[i] snake_case : Tuple = self.transformers[transformer_index]( UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , timestep=UpperCAmelCase__ , cross_attention_kwargs=UpperCAmelCase__ , return_dict=UpperCAmelCase__ , )[0] encoded_states.append(encoded_state - input_states ) tokens_start += self.condition_lengths[i] snake_case : Optional[int] = encoded_states[0] * self.mix_ratio + encoded_states[1] * (1 - self.mix_ratio) snake_case : Optional[int] = output_states + input_states if not return_dict: return (output_states,) return TransformeraDModelOutput(sample=UpperCAmelCase__ )
84
def a_ ( __magic_name__ ) -> bool: """simple docstring""" if p < 2: raise ValueError('''p should not be less than 2!''' ) elif p == 2: return True snake_case : int = 4 snake_case : Optional[Any] = (1 << p) - 1 for _ in range(p - 2 ): snake_case : Optional[Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
84
1
import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class a_ : def lowerCAmelCase( self : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Any ): """simple docstring""" return None class a_ : def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[Any] ): """simple docstring""" return None class a_ ( unittest.TestCase ): A__ : Optional[Any] = [ # (model_name, model_kwargs) ('bert-base-cased', {}), ('gpt2', {'use_cache': False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def lowerCAmelCase( self : List[Any] ): """simple docstring""" for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(UpperCAmelCase__ , '''tf''' , 12 , **UpperCAmelCase__ ) @require_torch @slow def lowerCAmelCase( self : int ): """simple docstring""" for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(UpperCAmelCase__ , '''pt''' , 12 , **UpperCAmelCase__ ) @require_torch @slow def lowerCAmelCase( self : Tuple ): """simple docstring""" from transformers import BertModel snake_case : Dict = ['''[UNK]''', '''[SEP]''', '''[CLS]''', '''[PAD]''', '''[MASK]''', '''some''', '''other''', '''words'''] with NamedTemporaryFile(mode='''w+t''' ) as vocab_file: vocab_file.write('''\n'''.join(UpperCAmelCase__ ) ) vocab_file.flush() snake_case : List[Any] = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: snake_case : Dict = BertModel(BertConfig(vocab_size=len(UpperCAmelCase__ ) ) ) model.save_pretrained(UpperCAmelCase__ ) self._test_export(UpperCAmelCase__ , '''pt''' , 12 , UpperCAmelCase__ ) @require_tf @slow def lowerCAmelCase( self : Dict ): """simple docstring""" for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: snake_case : Dict = self._test_export(UpperCAmelCase__ , '''tf''' , 12 , **UpperCAmelCase__ ) snake_case : Dict = quantize(Path(UpperCAmelCase__ ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(UpperCAmelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) @require_torch @slow def lowerCAmelCase( self : Optional[int] ): """simple docstring""" for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: snake_case : Optional[Any] = self._test_export(UpperCAmelCase__ , '''pt''' , 12 , **UpperCAmelCase__ ) snake_case : str = quantize(UpperCAmelCase__ ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(UpperCAmelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Any , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[Any]=None , **UpperCAmelCase__ : int ): """simple docstring""" try: # Compute path with TemporaryDirectory() as tempdir: snake_case : Tuple = Path(UpperCAmelCase__ ).joinpath('''model.onnx''' ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ) return path except Exception as e: self.fail(UpperCAmelCase__ ) @require_torch @require_tokenizers @slow def lowerCAmelCase( self : Optional[int] ): """simple docstring""" from transformers import BertModel snake_case : int = BertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) snake_case : Tuple = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(UpperCAmelCase__ , UpperCAmelCase__ , '''pt''' ) @require_tf @require_tokenizers @slow def lowerCAmelCase( self : str ): """simple docstring""" from transformers import TFBertModel snake_case : Union[str, Any] = TFBertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) snake_case : Optional[int] = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(UpperCAmelCase__ , UpperCAmelCase__ , '''tf''' ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : int ): """simple docstring""" snake_case : List[str] = FeatureExtractionPipeline(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : Tuple = ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''output_0''', '''output_1'''] snake_case , snake_case , snake_case , snake_case : Union[str, Any] = infer_shapes(UpperCAmelCase__ , UpperCAmelCase__ ) # Assert all variables are present self.assertEqual(len(UpperCAmelCase__ ) , len(UpperCAmelCase__ ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] , UpperCAmelCase__ ) self.assertSequenceEqual(variable_names[3:] , UpperCAmelCase__ ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] , {0: '''batch''', 1: '''sequence'''} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes['''output_0'''] , {0: '''batch''', 1: '''sequence'''} ) self.assertDictEqual(shapes['''output_1'''] , {0: '''batch'''} ) def lowerCAmelCase( self : int ): """simple docstring""" snake_case : int = ['''input_ids''', '''attention_mask''', '''token_type_ids'''] snake_case : Dict = {'''input_ids''': [1, 2, 3, 4], '''attention_mask''': [0, 0, 0, 0], '''token_type_ids''': [1, 1, 1, 1]} snake_case , snake_case : Optional[Any] = ensure_valid_input(FuncContiguousArgs() , UpperCAmelCase__ , UpperCAmelCase__ ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(UpperCAmelCase__ ) , 3 ) # Should have exactly the same input names self.assertEqual(set(UpperCAmelCase__ ) , set(UpperCAmelCase__ ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(UpperCAmelCase__ , (tokens['''input_ids'''], tokens['''token_type_ids'''], tokens['''attention_mask''']) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) snake_case , snake_case : Optional[int] = ensure_valid_input(FuncNonContiguousArgs() , UpperCAmelCase__ , UpperCAmelCase__ ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(UpperCAmelCase__ ) , 1 ) self.assertEqual(len(UpperCAmelCase__ ) , 1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] , tokens['''input_ids'''] ) self.assertEqual(ordered_input_names[0] , '''input_ids''' ) def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Tuple = generate_identified_filename(Path('''/home/something/my_fake_model.onnx''' ) , '''-test''' ) self.assertEqual('''/home/something/my_fake_model-test.onnx''' , generated.as_posix() )
84
from sklearn.metrics import fa_score import datasets _a : List[str] = '\nThe F1 score is the harmonic mean of the precision and recall. It can be computed with the equation:\nF1 = 2 * (precision * recall) / (precision + recall)\n' _a : Dict = '\nArgs:\n predictions (`list` of `int`): Predicted labels.\n references (`list` of `int`): Ground truth labels.\n labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None.\n pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1.\n average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`.\n\n - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary.\n - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives.\n - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.\n - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall.\n - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification).\n sample_weight (`list` of `float`): Sample weights Defaults to None.\n\nReturns:\n f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better.\n\nExamples:\n\n Example 1-A simple binary example\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0])\n >>> print(results)\n {\'f1\': 0.5}\n\n Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0)\n >>> print(round(results[\'f1\'], 2))\n 0.67\n\n Example 3-The same simple binary example as in Example 1, but with `sample_weight` included.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3])\n >>> print(round(results[\'f1\'], 2))\n 0.35\n\n Example 4-A multiclass example, with different values for the `average` input.\n >>> predictions = [0, 2, 1, 0, 0, 1]\n >>> references = [0, 1, 2, 0, 1, 2]\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro")\n >>> print(round(results[\'f1\'], 2))\n 0.33\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=None)\n >>> print(results)\n {\'f1\': array([0.8, 0. , 0. ])}\n' _a : List[Any] = '\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P.\n and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and\n Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.},\n journal={Journal of Machine Learning Research},\n volume={12},\n pages={2825--2830},\n year={2011}\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): def lowerCAmelCase( self : Any ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''int32''' ) ), '''references''': datasets.Sequence(datasets.Value('''int32''' ) ), } if self.config_name == '''multilabel''' else { '''predictions''': datasets.Value('''int32''' ), '''references''': datasets.Value('''int32''' ), } ) , reference_urls=['''https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html'''] , ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Optional[Any]=1 , UpperCAmelCase__ : List[str]="binary" , UpperCAmelCase__ : str=None ): """simple docstring""" snake_case : List[Any] = fa_score( UpperCAmelCase__ , UpperCAmelCase__ , labels=UpperCAmelCase__ , pos_label=UpperCAmelCase__ , average=UpperCAmelCase__ , sample_weight=UpperCAmelCase__ ) return {"f1": float(UpperCAmelCase__ ) if score.size == 1 else score}
84
1
import collections import os import re from pathlib import Path _a : Union[str, Any] = 'src/transformers' # Matches is_xxx_available() _a : Union[str, Any] = re.compile(R'is\_([a-z_]*)_available()') # Catches a one-line _import_struct = {xxx} _a : Optional[Any] = re.compile(R'^_import_structure\s+=\s+\{([^\}]+)\}') # Catches a line with a key-values pattern: "bla": ["foo", "bar"] _a : str = re.compile(R'\s+"\S*":\s+\[([^\]]*)\]') # Catches a line if not is_foo_available _a : Optional[int] = re.compile(R'^\s*if\s+not\s+is\_[a-z_]*\_available\(\)') # Catches a line _import_struct["bla"].append("foo") _a : Tuple = re.compile(R'^\s*_import_structure\["\S*"\]\.append\("(\S*)"\)') # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] _a : int = re.compile(R'^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]') # Catches a line with an object between quotes and a comma: "MyModel", _a : Any = re.compile(R'^\s+"([^"]+)",') # Catches a line with objects between brackets only: ["foo", "bar"], _a : List[str] = re.compile(R'^\s+\[([^\]]+)\]') # Catches a line with from foo import bar, bla, boo _a : List[str] = re.compile(R'\s+from\s+\S*\s+import\s+([^\(\s].*)\n') # Catches a line with try: _a : Optional[Any] = re.compile(R'^\s*try:') # Catches a line with else: _a : List[str] = re.compile(R'^\s*else:') def a_ ( __magic_name__ ) -> List[str]: """simple docstring""" if _re_test_backend.search(__magic_name__ ) is None: return None snake_case : Dict = [b[0] for b in _re_backend.findall(__magic_name__ )] backends.sort() return "_and_".join(__magic_name__ ) def a_ ( __magic_name__ ) -> str: """simple docstring""" with open(__magic_name__ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: snake_case : str = f.readlines() snake_case : List[str] = 0 while line_index < len(__magic_name__ ) and not lines[line_index].startswith('''_import_structure = {''' ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(__magic_name__ ): return None # First grab the objects without a specific backend in _import_structure snake_case : str = [] while not lines[line_index].startswith('''if TYPE_CHECKING''' ) and find_backend(lines[line_index] ) is None: snake_case : Any = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(__magic_name__ ): snake_case : Optional[Any] = _re_one_line_import_struct.search(__magic_name__ ).groups()[0] snake_case : Dict = re.findall(R'''\[([^\]]+)\]''' , __magic_name__ ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(''', ''' )] ) line_index += 1 continue snake_case : Optional[int] = _re_import_struct_key_value.search(__magic_name__ ) if single_line_import_search is not None: snake_case : Dict = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(''', ''' ) if len(__magic_name__ ) > 0] objects.extend(__magic_name__ ) elif line.startswith(''' ''' * 8 + '''"''' ): objects.append(line[9:-3] ) line_index += 1 snake_case : Dict = {'''none''': objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith('''if TYPE_CHECKING''' ): # If the line is an if not is_backend_available, we grab all objects associated. snake_case : Union[str, Any] = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: snake_case : Any = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 snake_case : str = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(''' ''' * 4 ): snake_case : Optional[int] = lines[line_index] if _re_import_struct_add_one.search(__magic_name__ ) is not None: objects.append(_re_import_struct_add_one.search(__magic_name__ ).groups()[0] ) elif _re_import_struct_add_many.search(__magic_name__ ) is not None: snake_case : Optional[int] = _re_import_struct_add_many.search(__magic_name__ ).groups()[0].split(''', ''' ) snake_case : List[Any] = [obj[1:-1] for obj in imports if len(__magic_name__ ) > 0] objects.extend(__magic_name__ ) elif _re_between_brackets.search(__magic_name__ ) is not None: snake_case : Optional[int] = _re_between_brackets.search(__magic_name__ ).groups()[0].split(''', ''' ) snake_case : Optional[Any] = [obj[1:-1] for obj in imports if len(__magic_name__ ) > 0] objects.extend(__magic_name__ ) elif _re_quote_object.search(__magic_name__ ) is not None: objects.append(_re_quote_object.search(__magic_name__ ).groups()[0] ) elif line.startswith(''' ''' * 8 + '''"''' ): objects.append(line[9:-3] ) elif line.startswith(''' ''' * 12 + '''"''' ): objects.append(line[13:-3] ) line_index += 1 snake_case : Dict = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend snake_case : Optional[int] = [] while ( line_index < len(__magic_name__ ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith('''else''' ) ): snake_case : int = lines[line_index] snake_case : List[Any] = _re_import.search(__magic_name__ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(''', ''' ) ) elif line.startswith(''' ''' * 8 ): objects.append(line[8:-2] ) line_index += 1 snake_case : Optional[Any] = {'''none''': objects} # Let's continue with backend-specific objects while line_index < len(__magic_name__ ): # If the line is an if is_backend_available, we grab all objects associated. snake_case : str = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: snake_case : Any = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 snake_case : Any = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(''' ''' * 8 ): snake_case : Tuple = lines[line_index] snake_case : Any = _re_import.search(__magic_name__ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(''', ''' ) ) elif line.startswith(''' ''' * 12 ): objects.append(line[12:-2] ) line_index += 1 snake_case : List[Any] = objects else: line_index += 1 return import_dict_objects, type_hint_objects def a_ ( __magic_name__ , __magic_name__ ) -> List[Any]: """simple docstring""" def find_duplicates(__magic_name__ ): return [k for k, v in collections.Counter(__magic_name__ ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] snake_case : Optional[Any] = [] for key in import_dict_objects.keys(): snake_case : Dict = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(F"Duplicate _import_structure definitions for: {duplicate_imports}" ) snake_case : List[str] = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(F"Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}" ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): snake_case : Tuple = '''base imports''' if key == '''none''' else F"{key} backend" errors.append(F"Differences for {name}:" ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(F" {a} in TYPE_HINT but not in _import_structure." ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(F" {a} in _import_structure but not in TYPE_HINT." ) return errors def a_ ( ) -> int: """simple docstring""" snake_case : str = [] for root, _, files in os.walk(__magic_name__ ): if "__init__.py" in files: snake_case : Optional[Any] = os.path.join(__magic_name__ , '''__init__.py''' ) snake_case : Optional[Any] = parse_init(__magic_name__ ) if objects is not None: snake_case : List[Any] = analyze_results(*__magic_name__ ) if len(__magic_name__ ) > 0: snake_case : int = F"Problem in {fname}, both halves do not define the same objects.\n{errors[0]}" failures.append('''\n'''.join(__magic_name__ ) ) if len(__magic_name__ ) > 0: raise ValueError('''\n\n'''.join(__magic_name__ ) ) def a_ ( ) -> str: """simple docstring""" snake_case : List[str] = [] for path, directories, files in os.walk(__magic_name__ ): for folder in directories: # Ignore private modules if folder.startswith('''_''' ): directories.remove(__magic_name__ ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(__magic_name__ ) / folder).glob('''*.py''' ) ) ) == 0: continue snake_case : Optional[int] = str((Path(__magic_name__ ) / folder).relative_to(__magic_name__ ) ) snake_case : Union[str, Any] = short_path.replace(os.path.sep , '''.''' ) submodules.append(__magic_name__ ) for fname in files: if fname == "__init__.py": continue snake_case : str = str((Path(__magic_name__ ) / fname).relative_to(__magic_name__ ) ) snake_case : Union[str, Any] = short_path.replace('''.py''' , '''''' ).replace(os.path.sep , '''.''' ) if len(submodule.split('''.''' ) ) == 1: submodules.append(__magic_name__ ) return submodules _a : int = [ 'convert_pytorch_checkpoint_to_tf2', 'modeling_flax_pytorch_utils', 'models.esm.openfold_utils', ] def a_ ( ) -> List[Any]: """simple docstring""" from transformers.utils import direct_transformers_import snake_case : Union[str, Any] = direct_transformers_import(__magic_name__ ) snake_case : Any = set(transformers._import_structure.keys() ) # This contains all the base keys of the _import_structure object defined in the init, but if the user is missing # some optional dependencies, they may not have all of them. Thus we read the init to read all additions and # (potentiall re-) add them. with open(os.path.join(__magic_name__ , '''__init__.py''' ) , '''r''' ) as f: snake_case : Dict = f.read() import_structure_keys.update(set(re.findall(R'''import_structure\[\"([^\"]*)\"\]''' , __magic_name__ ) ) ) snake_case : str = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in import_structure_keys ] if len(__magic_name__ ) > 0: snake_case : List[Any] = '''\n'''.join(F"- {module}" for module in module_not_registered ) raise ValueError( '''The following submodules are not properly registed in the main init of Transformers:\n''' F"{list_of_modules}\n" '''Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.''' ) if __name__ == "__main__": check_all_inits() check_submodules()
84
def a_ ( __magic_name__ ) -> int: """simple docstring""" if not isinstance(__magic_name__ , __magic_name__ ): raise TypeError('''only integers accepted as input''' ) else: snake_case : str = str(abs(__magic_name__ ) ) snake_case : Optional[Any] = [list(__magic_name__ ) for char in range(len(__magic_name__ ) )] for index in range(len(__magic_name__ ) ): num_transpositions[index].pop(__magic_name__ ) return max( int(''''''.join(list(__magic_name__ ) ) ) for transposition in num_transpositions ) if __name__ == "__main__": __import__('doctest').testmod()
84
1
from __future__ import annotations import collections import pprint from pathlib import Path def a_ ( __magic_name__ ) -> str: """simple docstring""" return "".join(sorted(__magic_name__ ) ) def a_ ( __magic_name__ ) -> list[str]: """simple docstring""" return word_by_signature[signature(__magic_name__ )] _a : str = Path(__file__).parent.joinpath('words.txt').read_text(encoding='utf-8') _a : Tuple = sorted({word.strip().lower() for word in data.splitlines()}) _a : Optional[int] = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": _a : Tuple = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open('anagrams.txt', 'w') as file: file.write('all_anagrams = \n ') file.write(pprint.pformat(all_anagrams))
84
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class a_ : def __init__( self : Union[str, Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any]=99 , UpperCAmelCase__ : Dict=13 , UpperCAmelCase__ : int=7 , UpperCAmelCase__ : Any=9 , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Tuple=False , UpperCAmelCase__ : Tuple=32 , UpperCAmelCase__ : Dict=5 , UpperCAmelCase__ : Optional[int]=4 , UpperCAmelCase__ : List[Any]=37 , UpperCAmelCase__ : Union[str, Any]=8 , UpperCAmelCase__ : Optional[Any]=0.1 , UpperCAmelCase__ : str=0.002 , UpperCAmelCase__ : str=1 , UpperCAmelCase__ : Any=0 , UpperCAmelCase__ : Union[str, Any]=0 , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Optional[Any]=None , ): """simple docstring""" snake_case : Union[str, Any] = parent snake_case : Union[str, Any] = batch_size snake_case : Any = encoder_seq_length snake_case : str = decoder_seq_length # For common tests snake_case : Optional[int] = self.decoder_seq_length snake_case : Optional[Any] = is_training snake_case : List[Any] = use_attention_mask snake_case : Union[str, Any] = use_labels snake_case : Any = vocab_size snake_case : Optional[int] = hidden_size snake_case : List[str] = num_hidden_layers snake_case : Union[str, Any] = num_attention_heads snake_case : Any = d_ff snake_case : Any = relative_attention_num_buckets snake_case : Optional[Any] = dropout_rate snake_case : int = initializer_factor snake_case : Optional[Any] = eos_token_id snake_case : Dict = pad_token_id snake_case : Optional[Any] = decoder_start_token_id snake_case : Union[str, Any] = None snake_case : List[str] = decoder_layers def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" return TaConfig.from_pretrained('''google/umt5-base''' ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Any , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : List[Any]=None , ): """simple docstring""" if attention_mask is None: snake_case : Union[str, Any] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: snake_case : Any = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: snake_case : List[Any] = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=UpperCAmelCase__ ) if decoder_head_mask is None: snake_case : Tuple = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase__ ) if cross_attn_head_mask is None: snake_case : Union[str, Any] = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase__ ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input snake_case : List[str] = input_ids.clamp(self.pad_token_id + 1 ) snake_case : List[str] = decoder_input_ids.clamp(self.pad_token_id + 1 ) snake_case : str = self.get_config() snake_case : Tuple = config.num_attention_heads snake_case : List[Any] = self.prepare_inputs_dict(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) return config, input_dict def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : List[str] = self.prepare_config_and_inputs() return config, inputs_dict def lowerCAmelCase( self : Dict ): """simple docstring""" return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def lowerCAmelCase( self : Tuple ): """simple docstring""" return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def lowerCAmelCase( self : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Tuple , ): """simple docstring""" snake_case : str = UMTaModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : str = model( input_ids=UpperCAmelCase__ , decoder_input_ids=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , decoder_attention_mask=UpperCAmelCase__ , ) snake_case : int = model(input_ids=UpperCAmelCase__ , decoder_input_ids=UpperCAmelCase__ ) snake_case : int = result.last_hidden_state snake_case : Dict = result.past_key_values snake_case : Dict = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(UpperCAmelCase__ ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def lowerCAmelCase( self : int , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[str] , ): """simple docstring""" snake_case : int = UMTaModel(config=UpperCAmelCase__ ).get_decoder().to(UpperCAmelCase__ ).eval() # first forward pass snake_case : List[Any] = model(UpperCAmelCase__ , use_cache=UpperCAmelCase__ ) snake_case : List[Any] = model(UpperCAmelCase__ ) snake_case : Any = model(UpperCAmelCase__ , use_cache=UpperCAmelCase__ ) self.parent.assertTrue(len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) ) self.parent.assertTrue(len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) + 1 ) snake_case , snake_case : List[str] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case : Any = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and snake_case : Union[str, Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case : Any = model(UpperCAmelCase__ )['''last_hidden_state'''] snake_case : Optional[Any] = model(UpperCAmelCase__ , past_key_values=UpperCAmelCase__ )['''last_hidden_state'''] # select random slice snake_case : Tuple = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case : Union[str, Any] = output_from_no_past[:, -1, random_slice_idx].detach() snake_case : Tuple = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1e-3 ) ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : int , ): """simple docstring""" snake_case : int = UMTaModel(config=UpperCAmelCase__ ).to(UpperCAmelCase__ ).half().eval() snake_case : str = model(**UpperCAmelCase__ )['''last_hidden_state'''] self.parent.assertFalse(torch.isnan(UpperCAmelCase__ ).any().item() ) @require_torch class a_ ( a , a , a , unittest.TestCase ): A__ : str = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) A__ : str = (UMTaForConditionalGeneration,) if is_torch_available() else () A__ : Any = ( { 'conversational': UMTaForConditionalGeneration, 'feature-extraction': UMTaModel, 'summarization': UMTaForConditionalGeneration, 'text2text-generation': UMTaForConditionalGeneration, 'translation': UMTaForConditionalGeneration, 'question-answering': UMTaForQuestionAnswering, } if is_torch_available() else {} ) A__ : Dict = True A__ : List[str] = False A__ : Optional[int] = False A__ : Optional[int] = True A__ : List[str] = True # The small UMT5 model needs higher percentages for CPU/MP tests A__ : int = [0.8, 0.9] def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Union[str, Any] = UMTaModelTester(self ) @unittest.skip('''Test has a segmentation fault on torch 1.8.0''' ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Tuple = self.model_tester.prepare_config_and_inputs() snake_case : Optional[Any] = UMTaModel(config_and_inputs[0] ).to(UpperCAmelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( UpperCAmelCase__ , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"{tmpdirname}/t5_test.onnx" , export_params=UpperCAmelCase__ , opset_version=9 , input_names=['''input_ids''', '''decoder_input_ids'''] , ) @unittest.skipIf(torch_device == '''cpu''' , '''Cant do half precision''' ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*UpperCAmelCase__ ) def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : Optional[int] = ['''encoder_attentions''', '''decoder_attentions''', '''cross_attentions'''] snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() snake_case : int = config_and_inputs[0] snake_case : Union[str, Any] = UMTaForConditionalGeneration(UpperCAmelCase__ ).eval() model.to(UpperCAmelCase__ ) snake_case : str = { '''head_mask''': torch.zeros(config.num_layers , config.num_heads , device=UpperCAmelCase__ ), '''decoder_head_mask''': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ), '''cross_attn_head_mask''': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ), } for attn_name, (name, mask) in zip(UpperCAmelCase__ , head_masking.items() ): snake_case : int = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": snake_case : List[str] = torch.ones( config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ) snake_case : Union[str, Any] = model.generate( config_and_inputs[1]['''input_ids'''] , num_beams=1 , max_length=3 , output_attentions=UpperCAmelCase__ , return_dict_in_generate=UpperCAmelCase__ , **UpperCAmelCase__ , ) # We check the state of decoder_attentions and cross_attentions just from the last step snake_case : List[str] = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('''Does not work on the tiny model as we keep hitting edge cases.''' ) def lowerCAmelCase( self : Any ): """simple docstring""" pass @require_torch @require_sentencepiece @require_tokenizers class a_ ( unittest.TestCase ): @slow @unittest.skip( '''Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged''' ) def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Optional[Any] = UMTaForConditionalGeneration.from_pretrained('''google/umt5-small''' , return_dict=UpperCAmelCase__ ).to(UpperCAmelCase__ ) snake_case : int = AutoTokenizer.from_pretrained('''google/umt5-small''' , use_fast=UpperCAmelCase__ , legacy=UpperCAmelCase__ ) snake_case : List[str] = [ '''Bonjour monsieur <extra_id_0> bien <extra_id_1>.''', '''No se como puedo <extra_id_0>.''', '''This is the reason why we <extra_id_0> them.''', '''The <extra_id_0> walks in <extra_id_1>, seats''', '''A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.''', ] snake_case : Dict = tokenizer(UpperCAmelCase__ , return_tensors='''pt''' , padding=UpperCAmelCase__ ).input_ids # fmt: off snake_case : Optional[Any] = torch.tensor( [ [ 38_530, 210_703, 256_299, 1_410, 256_298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25_922, 256_299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1_460, 339, 312, 19_014, 10_620, 758, 256_299, 2_355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256_299, 14_869, 281, 301, 256_298, 275, 119_983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256_299, 14_869, 281, 2_234, 289, 2_275, 333,61_391, 289, 256_298, 543, 256_297, 168_714, 329, 256_296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : List[Any] = model.generate(input_ids.to(UpperCAmelCase__ ) ) snake_case : int = [ '''<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>''', '''<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', ] snake_case : Tuple = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ )
84
1
from dataclasses import dataclass from typing import Optional, Tuple import torch from torch import nn from transformers import RobertaPreTrainedModel, XLMRobertaConfig, XLMRobertaModel from transformers.utils import ModelOutput @dataclass class a_ ( a ): A__ : Optional[torch.FloatTensor] = None A__ : torch.FloatTensor = None A__ : Optional[Tuple[torch.FloatTensor]] = None A__ : Optional[Tuple[torch.FloatTensor]] = None class a_ ( a ): def __init__( self : Tuple , UpperCAmelCase__ : Tuple=1 , UpperCAmelCase__ : str=0 , UpperCAmelCase__ : Tuple=2 , UpperCAmelCase__ : Optional[Any]=512 , UpperCAmelCase__ : Optional[int]="cls" , UpperCAmelCase__ : int=False , UpperCAmelCase__ : Optional[int]=True , **UpperCAmelCase__ : Any , ): """simple docstring""" super().__init__(pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , **UpperCAmelCase__ ) snake_case : Any = project_dim snake_case : Tuple = pooler_fn snake_case : Tuple = learn_encoder snake_case : List[Any] = use_attention_mask class a_ ( a ): A__ : Dict = [r'pooler', r'logit_scale'] A__ : Any = [r'position_ids', r'predictions.decoder.bias'] A__ : Optional[Any] = 'roberta' A__ : Optional[Any] = RobertaSeriesConfig def __init__( self : Tuple , UpperCAmelCase__ : Any ): """simple docstring""" super().__init__(UpperCAmelCase__ ) snake_case : int = XLMRobertaModel(UpperCAmelCase__ ) snake_case : List[Any] = nn.Linear(config.hidden_size , config.project_dim ) snake_case : Dict = getattr(UpperCAmelCase__ , '''has_pre_transformation''' , UpperCAmelCase__ ) if self.has_pre_transformation: snake_case : Dict = nn.Linear(config.hidden_size , config.project_dim ) snake_case : Optional[Any] = nn.LayerNorm(config.hidden_size , eps=config.layer_norm_eps ) self.post_init() def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : Optional[torch.Tensor] = None , UpperCAmelCase__ : Optional[torch.Tensor] = None , UpperCAmelCase__ : Optional[torch.Tensor] = None , UpperCAmelCase__ : Optional[torch.Tensor] = None , UpperCAmelCase__ : Optional[torch.Tensor] = None , UpperCAmelCase__ : Optional[torch.Tensor] = None , UpperCAmelCase__ : Optional[torch.Tensor] = None , UpperCAmelCase__ : Optional[torch.Tensor] = None , UpperCAmelCase__ : Optional[bool] = None , UpperCAmelCase__ : Optional[bool] = None , UpperCAmelCase__ : Optional[bool] = None , ): """simple docstring""" snake_case : List[str] = return_dict if return_dict is not None else self.config.use_return_dict snake_case : Optional[int] = self.base_model( input_ids=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , position_ids=UpperCAmelCase__ , head_mask=UpperCAmelCase__ , inputs_embeds=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , encoder_attention_mask=UpperCAmelCase__ , output_attentions=UpperCAmelCase__ , output_hidden_states=True if self.has_pre_transformation else output_hidden_states , return_dict=UpperCAmelCase__ , ) if self.has_pre_transformation: snake_case : Dict = outputs['''hidden_states'''][-2] snake_case : str = self.pre_LN(UpperCAmelCase__ ) snake_case : List[Any] = self.transformation_pre(UpperCAmelCase__ ) return TransformationModelOutput( projection_state=UpperCAmelCase__ , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , ) else: snake_case : List[str] = self.transformation(outputs.last_hidden_state ) return TransformationModelOutput( projection_state=UpperCAmelCase__ , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
84
import torch from diffusers import DiffusionPipeline class a_ ( a ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[Any] ): """simple docstring""" super().__init__() self.register_modules(unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ ) def __call__( self : Optional[int] ): """simple docstring""" snake_case : Any = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) snake_case : Dict = 1 snake_case : Optional[Any] = self.unet(UpperCAmelCase__ , UpperCAmelCase__ ).sample snake_case : List[Any] = self.scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample snake_case : List[Any] = scheduler_output - scheduler_output + torch.ones_like(UpperCAmelCase__ ) return result
84
1
# Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version _a : Optional[int] = get_logger(__name__) class a_ : A__ : Any = 'dummy_data' A__ : List[str] = 'datasets' A__ : Dict = False def __init__( self : str , UpperCAmelCase__ : str , UpperCAmelCase__ : str , UpperCAmelCase__ : Union[Version, str] , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Optional[List[Callable]] = None , ): """simple docstring""" snake_case : Optional[Any] = 0 snake_case : Optional[Any] = dataset_name snake_case : Tuple = cache_dir snake_case : Optional[Any] = use_local_dummy_data snake_case : Optional[Any] = config # download_callbacks take a single url as input snake_case : List[Callable] = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root snake_case : Optional[int] = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general snake_case : Dict = str(UpperCAmelCase__ ) # to be downloaded snake_case : str = None snake_case : str = None @property def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" if self._dummy_file is None: snake_case : Optional[Any] = self.download_dummy_data() return self._dummy_file @property def lowerCAmelCase( self : Tuple ): """simple docstring""" if self.config is not None: # structure is dummy / config_name / version_name return os.path.join('''dummy''' , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join('''dummy''' , self.version_name ) @property def lowerCAmelCase( self : List[Any] ): """simple docstring""" return os.path.join(self.dummy_data_folder , '''dummy_data.zip''' ) def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : str = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) snake_case : Optional[int] = cached_path( UpperCAmelCase__ , cache_dir=self.cache_dir , extract_compressed_file=UpperCAmelCase__ , force_extract=UpperCAmelCase__ ) return os.path.join(UpperCAmelCase__ , self.dummy_file_name ) @property def lowerCAmelCase( self : List[str] ): """simple docstring""" return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def lowerCAmelCase( self : int ): """simple docstring""" if self._bucket_url is None: snake_case : List[Any] = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , '''/''' ) ) return self._bucket_url @property def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , '''/''' ).split('''/''' )[:-1] ) def lowerCAmelCase( self : str , UpperCAmelCase__ : Optional[int] , *UpperCAmelCase__ : Any ): """simple docstring""" if self.load_existing_dummy_data: # dummy data is downloaded and tested snake_case : Union[str, Any] = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned snake_case : List[Any] = self.dummy_file_name # special case when data_url is a dict if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): return self.create_dummy_data_dict(UpperCAmelCase__ , UpperCAmelCase__ ) elif isinstance(UpperCAmelCase__ , (list, tuple) ): return self.create_dummy_data_list(UpperCAmelCase__ , UpperCAmelCase__ ) else: return self.create_dummy_data_single(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : List[Any] , *UpperCAmelCase__ : Optional[int] ): """simple docstring""" return self.download_and_extract(UpperCAmelCase__ ) def lowerCAmelCase( self : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[Any] ): """simple docstring""" return self.download_and_extract(UpperCAmelCase__ ) def lowerCAmelCase( self : int , UpperCAmelCase__ : Optional[Any] , *UpperCAmelCase__ : Dict , **UpperCAmelCase__ : Any ): """simple docstring""" return path def lowerCAmelCase( self : List[str] ): """simple docstring""" return {} def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : int , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Tuple = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): for single_url in single_urls: download_callback(UpperCAmelCase__ ) else: snake_case : int = single_urls download_callback(UpperCAmelCase__ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : List[str] = [os.path.join(UpperCAmelCase__ , urllib.parse.quote_plus(Path(UpperCAmelCase__ ).name ) ) for x in single_urls] else: snake_case : List[str] = single_urls snake_case : Dict = os.path.join(UpperCAmelCase__ , urllib.parse.quote_plus(Path(UpperCAmelCase__ ).name ) ) snake_case : Union[str, Any] = value # make sure that values are unique if all(isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique snake_case : List[str] = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" snake_case : Optional[int] = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one snake_case : Tuple = all(bool(re.findall('''[0-9]{3,}-of-[0-9]{3,}''' , UpperCAmelCase__ ) ) for url in data_url ) snake_case : Any = all( url.startswith('''https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed''' ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): snake_case : str = [data_url[0]] * len(UpperCAmelCase__ ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(UpperCAmelCase__ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus snake_case : int = os.path.join(UpperCAmelCase__ , urllib.parse.quote_plus(single_url.split('''/''' )[-1] ) ) dummy_data_list.append(UpperCAmelCase__ ) return dummy_data_list def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[str] ): """simple docstring""" for download_callback in self.download_callbacks: download_callback(UpperCAmelCase__ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus snake_case : int = os.path.join(UpperCAmelCase__ , urllib.parse.quote_plus(data_url.split('''/''' )[-1] ) ) if os.path.exists(UpperCAmelCase__ ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" pass def lowerCAmelCase( self : str ): """simple docstring""" pass def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" def _iter_archive_members(UpperCAmelCase__ : Optional[Any] ): # this preserves the order of the members inside the ZIP archive snake_case : Dict = Path(self.dummy_file ).parent snake_case : Union[str, Any] = path.relative_to(UpperCAmelCase__ ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: snake_case : Optional[int] = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(UpperCAmelCase__ ) snake_case : Dict = Path(UpperCAmelCase__ ) snake_case : Tuple = _iter_archive_members(UpperCAmelCase__ ) if self.use_local_dummy_data else path.rglob('''*''' ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith(('''.''', '''__''') ): yield file_path.relative_to(UpperCAmelCase__ ).as_posix(), file_path.open('''rb''' ) def lowerCAmelCase( self : Any , UpperCAmelCase__ : Dict ): """simple docstring""" if not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Optional[Any] = [paths] for path in paths: if os.path.isfile(UpperCAmelCase__ ): if os.path.basename(UpperCAmelCase__ ).startswith(('''.''', '''__''') ): return yield path else: for dirpath, dirnames, filenames in os.walk(UpperCAmelCase__ ): if os.path.basename(UpperCAmelCase__ ).startswith(('''.''', '''__''') ): continue dirnames.sort() for filename in sorted(UpperCAmelCase__ ): if filename.startswith(('''.''', '''__''') ): continue yield os.path.join(UpperCAmelCase__ , UpperCAmelCase__ )
84
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class a_ ( a ): A__ : List[str] = ['image_processor', 'tokenizer'] A__ : Any = 'CLIPImageProcessor' A__ : Optional[int] = ('CLIPTokenizer', 'CLIPTokenizerFast') def __init__( self : Union[str, Any] , UpperCAmelCase__ : str=None , UpperCAmelCase__ : Union[str, Any]=None , **UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : Optional[int] = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , UpperCAmelCase__ , ) snake_case : List[Any] = kwargs.pop('''feature_extractor''' ) snake_case : Optional[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(UpperCAmelCase__ , UpperCAmelCase__ ) def __call__( self : Any , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : int=None , **UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: snake_case : int = self.tokenizer(UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , **UpperCAmelCase__ ) if images is not None: snake_case : Dict = self.image_processor(UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , **UpperCAmelCase__ ) if text is not None and images is not None: snake_case : Tuple = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**UpperCAmelCase__ ) , tensor_type=UpperCAmelCase__ ) def lowerCAmelCase( self : List[str] , *UpperCAmelCase__ : Dict , **UpperCAmelCase__ : int ): """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : str ): """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : int = self.tokenizer.model_input_names snake_case : Tuple = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , UpperCAmelCase__ , ) return self.image_processor_class @property def lowerCAmelCase( self : Any ): """simple docstring""" warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , UpperCAmelCase__ , ) return self.image_processor
84
1
import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def a_ ( __magic_name__ = 8 ) -> str: """simple docstring""" snake_case : Tuple = ascii_letters + digits + punctuation return "".join(secrets.choice(__magic_name__ ) for _ in range(__magic_name__ ) ) def a_ ( __magic_name__ , __magic_name__ ) -> str: """simple docstring""" i -= len(__magic_name__ ) snake_case : str = i // 3 snake_case : List[Any] = i % 3 # chars = chars_incl + random_letters(ascii_letters, i / 3 + remainder) + # random_number(digits, i / 3) + random_characters(punctuation, i / 3) snake_case : List[Any] = ( chars_incl + random(__magic_name__ , quotient + remainder ) + random(__magic_name__ , __magic_name__ ) + random(__magic_name__ , __magic_name__ ) ) snake_case : str = list(__magic_name__ ) shuffle(__magic_name__ ) return "".join(__magic_name__ ) # random is a generalised function for letters, characters and numbers def a_ ( __magic_name__ , __magic_name__ ) -> str: """simple docstring""" return "".join(secrets.choice(__magic_name__ ) for _ in range(__magic_name__ ) ) def a_ ( __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" pass # Put your code here... def a_ ( __magic_name__ , __magic_name__ ) -> str: """simple docstring""" pass # Put your code here... def a_ ( __magic_name__ , __magic_name__ ) -> Optional[int]: """simple docstring""" pass # Put your code here... def a_ ( __magic_name__ , __magic_name__ = 8 ) -> bool: """simple docstring""" if len(__magic_name__ ) < min_length: # Your Password must be at least 8 characters long return False snake_case : str = any(char in ascii_uppercase for char in password ) snake_case : Tuple = any(char in ascii_lowercase for char in password ) snake_case : Dict = any(char in digits for char in password ) snake_case : Any = any(char in punctuation for char in password ) return upper and lower and num and spec_char # Passwords should contain UPPERCASE, lowerase # numbers, and special characters def a_ ( ) -> List[str]: """simple docstring""" snake_case : Union[str, Any] = int(input('''Please indicate the max length of your password: ''' ).strip() ) snake_case : Tuple = input( '''Please indicate the characters that must be in your password: ''' ).strip() print('''Password generated:''' , password_generator(__magic_name__ ) ) print( '''Alternative Password generated:''' , alternative_password_generator(__magic_name__ , __magic_name__ ) , ) print('''[If you are thinking of using this passsword, You better save it.]''' ) if __name__ == "__main__": main()
84
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets _a : Optional[Any] = '\\n@inproceedings{pillutla-etal:mauve:neurips2021,\n title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers},\n author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid},\n booktitle = {NeurIPS},\n year = {2021}\n}\n\n' _a : str = '\\nMAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure.\n\nMAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences.\n\nFor details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021).\n\nThis metrics is a wrapper around the official implementation of MAUVE:\nhttps://github.com/krishnap25/mauve\n' _a : List[Any] = '\nCalculates MAUVE scores between two lists of generated text and reference text.\nArgs:\n predictions: list of generated text to score. Each predictions\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\nOptional Args:\n num_buckets: the size of the histogram to quantize P and Q. Options: \'auto\' (default) or an integer\n pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1\n kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9\n kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5\n kmeans_max_iter: maximum number of k-means iterations. Default 500\n featurize_model_name: name of the model from which features are obtained. Default \'gpt2-large\' Use one of [\'gpt2\', \'gpt2-medium\', \'gpt2-large\', \'gpt2-xl\'].\n device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU\n max_text_length: maximum number of tokens to consider. Default 1024\n divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25\n mauve_scaling_factor: "c" from the paper. Default 5.\n verbose: If True (default), print running time updates\n seed: random seed to initialize k-means cluster assignments.\nReturns:\n mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer,\n frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer,\n divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve,\n p_hist: a discrete distribution, which is a quantized version of the text distribution p_text,\n q_hist: same as above, but with q_text.\nExamples:\n\n >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest\n >>> import datasets\n >>> mauve = datasets.load_metric(\'mauve\')\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP\n >>> print(out.mauve) # doctest: +SKIP\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): def lowerCAmelCase( self : Any ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://github.com/krishnap25/mauve''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/krishnap25/mauve'''] , reference_urls=[ '''https://arxiv.org/abs/2102.01454''', '''https://github.com/krishnap25/mauve''', ] , ) def lowerCAmelCase( self : int , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : List[str]="auto" , UpperCAmelCase__ : Tuple=-1 , UpperCAmelCase__ : Optional[int]=0.9 , UpperCAmelCase__ : List[Any]=5 , UpperCAmelCase__ : List[Any]=500 , UpperCAmelCase__ : Union[str, Any]="gpt2-large" , UpperCAmelCase__ : Optional[Any]=-1 , UpperCAmelCase__ : int=1_024 , UpperCAmelCase__ : List[Any]=25 , UpperCAmelCase__ : Union[str, Any]=5 , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : List[Any]=25 , ): """simple docstring""" snake_case : List[str] = compute_mauve( p_text=UpperCAmelCase__ , q_text=UpperCAmelCase__ , p_features=UpperCAmelCase__ , q_features=UpperCAmelCase__ , p_tokens=UpperCAmelCase__ , q_tokens=UpperCAmelCase__ , num_buckets=UpperCAmelCase__ , pca_max_data=UpperCAmelCase__ , kmeans_explained_var=UpperCAmelCase__ , kmeans_num_redo=UpperCAmelCase__ , kmeans_max_iter=UpperCAmelCase__ , featurize_model_name=UpperCAmelCase__ , device_id=UpperCAmelCase__ , max_text_length=UpperCAmelCase__ , divergence_curve_discretization_size=UpperCAmelCase__ , mauve_scaling_factor=UpperCAmelCase__ , verbose=UpperCAmelCase__ , seed=UpperCAmelCase__ , ) return out
84
1
import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class a_ ( a , unittest.TestCase ): A__ : List[str] = KandinskyVaaPriorPipeline A__ : Optional[int] = ['prompt'] A__ : Optional[Any] = ['prompt', 'negative_prompt'] A__ : str = [ 'num_images_per_prompt', 'generator', 'num_inference_steps', 'latents', 'negative_prompt', 'guidance_scale', 'output_type', 'return_dict', ] A__ : Dict = False @property def lowerCAmelCase( self : Dict ): """simple docstring""" return 32 @property def lowerCAmelCase( self : List[Any] ): """simple docstring""" return 32 @property def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" return self.time_input_dim @property def lowerCAmelCase( self : Tuple ): """simple docstring""" return self.time_input_dim * 4 @property def lowerCAmelCase( self : List[Any] ): """simple docstring""" return 100 @property def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : Optional[Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) return tokenizer @property def lowerCAmelCase( self : Any ): """simple docstring""" torch.manual_seed(0 ) snake_case : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) return CLIPTextModelWithProjection(UpperCAmelCase__ ) @property def lowerCAmelCase( self : Optional[int] ): """simple docstring""" torch.manual_seed(0 ) snake_case : Tuple = { '''num_attention_heads''': 2, '''attention_head_dim''': 12, '''embedding_dim''': self.text_embedder_hidden_size, '''num_layers''': 1, } snake_case : Union[str, Any] = PriorTransformer(**UpperCAmelCase__ ) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 snake_case : Optional[Any] = nn.Parameter(torch.ones(model.clip_std.shape ) ) return model @property def lowerCAmelCase( self : Optional[int] ): """simple docstring""" torch.manual_seed(0 ) snake_case : Any = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=224 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=14 , ) snake_case : int = CLIPVisionModelWithProjection(UpperCAmelCase__ ) return model @property def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Dict = CLIPImageProcessor( crop_size=224 , do_center_crop=UpperCAmelCase__ , do_normalize=UpperCAmelCase__ , do_resize=UpperCAmelCase__ , image_mean=[0.4814_5466, 0.457_8275, 0.4082_1073] , image_std=[0.2686_2954, 0.2613_0258, 0.2757_7711] , resample=3 , size=224 , ) return image_processor def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : str = self.dummy_prior snake_case : Tuple = self.dummy_image_encoder snake_case : Tuple = self.dummy_text_encoder snake_case : List[Any] = self.dummy_tokenizer snake_case : Optional[Any] = self.dummy_image_processor snake_case : int = UnCLIPScheduler( variance_type='''fixed_small_log''' , prediction_type='''sample''' , num_train_timesteps=1_000 , clip_sample=UpperCAmelCase__ , clip_sample_range=10.0 , ) snake_case : str = { '''prior''': prior, '''image_encoder''': image_encoder, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''scheduler''': scheduler, '''image_processor''': image_processor, } return components def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Tuple=0 ): """simple docstring""" if str(UpperCAmelCase__ ).startswith('''mps''' ): snake_case : Optional[int] = torch.manual_seed(UpperCAmelCase__ ) else: snake_case : Tuple = torch.Generator(device=UpperCAmelCase__ ).manual_seed(UpperCAmelCase__ ) snake_case : List[Any] = { '''prompt''': '''horse''', '''generator''': generator, '''guidance_scale''': 4.0, '''num_inference_steps''': 2, '''output_type''': '''np''', } return inputs def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : Dict = '''cpu''' snake_case : Union[str, Any] = self.get_dummy_components() snake_case : str = self.pipeline_class(**UpperCAmelCase__ ) snake_case : Tuple = pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) snake_case : str = pipe(**self.get_dummy_inputs(UpperCAmelCase__ ) ) snake_case : Union[str, Any] = output.image_embeds snake_case : Optional[Any] = pipe( **self.get_dummy_inputs(UpperCAmelCase__ ) , return_dict=UpperCAmelCase__ , )[0] snake_case : int = image[0, -10:] snake_case : Tuple = image_from_tuple[0, -10:] assert image.shape == (1, 32) snake_case : Optional[Any] = np.array( [-0.0532, 1.7120, 0.3656, -1.0852, -0.8946, -1.1756, 0.4348, 0.2482, 0.5146, -0.1156] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Optional[Any] = torch_device == '''cpu''' snake_case : Tuple = True snake_case : Union[str, Any] = False self._test_inference_batch_single_identical( test_max_difference=UpperCAmelCase__ , relax_max_difference=UpperCAmelCase__ , test_mean_pixel_difference=UpperCAmelCase__ , ) @skip_mps def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[Any] = torch_device == '''cpu''' snake_case : List[str] = False self._test_attention_slicing_forward_pass( test_max_difference=UpperCAmelCase__ , test_mean_pixel_difference=UpperCAmelCase__ , )
84
import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" if "cls_token" in name: snake_case : Tuple = name.replace('''cls_token''' , '''vit.embeddings.cls_token''' ) if "mask_token" in name: snake_case : Optional[int] = name.replace('''mask_token''' , '''decoder.mask_token''' ) if "decoder_pos_embed" in name: snake_case : List[str] = name.replace('''decoder_pos_embed''' , '''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: snake_case : List[str] = name.replace('''pos_embed''' , '''vit.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: snake_case : List[Any] = name.replace('''patch_embed.proj''' , '''vit.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: snake_case : int = name.replace('''patch_embed.norm''' , '''vit.embeddings.norm''' ) if "decoder_blocks" in name: snake_case : int = name.replace('''decoder_blocks''' , '''decoder.decoder_layers''' ) if "blocks" in name: snake_case : Optional[Any] = name.replace('''blocks''' , '''vit.encoder.layer''' ) if "attn.proj" in name: snake_case : str = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: snake_case : Any = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: snake_case : List[str] = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: snake_case : Tuple = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: snake_case : Tuple = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: snake_case : Tuple = name.replace('''mlp.fc2''' , '''output.dense''' ) if "decoder_embed" in name: snake_case : Dict = name.replace('''decoder_embed''' , '''decoder.decoder_embed''' ) if "decoder_norm" in name: snake_case : Dict = name.replace('''decoder_norm''' , '''decoder.decoder_norm''' ) if "decoder_pred" in name: snake_case : Dict = name.replace('''decoder_pred''' , '''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name: snake_case : Optional[int] = name.replace('''norm.weight''' , '''vit.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name: snake_case : List[str] = name.replace('''norm.bias''' , '''vit.layernorm.bias''' ) return name def a_ ( __magic_name__ , __magic_name__ ) -> str: """simple docstring""" for key in orig_state_dict.copy().keys(): snake_case : Union[str, Any] = orig_state_dict.pop(__magic_name__ ) if "qkv" in key: snake_case : Optional[int] = key.split('''.''' ) snake_case : int = int(key_split[1] ) if "decoder_blocks" in key: snake_case : List[str] = config.decoder_hidden_size snake_case : List[Any] = '''decoder.decoder_layers.''' if "weight" in key: snake_case : str = val[:dim, :] snake_case : Optional[Any] = val[dim : dim * 2, :] snake_case : Any = val[-dim:, :] elif "bias" in key: snake_case : Optional[Any] = val[:dim] snake_case : List[Any] = val[dim : dim * 2] snake_case : List[Any] = val[-dim:] else: snake_case : Optional[int] = config.hidden_size snake_case : Tuple = '''vit.encoder.layer.''' if "weight" in key: snake_case : Optional[Any] = val[:dim, :] snake_case : str = val[dim : dim * 2, :] snake_case : Union[str, Any] = val[-dim:, :] elif "bias" in key: snake_case : Tuple = val[:dim] snake_case : int = val[dim : dim * 2] snake_case : Optional[Any] = val[-dim:] else: snake_case : Optional[Any] = val return orig_state_dict def a_ ( __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" snake_case : List[str] = ViTMAEConfig() if "large" in checkpoint_url: snake_case : str = 1_024 snake_case : Tuple = 4_096 snake_case : Optional[Any] = 24 snake_case : List[Any] = 16 elif "huge" in checkpoint_url: snake_case : Tuple = 14 snake_case : int = 1_280 snake_case : Dict = 5_120 snake_case : Tuple = 32 snake_case : Optional[Any] = 16 snake_case : Optional[Any] = ViTMAEForPreTraining(__magic_name__ ) snake_case : Optional[Any] = torch.hub.load_state_dict_from_url(__magic_name__ , map_location='''cpu''' )['''model'''] snake_case : int = ViTMAEImageProcessor(size=config.image_size ) snake_case : Dict = convert_state_dict(__magic_name__ , __magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() snake_case : Tuple = '''https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg''' snake_case : List[Any] = Image.open(requests.get(__magic_name__ , stream=__magic_name__ ).raw ) snake_case : Dict = ViTMAEImageProcessor(size=config.image_size ) snake_case : str = image_processor(images=__magic_name__ , return_tensors='''pt''' ) # forward pass torch.manual_seed(2 ) snake_case : Union[str, Any] = model(**__magic_name__ ) snake_case : Optional[Any] = outputs.logits if "large" in checkpoint_url: snake_case : Any = torch.tensor( [[-0.7309, -0.7128, -1.0169], [-1.0161, -0.9058, -1.1878], [-1.0478, -0.9411, -1.1911]] ) elif "huge" in checkpoint_url: snake_case : List[Any] = torch.tensor( [[-1.1599, -0.9199, -1.2221], [-1.1952, -0.9269, -1.2307], [-1.2143, -0.9337, -1.2262]] ) else: snake_case : Dict = torch.tensor( [[-0.9192, -0.8481, -1.1259], [-1.1349, -1.0034, -1.2599], [-1.1757, -1.0429, -1.2726]] ) # verify logits assert torch.allclose(logits[0, :3, :3] , __magic_name__ , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(__magic_name__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__magic_name__ ) if __name__ == "__main__": _a : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth', type=str, help='URL of the checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _a : str = parser.parse_args() convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
84
1
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def a_ ( __magic_name__ , __magic_name__ ) -> np.array: """simple docstring""" snake_case : Dict = F"{sampling_rate}" snake_case : Optional[Any] = '''1''' snake_case : Union[str, Any] = '''f32le''' snake_case : Any = [ '''ffmpeg''', '''-i''', '''pipe:0''', '''-ac''', ac, '''-ar''', ar, '''-f''', format_for_conversion, '''-hide_banner''', '''-loglevel''', '''quiet''', '''pipe:1''', ] try: with subprocess.Popen(__magic_name__ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: snake_case : Optional[int] = ffmpeg_process.communicate(__magic_name__ ) except FileNotFoundError as error: raise ValueError('''ffmpeg was not found but is required to load audio files from filename''' ) from error snake_case : str = output_stream[0] snake_case : str = np.frombuffer(__magic_name__ , np.floataa ) if audio.shape[0] == 0: raise ValueError('''Malformed soundfile''' ) return audio def a_ ( __magic_name__ , __magic_name__ , __magic_name__ = "f32le" , ) -> Union[str, Any]: """simple docstring""" snake_case : List[Any] = F"{sampling_rate}" snake_case : List[str] = '''1''' if format_for_conversion == "s16le": snake_case : Dict = 2 elif format_for_conversion == "f32le": snake_case : int = 4 else: raise ValueError(F"Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`" ) snake_case : Optional[Any] = platform.system() if system == "Linux": snake_case : Union[str, Any] = '''alsa''' snake_case : Tuple = '''default''' elif system == "Darwin": snake_case : List[Any] = '''avfoundation''' snake_case : Optional[int] = ''':0''' elif system == "Windows": snake_case : str = '''dshow''' snake_case : str = '''default''' snake_case : Any = [ '''ffmpeg''', '''-f''', format_, '''-i''', input_, '''-ac''', ac, '''-ar''', ar, '''-f''', format_for_conversion, '''-fflags''', '''nobuffer''', '''-hide_banner''', '''-loglevel''', '''quiet''', '''pipe:1''', ] snake_case : Any = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample snake_case : Any = _ffmpeg_stream(__magic_name__ , __magic_name__ ) for item in iterator: yield item def a_ ( __magic_name__ , __magic_name__ , __magic_name__ = None , __magic_name__ = None , __magic_name__ = "f32le" , ) -> Union[str, Any]: """simple docstring""" if stream_chunk_s is not None: snake_case : Dict = stream_chunk_s else: snake_case : Dict = chunk_length_s snake_case : int = ffmpeg_microphone(__magic_name__ , __magic_name__ , format_for_conversion=__magic_name__ ) if format_for_conversion == "s16le": snake_case : Dict = np.intaa snake_case : Optional[Any] = 2 elif format_for_conversion == "f32le": snake_case : Optional[Any] = np.floataa snake_case : Dict = 4 else: raise ValueError(F"Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`" ) if stride_length_s is None: snake_case : List[Any] = chunk_length_s / 6 snake_case : str = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(__magic_name__ , (int, float) ): snake_case : Optional[int] = [stride_length_s, stride_length_s] snake_case : Tuple = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample snake_case : str = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample snake_case : Union[str, Any] = datetime.datetime.now() snake_case : int = datetime.timedelta(seconds=__magic_name__ ) for item in chunk_bytes_iter(__magic_name__ , __magic_name__ , stride=(stride_left, stride_right) , stream=__magic_name__ ): # Put everything back in numpy scale snake_case : Dict = np.frombuffer(item['''raw'''] , dtype=__magic_name__ ) snake_case : List[str] = ( item['''stride'''][0] // size_of_sample, item['''stride'''][1] // size_of_sample, ) snake_case : Optional[int] = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = False ) -> Optional[Any]: """simple docstring""" snake_case : str = b'''''' snake_case , snake_case : List[Any] = stride if stride_left + stride_right >= chunk_len: raise ValueError( F"Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}" ) snake_case : List[str] = 0 for raw in iterator: acc += raw if stream and len(__magic_name__ ) < chunk_len: snake_case : Optional[Any] = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(__magic_name__ ) >= chunk_len: # We are flushing the accumulator snake_case : Dict = (_stride_left, stride_right) snake_case : Optional[Any] = {'''raw''': acc[:chunk_len], '''stride''': stride} if stream: snake_case : Tuple = False yield item snake_case : Union[str, Any] = stride_left snake_case : Tuple = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(__magic_name__ ) > stride_left: snake_case : Tuple = {'''raw''': acc, '''stride''': (_stride_left, 0)} if stream: snake_case : Dict = False yield item def a_ ( __magic_name__ , __magic_name__ ) -> List[str]: """simple docstring""" snake_case : Tuple = 2**24 # 16Mo try: with subprocess.Popen(__magic_name__ , stdout=subprocess.PIPE , bufsize=__magic_name__ ) as ffmpeg_process: while True: snake_case : Optional[int] = ffmpeg_process.stdout.read(__magic_name__ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('''ffmpeg was not found but is required to stream audio files from filename''' ) from error
84
import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _a : Optional[Any] = 16 _a : Union[str, Any] = 32 def a_ ( __magic_name__ , __magic_name__ = 16 ) -> Dict: """simple docstring""" snake_case : Tuple = AutoTokenizer.from_pretrained('''bert-base-cased''' ) snake_case : Any = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(__magic_name__ ): # max_length=None => use the model max length (it's actually the default) snake_case : Union[str, Any] = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=__magic_name__ , max_length=__magic_name__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): snake_case : Union[str, Any] = datasets.map( __magic_name__ , batched=__magic_name__ , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library snake_case : Optional[Any] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(__magic_name__ ): # On TPU it's best to pad everything to the same length or training will be very slow. snake_case : str = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": snake_case : Tuple = 16 elif accelerator.mixed_precision != "no": snake_case : Dict = 8 else: snake_case : Union[str, Any] = None return tokenizer.pad( __magic_name__ , padding='''longest''' , max_length=__magic_name__ , pad_to_multiple_of=__magic_name__ , return_tensors='''pt''' , ) # Instantiate dataloaders. snake_case : str = DataLoader( tokenized_datasets['''train'''] , shuffle=__magic_name__ , collate_fn=__magic_name__ , batch_size=__magic_name__ ) snake_case : List[str] = DataLoader( tokenized_datasets['''validation'''] , shuffle=__magic_name__ , collate_fn=__magic_name__ , batch_size=__magic_name__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders _a : Optional[int] = mocked_dataloaders # noqa: F811 def a_ ( __magic_name__ , __magic_name__ ) -> Optional[Any]: """simple docstring""" if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , __magic_name__ ) == "1": snake_case : Optional[int] = 2 # Initialize accelerator snake_case : int = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs snake_case : Dict = config['''lr'''] snake_case : Any = int(config['''num_epochs'''] ) snake_case : List[str] = int(config['''seed'''] ) snake_case : List[Any] = int(config['''batch_size'''] ) snake_case : Tuple = evaluate.load('''glue''' , '''mrpc''' ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=__magic_name__ ) def inner_training_loop(__magic_name__ ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(__magic_name__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) snake_case : str = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=__magic_name__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). snake_case : Optional[int] = model.to(accelerator.device ) # Instantiate optimizer snake_case : Optional[int] = AdamW(params=model.parameters() , lr=__magic_name__ ) snake_case , snake_case : List[Any] = get_dataloaders(__magic_name__ , __magic_name__ ) # Instantiate scheduler snake_case : int = get_linear_schedule_with_warmup( optimizer=__magic_name__ , num_warmup_steps=100 , num_training_steps=(len(__magic_name__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. snake_case , snake_case , snake_case , snake_case , snake_case : Tuple = accelerator.prepare( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # Now we train the model for epoch in range(__magic_name__ ): model.train() for step, batch in enumerate(__magic_name__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) snake_case : int = model(**__magic_name__ ) snake_case : Optional[int] = outputs.loss accelerator.backward(__magic_name__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__magic_name__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): snake_case : List[str] = model(**__magic_name__ ) snake_case : List[Any] = outputs.logits.argmax(dim=-1 ) snake_case , snake_case : Dict = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=__magic_name__ , references=__magic_name__ , ) snake_case : Tuple = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"epoch {epoch}:" , __magic_name__ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def a_ ( ) -> Union[str, Any]: """simple docstring""" snake_case : int = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=__magic_name__ , default=__magic_name__ , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) snake_case : Optional[Any] = parser.parse_args() snake_case : Optional[int] = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(__magic_name__ , __magic_name__ ) if __name__ == "__main__": main()
84
1
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _a : str = logging.get_logger(__name__) _a : Optional[int] = {'tokenizer_file': 'tokenizer.json'} _a : List[Any] = { 'tokenizer_file': { 'bigscience/tokenizer': 'https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json', 'bigscience/bloom-560m': 'https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json', 'bigscience/bloom-1b1': 'https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json', 'bigscience/bloom-1b7': 'https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json', 'bigscience/bloom-3b': 'https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json', 'bigscience/bloom-7b1': 'https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json', 'bigscience/bloom': 'https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json', }, } class a_ ( a ): A__ : Optional[int] = VOCAB_FILES_NAMES A__ : str = PRETRAINED_VOCAB_FILES_MAP A__ : Union[str, Any] = ['input_ids', 'attention_mask'] A__ : str = None def __init__( self : List[str] , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : Optional[Any]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : List[str]="<unk>" , UpperCAmelCase__ : List[Any]="<s>" , UpperCAmelCase__ : Any="</s>" , UpperCAmelCase__ : int="<pad>" , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : int=False , **UpperCAmelCase__ : Tuple , ): """simple docstring""" super().__init__( UpperCAmelCase__ , UpperCAmelCase__ , tokenizer_file=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , add_prefix_space=UpperCAmelCase__ , clean_up_tokenization_spaces=UpperCAmelCase__ , **UpperCAmelCase__ , ) snake_case : Dict = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , UpperCAmelCase__ ) != add_prefix_space: snake_case : Tuple = getattr(UpperCAmelCase__ , pre_tok_state.pop('''type''' ) ) snake_case : Optional[int] = add_prefix_space snake_case : List[str] = pre_tok_class(**UpperCAmelCase__ ) snake_case : Any = add_prefix_space def lowerCAmelCase( self : Tuple , *UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : Optional[Any] ): """simple docstring""" snake_case : str = kwargs.get('''is_split_into_words''' , UpperCAmelCase__ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with" ''' pretokenized inputs.''' ) return super()._batch_encode_plus(*UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : str , *UpperCAmelCase__ : List[str] , **UpperCAmelCase__ : List[Any] ): """simple docstring""" snake_case : List[str] = kwargs.get('''is_split_into_words''' , UpperCAmelCase__ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with" ''' pretokenized inputs.''' ) return super()._encode_plus(*UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ): """simple docstring""" snake_case : Union[str, Any] = self._tokenizer.model.save(UpperCAmelCase__ , name=UpperCAmelCase__ ) return tuple(UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : "Conversation" ): """simple docstring""" snake_case : Dict = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) + [self.eos_token_id] ) if len(UpperCAmelCase__ ) > self.model_max_length: snake_case : int = input_ids[-self.model_max_length :] return input_ids
84
from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract _a : Dict = logging.get_logger(__name__) def a_ ( __magic_name__ , __magic_name__ , __magic_name__ ) -> List[Any]: """simple docstring""" return [ int(1_000 * (box[0] / width) ), int(1_000 * (box[1] / height) ), int(1_000 * (box[2] / width) ), int(1_000 * (box[3] / height) ), ] def a_ ( __magic_name__ , __magic_name__ , __magic_name__ = None ) -> str: """simple docstring""" snake_case : Any = tesseract_config if tesseract_config is not None else '''''' # apply OCR snake_case : str = to_pil_image(__magic_name__ ) snake_case , snake_case : Union[str, Any] = pil_image.size snake_case : List[Any] = pytesseract.image_to_data(__magic_name__ , lang=__magic_name__ , output_type='''dict''' , config=__magic_name__ ) snake_case , snake_case , snake_case , snake_case , snake_case : Optional[Any] = data['''text'''], data['''left'''], data['''top'''], data['''width'''], data['''height'''] # filter empty words and corresponding coordinates snake_case : Union[str, Any] = [idx for idx, word in enumerate(__magic_name__ ) if not word.strip()] snake_case : Union[str, Any] = [word for idx, word in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : Optional[Any] = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format snake_case : List[Any] = [] for x, y, w, h in zip(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): snake_case : Optional[int] = [x, y, x + w, y + h] actual_boxes.append(__magic_name__ ) # finally, normalize the bounding boxes snake_case : List[Any] = [] for box in actual_boxes: normalized_boxes.append(normalize_box(__magic_name__ , __magic_name__ , __magic_name__ ) ) assert len(__magic_name__ ) == len(__magic_name__ ), "Not as many words as there are bounding boxes" return words, normalized_boxes class a_ ( a ): A__ : int = ['pixel_values'] def __init__( self : Optional[int] , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : PILImageResampling = PILImageResampling.BILINEAR , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[str] = "" , **UpperCAmelCase__ : int , ): """simple docstring""" super().__init__(**UpperCAmelCase__ ) snake_case : Any = size if size is not None else {'''height''': 224, '''width''': 224} snake_case : Tuple = get_size_dict(UpperCAmelCase__ ) snake_case : Dict = do_resize snake_case : str = size snake_case : Optional[int] = resample snake_case : Union[str, Any] = apply_ocr snake_case : int = ocr_lang snake_case : Union[str, Any] = tesseract_config def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : np.ndarray , UpperCAmelCase__ : Dict[str, int] , UpperCAmelCase__ : PILImageResampling = PILImageResampling.BILINEAR , UpperCAmelCase__ : Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase__ : Any , ): """simple docstring""" snake_case : Dict = get_size_dict(UpperCAmelCase__ ) if "height" not in size or "width" not in size: raise ValueError(F"The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}" ) snake_case : Tuple = (size['''height'''], size['''width''']) return resize(UpperCAmelCase__ , size=UpperCAmelCase__ , resample=UpperCAmelCase__ , data_format=UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : ImageInput , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : PILImageResampling = None , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[Union[str, TensorType]] = None , UpperCAmelCase__ : ChannelDimension = ChannelDimension.FIRST , **UpperCAmelCase__ : List[Any] , ): """simple docstring""" snake_case : Tuple = do_resize if do_resize is not None else self.do_resize snake_case : List[Any] = size if size is not None else self.size snake_case : Tuple = get_size_dict(UpperCAmelCase__ ) snake_case : str = resample if resample is not None else self.resample snake_case : Optional[int] = apply_ocr if apply_ocr is not None else self.apply_ocr snake_case : Any = ocr_lang if ocr_lang is not None else self.ocr_lang snake_case : Optional[int] = tesseract_config if tesseract_config is not None else self.tesseract_config snake_case : List[str] = make_list_of_images(UpperCAmelCase__ ) if not valid_images(UpperCAmelCase__ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) # All transformations expect numpy arrays. snake_case : Any = [to_numpy_array(UpperCAmelCase__ ) for image in images] if apply_ocr: requires_backends(self , '''pytesseract''' ) snake_case : Optional[int] = [] snake_case : Union[str, Any] = [] for image in images: snake_case , snake_case : List[Any] = apply_tesseract(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) words_batch.append(UpperCAmelCase__ ) boxes_batch.append(UpperCAmelCase__ ) if do_resize: snake_case : Any = [self.resize(image=UpperCAmelCase__ , size=UpperCAmelCase__ , resample=UpperCAmelCase__ ) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) snake_case : int = [flip_channel_order(UpperCAmelCase__ ) for image in images] snake_case : Optional[Any] = [to_channel_dimension_format(UpperCAmelCase__ , UpperCAmelCase__ ) for image in images] snake_case : List[Any] = BatchFeature(data={'''pixel_values''': images} , tensor_type=UpperCAmelCase__ ) if apply_ocr: snake_case : Dict = words_batch snake_case : Dict = boxes_batch return data
84
1
from math import factorial def a_ ( __magic_name__ = 100 ) -> int: """simple docstring""" return sum(map(__magic_name__ , str(factorial(__magic_name__ ) ) ) ) if __name__ == "__main__": print(solution(int(input('Enter the Number: ').strip())))
84
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class a_ : def __init__( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Union[str, Any]=13 , UpperCAmelCase__ : Union[str, Any]=2 , UpperCAmelCase__ : List[Any]=24 , UpperCAmelCase__ : Union[str, Any]=16 , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : str=True , UpperCAmelCase__ : int=32 , UpperCAmelCase__ : Tuple=5 , UpperCAmelCase__ : Dict=4 , UpperCAmelCase__ : Optional[int]=37 , UpperCAmelCase__ : Optional[int]="gelu" , UpperCAmelCase__ : Dict=0.1 , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : Optional[int]=10 , UpperCAmelCase__ : Any=0.02 , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : List[Any]=2 , UpperCAmelCase__ : Optional[Any]=2 , ): """simple docstring""" snake_case : Tuple = parent snake_case : Dict = batch_size snake_case : str = patch_size snake_case : Union[str, Any] = max_length snake_case : str = num_mel_bins snake_case : Any = is_training snake_case : Union[str, Any] = use_labels snake_case : Tuple = hidden_size snake_case : Dict = num_hidden_layers snake_case : Any = num_attention_heads snake_case : Any = intermediate_size snake_case : List[Any] = hidden_act snake_case : str = hidden_dropout_prob snake_case : str = attention_probs_dropout_prob snake_case : str = type_sequence_label_size snake_case : Optional[int] = initializer_range snake_case : str = scope snake_case : int = frequency_stride snake_case : Union[str, Any] = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) snake_case : Optional[int] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 snake_case : Any = (self.max_length - self.patch_size) // self.time_stride + 1 snake_case : Union[str, Any] = frequency_out_dimension * time_out_dimension snake_case : Union[str, Any] = num_patches + 2 def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Optional[int] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) snake_case : str = None if self.use_labels: snake_case : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case : List[str] = self.get_config() return config, input_values, labels def lowerCAmelCase( self : Any ): """simple docstring""" return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCAmelCase__ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : str = ASTModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : Any = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : int = self.prepare_config_and_inputs() ( ( snake_case ) , ( snake_case ) , ( snake_case ) , ) : int = config_and_inputs snake_case : Tuple = {'''input_values''': input_values} return config, inputs_dict @require_torch class a_ ( a , a , unittest.TestCase ): A__ : List[Any] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) A__ : int = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) A__ : Optional[Any] = False A__ : Dict = False A__ : int = False A__ : Optional[int] = False def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : int ): """simple docstring""" if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[int] = ASTModelTester(self ) snake_case : Optional[int] = ConfigTester(self , config_class=UpperCAmelCase__ , has_text_modality=UpperCAmelCase__ , hidden_size=37 ) def lowerCAmelCase( self : List[str] ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def lowerCAmelCase( self : Tuple ): """simple docstring""" pass def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Optional[Any] = model_class(UpperCAmelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase__ , nn.Linear ) ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Any = model_class(UpperCAmelCase__ ) snake_case : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case : str = [*signature.parameters.keys()] snake_case : List[str] = ['''input_values'''] self.assertListEqual(arg_names[:1] , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case : List[str] = ASTModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) def a_ ( ) -> Dict: """simple docstring""" snake_case : Dict = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) snake_case , snake_case : int = torchaudio.load(__magic_name__ ) return audio, sampling_rate @require_torch @require_torchaudio class a_ ( unittest.TestCase ): @cached_property def lowerCAmelCase( self : Any ): """simple docstring""" return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : List[str] = self.default_feature_extractor snake_case : str = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(UpperCAmelCase__ ) snake_case : str = self.default_feature_extractor snake_case , snake_case : int = prepare_audio() snake_case : Optional[int] = audio.squeeze().numpy() snake_case : Optional[Any] = feature_extractor(UpperCAmelCase__ , sampling_rate=UpperCAmelCase__ , return_tensors='''pt''' ).to(UpperCAmelCase__ ) # forward pass with torch.no_grad(): snake_case : Union[str, Any] = model(**UpperCAmelCase__ ) # verify the logits snake_case : Any = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase__ ) snake_case : str = torch.tensor([-0.8760, -7.0042, -8.6602] ).to(UpperCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase__ , atol=1e-4 ) )
84
1
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class a_ : def __init__( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Union[str, Any]=13 , UpperCAmelCase__ : Union[str, Any]=2 , UpperCAmelCase__ : List[Any]=24 , UpperCAmelCase__ : Union[str, Any]=16 , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : str=True , UpperCAmelCase__ : int=32 , UpperCAmelCase__ : Tuple=5 , UpperCAmelCase__ : Dict=4 , UpperCAmelCase__ : Optional[int]=37 , UpperCAmelCase__ : Optional[int]="gelu" , UpperCAmelCase__ : Dict=0.1 , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : Optional[int]=10 , UpperCAmelCase__ : Any=0.02 , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : List[Any]=2 , UpperCAmelCase__ : Optional[Any]=2 , ): """simple docstring""" snake_case : Tuple = parent snake_case : Dict = batch_size snake_case : str = patch_size snake_case : Union[str, Any] = max_length snake_case : str = num_mel_bins snake_case : Any = is_training snake_case : Union[str, Any] = use_labels snake_case : Tuple = hidden_size snake_case : Dict = num_hidden_layers snake_case : Any = num_attention_heads snake_case : Any = intermediate_size snake_case : List[Any] = hidden_act snake_case : str = hidden_dropout_prob snake_case : str = attention_probs_dropout_prob snake_case : str = type_sequence_label_size snake_case : Optional[int] = initializer_range snake_case : str = scope snake_case : int = frequency_stride snake_case : Union[str, Any] = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) snake_case : Optional[int] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 snake_case : Any = (self.max_length - self.patch_size) // self.time_stride + 1 snake_case : Union[str, Any] = frequency_out_dimension * time_out_dimension snake_case : Union[str, Any] = num_patches + 2 def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Optional[int] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) snake_case : str = None if self.use_labels: snake_case : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case : List[str] = self.get_config() return config, input_values, labels def lowerCAmelCase( self : Any ): """simple docstring""" return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCAmelCase__ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : str = ASTModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : Any = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : int = self.prepare_config_and_inputs() ( ( snake_case ) , ( snake_case ) , ( snake_case ) , ) : int = config_and_inputs snake_case : Tuple = {'''input_values''': input_values} return config, inputs_dict @require_torch class a_ ( a , a , unittest.TestCase ): A__ : List[Any] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) A__ : int = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) A__ : Optional[Any] = False A__ : Dict = False A__ : int = False A__ : Optional[int] = False def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : int ): """simple docstring""" if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[int] = ASTModelTester(self ) snake_case : Optional[int] = ConfigTester(self , config_class=UpperCAmelCase__ , has_text_modality=UpperCAmelCase__ , hidden_size=37 ) def lowerCAmelCase( self : List[str] ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def lowerCAmelCase( self : Tuple ): """simple docstring""" pass def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Optional[Any] = model_class(UpperCAmelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase__ , nn.Linear ) ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Any = model_class(UpperCAmelCase__ ) snake_case : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case : str = [*signature.parameters.keys()] snake_case : List[str] = ['''input_values'''] self.assertListEqual(arg_names[:1] , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case : List[str] = ASTModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) def a_ ( ) -> Dict: """simple docstring""" snake_case : Dict = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) snake_case , snake_case : int = torchaudio.load(__magic_name__ ) return audio, sampling_rate @require_torch @require_torchaudio class a_ ( unittest.TestCase ): @cached_property def lowerCAmelCase( self : Any ): """simple docstring""" return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : List[str] = self.default_feature_extractor snake_case : str = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(UpperCAmelCase__ ) snake_case : str = self.default_feature_extractor snake_case , snake_case : int = prepare_audio() snake_case : Optional[int] = audio.squeeze().numpy() snake_case : Optional[Any] = feature_extractor(UpperCAmelCase__ , sampling_rate=UpperCAmelCase__ , return_tensors='''pt''' ).to(UpperCAmelCase__ ) # forward pass with torch.no_grad(): snake_case : Union[str, Any] = model(**UpperCAmelCase__ ) # verify the logits snake_case : Any = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase__ ) snake_case : str = torch.tensor([-0.8760, -7.0042, -8.6602] ).to(UpperCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase__ , atol=1e-4 ) )
84
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _a : Union[str, Any] = logging.getLogger(__name__) def a_ ( __magic_name__ , __magic_name__ ) -> Optional[int]: """simple docstring""" return (preds == labels).mean() @dataclass class a_ : A__ : str = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) @dataclass class a_ : A__ : str = field(metadata={'help': 'The name of the task to train on: ' + ', '.join(processors.keys() )} ) A__ : str = field(metadata={'help': 'Should contain the data files for the task.'} ) A__ : int = field( default=128 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) A__ : bool = field( default=a , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) def a_ ( ) -> Dict: """simple docstring""" snake_case : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) snake_case , snake_case , snake_case : Tuple = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"Output directory ({training_args.output_dir}) already exists and is not empty. Use" ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , __magic_name__ ) # Set seed set_seed(training_args.seed ) try: snake_case : int = processors[data_args.task_name]() snake_case : List[str] = processor.get_labels() snake_case : str = len(__magic_name__ ) except KeyError: raise ValueError('''Task not found: %s''' % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=__magic_name__ , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) snake_case : str = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) snake_case : Any = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=__magic_name__ , cache_dir=model_args.cache_dir , ) # Get datasets snake_case : Optional[int] = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=__magic_name__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) snake_case : Any = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=__magic_name__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(__magic_name__ ) -> Dict: snake_case : str = np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(__magic_name__ , p.label_ids )} # Data collator snake_case : Dict = DataCollatorWithPadding(__magic_name__ , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer snake_case : List[Any] = Trainer( model=__magic_name__ , args=__magic_name__ , train_dataset=__magic_name__ , eval_dataset=__magic_name__ , compute_metrics=__magic_name__ , data_collator=__magic_name__ , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation snake_case : Optional[int] = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) snake_case : Optional[Any] = trainer.evaluate() snake_case : Union[str, Any] = os.path.join(training_args.output_dir , '''eval_results.txt''' ) if trainer.is_world_master(): with open(__magic_name__ , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(''' %s = %s''' , __magic_name__ , __magic_name__ ) writer.write('''%s = %s\n''' % (key, value) ) results.update(__magic_name__ ) return results def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" main() if __name__ == "__main__": main()
84
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _a : str = logging.get_logger(__name__) _a : str = { 'facebook/vit-mae-base': 'https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json', # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class a_ ( a ): A__ : Dict = 'vit_mae' def __init__( self : List[Any] , UpperCAmelCase__ : Any=768 , UpperCAmelCase__ : Optional[int]=12 , UpperCAmelCase__ : Optional[int]=12 , UpperCAmelCase__ : Any=3_072 , UpperCAmelCase__ : Tuple="gelu" , UpperCAmelCase__ : List[Any]=0.0 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Optional[Any]=0.02 , UpperCAmelCase__ : Any=1e-1_2 , UpperCAmelCase__ : Dict=224 , UpperCAmelCase__ : Optional[int]=16 , UpperCAmelCase__ : Union[str, Any]=3 , UpperCAmelCase__ : int=True , UpperCAmelCase__ : List[str]=16 , UpperCAmelCase__ : Dict=512 , UpperCAmelCase__ : Optional[int]=8 , UpperCAmelCase__ : Dict=2_048 , UpperCAmelCase__ : Optional[int]=0.75 , UpperCAmelCase__ : Optional[Any]=False , **UpperCAmelCase__ : Optional[int] , ): """simple docstring""" super().__init__(**UpperCAmelCase__ ) snake_case : Any = hidden_size snake_case : Union[str, Any] = num_hidden_layers snake_case : Dict = num_attention_heads snake_case : Dict = intermediate_size snake_case : Union[str, Any] = hidden_act snake_case : Optional[int] = hidden_dropout_prob snake_case : str = attention_probs_dropout_prob snake_case : Dict = initializer_range snake_case : List[str] = layer_norm_eps snake_case : List[str] = image_size snake_case : List[str] = patch_size snake_case : Tuple = num_channels snake_case : List[Any] = qkv_bias snake_case : Optional[Any] = decoder_num_attention_heads snake_case : Union[str, Any] = decoder_hidden_size snake_case : str = decoder_num_hidden_layers snake_case : Optional[Any] = decoder_intermediate_size snake_case : Optional[int] = mask_ratio snake_case : Optional[int] = norm_pix_loss
84
import re def a_ ( __magic_name__ ) -> bool: """simple docstring""" snake_case : List[str] = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(__magic_name__ , __magic_name__ ) ) if __name__ == "__main__": _a : Any = '0094702343221' print(is_sri_lankan_phone_number(phone))
84
1
import os from typing import List, Optional, Union from ...tokenization_utils import PreTrainedTokenizer from ...tokenization_utils_base import AddedToken from ...utils import logging _a : str = logging.get_logger(__name__) _a : Union[str, Any] = {'vocab_file': 'vocab.txt'} _a : int = { 'vocab_file': { 'facebook/esm2_t6_8M_UR50D': 'https://huggingface.co/facebook/esm2_t6_8M_UR50D/resolve/main/vocab.txt', 'facebook/esm2_t12_35M_UR50D': 'https://huggingface.co/facebook/esm2_t12_35M_UR50D/resolve/main/vocab.txt', }, } _a : List[str] = { 'facebook/esm2_t6_8M_UR50D': 1_024, 'facebook/esm2_t12_35M_UR50D': 1_024, } def a_ ( __magic_name__ ) -> Any: """simple docstring""" with open(__magic_name__ , '''r''' ) as f: snake_case : Optional[int] = f.read().splitlines() return [l.strip() for l in lines] class a_ ( a ): A__ : Any = VOCAB_FILES_NAMES A__ : str = PRETRAINED_VOCAB_FILES_MAP A__ : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Union[str, Any] = ['input_ids', 'attention_mask'] def __init__( self : Optional[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Any="<unk>" , UpperCAmelCase__ : Dict="<cls>" , UpperCAmelCase__ : Dict="<pad>" , UpperCAmelCase__ : Optional[Any]="<mask>" , UpperCAmelCase__ : int="<eos>" , **UpperCAmelCase__ : Optional[Any] , ): """simple docstring""" super().__init__(**UpperCAmelCase__ ) snake_case : Tuple = load_vocab_file(UpperCAmelCase__ ) snake_case : int = dict(enumerate(self.all_tokens ) ) snake_case : List[str] = {tok: ind for ind, tok in enumerate(self.all_tokens )} snake_case : List[Any] = unk_token snake_case : List[Any] = cls_token snake_case : Optional[Any] = pad_token snake_case : Dict = mask_token snake_case : Dict = eos_token snake_case : Union[str, Any] = self.all_tokens self._create_trie(self.unique_no_split_tokens ) def lowerCAmelCase( self : int , UpperCAmelCase__ : int ): """simple docstring""" return self._id_to_token.get(UpperCAmelCase__ , self.unk_token ) def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : str ): """simple docstring""" return self._token_to_id.get(UpperCAmelCase__ , self._token_to_id.get(self.unk_token ) ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : int , **UpperCAmelCase__ : Tuple ): """simple docstring""" return text.split() def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : Optional[int]=False ): """simple docstring""" return len(self._id_to_token ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" return {token: i for i, token in enumerate(self.all_tokens )} def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : str ): """simple docstring""" return self._token_to_id.get(UpperCAmelCase__ , self._token_to_id.get(self.unk_token ) ) def lowerCAmelCase( self : str , UpperCAmelCase__ : int ): """simple docstring""" return self._id_to_token.get(UpperCAmelCase__ , self.unk_token ) def lowerCAmelCase( self : Any , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ): """simple docstring""" snake_case : str = [self.cls_token_id] snake_case : Tuple = [self.eos_token_id] # No sep token in ESM vocabulary if token_ids_a is None: if self.eos_token_id is None: return cls + token_ids_a else: return cls + token_ids_a + sep elif self.eos_token_id is None: raise ValueError('''Cannot tokenize multiple sequences when EOS token is not set!''' ) return cls + token_ids_a + sep + token_ids_a + sep # Multiple inputs always have an EOS token def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : List , UpperCAmelCase__ : Optional[List] = None , UpperCAmelCase__ : bool = False ): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if token in self.all_special_ids else 0 for token in token_ids_a] snake_case : Optional[Any] = [1] + ([0] * len(UpperCAmelCase__ )) + [1] if token_ids_a is not None: mask += [0] * len(UpperCAmelCase__ ) + [1] return mask def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Tuple ): """simple docstring""" snake_case : Dict = os.path.join(UpperCAmelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + '''vocab.txt''' ) with open(UpperCAmelCase__ , '''w''' ) as f: f.write('''\n'''.join(self.all_tokens ) ) return (vocab_file,) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" return self.get_vocab_size(with_added_tokens=UpperCAmelCase__ ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : Union[List[str], List[AddedToken]] , UpperCAmelCase__ : bool = False ): """simple docstring""" return super()._add_tokens(UpperCAmelCase__ , special_tokens=UpperCAmelCase__ )
84
import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class a_ ( unittest.TestCase ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[int]=7 , UpperCAmelCase__ : Union[str, Any]=3 , UpperCAmelCase__ : int=18 , UpperCAmelCase__ : Optional[int]=30 , UpperCAmelCase__ : Optional[int]=400 , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : int=True , ): """simple docstring""" snake_case : int = size if size is not None else {'''height''': 18, '''width''': 18} snake_case : Optional[Any] = parent snake_case : Any = batch_size snake_case : Any = num_channels snake_case : Union[str, Any] = image_size snake_case : Dict = min_resolution snake_case : Dict = max_resolution snake_case : int = do_resize snake_case : List[str] = size snake_case : List[Any] = apply_ocr def lowerCAmelCase( self : int ): """simple docstring""" return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class a_ ( a , unittest.TestCase ): A__ : List[Any] = LayoutLMvaImageProcessor if is_pytesseract_available() else None def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : Optional[Any] = LayoutLMvaImageProcessingTester(self ) @property def lowerCAmelCase( self : Dict ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[str] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase__ , '''do_resize''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''size''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''apply_ocr''' ) ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 18, '''width''': 18} ) snake_case : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) def lowerCAmelCase( self : str ): """simple docstring""" pass def lowerCAmelCase( self : Dict ): """simple docstring""" # Initialize image_processing snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , Image.Image ) # Test not batched input snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) self.assertIsInstance(encoding.words , UpperCAmelCase__ ) self.assertIsInstance(encoding.boxes , UpperCAmelCase__ ) # Test batched snake_case : Dict = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" # Initialize image_processing snake_case : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , numpify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , np.ndarray ) # Test not batched input snake_case : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched snake_case : List[Any] = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" # Initialize image_processing snake_case : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , torchify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , torch.Tensor ) # Test not batched input snake_case : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched snake_case : Tuple = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" # with apply_OCR = True snake_case : int = LayoutLMvaImageProcessor() from datasets import load_dataset snake_case : List[Any] = load_dataset('''hf-internal-testing/fixtures_docvqa''' , split='''test''' ) snake_case : List[Any] = Image.open(ds[0]['''file'''] ).convert('''RGB''' ) snake_case : Any = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 snake_case : Optional[Any] = [['''11:14''', '''to''', '''11:39''', '''a.m''', '''11:39''', '''to''', '''11:44''', '''a.m.''', '''11:44''', '''a.m.''', '''to''', '''12:25''', '''p.m.''', '''12:25''', '''to''', '''12:58''', '''p.m.''', '''12:58''', '''to''', '''4:00''', '''p.m.''', '''2:00''', '''to''', '''5:00''', '''p.m.''', '''Coffee''', '''Break''', '''Coffee''', '''will''', '''be''', '''served''', '''for''', '''men''', '''and''', '''women''', '''in''', '''the''', '''lobby''', '''adjacent''', '''to''', '''exhibit''', '''area.''', '''Please''', '''move''', '''into''', '''exhibit''', '''area.''', '''(Exhibits''', '''Open)''', '''TRRF''', '''GENERAL''', '''SESSION''', '''(PART''', '''|)''', '''Presiding:''', '''Lee''', '''A.''', '''Waller''', '''TRRF''', '''Vice''', '''President''', '''“Introductory''', '''Remarks”''', '''Lee''', '''A.''', '''Waller,''', '''TRRF''', '''Vice''', '''Presi-''', '''dent''', '''Individual''', '''Interviews''', '''with''', '''TRRF''', '''Public''', '''Board''', '''Members''', '''and''', '''Sci-''', '''entific''', '''Advisory''', '''Council''', '''Mem-''', '''bers''', '''Conducted''', '''by''', '''TRRF''', '''Treasurer''', '''Philip''', '''G.''', '''Kuehn''', '''to''', '''get''', '''answers''', '''which''', '''the''', '''public''', '''refrigerated''', '''warehousing''', '''industry''', '''is''', '''looking''', '''for.''', '''Plus''', '''questions''', '''from''', '''the''', '''floor.''', '''Dr.''', '''Emil''', '''M.''', '''Mrak,''', '''University''', '''of''', '''Cal-''', '''ifornia,''', '''Chairman,''', '''TRRF''', '''Board;''', '''Sam''', '''R.''', '''Cecil,''', '''University''', '''of''', '''Georgia''', '''College''', '''of''', '''Agriculture;''', '''Dr.''', '''Stanley''', '''Charm,''', '''Tufts''', '''University''', '''School''', '''of''', '''Medicine;''', '''Dr.''', '''Robert''', '''H.''', '''Cotton,''', '''ITT''', '''Continental''', '''Baking''', '''Company;''', '''Dr.''', '''Owen''', '''Fennema,''', '''University''', '''of''', '''Wis-''', '''consin;''', '''Dr.''', '''Robert''', '''E.''', '''Hardenburg,''', '''USDA.''', '''Questions''', '''and''', '''Answers''', '''Exhibits''', '''Open''', '''Capt.''', '''Jack''', '''Stoney''', '''Room''', '''TRRF''', '''Scientific''', '''Advisory''', '''Council''', '''Meeting''', '''Ballroom''', '''Foyer''']] # noqa: E231 snake_case : Union[str, Any] = [[[141, 57, 214, 69], [228, 58, 252, 69], [141, 75, 216, 88], [230, 79, 280, 88], [142, 260, 218, 273], [230, 261, 255, 273], [143, 279, 218, 290], [231, 282, 290, 291], [143, 342, 218, 354], [231, 345, 289, 355], [202, 362, 227, 373], [143, 379, 220, 392], [231, 382, 291, 394], [144, 714, 220, 726], [231, 715, 256, 726], [144, 732, 220, 745], [232, 736, 291, 747], [144, 769, 218, 782], [231, 770, 256, 782], [141, 788, 202, 801], [215, 791, 274, 804], [143, 826, 204, 838], [215, 826, 240, 838], [142, 844, 202, 857], [215, 847, 274, 859], [334, 57, 427, 69], [440, 57, 522, 69], [369, 75, 461, 88], [469, 75, 516, 88], [528, 76, 562, 88], [570, 76, 667, 88], [675, 75, 711, 87], [721, 79, 778, 88], [789, 75, 840, 88], [369, 97, 470, 107], [484, 94, 507, 106], [518, 94, 562, 107], [576, 94, 655, 110], [668, 94, 792, 109], [804, 95, 829, 107], [369, 113, 465, 125], [477, 116, 547, 125], [562, 113, 658, 125], [671, 116, 748, 125], [761, 113, 811, 125], [369, 131, 465, 143], [477, 133, 548, 143], [563, 130, 698, 145], [710, 130, 802, 146], [336, 171, 412, 183], [423, 171, 572, 183], [582, 170, 716, 184], [728, 171, 817, 187], [829, 171, 844, 186], [338, 197, 482, 212], [507, 196, 557, 209], [569, 196, 595, 208], [610, 196, 702, 209], [505, 214, 583, 226], [595, 214, 656, 227], [670, 215, 807, 227], [335, 259, 543, 274], [556, 259, 708, 272], [372, 279, 422, 291], [435, 279, 460, 291], [474, 279, 574, 292], [587, 278, 664, 291], [676, 278, 738, 291], [751, 279, 834, 291], [372, 298, 434, 310], [335, 341, 483, 354], [497, 341, 655, 354], [667, 341, 728, 354], [740, 341, 825, 354], [335, 360, 430, 372], [442, 360, 534, 372], [545, 359, 687, 372], [697, 360, 754, 372], [765, 360, 823, 373], [334, 378, 428, 391], [440, 378, 577, 394], [590, 378, 705, 391], [720, 378, 801, 391], [334, 397, 400, 409], [370, 416, 529, 429], [544, 416, 576, 432], [587, 416, 665, 428], [677, 416, 814, 429], [372, 435, 452, 450], [465, 434, 495, 447], [511, 434, 600, 447], [611, 436, 637, 447], [649, 436, 694, 451], [705, 438, 824, 447], [369, 453, 452, 466], [464, 454, 509, 466], [522, 453, 611, 469], [625, 453, 792, 469], [370, 472, 556, 488], [570, 472, 684, 487], [697, 472, 718, 485], [732, 472, 835, 488], [369, 490, 411, 503], [425, 490, 484, 503], [496, 490, 635, 506], [645, 490, 707, 503], [718, 491, 761, 503], [771, 490, 840, 503], [336, 510, 374, 521], [388, 510, 447, 522], [460, 510, 489, 521], [503, 510, 580, 522], [592, 509, 736, 525], [745, 509, 770, 522], [781, 509, 840, 522], [338, 528, 434, 541], [448, 528, 596, 541], [609, 527, 687, 540], [700, 528, 792, 541], [336, 546, 397, 559], [407, 546, 431, 559], [443, 546, 525, 560], [537, 546, 680, 562], [688, 546, 714, 559], [722, 546, 837, 562], [336, 565, 449, 581], [461, 565, 485, 577], [497, 565, 665, 581], [681, 565, 718, 577], [732, 565, 837, 580], [337, 584, 438, 597], [452, 583, 521, 596], [535, 584, 677, 599], [690, 583, 787, 596], [801, 583, 825, 596], [338, 602, 478, 615], [492, 602, 530, 614], [543, 602, 638, 615], [650, 602, 676, 614], [688, 602, 788, 615], [802, 602, 843, 614], [337, 621, 502, 633], [516, 621, 615, 637], [629, 621, 774, 636], [789, 621, 827, 633], [337, 639, 418, 652], [432, 640, 571, 653], [587, 639, 731, 655], [743, 639, 769, 652], [780, 639, 841, 652], [338, 658, 440, 673], [455, 658, 491, 670], [508, 658, 602, 671], [616, 658, 638, 670], [654, 658, 835, 674], [337, 677, 429, 689], [337, 714, 482, 726], [495, 714, 548, 726], [561, 714, 683, 726], [338, 770, 461, 782], [474, 769, 554, 785], [489, 788, 562, 803], [576, 788, 643, 801], [656, 787, 751, 804], [764, 788, 844, 801], [334, 825, 421, 838], [430, 824, 574, 838], [584, 824, 723, 841], [335, 844, 450, 857], [464, 843, 583, 860], [628, 862, 755, 875], [769, 861, 848, 878]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , UpperCAmelCase__ ) self.assertListEqual(encoding.boxes , UpperCAmelCase__ ) # with apply_OCR = False snake_case : str = LayoutLMvaImageProcessor(apply_ocr=UpperCAmelCase__ ) snake_case : Optional[Any] = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) )
84
1
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _a : List[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name _a : Dict = '\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained("kandinsky-community/kandinsky-2-2-prior")\n >>> pipe_prior.to("cuda")\n >>> prompt = "red cat, 4k photo"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained("kandinsky-community/kandinsky-2-2-decoder")\n >>> pipe.to("cuda")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save("cat.png")\n ```\n' def a_ ( __magic_name__ , __magic_name__ , __magic_name__=8 ) -> Union[str, Any]: """simple docstring""" snake_case : Optional[int] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 snake_case : Optional[int] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class a_ ( a ): def __init__( self : Optional[Any] , UpperCAmelCase__ : UNetaDConditionModel , UpperCAmelCase__ : DDPMScheduler , UpperCAmelCase__ : VQModel , ): """simple docstring""" super().__init__() self.register_modules( unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ , movq=UpperCAmelCase__ , ) snake_case : int = 2 ** (len(self.movq.config.block_out_channels ) - 1) def lowerCAmelCase( self : int , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[Any] ): """simple docstring""" if latents is None: snake_case : str = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=UpperCAmelCase__ , dtype=UpperCAmelCase__ ) else: if latents.shape != shape: raise ValueError(F"Unexpected latents shape, got {latents.shape}, expected {shape}" ) snake_case : Union[str, Any] = latents.to(UpperCAmelCase__ ) snake_case : Optional[Any] = latents * scheduler.init_noise_sigma return latents def lowerCAmelCase( self : int , UpperCAmelCase__ : List[str]=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) snake_case : Tuple = torch.device(F"cuda:{gpu_id}" ) snake_case : Any = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : int=0 ): """simple docstring""" if is_accelerate_available() and is_accelerate_version('''>=''' , '''0.17.0.dev0''' ): from accelerate import cpu_offload_with_hook else: raise ImportError('''`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.''' ) snake_case : List[Any] = torch.device(F"cuda:{gpu_id}" ) if self.device.type != "cpu": self.to('''cpu''' , silence_dtype_warnings=UpperCAmelCase__ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) snake_case : Tuple = None for cpu_offloaded_model in [self.unet, self.movq]: snake_case , snake_case : Tuple = cpu_offload_with_hook(UpperCAmelCase__ , UpperCAmelCase__ , prev_module_hook=UpperCAmelCase__ ) # We'll offload the last model manually. snake_case : Dict = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def lowerCAmelCase( self : List[Any] ): """simple docstring""" if not hasattr(self.unet , '''_hf_hook''' ): return self.device for module in self.unet.modules(): if ( hasattr(UpperCAmelCase__ , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(UpperCAmelCase__ ) def __call__( self : Any , UpperCAmelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase__ : int = 512 , UpperCAmelCase__ : int = 512 , UpperCAmelCase__ : int = 100 , UpperCAmelCase__ : float = 4.0 , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase__ : Optional[torch.FloatTensor] = None , UpperCAmelCase__ : Optional[str] = "pil" , UpperCAmelCase__ : bool = True , ): """simple docstring""" snake_case : Optional[int] = self._execution_device snake_case : int = guidance_scale > 1.0 if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Tuple = torch.cat(UpperCAmelCase__ , dim=0 ) snake_case : List[str] = image_embeds.shape[0] * num_images_per_prompt if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : str = torch.cat(UpperCAmelCase__ , dim=0 ) if do_classifier_free_guidance: snake_case : str = image_embeds.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : str = negative_image_embeds.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Union[str, Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=UpperCAmelCase__ ) self.scheduler.set_timesteps(UpperCAmelCase__ , device=UpperCAmelCase__ ) snake_case : List[Any] = self.scheduler.timesteps snake_case : Optional[int] = self.unet.config.in_channels snake_case , snake_case : Dict = downscale_height_and_width(UpperCAmelCase__ , UpperCAmelCase__ , self.movq_scale_factor ) # create initial latent snake_case : str = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , self.scheduler , ) for i, t in enumerate(self.progress_bar(UpperCAmelCase__ ) ): # expand the latents if we are doing classifier free guidance snake_case : str = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents snake_case : Optional[Any] = {'''image_embeds''': image_embeds} snake_case : Any = self.unet( sample=UpperCAmelCase__ , timestep=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , added_cond_kwargs=UpperCAmelCase__ , return_dict=UpperCAmelCase__ , )[0] if do_classifier_free_guidance: snake_case , snake_case : Any = noise_pred.split(latents.shape[1] , dim=1 ) snake_case , snake_case : Tuple = noise_pred.chunk(2 ) snake_case , snake_case : str = variance_pred.chunk(2 ) snake_case : List[Any] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) snake_case : Dict = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , '''variance_type''' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): snake_case , snake_case : List[Any] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 snake_case : Optional[int] = self.scheduler.step( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , generator=UpperCAmelCase__ , )[0] # post-processing snake_case : List[str] = self.movq.decode(UpperCAmelCase__ , force_not_quantize=UpperCAmelCase__ )['''sample'''] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}" ) if output_type in ["np", "pil"]: snake_case : List[Any] = image * 0.5 + 0.5 snake_case : Optional[int] = image.clamp(0 , 1 ) snake_case : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": snake_case : Union[str, Any] = self.numpy_to_pil(UpperCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase__ )
84
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _a : Tuple = logging.get_logger(__name__) # pylint: disable=invalid-name _a : Dict = '\n Examples:\n ```py\n >>> import torch\n >>> import numpy as np\n\n >>> from diffusers import KandinskyV22PriorPipeline, KandinskyV22ControlnetPipeline\n >>> from transformers import pipeline\n >>> from diffusers.utils import load_image\n\n\n >>> def make_hint(image, depth_estimator):\n ... image = depth_estimator(image)["depth"]\n ... image = np.array(image)\n ... image = image[:, :, None]\n ... image = np.concatenate([image, image, image], axis=2)\n ... detected_map = torch.from_numpy(image).float() / 255.0\n ... hint = detected_map.permute(2, 0, 1)\n ... return hint\n\n\n >>> depth_estimator = pipeline("depth-estimation")\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16\n ... )\n >>> pipe_prior = pipe_prior.to("cuda")\n\n >>> pipe = KandinskyV22ControlnetPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-controlnet-depth", torch_dtype=torch.float16\n ... )\n >>> pipe = pipe.to("cuda")\n\n\n >>> img = load_image(\n ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main"\n ... "/kandinsky/cat.png"\n ... ).resize((768, 768))\n\n >>> hint = make_hint(img, depth_estimator).unsqueeze(0).half().to("cuda")\n\n >>> prompt = "A robot, 4k photo"\n >>> negative_prior_prompt = "lowres, text, error, cropped, worst quality, low quality, jpeg artifacts, ugly, duplicate, morbid, mutilated, out of frame, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, mutation, deformed, blurry, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, username, watermark, signature"\n\n >>> generator = torch.Generator(device="cuda").manual_seed(43)\n\n >>> image_emb, zero_image_emb = pipe_prior(\n ... prompt=prompt, negative_prompt=negative_prior_prompt, generator=generator\n ... ).to_tuple()\n\n >>> images = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... hint=hint,\n ... num_inference_steps=50,\n ... generator=generator,\n ... height=768,\n ... width=768,\n ... ).images\n\n >>> images[0].save("robot_cat.png")\n ```\n' def a_ ( __magic_name__ , __magic_name__ , __magic_name__=8 ) -> str: """simple docstring""" snake_case : List[Any] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 snake_case : Tuple = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class a_ ( a ): def __init__( self : Optional[int] , UpperCAmelCase__ : UNetaDConditionModel , UpperCAmelCase__ : DDPMScheduler , UpperCAmelCase__ : VQModel , ): """simple docstring""" super().__init__() self.register_modules( unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ , movq=UpperCAmelCase__ , ) snake_case : List[Any] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def lowerCAmelCase( self : int , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Any ): """simple docstring""" if latents is None: snake_case : int = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=UpperCAmelCase__ , dtype=UpperCAmelCase__ ) else: if latents.shape != shape: raise ValueError(F"Unexpected latents shape, got {latents.shape}, expected {shape}" ) snake_case : Optional[Any] = latents.to(UpperCAmelCase__ ) snake_case : List[Any] = latents * scheduler.init_noise_sigma return latents def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Optional[int]=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) snake_case : Union[str, Any] = torch.device(F"cuda:{gpu_id}" ) snake_case : Dict = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : Any=0 ): """simple docstring""" if is_accelerate_available() and is_accelerate_version('''>=''' , '''0.17.0.dev0''' ): from accelerate import cpu_offload_with_hook else: raise ImportError('''`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.''' ) snake_case : Optional[int] = torch.device(F"cuda:{gpu_id}" ) if self.device.type != "cpu": self.to('''cpu''' , silence_dtype_warnings=UpperCAmelCase__ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) snake_case : List[str] = None for cpu_offloaded_model in [self.unet, self.movq]: snake_case , snake_case : Optional[int] = cpu_offload_with_hook(UpperCAmelCase__ , UpperCAmelCase__ , prev_module_hook=UpperCAmelCase__ ) # We'll offload the last model manually. snake_case : Tuple = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" if not hasattr(self.unet , '''_hf_hook''' ): return self.device for module in self.unet.modules(): if ( hasattr(UpperCAmelCase__ , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(UpperCAmelCase__ ) def __call__( self : List[str] , UpperCAmelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase__ : torch.FloatTensor , UpperCAmelCase__ : int = 512 , UpperCAmelCase__ : int = 512 , UpperCAmelCase__ : int = 100 , UpperCAmelCase__ : float = 4.0 , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase__ : Optional[torch.FloatTensor] = None , UpperCAmelCase__ : Optional[str] = "pil" , UpperCAmelCase__ : bool = True , ): """simple docstring""" snake_case : Optional[int] = self._execution_device snake_case : Union[str, Any] = guidance_scale > 1.0 if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Any = torch.cat(UpperCAmelCase__ , dim=0 ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Union[str, Any] = torch.cat(UpperCAmelCase__ , dim=0 ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : int = torch.cat(UpperCAmelCase__ , dim=0 ) snake_case : List[Any] = image_embeds.shape[0] * num_images_per_prompt if do_classifier_free_guidance: snake_case : Dict = image_embeds.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Optional[Any] = negative_image_embeds.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Tuple = hint.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Optional[Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=UpperCAmelCase__ ) snake_case : List[str] = torch.cat([hint, hint] , dim=0 ).to(dtype=self.unet.dtype , device=UpperCAmelCase__ ) self.scheduler.set_timesteps(UpperCAmelCase__ , device=UpperCAmelCase__ ) snake_case : str = self.scheduler.timesteps snake_case : Optional[Any] = self.movq.config.latent_channels snake_case , snake_case : Optional[Any] = downscale_height_and_width(UpperCAmelCase__ , UpperCAmelCase__ , self.movq_scale_factor ) # create initial latent snake_case : Dict = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , self.scheduler , ) for i, t in enumerate(self.progress_bar(UpperCAmelCase__ ) ): # expand the latents if we are doing classifier free guidance snake_case : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents snake_case : Optional[int] = {'''image_embeds''': image_embeds, '''hint''': hint} snake_case : Any = self.unet( sample=UpperCAmelCase__ , timestep=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , added_cond_kwargs=UpperCAmelCase__ , return_dict=UpperCAmelCase__ , )[0] if do_classifier_free_guidance: snake_case , snake_case : Dict = noise_pred.split(latents.shape[1] , dim=1 ) snake_case , snake_case : Any = noise_pred.chunk(2 ) snake_case , snake_case : Dict = variance_pred.chunk(2 ) snake_case : str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) snake_case : List[str] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , '''variance_type''' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): snake_case , snake_case : Tuple = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 snake_case : List[Any] = self.scheduler.step( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , generator=UpperCAmelCase__ , )[0] # post-processing snake_case : List[Any] = self.movq.decode(UpperCAmelCase__ , force_not_quantize=UpperCAmelCase__ )['''sample'''] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}" ) if output_type in ["np", "pil"]: snake_case : Optional[Any] = image * 0.5 + 0.5 snake_case : int = image.clamp(0 , 1 ) snake_case : List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": snake_case : str = self.numpy_to_pil(UpperCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase__ )
84
1
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _a : Optional[Any] = logging.get_logger(__name__) _a : Any = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} _a : Optional[Any] = { 'tokenizer_file': { 'EleutherAI/gpt-neox-20b': 'https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/tokenizer.json', }, } _a : List[Any] = { 'gpt-neox-20b': 2_048, } class a_ ( a ): A__ : str = VOCAB_FILES_NAMES A__ : str = PRETRAINED_VOCAB_FILES_MAP A__ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : List[Any] = ['input_ids', 'attention_mask'] def __init__( self : Any , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : str=None , UpperCAmelCase__ : str="<|endoftext|>" , UpperCAmelCase__ : Dict="<|endoftext|>" , UpperCAmelCase__ : str="<|endoftext|>" , UpperCAmelCase__ : int=False , **UpperCAmelCase__ : Optional[int] , ): """simple docstring""" super().__init__( UpperCAmelCase__ , UpperCAmelCase__ , tokenizer_file=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , add_prefix_space=UpperCAmelCase__ , **UpperCAmelCase__ , ) snake_case : Optional[int] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , UpperCAmelCase__ ) != add_prefix_space: snake_case : List[str] = getattr(UpperCAmelCase__ , pre_tok_state.pop('''type''' ) ) snake_case : Union[str, Any] = add_prefix_space snake_case : List[Any] = pre_tok_class(**UpperCAmelCase__ ) snake_case : Optional[Any] = add_prefix_space def lowerCAmelCase( self : int , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ): """simple docstring""" snake_case : List[str] = self._tokenizer.model.save(UpperCAmelCase__ , name=UpperCAmelCase__ ) return tuple(UpperCAmelCase__ ) def lowerCAmelCase( self : int , UpperCAmelCase__ : "Conversation" ): """simple docstring""" snake_case : Optional[Any] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) + [self.eos_token_id] ) if len(UpperCAmelCase__ ) > self.model_max_length: snake_case : Union[str, Any] = input_ids[-self.model_max_length :] return input_ids
84
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _a : List[Any] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class a_ ( a , unittest.TestCase ): A__ : Dict = ReformerTokenizer A__ : Optional[int] = ReformerTokenizerFast A__ : str = True A__ : Tuple = False A__ : str = True def lowerCAmelCase( self : List[Any] ): """simple docstring""" super().setUp() snake_case : str = ReformerTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : int = '''<s>''' snake_case : List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase__ ) , UpperCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase__ ) , UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Any = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(UpperCAmelCase__ ) , 1_000 ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def lowerCAmelCase( self : Dict ): """simple docstring""" if not self.test_rust_tokenizer: return snake_case : Any = self.get_tokenizer() snake_case : str = self.get_rust_tokenizer() snake_case : Tuple = '''I was born in 92000, and this is falsé.''' snake_case : str = tokenizer.tokenize(UpperCAmelCase__ ) snake_case : int = rust_tokenizer.tokenize(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : Union[str, Any] = tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) snake_case : List[str] = rust_tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : List[str] = self.get_rust_tokenizer() snake_case : Optional[int] = tokenizer.encode(UpperCAmelCase__ ) snake_case : Optional[Any] = rust_tokenizer.encode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : List[Any]=15 ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): snake_case : str = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) # Simple input snake_case : Union[str, Any] = '''This is a simple input''' snake_case : List[str] = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case : int = ('''This is a simple input''', '''This is a pair''') snake_case : int = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Simple input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Simple input self.assertRaises( UpperCAmelCase__ , tokenizer_r.batch_encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' , ) # Pair input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Pair input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Pair input self.assertRaises( UpperCAmelCase__ , tokenizer_r.batch_encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' , ) def lowerCAmelCase( self : str ): """simple docstring""" pass def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Union[str, Any] = ReformerTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) snake_case : List[str] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(UpperCAmelCase__ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , [285, 46, 10, 170, 382] , ) snake_case : int = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) snake_case : int = tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) snake_case : List[Any] = tokenizer.convert_ids_to_tokens(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def lowerCAmelCase( self : Tuple ): """simple docstring""" return ReformerTokenizer.from_pretrained('''google/reformer-crime-and-punishment''' ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : Any = '''Hello World!''' snake_case : Optional[Any] = [126, 32, 262, 152, 38, 72, 287] self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @slow def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[Any] = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) snake_case : Dict = [ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @require_torch @slow def lowerCAmelCase( self : List[Any] ): """simple docstring""" import torch from transformers import ReformerConfig, ReformerModel # Build sequence snake_case : Any = list(self.big_tokenizer.get_vocab().keys() )[:10] snake_case : Union[str, Any] = ''' '''.join(UpperCAmelCase__ ) snake_case : Optional[int] = self.big_tokenizer.encode_plus(UpperCAmelCase__ , return_tensors='''pt''' ) snake_case : List[str] = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors='''pt''' ) snake_case : Optional[Any] = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) snake_case : Tuple = encoded_sequence['''input_ids'''].shape snake_case : List[Any] = ReformerModel(UpperCAmelCase__ ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**UpperCAmelCase__ ) model(**UpperCAmelCase__ ) @slow def lowerCAmelCase( self : Optional[int] ): """simple docstring""" # fmt: off snake_case : Tuple = {'''input_ids''': [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 snake_case : Tuple = [ '''This is a very simple sentence.''', '''The quick brown fox jumps over the lazy dog.''', ] self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase__ , model_name='''google/reformer-crime-and-punishment''' , revision='''0e6c3decb8211d49bf881013425dc8b0448b3f5a''' , padding=UpperCAmelCase__ , sequences=UpperCAmelCase__ , )
84
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a : Dict = logging.get_logger(__name__) _a : str = { 'facebook/data2vec-vision-base-ft': ( 'https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json' ), } class a_ ( a ): A__ : Tuple = 'data2vec-vision' def __init__( self : Optional[Any] , UpperCAmelCase__ : Optional[int]=768 , UpperCAmelCase__ : List[Any]=12 , UpperCAmelCase__ : Optional[int]=12 , UpperCAmelCase__ : int=3_072 , UpperCAmelCase__ : Union[str, Any]="gelu" , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Union[str, Any]=0.0 , UpperCAmelCase__ : List[str]=0.02 , UpperCAmelCase__ : List[str]=1e-1_2 , UpperCAmelCase__ : int=224 , UpperCAmelCase__ : Dict=16 , UpperCAmelCase__ : Tuple=3 , UpperCAmelCase__ : Dict=False , UpperCAmelCase__ : Tuple=False , UpperCAmelCase__ : List[str]=False , UpperCAmelCase__ : Any=False , UpperCAmelCase__ : Optional[Any]=0.1 , UpperCAmelCase__ : Optional[int]=0.1 , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : str=[3, 5, 7, 11] , UpperCAmelCase__ : Optional[int]=[1, 2, 3, 6] , UpperCAmelCase__ : str=True , UpperCAmelCase__ : Tuple=0.4 , UpperCAmelCase__ : List[str]=256 , UpperCAmelCase__ : Optional[Any]=1 , UpperCAmelCase__ : Any=False , UpperCAmelCase__ : Union[str, Any]=255 , **UpperCAmelCase__ : Union[str, Any] , ): """simple docstring""" super().__init__(**UpperCAmelCase__ ) snake_case : Dict = hidden_size snake_case : Tuple = num_hidden_layers snake_case : Dict = num_attention_heads snake_case : int = intermediate_size snake_case : Dict = hidden_act snake_case : Tuple = hidden_dropout_prob snake_case : Union[str, Any] = attention_probs_dropout_prob snake_case : Dict = initializer_range snake_case : Tuple = layer_norm_eps snake_case : List[Any] = image_size snake_case : Any = patch_size snake_case : List[Any] = num_channels snake_case : Any = use_mask_token snake_case : Any = use_absolute_position_embeddings snake_case : List[Any] = use_relative_position_bias snake_case : Any = use_shared_relative_position_bias snake_case : Any = layer_scale_init_value snake_case : Optional[int] = drop_path_rate snake_case : str = use_mean_pooling # decode head attributes (semantic segmentation) snake_case : Any = out_indices snake_case : List[str] = pool_scales # auxiliary head attributes (semantic segmentation) snake_case : Union[str, Any] = use_auxiliary_head snake_case : Optional[int] = auxiliary_loss_weight snake_case : int = auxiliary_channels snake_case : int = auxiliary_num_convs snake_case : List[str] = auxiliary_concat_input snake_case : int = semantic_loss_ignore_index class a_ ( a ): A__ : Optional[int] = version.parse('1.11' ) @property def lowerCAmelCase( self : int ): """simple docstring""" return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def lowerCAmelCase( self : str ): """simple docstring""" return 1e-4
84
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def a_ ( __magic_name__ ) -> Tuple: """simple docstring""" snake_case , snake_case : Any = image.size snake_case , snake_case : List[str] = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 snake_case : List[str] = image.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) snake_case : Dict = np.array(__magic_name__ ).astype(np.floataa ) / 255.0 snake_case : Optional[Any] = image[None].transpose(0 , 3 , 1 , 2 ) snake_case : Tuple = torch.from_numpy(__magic_name__ ) return 2.0 * image - 1.0 class a_ ( a ): def __init__( self : Optional[Any] , UpperCAmelCase__ : VQModel , UpperCAmelCase__ : UNetaDModel , UpperCAmelCase__ : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): """simple docstring""" super().__init__() self.register_modules(vqvae=UpperCAmelCase__ , unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ ) @torch.no_grad() def __call__( self : Any , UpperCAmelCase__ : Union[torch.Tensor, PIL.Image.Image] = None , UpperCAmelCase__ : Optional[int] = 1 , UpperCAmelCase__ : Optional[int] = 100 , UpperCAmelCase__ : Optional[float] = 0.0 , UpperCAmelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase__ : Optional[str] = "pil" , UpperCAmelCase__ : bool = True , ): """simple docstring""" if isinstance(UpperCAmelCase__ , PIL.Image.Image ): snake_case : Optional[int] = 1 elif isinstance(UpperCAmelCase__ , torch.Tensor ): snake_case : Any = image.shape[0] else: raise ValueError(F"`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(UpperCAmelCase__ )}" ) if isinstance(UpperCAmelCase__ , PIL.Image.Image ): snake_case : Optional[Any] = preprocess(UpperCAmelCase__ ) snake_case , snake_case : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image snake_case : List[Any] = (batch_size, self.unet.config.in_channels // 2, height, width) snake_case : str = next(self.unet.parameters() ).dtype snake_case : Dict = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=self.device , dtype=UpperCAmelCase__ ) snake_case : Any = image.to(device=self.device , dtype=UpperCAmelCase__ ) # set timesteps and move to the correct device self.scheduler.set_timesteps(UpperCAmelCase__ , device=self.device ) snake_case : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler snake_case : Union[str, Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] snake_case : Any = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) snake_case : Optional[Any] = {} if accepts_eta: snake_case : Dict = eta for t in self.progress_bar(UpperCAmelCase__ ): # concat latents and low resolution image in the channel dimension. snake_case : Optional[int] = torch.cat([latents, image] , dim=1 ) snake_case : str = self.scheduler.scale_model_input(UpperCAmelCase__ , UpperCAmelCase__ ) # predict the noise residual snake_case : int = self.unet(UpperCAmelCase__ , UpperCAmelCase__ ).sample # compute the previous noisy sample x_t -> x_t-1 snake_case : Any = self.scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample # decode the image latents with the VQVAE snake_case : Optional[int] = self.vqvae.decode(UpperCAmelCase__ ).sample snake_case : int = torch.clamp(UpperCAmelCase__ , -1.0 , 1.0 ) snake_case : Dict = image / 2 + 0.5 snake_case : int = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case : Any = self.numpy_to_pil(UpperCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase__ )
84
1
import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class a_ ( unittest.TestCase ): def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : int = ['''a''', '''b''', '''c'''] # Defaults to last layer if both are None snake_case , snake_case : Tuple = get_aligned_output_features_output_indices(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , ['''c'''] ) self.assertEqual(UpperCAmelCase__ , [2] ) # Out indices set to match out features snake_case , snake_case : Dict = get_aligned_output_features_output_indices(['''a''', '''c'''] , UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , ['''a''', '''c'''] ) self.assertEqual(UpperCAmelCase__ , [0, 2] ) # Out features set to match out indices snake_case , snake_case : Any = get_aligned_output_features_output_indices(UpperCAmelCase__ , [0, 2] , UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , ['''a''', '''c'''] ) self.assertEqual(UpperCAmelCase__ , [0, 2] ) # Out features selected from negative indices snake_case , snake_case : Optional[Any] = get_aligned_output_features_output_indices(UpperCAmelCase__ , [-3, -1] , UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , ['''a''', '''c'''] ) self.assertEqual(UpperCAmelCase__ , [-3, -1] ) def lowerCAmelCase( self : str ): """simple docstring""" # Stage names must be set with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(['''a''', '''b'''] , (0, 1) , UpperCAmelCase__ ) # Out features must be a list with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(('''a''', '''b''') , (0, 1) , ['''a''', '''b'''] ) # Out features must be a subset of stage names with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(['''a''', '''b'''] , (0, 1) , ['''a'''] ) # Out indices must be a list or tuple with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(UpperCAmelCase__ , 0 , ['''a''', '''b'''] ) # Out indices must be a subset of stage names with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(UpperCAmelCase__ , (0, 1) , ['''a'''] ) # Out features and out indices must be the same length with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(['''a''', '''b'''] , (0,) , ['''a''', '''b''', '''c'''] ) # Out features should match out indices with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(['''a''', '''b'''] , (0, 2) , ['''a''', '''b''', '''c'''] ) # Out features and out indices should be in order with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(['''b''', '''a'''] , (0, 1) , ['''a''', '''b'''] ) # Check passes with valid inputs verify_out_features_out_indices(['''a''', '''b''', '''d'''] , (0, 1, -1) , ['''a''', '''b''', '''c''', '''d'''] ) def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : Union[str, Any] = BackboneMixin() snake_case : List[Any] = ['''a''', '''b''', '''c'''] snake_case : int = ['''a''', '''c'''] snake_case : Tuple = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features , ['''a''', '''c'''] ) self.assertEqual(backbone.out_indices , [0, 2] ) # Check out features and indices are updated correctly snake_case : Optional[int] = ['''a''', '''b'''] self.assertEqual(backbone.out_features , ['''a''', '''b'''] ) self.assertEqual(backbone.out_indices , [0, 1] ) snake_case : List[Any] = [-3, -1] self.assertEqual(backbone.out_features , ['''a''', '''c'''] ) self.assertEqual(backbone.out_indices , [-3, -1] )
84
import os import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers.models.realm.configuration_realm import RealmConfig from transformers.models.realm.retrieval_realm import _REALM_BLOCK_RECORDS_FILENAME, RealmRetriever from transformers.models.realm.tokenization_realm import VOCAB_FILES_NAMES, RealmTokenizer class a_ ( a ): def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[Any] = tempfile.mkdtemp() snake_case : Dict = 5 # Realm tok snake_case : str = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''test''', '''question''', '''this''', '''is''', '''the''', '''first''', '''second''', '''third''', '''fourth''', '''fifth''', '''record''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] snake_case : Tuple = os.path.join(self.tmpdirname , '''realm_tokenizer''' ) os.makedirs(UpperCAmelCase__ , exist_ok=UpperCAmelCase__ ) snake_case : Any = os.path.join(UpperCAmelCase__ , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) snake_case : Tuple = os.path.join(self.tmpdirname , '''realm_block_records''' ) os.makedirs(UpperCAmelCase__ , exist_ok=UpperCAmelCase__ ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" return RealmTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''realm_tokenizer''' ) ) def lowerCAmelCase( self : Any ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Any = RealmConfig(num_block_records=self.num_block_records ) return config def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Optional[int] = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''question''': ['''foo''', '''bar'''], '''answers''': [['''Foo''', '''Bar'''], ['''Bar''']], } ) return dataset def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Dict = np.array( [ b'''This is the first record''', b'''This is the second record''', b'''This is the third record''', b'''This is the fourth record''', b'''This is the fifth record''', b'''This is a longer longer longer record''', ] , dtype=UpperCAmelCase__ , ) return block_records def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Tuple = RealmRetriever( block_records=self.get_dummy_block_records() , tokenizer=self.get_tokenizer() , ) return retriever def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[str] = self.get_config() snake_case : Optional[Any] = self.get_dummy_retriever() snake_case : Optional[int] = retriever.tokenizer snake_case : Dict = np.array([0, 3] , dtype='''long''' ) snake_case : Optional[int] = tokenizer(['''Test question'''] ).input_ids snake_case : Union[str, Any] = tokenizer( ['''the fourth'''] , add_special_tokens=UpperCAmelCase__ , return_token_type_ids=UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , ).input_ids snake_case : Optional[Any] = config.reader_seq_len snake_case , snake_case , snake_case , snake_case : List[str] = retriever( UpperCAmelCase__ , UpperCAmelCase__ , answer_ids=UpperCAmelCase__ , max_length=UpperCAmelCase__ , return_tensors='''np''' ) self.assertEqual(len(UpperCAmelCase__ ) , 2 ) self.assertEqual(len(UpperCAmelCase__ ) , 2 ) self.assertEqual(len(UpperCAmelCase__ ) , 2 ) self.assertEqual(concat_inputs.input_ids.shape , (2, 10) ) self.assertEqual(concat_inputs.attention_mask.shape , (2, 10) ) self.assertEqual(concat_inputs.token_type_ids.shape , (2, 10) ) self.assertEqual(concat_inputs.special_tokens_mask.shape , (2, 10) ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[0] ) , ['''[CLS]''', '''test''', '''question''', '''[SEP]''', '''this''', '''is''', '''the''', '''first''', '''record''', '''[SEP]'''] , ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[1] ) , ['''[CLS]''', '''test''', '''question''', '''[SEP]''', '''this''', '''is''', '''the''', '''fourth''', '''record''', '''[SEP]'''] , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[Any] = self.get_config() snake_case : Optional[int] = self.get_dummy_retriever() snake_case : List[str] = retriever.tokenizer snake_case : Optional[Any] = np.array([0, 3, 5] , dtype='''long''' ) snake_case : Optional[int] = tokenizer(['''Test question'''] ).input_ids snake_case : Any = tokenizer( ['''the fourth''', '''longer longer'''] , add_special_tokens=UpperCAmelCase__ , return_token_type_ids=UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , ).input_ids snake_case : List[Any] = config.reader_seq_len snake_case , snake_case , snake_case , snake_case : Union[str, Any] = retriever( UpperCAmelCase__ , UpperCAmelCase__ , answer_ids=UpperCAmelCase__ , max_length=UpperCAmelCase__ , return_tensors='''np''' ) self.assertEqual([False, True, True] , UpperCAmelCase__ ) self.assertEqual([[-1, -1, -1], [6, -1, -1], [6, 7, 8]] , UpperCAmelCase__ ) self.assertEqual([[-1, -1, -1], [7, -1, -1], [7, 8, 9]] , UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : int = self.get_dummy_retriever() retriever.save_pretrained(os.path.join(self.tmpdirname , '''realm_block_records''' ) ) # Test local path snake_case : Optional[Any] = retriever.from_pretrained(os.path.join(self.tmpdirname , '''realm_block_records''' ) ) self.assertEqual(retriever.block_records[0] , b'''This is the first record''' ) # Test mocked remote path with patch('''transformers.models.realm.retrieval_realm.hf_hub_download''' ) as mock_hf_hub_download: snake_case : Any = os.path.join( os.path.join(self.tmpdirname , '''realm_block_records''' ) , _REALM_BLOCK_RECORDS_FILENAME ) snake_case : Any = RealmRetriever.from_pretrained('''google/realm-cc-news-pretrained-openqa''' ) self.assertEqual(retriever.block_records[0] , b'''This is the first record''' )
84
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _a : Union[str, Any] = { 'configuration_gpt_neo': ['GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GPTNeoConfig', 'GPTNeoOnnxConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[str] = [ 'GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST', 'GPTNeoForCausalLM', 'GPTNeoForQuestionAnswering', 'GPTNeoForSequenceClassification', 'GPTNeoForTokenClassification', 'GPTNeoModel', 'GPTNeoPreTrainedModel', 'load_tf_weights_in_gpt_neo', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = [ 'FlaxGPTNeoForCausalLM', 'FlaxGPTNeoModel', 'FlaxGPTNeoPreTrainedModel', ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys _a : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
84
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _a : str = {'configuration_encoder_decoder': ['EncoderDecoderConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = ['EncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Dict = ['TFEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = ['FlaxEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys _a : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
84
1
import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() _a : List[str] = logging.get_logger(__name__) _a : str = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'lm_head', 'mask_emb': 'masked_spec_embed', } _a : Tuple = [ 'lm_head', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" for attribute in key.split('''.''' ): snake_case : List[str] = getattr(__magic_name__ , __magic_name__ ) if weight_type is not None: snake_case : Tuple = getattr(__magic_name__ , __magic_name__ ).shape else: snake_case : List[str] = hf_pointer.shape assert hf_shape == value.shape, ( F"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" F" {value.shape} for {full_name}" ) if weight_type == "weight": snake_case : Optional[int] = value elif weight_type == "weight_g": snake_case : Union[str, Any] = value elif weight_type == "weight_v": snake_case : Optional[int] = value elif weight_type == "bias": snake_case : Optional[Any] = value else: snake_case : str = value logger.info(F"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def a_ ( __magic_name__ , __magic_name__ ) -> Optional[Any]: """simple docstring""" snake_case : Optional[Any] = [] snake_case : Optional[int] = fairseq_model.state_dict() snake_case : Optional[Any] = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight snake_case : Any = None for name, value in fairseq_dict.items(): snake_case : Dict = False if "conv_layers" in name: load_conv_layer( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , hf_model.config.feat_extract_norm == '''group''' , ) snake_case : List[str] = True elif name.split('''.''' )[0] == "proj": snake_case : Tuple = fairseq_model.proj snake_case : Any = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: snake_case : int = True if "*" in mapped_key: snake_case : Dict = name.split(__magic_name__ )[0].split('''.''' )[-2] snake_case : Optional[int] = mapped_key.replace('''*''' , __magic_name__ ) if "weight_g" in name: snake_case : Any = '''weight_g''' elif "weight_v" in name: snake_case : str = '''weight_v''' elif "bias" in name: snake_case : Optional[Any] = '''bias''' elif "weight" in name: snake_case : Tuple = '''weight''' else: snake_case : Dict = None set_recursively(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) continue if not is_used: unused_weights.append(__magic_name__ ) logger.warning(F"Unused weights: {unused_weights}" ) return proj_weight def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Union[str, Any]: """simple docstring""" snake_case : Any = full_name.split('''conv_layers.''' )[-1] snake_case : List[str] = name.split('''.''' ) snake_case : int = int(items[0] ) snake_case : List[Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) snake_case : str = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) snake_case : Optional[int] = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was" " found." ) snake_case : Any = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found." ) snake_case : str = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(__magic_name__ ) def a_ ( __magic_name__ ) -> Tuple: """simple docstring""" snake_case , snake_case : int = emb.weight.shape snake_case : Optional[Any] = nn.Linear(__magic_name__ , __magic_name__ , bias=__magic_name__ ) snake_case : List[Any] = emb.weight.data return lin_layer def a_ ( __magic_name__ ) -> Optional[int]: """simple docstring""" with open(__magic_name__ , '''r''' , encoding='''utf-8''' ) as f: snake_case : int = f.readlines() snake_case : str = [line.split(''' ''' )[0] for line in lines] snake_case : List[str] = len(__magic_name__ ) snake_case : str = { '''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3, } vocab_dict.update(dict(zip(__magic_name__ , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , ) -> Any: """simple docstring""" snake_case : List[Any] = WavaVecaConfig.from_pretrained(__magic_name__ ) snake_case : Union[str, Any] = SpeechaTextaConfig.from_pretrained( __magic_name__ , vocab_size=__magic_name__ , decoder_layers=__magic_name__ , do_stable_layer_norm=__magic_name__ ) snake_case : Optional[int] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0 , do_normalize=__magic_name__ , return_attention_mask=__magic_name__ , ) snake_case , snake_case , snake_case : int = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) snake_case : Optional[int] = model[0].eval() # set weights for wav2vec2 encoder snake_case : int = WavaVecaModel(__magic_name__ ) snake_case : Dict = recursively_load_weights_wavaveca(model.encoder , __magic_name__ ) snake_case : Union[str, Any] = SpeechaTextaForCausalLM(__magic_name__ ) snake_case , snake_case : List[Any] = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=__magic_name__ ) # set output linear layer unexpected_keys.remove('''embed_out''' ) snake_case : Optional[int] = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(F"The following keys are missing when loading the decoder weights: {missing_keys}" ) logger.warning(F"The following keys are unexpected when loading the decoder weights: {unexpected_keys}" ) snake_case : Tuple = SpeechEncoderDecoderModel(encoder=__magic_name__ , decoder=__magic_name__ ) snake_case : Optional[int] = False # add projection layer snake_case : Any = nn.Parameter(projection_layer.weight ) snake_case : List[str] = nn.Parameter(projection_layer.bias ) snake_case : int = create_vocab_dict(__magic_name__ ) with open(os.path.join(__magic_name__ , '''vocab.json''' ) , '''w''' ) as fp: json.dump(__magic_name__ , __magic_name__ ) snake_case : Tuple = SpeechaTextaTokenizer(os.path.join(__magic_name__ , '''vocab.json''' ) ) tokenizer.save_pretrained(__magic_name__ ) snake_case : Union[str, Any] = hf_wavavec.config.to_dict() snake_case : Any = tokenizer.pad_token_id snake_case : Tuple = tokenizer.bos_token_id snake_case : List[Any] = tokenizer.eos_token_id snake_case : Optional[int] = '''speech_to_text_2''' snake_case : str = '''wav2vec2''' snake_case : List[Any] = SpeechEncoderDecoderConfig.from_dict(__magic_name__ ) hf_wavavec.save_pretrained(__magic_name__ ) feature_extractor.save_pretrained(__magic_name__ ) if __name__ == "__main__": _a : Tuple = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument( '--encoder_config_path', default='facebook/wav2vec2-large-lv60', type=str, help='Path to hf encoder wav2vec2 checkpoint config', ) parser.add_argument( '--decoder_config_path', default='facebook/s2t-small-mustc-en-fr-st', type=str, help='Path to hf decoder s2t checkpoint config', ) parser.add_argument('--vocab_size', default=10_224, type=int, help='Vocab size of decoder') parser.add_argument('--num_decoder_layers', default=7, type=int, help='Number of decoder layers') _a : Any = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
84
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer _a : str = logging.get_logger(__name__) _a : Optional[int] = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} _a : Optional[Any] = { 'vocab_file': { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json' ), }, } _a : Union[str, Any] = { 'yjernite/retribert-base-uncased': 512, } _a : Tuple = { 'yjernite/retribert-base-uncased': {'do_lower_case': True}, } class a_ ( a ): A__ : List[str] = VOCAB_FILES_NAMES A__ : Any = PRETRAINED_VOCAB_FILES_MAP A__ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Any = PRETRAINED_INIT_CONFIGURATION A__ : Optional[Any] = RetriBertTokenizer A__ : Any = ['input_ids', 'attention_mask'] def __init__( self : Optional[int] , UpperCAmelCase__ : Any=None , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Dict="[UNK]" , UpperCAmelCase__ : str="[SEP]" , UpperCAmelCase__ : Union[str, Any]="[PAD]" , UpperCAmelCase__ : Dict="[CLS]" , UpperCAmelCase__ : Optional[Any]="[MASK]" , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Optional[int]=None , **UpperCAmelCase__ : Dict , ): """simple docstring""" super().__init__( UpperCAmelCase__ , tokenizer_file=UpperCAmelCase__ , do_lower_case=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , tokenize_chinese_chars=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ , **UpperCAmelCase__ , ) snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , UpperCAmelCase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' , UpperCAmelCase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , UpperCAmelCase__ ) != tokenize_chinese_chars ): snake_case : int = getattr(UpperCAmelCase__ , normalizer_state.pop('''type''' ) ) snake_case : List[Any] = do_lower_case snake_case : Union[str, Any] = strip_accents snake_case : int = tokenize_chinese_chars snake_case : int = normalizer_class(**UpperCAmelCase__ ) snake_case : Union[str, Any] = do_lower_case def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Union[str, Any]=None ): """simple docstring""" snake_case : str = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ): """simple docstring""" snake_case : List[Any] = [self.sep_token_id] snake_case : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCAmelCase( self : str , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ): """simple docstring""" snake_case : Tuple = self._tokenizer.model.save(UpperCAmelCase__ , name=UpperCAmelCase__ ) return tuple(UpperCAmelCase__ )
84
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _a : Optional[int] = { 'configuration_efficientnet': [ 'EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'EfficientNetConfig', 'EfficientNetOnnxConfig', ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[Any] = ['EfficientNetImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[Any] = [ 'EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'EfficientNetForImageClassification', 'EfficientNetModel', 'EfficientNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_efficientnet import ( EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientNetConfig, EfficientNetOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientnet import EfficientNetImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientnet import ( EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientNetForImageClassification, EfficientNetModel, EfficientNetPreTrainedModel, ) else: import sys _a : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure)
84
import string import numpy def a_ ( __magic_name__ , __magic_name__ ) -> int: """simple docstring""" return b if a == 0 else greatest_common_divisor(b % a , __magic_name__ ) class a_ : A__ : List[Any] = string.ascii_uppercase + string.digits # This cipher takes alphanumerics into account # i.e. a total of 36 characters # take x and return x % len(key_string) A__ : List[str] = numpy.vectorize(lambda a : x % 36 ) A__ : Dict = numpy.vectorize(a ) def __init__( self : List[str] , UpperCAmelCase__ : numpy.ndarray ): """simple docstring""" snake_case : int = self.modulus(UpperCAmelCase__ ) # mod36 calc's on the encrypt key self.check_determinant() # validate the determinant of the encryption key snake_case : List[str] = encrypt_key.shape[0] def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : str ): """simple docstring""" return self.key_string.index(UpperCAmelCase__ ) def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : int ): """simple docstring""" return self.key_string[round(UpperCAmelCase__ )] def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[Any] = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: snake_case : Tuple = det % len(self.key_string ) snake_case : Tuple = len(self.key_string ) if greatest_common_divisor(UpperCAmelCase__ , len(self.key_string ) ) != 1: snake_case : List[Any] = ( F"determinant modular {req_l} of encryption key({det}) " F"is not co prime w.r.t {req_l}.\nTry another key." ) raise ValueError(UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Optional[int] = [char for char in text.upper() if char in self.key_string] snake_case : Optional[int] = chars[-1] while len(UpperCAmelCase__ ) % self.break_key != 0: chars.append(UpperCAmelCase__ ) return "".join(UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Optional[int] = self.process_text(text.upper() ) snake_case : Optional[int] = '''''' for i in range(0 , len(UpperCAmelCase__ ) - self.break_key + 1 , self.break_key ): snake_case : int = text[i : i + self.break_key] snake_case : int = [self.replace_letters(UpperCAmelCase__ ) for char in batch] snake_case : Tuple = numpy.array([vec] ).T snake_case : Optional[Any] = self.modulus(self.encrypt_key.dot(UpperCAmelCase__ ) ).T.tolist()[ 0 ] snake_case : Dict = ''''''.join( self.replace_digits(UpperCAmelCase__ ) for num in batch_encrypted ) encrypted += encrypted_batch return encrypted def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Optional[int] = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: snake_case : int = det % len(self.key_string ) snake_case : Dict = None for i in range(len(self.key_string ) ): if (det * i) % len(self.key_string ) == 1: snake_case : Any = i break snake_case : Any = ( det_inv * numpy.linalg.det(self.encrypt_key ) * numpy.linalg.inv(self.encrypt_key ) ) return self.to_int(self.modulus(UpperCAmelCase__ ) ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Any = self.make_decrypt_key() snake_case : Optional[Any] = self.process_text(text.upper() ) snake_case : int = '''''' for i in range(0 , len(UpperCAmelCase__ ) - self.break_key + 1 , self.break_key ): snake_case : Any = text[i : i + self.break_key] snake_case : int = [self.replace_letters(UpperCAmelCase__ ) for char in batch] snake_case : List[str] = numpy.array([vec] ).T snake_case : Optional[Any] = self.modulus(decrypt_key.dot(UpperCAmelCase__ ) ).T.tolist()[0] snake_case : int = ''''''.join( self.replace_digits(UpperCAmelCase__ ) for num in batch_decrypted ) decrypted += decrypted_batch return decrypted def a_ ( ) -> None: """simple docstring""" snake_case : Any = int(input('''Enter the order of the encryption key: ''' ) ) snake_case : List[Any] = [] print('''Enter each row of the encryption key with space separated integers''' ) for _ in range(__magic_name__ ): snake_case : Optional[Any] = [int(__magic_name__ ) for x in input().split()] hill_matrix.append(__magic_name__ ) snake_case : List[str] = HillCipher(numpy.array(__magic_name__ ) ) print('''Would you like to encrypt or decrypt some text? (1 or 2)''' ) snake_case : int = input('''\n1. Encrypt\n2. Decrypt\n''' ) if option == "1": snake_case : List[Any] = input('''What text would you like to encrypt?: ''' ) print('''Your encrypted text is:''' ) print(hc.encrypt(__magic_name__ ) ) elif option == "2": snake_case : int = input('''What text would you like to decrypt?: ''' ) print('''Your decrypted text is:''' ) print(hc.decrypt(__magic_name__ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
84
1
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class a_ : def __init__( self : Union[str, Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any]=99 , UpperCAmelCase__ : Dict=13 , UpperCAmelCase__ : int=7 , UpperCAmelCase__ : Any=9 , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Tuple=False , UpperCAmelCase__ : Tuple=32 , UpperCAmelCase__ : Dict=5 , UpperCAmelCase__ : Optional[int]=4 , UpperCAmelCase__ : List[Any]=37 , UpperCAmelCase__ : Union[str, Any]=8 , UpperCAmelCase__ : Optional[Any]=0.1 , UpperCAmelCase__ : str=0.002 , UpperCAmelCase__ : str=1 , UpperCAmelCase__ : Any=0 , UpperCAmelCase__ : Union[str, Any]=0 , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Optional[Any]=None , ): """simple docstring""" snake_case : Union[str, Any] = parent snake_case : Union[str, Any] = batch_size snake_case : Any = encoder_seq_length snake_case : str = decoder_seq_length # For common tests snake_case : Optional[int] = self.decoder_seq_length snake_case : Optional[Any] = is_training snake_case : List[Any] = use_attention_mask snake_case : Union[str, Any] = use_labels snake_case : Any = vocab_size snake_case : Optional[int] = hidden_size snake_case : List[str] = num_hidden_layers snake_case : Union[str, Any] = num_attention_heads snake_case : Any = d_ff snake_case : Any = relative_attention_num_buckets snake_case : Optional[Any] = dropout_rate snake_case : int = initializer_factor snake_case : Optional[Any] = eos_token_id snake_case : Dict = pad_token_id snake_case : Optional[Any] = decoder_start_token_id snake_case : Union[str, Any] = None snake_case : List[str] = decoder_layers def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" return TaConfig.from_pretrained('''google/umt5-base''' ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Any , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : List[Any]=None , ): """simple docstring""" if attention_mask is None: snake_case : Union[str, Any] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: snake_case : Any = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: snake_case : List[Any] = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=UpperCAmelCase__ ) if decoder_head_mask is None: snake_case : Tuple = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase__ ) if cross_attn_head_mask is None: snake_case : Union[str, Any] = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase__ ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input snake_case : List[str] = input_ids.clamp(self.pad_token_id + 1 ) snake_case : List[str] = decoder_input_ids.clamp(self.pad_token_id + 1 ) snake_case : str = self.get_config() snake_case : Tuple = config.num_attention_heads snake_case : List[Any] = self.prepare_inputs_dict(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) return config, input_dict def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : List[str] = self.prepare_config_and_inputs() return config, inputs_dict def lowerCAmelCase( self : Dict ): """simple docstring""" return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def lowerCAmelCase( self : Tuple ): """simple docstring""" return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def lowerCAmelCase( self : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Tuple , ): """simple docstring""" snake_case : str = UMTaModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : str = model( input_ids=UpperCAmelCase__ , decoder_input_ids=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , decoder_attention_mask=UpperCAmelCase__ , ) snake_case : int = model(input_ids=UpperCAmelCase__ , decoder_input_ids=UpperCAmelCase__ ) snake_case : int = result.last_hidden_state snake_case : Dict = result.past_key_values snake_case : Dict = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(UpperCAmelCase__ ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def lowerCAmelCase( self : int , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[str] , ): """simple docstring""" snake_case : int = UMTaModel(config=UpperCAmelCase__ ).get_decoder().to(UpperCAmelCase__ ).eval() # first forward pass snake_case : List[Any] = model(UpperCAmelCase__ , use_cache=UpperCAmelCase__ ) snake_case : List[Any] = model(UpperCAmelCase__ ) snake_case : Any = model(UpperCAmelCase__ , use_cache=UpperCAmelCase__ ) self.parent.assertTrue(len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) ) self.parent.assertTrue(len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) + 1 ) snake_case , snake_case : List[str] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case : Any = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and snake_case : Union[str, Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case : Any = model(UpperCAmelCase__ )['''last_hidden_state'''] snake_case : Optional[Any] = model(UpperCAmelCase__ , past_key_values=UpperCAmelCase__ )['''last_hidden_state'''] # select random slice snake_case : Tuple = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case : Union[str, Any] = output_from_no_past[:, -1, random_slice_idx].detach() snake_case : Tuple = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1e-3 ) ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : int , ): """simple docstring""" snake_case : int = UMTaModel(config=UpperCAmelCase__ ).to(UpperCAmelCase__ ).half().eval() snake_case : str = model(**UpperCAmelCase__ )['''last_hidden_state'''] self.parent.assertFalse(torch.isnan(UpperCAmelCase__ ).any().item() ) @require_torch class a_ ( a , a , a , unittest.TestCase ): A__ : str = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) A__ : str = (UMTaForConditionalGeneration,) if is_torch_available() else () A__ : Any = ( { 'conversational': UMTaForConditionalGeneration, 'feature-extraction': UMTaModel, 'summarization': UMTaForConditionalGeneration, 'text2text-generation': UMTaForConditionalGeneration, 'translation': UMTaForConditionalGeneration, 'question-answering': UMTaForQuestionAnswering, } if is_torch_available() else {} ) A__ : Dict = True A__ : List[str] = False A__ : Optional[int] = False A__ : Optional[int] = True A__ : List[str] = True # The small UMT5 model needs higher percentages for CPU/MP tests A__ : int = [0.8, 0.9] def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Union[str, Any] = UMTaModelTester(self ) @unittest.skip('''Test has a segmentation fault on torch 1.8.0''' ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Tuple = self.model_tester.prepare_config_and_inputs() snake_case : Optional[Any] = UMTaModel(config_and_inputs[0] ).to(UpperCAmelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( UpperCAmelCase__ , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"{tmpdirname}/t5_test.onnx" , export_params=UpperCAmelCase__ , opset_version=9 , input_names=['''input_ids''', '''decoder_input_ids'''] , ) @unittest.skipIf(torch_device == '''cpu''' , '''Cant do half precision''' ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*UpperCAmelCase__ ) def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : Optional[int] = ['''encoder_attentions''', '''decoder_attentions''', '''cross_attentions'''] snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() snake_case : int = config_and_inputs[0] snake_case : Union[str, Any] = UMTaForConditionalGeneration(UpperCAmelCase__ ).eval() model.to(UpperCAmelCase__ ) snake_case : str = { '''head_mask''': torch.zeros(config.num_layers , config.num_heads , device=UpperCAmelCase__ ), '''decoder_head_mask''': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ), '''cross_attn_head_mask''': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ), } for attn_name, (name, mask) in zip(UpperCAmelCase__ , head_masking.items() ): snake_case : int = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": snake_case : List[str] = torch.ones( config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ) snake_case : Union[str, Any] = model.generate( config_and_inputs[1]['''input_ids'''] , num_beams=1 , max_length=3 , output_attentions=UpperCAmelCase__ , return_dict_in_generate=UpperCAmelCase__ , **UpperCAmelCase__ , ) # We check the state of decoder_attentions and cross_attentions just from the last step snake_case : List[str] = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('''Does not work on the tiny model as we keep hitting edge cases.''' ) def lowerCAmelCase( self : Any ): """simple docstring""" pass @require_torch @require_sentencepiece @require_tokenizers class a_ ( unittest.TestCase ): @slow @unittest.skip( '''Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged''' ) def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Optional[Any] = UMTaForConditionalGeneration.from_pretrained('''google/umt5-small''' , return_dict=UpperCAmelCase__ ).to(UpperCAmelCase__ ) snake_case : int = AutoTokenizer.from_pretrained('''google/umt5-small''' , use_fast=UpperCAmelCase__ , legacy=UpperCAmelCase__ ) snake_case : List[str] = [ '''Bonjour monsieur <extra_id_0> bien <extra_id_1>.''', '''No se como puedo <extra_id_0>.''', '''This is the reason why we <extra_id_0> them.''', '''The <extra_id_0> walks in <extra_id_1>, seats''', '''A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.''', ] snake_case : Dict = tokenizer(UpperCAmelCase__ , return_tensors='''pt''' , padding=UpperCAmelCase__ ).input_ids # fmt: off snake_case : Optional[Any] = torch.tensor( [ [ 38_530, 210_703, 256_299, 1_410, 256_298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25_922, 256_299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1_460, 339, 312, 19_014, 10_620, 758, 256_299, 2_355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256_299, 14_869, 281, 301, 256_298, 275, 119_983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256_299, 14_869, 281, 2_234, 289, 2_275, 333,61_391, 289, 256_298, 543, 256_297, 168_714, 329, 256_296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : List[Any] = model.generate(input_ids.to(UpperCAmelCase__ ) ) snake_case : int = [ '''<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>''', '''<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', ] snake_case : Tuple = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ )
84
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class a_ ( a ): A__ : List[Any] = 'Salesforce/blip-image-captioning-base' A__ : Dict = ( 'This is a tool that generates a description of an image. It takes an input named `image` which should be the ' 'image to caption, and returns a text that contains the description in English.' ) A__ : str = 'image_captioner' A__ : Dict = AutoModelForVisionaSeq A__ : Optional[Any] = ['image'] A__ : List[str] = ['text'] def __init__( self : List[str] , *UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" requires_backends(self , ['''vision'''] ) super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : "Image" ): """simple docstring""" return self.pre_processor(images=UpperCAmelCase__ , return_tensors='''pt''' ) def lowerCAmelCase( self : Any , UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" return self.model.generate(**UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : List[Any] ): """simple docstring""" return self.pre_processor.batch_decode(UpperCAmelCase__ , skip_special_tokens=UpperCAmelCase__ )[0].strip()
84
1
import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def a_ ( __magic_name__ , __magic_name__ ) -> List[Any]: """simple docstring""" snake_case : Optional[int] = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg''' snake_case : Tuple = Image.open(requests.get(__magic_name__ , stream=__magic_name__ ).raw ).convert('''RGB''' ) snake_case : List[str] = transforms.Compose( [ transforms.Resize((image_size, image_size) , interpolation=InterpolationMode.BICUBIC ), transforms.ToTensor(), transforms.Normalize((0.4814_5466, 0.457_8275, 0.4082_1073) , (0.2686_2954, 0.2613_0258, 0.2757_7711) ), ] ) snake_case : Any = transform(__magic_name__ ).unsqueeze(0 ).to(__magic_name__ ) return image def a_ ( __magic_name__ ) -> Optional[int]: """simple docstring""" if "visual_encoder" in key: snake_case : Tuple = re.sub('''visual_encoder*''' , '''vision_model.encoder''' , __magic_name__ ) if "blocks" in key: snake_case : Dict = re.sub(R'''blocks''' , '''layers''' , __magic_name__ ) if "attn" in key: snake_case : int = re.sub(R'''attn''' , '''self_attn''' , __magic_name__ ) if "norm1" in key: snake_case : Dict = re.sub(R'''norm1''' , '''layer_norm1''' , __magic_name__ ) if "norm2" in key: snake_case : List[Any] = re.sub(R'''norm2''' , '''layer_norm2''' , __magic_name__ ) if "encoder.norm" in key: snake_case : Optional[Any] = re.sub(R'''encoder.norm''' , '''post_layernorm''' , __magic_name__ ) if "encoder.patch_embed.proj" in key: snake_case : str = re.sub(R'''encoder.patch_embed.proj''' , '''embeddings.patch_embedding''' , __magic_name__ ) if "encoder.pos_embed" in key: snake_case : Union[str, Any] = re.sub(R'''encoder.pos_embed''' , '''embeddings.position_embedding''' , __magic_name__ ) if "encoder.cls_token" in key: snake_case : Union[str, Any] = re.sub(R'''encoder.cls_token''' , '''embeddings.class_embedding''' , __magic_name__ ) if "self_attn" in key: snake_case : str = re.sub(R'''self_attn.proj''' , '''self_attn.projection''' , __magic_name__ ) return key @torch.no_grad() def a_ ( __magic_name__ , __magic_name__=None ) -> Dict: """simple docstring""" if config_path is not None: snake_case : Tuple = BlipConfig.from_pretrained(__magic_name__ ) else: snake_case : Optional[int] = BlipConfig(projection_dim=512 , text_config={} , vision_config={} ) snake_case : Union[str, Any] = BlipForConditionalGeneration(__magic_name__ ).eval() snake_case : List[Any] = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth''' snake_case : Optional[Any] = blip_decoder(pretrained=__magic_name__ , image_size=384 , vit='''base''' ) snake_case : Dict = pt_model.eval() snake_case : Union[str, Any] = pt_model.state_dict() for key in modified_state_dict.copy(): snake_case : Tuple = modified_state_dict.pop(__magic_name__ ) snake_case : Tuple = rename_key(__magic_name__ ) snake_case : Dict = value hf_model.load_state_dict(__magic_name__ ) snake_case : Union[str, Any] = 384 snake_case : str = load_demo_image(image_size=__magic_name__ , device='''cpu''' ) snake_case : List[str] = BertTokenizer.from_pretrained('''bert-base-uncased''' ) snake_case : Dict = tokenizer(['''a picture of'''] ).input_ids snake_case : Optional[Any] = hf_model.generate(__magic_name__ , __magic_name__ ) assert out[0].tolist() == [30_522, 1_037, 3_861, 1_997, 1_037, 2_450, 3_564, 2_006, 1_996, 3_509, 2_007, 2_014, 3_899, 102] snake_case : str = hf_model.generate(__magic_name__ ) assert out[0].tolist() == [30_522, 1_037, 2_450, 3_564, 2_006, 1_996, 3_509, 2_007, 2_014, 3_899, 102] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(__magic_name__ ) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' snake_case : Optional[int] = ( '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth''' ) snake_case : Optional[Any] = blip_vqa(pretrained=__magic_name__ , image_size=__magic_name__ , vit='''base''' ) vqa_model.eval() snake_case : int = vqa_model.state_dict() for key in modified_state_dict.copy(): snake_case : Tuple = modified_state_dict.pop(__magic_name__ ) snake_case : Union[str, Any] = rename_key(__magic_name__ ) snake_case : Union[str, Any] = value snake_case : Any = BlipForQuestionAnswering(__magic_name__ ) hf_vqa_model.load_state_dict(__magic_name__ ) snake_case : int = ['''How many dogs are in this image?'''] snake_case : Any = tokenizer(__magic_name__ , return_tensors='''pt''' ).input_ids snake_case : Dict = hf_vqa_model.generate(__magic_name__ , __magic_name__ ) print(tokenizer.decode(answer[0] ) ) assert tokenizer.decode(answer[0] ) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + '''_vqa''' ) snake_case : Optional[Any] = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth''' snake_case : Dict = blip_itm(pretrained=__magic_name__ , image_size=__magic_name__ , vit='''base''' ) itm_model.eval() snake_case : Union[str, Any] = itm_model.state_dict() for key in modified_state_dict.copy(): snake_case : Optional[Any] = modified_state_dict.pop(__magic_name__ ) snake_case : Optional[int] = rename_key(__magic_name__ ) snake_case : List[Any] = value snake_case : Optional[Any] = BlipForImageTextRetrieval(__magic_name__ ) snake_case : str = ['''A picture of a woman with a dog sitting in a beach'''] snake_case : List[str] = tokenizer( __magic_name__ , return_tensors='''pt''' , padding='''max_length''' , truncation=__magic_name__ , max_length=35 , ).input_ids hf_itm_model.load_state_dict(__magic_name__ ) hf_itm_model.eval() snake_case : Optional[Any] = hf_itm_model(__magic_name__ , __magic_name__ , use_itm_head=__magic_name__ ) snake_case : Optional[int] = hf_itm_model(__magic_name__ , __magic_name__ , use_itm_head=__magic_name__ ) assert out[0].item() == 0.2110_6874_9427_7954 assert torch.nn.functional.softmax(out_itm[0] , dim=1 )[:, 1].item() == 0.4_5698_8453_8650_5127 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + '''_itm''' ) if __name__ == "__main__": _a : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') _a : str = parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
84
def a_ ( __magic_name__ ) -> bool: """simple docstring""" if p < 2: raise ValueError('''p should not be less than 2!''' ) elif p == 2: return True snake_case : int = 4 snake_case : Optional[Any] = (1 << p) - 1 for _ in range(p - 2 ): snake_case : Optional[Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
84
1
import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy as np from .import_utils import is_flax_available, is_tf_available, is_torch_available, is_torch_fx_proxy if is_flax_available(): import jax.numpy as jnp class a_ ( a ): def __get__( self : List[str] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[str]=None ): """simple docstring""" # See docs.python.org/3/howto/descriptor.html#properties if obj is None: return self if self.fget is None: raise AttributeError('''unreadable attribute''' ) snake_case : Optional[int] = '''__cached_''' + self.fget.__name__ snake_case : str = getattr(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) if cached is None: snake_case : List[Any] = self.fget(UpperCAmelCase__ ) setattr(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) return cached def a_ ( __magic_name__ ) -> str: """simple docstring""" snake_case : str = val.lower() if val in {"y", "yes", "t", "true", "on", "1"}: return 1 if val in {"n", "no", "f", "false", "off", "0"}: return 0 raise ValueError(F"invalid truth value {val!r}" ) def a_ ( __magic_name__ ) -> List[str]: """simple docstring""" if is_torch_fx_proxy(__magic_name__ ): return True if is_torch_available(): import torch if isinstance(__magic_name__ , torch.Tensor ): return True if is_tf_available(): import tensorflow as tf if isinstance(__magic_name__ , tf.Tensor ): return True if is_flax_available(): import jax.numpy as jnp from jax.core import Tracer if isinstance(__magic_name__ , (jnp.ndarray, Tracer) ): return True return isinstance(__magic_name__ , np.ndarray ) def a_ ( __magic_name__ ) -> str: """simple docstring""" return isinstance(__magic_name__ , np.ndarray ) def a_ ( __magic_name__ ) -> Dict: """simple docstring""" return _is_numpy(__magic_name__ ) def a_ ( __magic_name__ ) -> Optional[Any]: """simple docstring""" import torch return isinstance(__magic_name__ , torch.Tensor ) def a_ ( __magic_name__ ) -> int: """simple docstring""" return False if not is_torch_available() else _is_torch(__magic_name__ ) def a_ ( __magic_name__ ) -> Any: """simple docstring""" import torch return isinstance(__magic_name__ , torch.device ) def a_ ( __magic_name__ ) -> Union[str, Any]: """simple docstring""" return False if not is_torch_available() else _is_torch_device(__magic_name__ ) def a_ ( __magic_name__ ) -> Optional[Any]: """simple docstring""" import torch if isinstance(__magic_name__ , __magic_name__ ): if hasattr(__magic_name__ , __magic_name__ ): snake_case : Dict = getattr(__magic_name__ , __magic_name__ ) else: return False return isinstance(__magic_name__ , torch.dtype ) def a_ ( __magic_name__ ) -> Optional[int]: """simple docstring""" return False if not is_torch_available() else _is_torch_dtype(__magic_name__ ) def a_ ( __magic_name__ ) -> Optional[Any]: """simple docstring""" import tensorflow as tf return isinstance(__magic_name__ , tf.Tensor ) def a_ ( __magic_name__ ) -> Dict: """simple docstring""" return False if not is_tf_available() else _is_tensorflow(__magic_name__ ) def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" import tensorflow as tf # the `is_symbolic_tensor` predicate is only available starting with TF 2.14 if hasattr(__magic_name__ , '''is_symbolic_tensor''' ): return tf.is_symbolic_tensor(__magic_name__ ) return type(__magic_name__ ) == tf.Tensor def a_ ( __magic_name__ ) -> Optional[Any]: """simple docstring""" return False if not is_tf_available() else _is_tf_symbolic_tensor(__magic_name__ ) def a_ ( __magic_name__ ) -> Optional[Any]: """simple docstring""" import jax.numpy as jnp # noqa: F811 return isinstance(__magic_name__ , jnp.ndarray ) def a_ ( __magic_name__ ) -> List[str]: """simple docstring""" return False if not is_flax_available() else _is_jax(__magic_name__ ) def a_ ( __magic_name__ ) -> Optional[Any]: """simple docstring""" if isinstance(__magic_name__ , (dict, UserDict) ): return {k: to_py_obj(__magic_name__ ) for k, v in obj.items()} elif isinstance(__magic_name__ , (list, tuple) ): return [to_py_obj(__magic_name__ ) for o in obj] elif is_tf_tensor(__magic_name__ ): return obj.numpy().tolist() elif is_torch_tensor(__magic_name__ ): return obj.detach().cpu().tolist() elif is_jax_tensor(__magic_name__ ): return np.asarray(__magic_name__ ).tolist() elif isinstance(__magic_name__ , (np.ndarray, np.number) ): # tolist also works on 0d np arrays return obj.tolist() else: return obj def a_ ( __magic_name__ ) -> Dict: """simple docstring""" if isinstance(__magic_name__ , (dict, UserDict) ): return {k: to_numpy(__magic_name__ ) for k, v in obj.items()} elif isinstance(__magic_name__ , (list, tuple) ): return np.array(__magic_name__ ) elif is_tf_tensor(__magic_name__ ): return obj.numpy() elif is_torch_tensor(__magic_name__ ): return obj.detach().cpu().numpy() elif is_jax_tensor(__magic_name__ ): return np.asarray(__magic_name__ ) else: return obj class a_ ( a ): def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : List[str] = fields(self ) # Safety and consistency checks if not len(UpperCAmelCase__ ): raise ValueError(F"{self.__class__.__name__} has no fields." ) if not all(field.default is None for field in class_fields[1:] ): raise ValueError(F"{self.__class__.__name__} should not have more than one required field." ) snake_case : str = getattr(self , class_fields[0].name ) snake_case : str = all(getattr(self , field.name ) is None for field in class_fields[1:] ) if other_fields_are_none and not is_tensor(UpperCAmelCase__ ): if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Tuple = first_field.items() snake_case : List[str] = True else: try: snake_case : Optional[int] = iter(UpperCAmelCase__ ) snake_case : Optional[Any] = True except TypeError: snake_case : List[Any] = False # if we provided an iterator as first field and the iterator is a (key, value) iterator # set the associated fields if first_field_iterator: for idx, element in enumerate(UpperCAmelCase__ ): if ( not isinstance(UpperCAmelCase__ , (list, tuple) ) or not len(UpperCAmelCase__ ) == 2 or not isinstance(element[0] , UpperCAmelCase__ ) ): if idx == 0: # If we do not have an iterator of key/values, set it as attribute snake_case : List[Any] = first_field else: # If we have a mixed iterator, raise an error raise ValueError( F"Cannot set key/value for {element}. It needs to be a tuple (key, value)." ) break setattr(self , element[0] , element[1] ) if element[1] is not None: snake_case : Union[str, Any] = element[1] elif first_field is not None: snake_case : Any = first_field else: for field in class_fields: snake_case : int = getattr(self , field.name ) if v is not None: snake_case : Tuple = v def __delitem__( self : Tuple , *UpperCAmelCase__ : Any , **UpperCAmelCase__ : Tuple ): """simple docstring""" raise Exception(F"You cannot use ``__delitem__`` on a {self.__class__.__name__} instance." ) def lowerCAmelCase( self : Optional[Any] , *UpperCAmelCase__ : Any , **UpperCAmelCase__ : List[Any] ): """simple docstring""" raise Exception(F"You cannot use ``setdefault`` on a {self.__class__.__name__} instance." ) def lowerCAmelCase( self : int , *UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : Tuple ): """simple docstring""" raise Exception(F"You cannot use ``pop`` on a {self.__class__.__name__} instance." ) def lowerCAmelCase( self : Optional[int] , *UpperCAmelCase__ : Tuple , **UpperCAmelCase__ : int ): """simple docstring""" raise Exception(F"You cannot use ``update`` on a {self.__class__.__name__} instance." ) def __getitem__( self : Optional[Any] , UpperCAmelCase__ : Tuple ): """simple docstring""" if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Optional[Any] = dict(self.items() ) return inner_dict[k] else: return self.to_tuple()[k] def __setattr__( self : str , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : str ): """simple docstring""" if name in self.keys() and value is not None: # Don't call self.__setitem__ to avoid recursion errors super().__setitem__(UpperCAmelCase__ , UpperCAmelCase__ ) super().__setattr__(UpperCAmelCase__ , UpperCAmelCase__ ) def __setitem__( self : str , UpperCAmelCase__ : Any , UpperCAmelCase__ : List[Any] ): """simple docstring""" # Will raise a KeyException if needed super().__setitem__(UpperCAmelCase__ , UpperCAmelCase__ ) # Don't call self.__setattr__ to avoid recursion errors super().__setattr__(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" return tuple(self[k] for k in self.keys() ) class a_ ( a , a ): @classmethod def lowerCAmelCase( cls : str , UpperCAmelCase__ : str ): """simple docstring""" raise ValueError( F"{value} is not a valid {cls.__name__}, please select one of {list(cls._valueamember_map_.keys() )}" ) class a_ ( a ): A__ : Optional[Any] = 'longest' A__ : Tuple = 'max_length' A__ : Any = 'do_not_pad' class a_ ( a ): A__ : Optional[int] = 'pt' A__ : List[Any] = 'tf' A__ : List[str] = 'np' A__ : Optional[Any] = 'jax' class a_ : def __init__( self : List[str] , UpperCAmelCase__ : List[ContextManager] ): """simple docstring""" snake_case : Optional[Any] = context_managers snake_case : Tuple = ExitStack() def __enter__( self : Tuple ): """simple docstring""" for context_manager in self.context_managers: self.stack.enter_context(UpperCAmelCase__ ) def __exit__( self : int , *UpperCAmelCase__ : List[Any] , **UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" self.stack.__exit__(*UpperCAmelCase__ , **UpperCAmelCase__ ) def a_ ( __magic_name__ ) -> int: """simple docstring""" snake_case : Any = infer_framework(__magic_name__ ) if framework == "tf": snake_case : str = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": snake_case : int = inspect.signature(model_class.forward ) # PyTorch models else: snake_case : int = inspect.signature(model_class.__call__ ) # Flax models for p in signature.parameters: if p == "return_loss" and signature.parameters[p].default is True: return True return False def a_ ( __magic_name__ ) -> str: """simple docstring""" snake_case : Optional[Any] = model_class.__name__ snake_case : Any = infer_framework(__magic_name__ ) if framework == "tf": snake_case : Dict = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": snake_case : Optional[int] = inspect.signature(model_class.forward ) # PyTorch models else: snake_case : List[str] = inspect.signature(model_class.__call__ ) # Flax models if "QuestionAnswering" in model_name: return [p for p in signature.parameters if "label" in p or p in ("start_positions", "end_positions")] else: return [p for p in signature.parameters if "label" in p] def a_ ( __magic_name__ , __magic_name__ = "" , __magic_name__ = "." ) -> Tuple: """simple docstring""" def _flatten_dict(__magic_name__ , __magic_name__="" , __magic_name__="." ): for k, v in d.items(): snake_case : Dict = str(__magic_name__ ) + delimiter + str(__magic_name__ ) if parent_key else k if v and isinstance(__magic_name__ , __magic_name__ ): yield from flatten_dict(__magic_name__ , __magic_name__ , delimiter=__magic_name__ ).items() else: yield key, v return dict(_flatten_dict(__magic_name__ , __magic_name__ , __magic_name__ ) ) @contextmanager def a_ ( __magic_name__ , __magic_name__ = False ) -> Tuple: """simple docstring""" if use_temp_dir: with tempfile.TemporaryDirectory() as tmp_dir: yield tmp_dir else: yield working_dir def a_ ( __magic_name__ , __magic_name__=None ) -> Tuple: """simple docstring""" if is_numpy_array(__magic_name__ ): return np.transpose(__magic_name__ , axes=__magic_name__ ) elif is_torch_tensor(__magic_name__ ): return array.T if axes is None else array.permute(*__magic_name__ ) elif is_tf_tensor(__magic_name__ ): import tensorflow as tf return tf.transpose(__magic_name__ , perm=__magic_name__ ) elif is_jax_tensor(__magic_name__ ): return jnp.transpose(__magic_name__ , axes=__magic_name__ ) else: raise ValueError(F"Type not supported for transpose: {type(__magic_name__ )}." ) def a_ ( __magic_name__ , __magic_name__ ) -> List[Any]: """simple docstring""" if is_numpy_array(__magic_name__ ): return np.reshape(__magic_name__ , __magic_name__ ) elif is_torch_tensor(__magic_name__ ): return array.reshape(*__magic_name__ ) elif is_tf_tensor(__magic_name__ ): import tensorflow as tf return tf.reshape(__magic_name__ , __magic_name__ ) elif is_jax_tensor(__magic_name__ ): return jnp.reshape(__magic_name__ , __magic_name__ ) else: raise ValueError(F"Type not supported for reshape: {type(__magic_name__ )}." ) def a_ ( __magic_name__ , __magic_name__=None ) -> Optional[int]: """simple docstring""" if is_numpy_array(__magic_name__ ): return np.squeeze(__magic_name__ , axis=__magic_name__ ) elif is_torch_tensor(__magic_name__ ): return array.squeeze() if axis is None else array.squeeze(dim=__magic_name__ ) elif is_tf_tensor(__magic_name__ ): import tensorflow as tf return tf.squeeze(__magic_name__ , axis=__magic_name__ ) elif is_jax_tensor(__magic_name__ ): return jnp.squeeze(__magic_name__ , axis=__magic_name__ ) else: raise ValueError(F"Type not supported for squeeze: {type(__magic_name__ )}." ) def a_ ( __magic_name__ , __magic_name__ ) -> Dict: """simple docstring""" if is_numpy_array(__magic_name__ ): return np.expand_dims(__magic_name__ , __magic_name__ ) elif is_torch_tensor(__magic_name__ ): return array.unsqueeze(dim=__magic_name__ ) elif is_tf_tensor(__magic_name__ ): import tensorflow as tf return tf.expand_dims(__magic_name__ , axis=__magic_name__ ) elif is_jax_tensor(__magic_name__ ): return jnp.expand_dims(__magic_name__ , axis=__magic_name__ ) else: raise ValueError(F"Type not supported for expand_dims: {type(__magic_name__ )}." ) def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" if is_numpy_array(__magic_name__ ): return np.size(__magic_name__ ) elif is_torch_tensor(__magic_name__ ): return array.numel() elif is_tf_tensor(__magic_name__ ): import tensorflow as tf return tf.size(__magic_name__ ) elif is_jax_tensor(__magic_name__ ): return array.size else: raise ValueError(F"Type not supported for expand_dims: {type(__magic_name__ )}." ) def a_ ( __magic_name__ , __magic_name__ ) -> Tuple: """simple docstring""" for key, value in auto_map.items(): if isinstance(__magic_name__ , (tuple, list) ): snake_case : int = [F"{repo_id}--{v}" if (v is not None and '''--''' not in v) else v for v in value] elif value is not None and "--" not in value: snake_case : int = F"{repo_id}--{value}" return auto_map def a_ ( __magic_name__ ) -> Optional[Any]: """simple docstring""" for base_class in inspect.getmro(__magic_name__ ): snake_case : str = base_class.__module__ snake_case : Union[str, Any] = base_class.__name__ if module.startswith('''tensorflow''' ) or module.startswith('''keras''' ) or name == "TFPreTrainedModel": return "tf" elif module.startswith('''torch''' ) or name == "PreTrainedModel": return "pt" elif module.startswith('''flax''' ) or module.startswith('''jax''' ) or name == "FlaxPreTrainedModel": return "flax" else: raise TypeError(F"Could not infer framework from class {model_class}." )
84
from sklearn.metrics import fa_score import datasets _a : List[str] = '\nThe F1 score is the harmonic mean of the precision and recall. It can be computed with the equation:\nF1 = 2 * (precision * recall) / (precision + recall)\n' _a : Dict = '\nArgs:\n predictions (`list` of `int`): Predicted labels.\n references (`list` of `int`): Ground truth labels.\n labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None.\n pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1.\n average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`.\n\n - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary.\n - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives.\n - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.\n - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall.\n - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification).\n sample_weight (`list` of `float`): Sample weights Defaults to None.\n\nReturns:\n f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better.\n\nExamples:\n\n Example 1-A simple binary example\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0])\n >>> print(results)\n {\'f1\': 0.5}\n\n Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0)\n >>> print(round(results[\'f1\'], 2))\n 0.67\n\n Example 3-The same simple binary example as in Example 1, but with `sample_weight` included.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3])\n >>> print(round(results[\'f1\'], 2))\n 0.35\n\n Example 4-A multiclass example, with different values for the `average` input.\n >>> predictions = [0, 2, 1, 0, 0, 1]\n >>> references = [0, 1, 2, 0, 1, 2]\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro")\n >>> print(round(results[\'f1\'], 2))\n 0.33\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=None)\n >>> print(results)\n {\'f1\': array([0.8, 0. , 0. ])}\n' _a : List[Any] = '\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P.\n and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and\n Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.},\n journal={Journal of Machine Learning Research},\n volume={12},\n pages={2825--2830},\n year={2011}\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): def lowerCAmelCase( self : Any ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''int32''' ) ), '''references''': datasets.Sequence(datasets.Value('''int32''' ) ), } if self.config_name == '''multilabel''' else { '''predictions''': datasets.Value('''int32''' ), '''references''': datasets.Value('''int32''' ), } ) , reference_urls=['''https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html'''] , ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Optional[Any]=1 , UpperCAmelCase__ : List[str]="binary" , UpperCAmelCase__ : str=None ): """simple docstring""" snake_case : List[Any] = fa_score( UpperCAmelCase__ , UpperCAmelCase__ , labels=UpperCAmelCase__ , pos_label=UpperCAmelCase__ , average=UpperCAmelCase__ , sample_weight=UpperCAmelCase__ ) return {"f1": float(UpperCAmelCase__ ) if score.size == 1 else score}
84
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _a : Optional[Any] = logging.get_logger(__name__) _a : Optional[int] = { 'facebook/nllb-moe-54B': 'https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json', } class a_ ( a ): A__ : str = 'nllb-moe' A__ : Any = ['past_key_values'] A__ : Optional[int] = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self : List[str] , UpperCAmelCase__ : Dict=128_112 , UpperCAmelCase__ : Any=1_024 , UpperCAmelCase__ : str=12 , UpperCAmelCase__ : Union[str, Any]=4_096 , UpperCAmelCase__ : Tuple=16 , UpperCAmelCase__ : Union[str, Any]=12 , UpperCAmelCase__ : int=4_096 , UpperCAmelCase__ : Optional[Any]=16 , UpperCAmelCase__ : List[Any]=0.05 , UpperCAmelCase__ : Dict=0.05 , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : List[Any]="relu" , UpperCAmelCase__ : Union[str, Any]=1_024 , UpperCAmelCase__ : int=0.1 , UpperCAmelCase__ : Tuple=0.1 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : str=0.02 , UpperCAmelCase__ : Dict=2 , UpperCAmelCase__ : List[str]=True , UpperCAmelCase__ : Any=False , UpperCAmelCase__ : Union[str, Any]="float32" , UpperCAmelCase__ : int=False , UpperCAmelCase__ : List[Any]=128 , UpperCAmelCase__ : Optional[int]=64 , UpperCAmelCase__ : Optional[Any]=4 , UpperCAmelCase__ : List[str]=4 , UpperCAmelCase__ : Optional[Any]=0.001 , UpperCAmelCase__ : Tuple=0.001 , UpperCAmelCase__ : Tuple="all" , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : int=False , UpperCAmelCase__ : List[Any]=1.0 , UpperCAmelCase__ : List[str]=0.2 , UpperCAmelCase__ : List[str]=1 , UpperCAmelCase__ : Tuple=0 , UpperCAmelCase__ : List[str]=2 , UpperCAmelCase__ : Tuple=False , **UpperCAmelCase__ : Dict , ): """simple docstring""" snake_case : int = vocab_size snake_case : List[str] = max_position_embeddings snake_case : List[Any] = d_model snake_case : List[Any] = encoder_ffn_dim snake_case : Optional[Any] = encoder_layers snake_case : List[Any] = encoder_attention_heads snake_case : Union[str, Any] = decoder_ffn_dim snake_case : str = decoder_layers snake_case : Optional[int] = decoder_attention_heads snake_case : str = dropout snake_case : Any = attention_dropout snake_case : Any = activation_dropout snake_case : List[Any] = activation_function snake_case : Any = init_std snake_case : Optional[Any] = encoder_layerdrop snake_case : Dict = decoder_layerdrop snake_case : Any = use_cache snake_case : Optional[int] = encoder_layers snake_case : Union[str, Any] = scale_embedding # scale factor will be sqrt(d_model) if True snake_case : List[str] = router_z_loss_coef snake_case : str = router_aux_loss_coef snake_case : Union[str, Any] = decoder_sparse_step snake_case : Optional[int] = encoder_sparse_step snake_case : int = num_experts snake_case : int = expert_capacity snake_case : int = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}" ) snake_case : Union[str, Any] = router_dtype snake_case : Optional[Any] = router_ignore_padding_tokens snake_case : str = batch_prioritized_routing snake_case : Optional[Any] = second_expert_policy snake_case : Dict = normalize_router_prob_before_dropping snake_case : Optional[int] = moe_eval_capacity_token_fraction snake_case : Tuple = moe_token_dropout snake_case : List[Any] = output_router_logits super().__init__( pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , is_encoder_decoder=UpperCAmelCase__ , decoder_start_token_id=UpperCAmelCase__ , **UpperCAmelCase__ , )
84
def a_ ( __magic_name__ ) -> int: """simple docstring""" if not isinstance(__magic_name__ , __magic_name__ ): raise TypeError('''only integers accepted as input''' ) else: snake_case : str = str(abs(__magic_name__ ) ) snake_case : Optional[Any] = [list(__magic_name__ ) for char in range(len(__magic_name__ ) )] for index in range(len(__magic_name__ ) ): num_transpositions[index].pop(__magic_name__ ) return max( int(''''''.join(list(__magic_name__ ) ) ) for transposition in num_transpositions ) if __name__ == "__main__": __import__('doctest').testmod()
84
1
import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class a_ : def __init__( self : Dict , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : str=13 , UpperCAmelCase__ : List[Any]=7 , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : Any=99 , UpperCAmelCase__ : List[Any]=64 , UpperCAmelCase__ : Optional[Any]=5 , UpperCAmelCase__ : str=4 , UpperCAmelCase__ : Union[str, Any]=64 , UpperCAmelCase__ : Dict="gelu" , UpperCAmelCase__ : Dict=0.1 , UpperCAmelCase__ : int=0.1 , UpperCAmelCase__ : Any=512 , UpperCAmelCase__ : int=16 , UpperCAmelCase__ : Optional[int]=2 , UpperCAmelCase__ : Tuple=0.02 , UpperCAmelCase__ : Optional[Any]=3 , UpperCAmelCase__ : List[Any]=4 , UpperCAmelCase__ : str=None , ): """simple docstring""" snake_case : Dict = parent snake_case : List[Any] = batch_size snake_case : Tuple = seq_length snake_case : Union[str, Any] = is_training snake_case : Dict = use_input_mask snake_case : List[str] = use_token_type_ids snake_case : Dict = use_labels snake_case : str = vocab_size snake_case : Tuple = hidden_size snake_case : str = num_hidden_layers snake_case : str = num_attention_heads snake_case : int = intermediate_size snake_case : List[str] = hidden_act snake_case : List[Any] = hidden_dropout_prob snake_case : Tuple = attention_probs_dropout_prob snake_case : Tuple = max_position_embeddings snake_case : Tuple = type_vocab_size snake_case : List[str] = type_sequence_label_size snake_case : Optional[int] = initializer_range snake_case : Dict = num_labels snake_case : Any = num_choices snake_case : Dict = scope def lowerCAmelCase( self : Optional[int] ): """simple docstring""" return MPNetConfig.from_pretrained('''microsoft/mpnet-base''' ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case : List[str] = None if self.use_input_mask: snake_case : Dict = random_attention_mask([self.batch_size, self.seq_length] ) snake_case : Tuple = None snake_case : int = None snake_case : str = None if self.use_labels: snake_case : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case : int = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case : Dict = ids_tensor([self.batch_size] , self.num_choices ) snake_case : Union[str, Any] = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCAmelCase( self : str ): """simple docstring""" return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def lowerCAmelCase( self : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : str = MPNetModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : Any = model(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : Dict = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Dict ): """simple docstring""" snake_case : Optional[int] = MPNetForQuestionAnswering(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : str = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , start_positions=UpperCAmelCase__ , end_positions=UpperCAmelCase__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCAmelCase( self : Any , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : str ): """simple docstring""" snake_case : str = self.num_labels snake_case : int = MPNetForSequenceClassification(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : List[str] = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Dict ): """simple docstring""" snake_case : str = self.num_choices snake_case : Optional[Any] = MPNetForMultipleChoice(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : Union[str, Any] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case : Optional[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case : int = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , labels=UpperCAmelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : int ): """simple docstring""" snake_case : Optional[int] = self.num_labels snake_case : List[Any] = MPNetForTokenClassification(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : Dict = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : Optional[Any] = self.prepare_config_and_inputs() ((snake_case) , (snake_case) , (snake_case) , (snake_case) , (snake_case) , (snake_case)) : List[Any] = config_and_inputs snake_case : Tuple = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class a_ ( a , a , unittest.TestCase ): A__ : Optional[int] = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) A__ : Optional[Any] = ( { 'feature-extraction': MPNetModel, 'fill-mask': MPNetForMaskedLM, 'question-answering': MPNetForQuestionAnswering, 'text-classification': MPNetForSequenceClassification, 'token-classification': MPNetForTokenClassification, 'zero-shot': MPNetForSequenceClassification, } if is_torch_available() else {} ) A__ : int = False A__ : Optional[int] = True def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : Dict = MPNetModelTester(self ) snake_case : Optional[int] = ConfigTester(self , config_class=UpperCAmelCase__ , hidden_size=37 ) def lowerCAmelCase( self : int ): """simple docstring""" self.config_tester.run_common_tests() def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*UpperCAmelCase__ ) def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*UpperCAmelCase__ ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*UpperCAmelCase__ ) def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*UpperCAmelCase__ ) @require_torch class a_ ( unittest.TestCase ): @slow def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : Dict = MPNetModel.from_pretrained('''microsoft/mpnet-base''' ) snake_case : List[Any] = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) snake_case : List[str] = model(UpperCAmelCase__ )[0] snake_case : List[Any] = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , UpperCAmelCase__ ) snake_case : List[Any] = torch.tensor( [[[-0.0550, 0.1943, -0.0740], [-0.0562, 0.2211, -0.0579], [-0.0437, 0.3337, -0.0641]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase__ , atol=1e-4 ) )
84
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class a_ : def __init__( self : Union[str, Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any]=99 , UpperCAmelCase__ : Dict=13 , UpperCAmelCase__ : int=7 , UpperCAmelCase__ : Any=9 , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Tuple=False , UpperCAmelCase__ : Tuple=32 , UpperCAmelCase__ : Dict=5 , UpperCAmelCase__ : Optional[int]=4 , UpperCAmelCase__ : List[Any]=37 , UpperCAmelCase__ : Union[str, Any]=8 , UpperCAmelCase__ : Optional[Any]=0.1 , UpperCAmelCase__ : str=0.002 , UpperCAmelCase__ : str=1 , UpperCAmelCase__ : Any=0 , UpperCAmelCase__ : Union[str, Any]=0 , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Optional[Any]=None , ): """simple docstring""" snake_case : Union[str, Any] = parent snake_case : Union[str, Any] = batch_size snake_case : Any = encoder_seq_length snake_case : str = decoder_seq_length # For common tests snake_case : Optional[int] = self.decoder_seq_length snake_case : Optional[Any] = is_training snake_case : List[Any] = use_attention_mask snake_case : Union[str, Any] = use_labels snake_case : Any = vocab_size snake_case : Optional[int] = hidden_size snake_case : List[str] = num_hidden_layers snake_case : Union[str, Any] = num_attention_heads snake_case : Any = d_ff snake_case : Any = relative_attention_num_buckets snake_case : Optional[Any] = dropout_rate snake_case : int = initializer_factor snake_case : Optional[Any] = eos_token_id snake_case : Dict = pad_token_id snake_case : Optional[Any] = decoder_start_token_id snake_case : Union[str, Any] = None snake_case : List[str] = decoder_layers def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" return TaConfig.from_pretrained('''google/umt5-base''' ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Any , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : List[Any]=None , ): """simple docstring""" if attention_mask is None: snake_case : Union[str, Any] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: snake_case : Any = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: snake_case : List[Any] = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=UpperCAmelCase__ ) if decoder_head_mask is None: snake_case : Tuple = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase__ ) if cross_attn_head_mask is None: snake_case : Union[str, Any] = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase__ ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input snake_case : List[str] = input_ids.clamp(self.pad_token_id + 1 ) snake_case : List[str] = decoder_input_ids.clamp(self.pad_token_id + 1 ) snake_case : str = self.get_config() snake_case : Tuple = config.num_attention_heads snake_case : List[Any] = self.prepare_inputs_dict(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) return config, input_dict def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : List[str] = self.prepare_config_and_inputs() return config, inputs_dict def lowerCAmelCase( self : Dict ): """simple docstring""" return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def lowerCAmelCase( self : Tuple ): """simple docstring""" return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def lowerCAmelCase( self : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Tuple , ): """simple docstring""" snake_case : str = UMTaModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : str = model( input_ids=UpperCAmelCase__ , decoder_input_ids=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , decoder_attention_mask=UpperCAmelCase__ , ) snake_case : int = model(input_ids=UpperCAmelCase__ , decoder_input_ids=UpperCAmelCase__ ) snake_case : int = result.last_hidden_state snake_case : Dict = result.past_key_values snake_case : Dict = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(UpperCAmelCase__ ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def lowerCAmelCase( self : int , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[str] , ): """simple docstring""" snake_case : int = UMTaModel(config=UpperCAmelCase__ ).get_decoder().to(UpperCAmelCase__ ).eval() # first forward pass snake_case : List[Any] = model(UpperCAmelCase__ , use_cache=UpperCAmelCase__ ) snake_case : List[Any] = model(UpperCAmelCase__ ) snake_case : Any = model(UpperCAmelCase__ , use_cache=UpperCAmelCase__ ) self.parent.assertTrue(len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) ) self.parent.assertTrue(len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) + 1 ) snake_case , snake_case : List[str] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case : Any = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and snake_case : Union[str, Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case : Any = model(UpperCAmelCase__ )['''last_hidden_state'''] snake_case : Optional[Any] = model(UpperCAmelCase__ , past_key_values=UpperCAmelCase__ )['''last_hidden_state'''] # select random slice snake_case : Tuple = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case : Union[str, Any] = output_from_no_past[:, -1, random_slice_idx].detach() snake_case : Tuple = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1e-3 ) ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : int , ): """simple docstring""" snake_case : int = UMTaModel(config=UpperCAmelCase__ ).to(UpperCAmelCase__ ).half().eval() snake_case : str = model(**UpperCAmelCase__ )['''last_hidden_state'''] self.parent.assertFalse(torch.isnan(UpperCAmelCase__ ).any().item() ) @require_torch class a_ ( a , a , a , unittest.TestCase ): A__ : str = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) A__ : str = (UMTaForConditionalGeneration,) if is_torch_available() else () A__ : Any = ( { 'conversational': UMTaForConditionalGeneration, 'feature-extraction': UMTaModel, 'summarization': UMTaForConditionalGeneration, 'text2text-generation': UMTaForConditionalGeneration, 'translation': UMTaForConditionalGeneration, 'question-answering': UMTaForQuestionAnswering, } if is_torch_available() else {} ) A__ : Dict = True A__ : List[str] = False A__ : Optional[int] = False A__ : Optional[int] = True A__ : List[str] = True # The small UMT5 model needs higher percentages for CPU/MP tests A__ : int = [0.8, 0.9] def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Union[str, Any] = UMTaModelTester(self ) @unittest.skip('''Test has a segmentation fault on torch 1.8.0''' ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Tuple = self.model_tester.prepare_config_and_inputs() snake_case : Optional[Any] = UMTaModel(config_and_inputs[0] ).to(UpperCAmelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( UpperCAmelCase__ , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"{tmpdirname}/t5_test.onnx" , export_params=UpperCAmelCase__ , opset_version=9 , input_names=['''input_ids''', '''decoder_input_ids'''] , ) @unittest.skipIf(torch_device == '''cpu''' , '''Cant do half precision''' ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*UpperCAmelCase__ ) def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : Optional[int] = ['''encoder_attentions''', '''decoder_attentions''', '''cross_attentions'''] snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() snake_case : int = config_and_inputs[0] snake_case : Union[str, Any] = UMTaForConditionalGeneration(UpperCAmelCase__ ).eval() model.to(UpperCAmelCase__ ) snake_case : str = { '''head_mask''': torch.zeros(config.num_layers , config.num_heads , device=UpperCAmelCase__ ), '''decoder_head_mask''': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ), '''cross_attn_head_mask''': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ), } for attn_name, (name, mask) in zip(UpperCAmelCase__ , head_masking.items() ): snake_case : int = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": snake_case : List[str] = torch.ones( config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ) snake_case : Union[str, Any] = model.generate( config_and_inputs[1]['''input_ids'''] , num_beams=1 , max_length=3 , output_attentions=UpperCAmelCase__ , return_dict_in_generate=UpperCAmelCase__ , **UpperCAmelCase__ , ) # We check the state of decoder_attentions and cross_attentions just from the last step snake_case : List[str] = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('''Does not work on the tiny model as we keep hitting edge cases.''' ) def lowerCAmelCase( self : Any ): """simple docstring""" pass @require_torch @require_sentencepiece @require_tokenizers class a_ ( unittest.TestCase ): @slow @unittest.skip( '''Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged''' ) def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Optional[Any] = UMTaForConditionalGeneration.from_pretrained('''google/umt5-small''' , return_dict=UpperCAmelCase__ ).to(UpperCAmelCase__ ) snake_case : int = AutoTokenizer.from_pretrained('''google/umt5-small''' , use_fast=UpperCAmelCase__ , legacy=UpperCAmelCase__ ) snake_case : List[str] = [ '''Bonjour monsieur <extra_id_0> bien <extra_id_1>.''', '''No se como puedo <extra_id_0>.''', '''This is the reason why we <extra_id_0> them.''', '''The <extra_id_0> walks in <extra_id_1>, seats''', '''A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.''', ] snake_case : Dict = tokenizer(UpperCAmelCase__ , return_tensors='''pt''' , padding=UpperCAmelCase__ ).input_ids # fmt: off snake_case : Optional[Any] = torch.tensor( [ [ 38_530, 210_703, 256_299, 1_410, 256_298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25_922, 256_299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1_460, 339, 312, 19_014, 10_620, 758, 256_299, 2_355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256_299, 14_869, 281, 301, 256_298, 275, 119_983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256_299, 14_869, 281, 2_234, 289, 2_275, 333,61_391, 289, 256_298, 543, 256_297, 168_714, 329, 256_296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : List[Any] = model.generate(input_ids.to(UpperCAmelCase__ ) ) snake_case : int = [ '''<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>''', '''<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', ] snake_case : Tuple = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ )
84
1
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _a : Any = logging.get_logger(__name__) _a : Tuple = { 'Salesforce/blip-vqa-base': 'https://huggingface.co/Salesforce/blip-vqa-base/resolve/main/config.json', 'Salesforce/blip-vqa-capfit-large': ( 'https://huggingface.co/Salesforce/blip-vqa-base-capfit/resolve/main/config.json' ), 'Salesforce/blip-image-captioning-base': ( 'https://huggingface.co/Salesforce/blip-image-captioning-base/resolve/main/config.json' ), 'Salesforce/blip-image-captioning-large': ( 'https://huggingface.co/Salesforce/blip-image-captioning-large/resolve/main/config.json' ), 'Salesforce/blip-itm-base-coco': 'https://huggingface.co/Salesforce/blip-itm-base-coco/resolve/main/config.json', 'Salesforce/blip-itm-large-coco': 'https://huggingface.co/Salesforce/blip-itm-large-coco/resolve/main/config.json', 'Salesforce/blip-itm-base-flikr': 'https://huggingface.co/Salesforce/blip-itm-base-flikr/resolve/main/config.json', 'Salesforce/blip-itm-large-flikr': ( 'https://huggingface.co/Salesforce/blip-itm-large-flikr/resolve/main/config.json' ), } class a_ ( a ): A__ : Optional[int] = 'blip_text_model' def __init__( self : str , UpperCAmelCase__ : Optional[Any]=30_524 , UpperCAmelCase__ : Tuple=768 , UpperCAmelCase__ : int=768 , UpperCAmelCase__ : Optional[int]=3_072 , UpperCAmelCase__ : Optional[int]=768 , UpperCAmelCase__ : str=12 , UpperCAmelCase__ : Optional[int]=8 , UpperCAmelCase__ : Tuple=512 , UpperCAmelCase__ : Optional[Any]="gelu" , UpperCAmelCase__ : Dict=1e-1_2 , UpperCAmelCase__ : Any=0.0 , UpperCAmelCase__ : Optional[Any]=0.0 , UpperCAmelCase__ : Optional[Any]=0.02 , UpperCAmelCase__ : Optional[int]=30_522 , UpperCAmelCase__ : List[Any]=2 , UpperCAmelCase__ : Dict=0 , UpperCAmelCase__ : Union[str, Any]=102 , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : Union[str, Any]=True , **UpperCAmelCase__ : Tuple , ): """simple docstring""" super().__init__( pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , sep_token_id=UpperCAmelCase__ , **UpperCAmelCase__ , ) snake_case : str = vocab_size snake_case : List[Any] = hidden_size snake_case : int = encoder_hidden_size snake_case : Optional[Any] = intermediate_size snake_case : Optional[int] = projection_dim snake_case : str = hidden_dropout_prob snake_case : Dict = num_hidden_layers snake_case : Dict = num_attention_heads snake_case : Tuple = max_position_embeddings snake_case : Optional[int] = layer_norm_eps snake_case : Optional[Any] = hidden_act snake_case : Optional[int] = initializer_range snake_case : Optional[Any] = attention_probs_dropout_prob snake_case : List[Any] = is_decoder snake_case : Tuple = use_cache @classmethod def lowerCAmelCase( cls : Dict , UpperCAmelCase__ : Union[str, os.PathLike] , **UpperCAmelCase__ : str ): """simple docstring""" cls._set_token_in_kwargs(UpperCAmelCase__ ) snake_case , snake_case : int = cls.get_config_dict(UpperCAmelCase__ , **UpperCAmelCase__ ) # get the text config dict if we are loading from BlipConfig if config_dict.get('''model_type''' ) == "blip": snake_case : Optional[Any] = config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"You are using a model of type {config_dict['model_type']} to instantiate a model of type " F"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(UpperCAmelCase__ , **UpperCAmelCase__ ) class a_ ( a ): A__ : Tuple = 'blip_vision_model' def __init__( self : str , UpperCAmelCase__ : Optional[int]=768 , UpperCAmelCase__ : Optional[Any]=3_072 , UpperCAmelCase__ : List[Any]=512 , UpperCAmelCase__ : Any=12 , UpperCAmelCase__ : Optional[int]=12 , UpperCAmelCase__ : Any=384 , UpperCAmelCase__ : Optional[Any]=16 , UpperCAmelCase__ : List[str]="gelu" , UpperCAmelCase__ : int=1e-5 , UpperCAmelCase__ : Optional[int]=0.0 , UpperCAmelCase__ : Any=1e-1_0 , **UpperCAmelCase__ : Dict , ): """simple docstring""" super().__init__(**UpperCAmelCase__ ) snake_case : List[str] = hidden_size snake_case : Union[str, Any] = intermediate_size snake_case : Tuple = projection_dim snake_case : Dict = num_hidden_layers snake_case : str = num_attention_heads snake_case : Dict = patch_size snake_case : Optional[Any] = image_size snake_case : Union[str, Any] = initializer_range snake_case : Dict = attention_dropout snake_case : Union[str, Any] = layer_norm_eps snake_case : Optional[int] = hidden_act @classmethod def lowerCAmelCase( cls : Union[str, Any] , UpperCAmelCase__ : Union[str, os.PathLike] , **UpperCAmelCase__ : Optional[Any] ): """simple docstring""" cls._set_token_in_kwargs(UpperCAmelCase__ ) snake_case , snake_case : Union[str, Any] = cls.get_config_dict(UpperCAmelCase__ , **UpperCAmelCase__ ) # get the vision config dict if we are loading from BlipConfig if config_dict.get('''model_type''' ) == "blip": snake_case : Tuple = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"You are using a model of type {config_dict['model_type']} to instantiate a model of type " F"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(UpperCAmelCase__ , **UpperCAmelCase__ ) class a_ ( a ): A__ : List[str] = 'blip' A__ : Tuple = True def __init__( self : Optional[Any] , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Any=None , UpperCAmelCase__ : Optional[int]=512 , UpperCAmelCase__ : Any=2.6592 , UpperCAmelCase__ : Any=256 , **UpperCAmelCase__ : Any , ): """simple docstring""" super().__init__(**UpperCAmelCase__ ) if text_config is None: snake_case : Union[str, Any] = {} logger.info('''`text_config` is `None`. Initializing the `BlipTextConfig` with default values.''' ) if vision_config is None: snake_case : Tuple = {} logger.info('''`vision_config` is `None`. Initializing the `BlipVisionConfig` with default values.''' ) snake_case : List[Any] = BlipTextConfig(**UpperCAmelCase__ ) snake_case : List[str] = BlipVisionConfig(**UpperCAmelCase__ ) snake_case : Dict = self.vision_config.hidden_size snake_case : Any = projection_dim snake_case : Optional[int] = logit_scale_init_value snake_case : List[Any] = 1.0 snake_case : int = 0.02 snake_case : Optional[Any] = image_text_hidden_size @classmethod def lowerCAmelCase( cls : Union[str, Any] , UpperCAmelCase__ : BlipTextConfig , UpperCAmelCase__ : BlipVisionConfig , **UpperCAmelCase__ : int ): """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **UpperCAmelCase__ ) def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Optional[int] = copy.deepcopy(self.__dict__ ) snake_case : Tuple = self.text_config.to_dict() snake_case : Optional[int] = self.vision_config.to_dict() snake_case : List[str] = self.__class__.model_type return output
84
import torch from diffusers import DiffusionPipeline class a_ ( a ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[Any] ): """simple docstring""" super().__init__() self.register_modules(unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ ) def __call__( self : Optional[int] ): """simple docstring""" snake_case : Any = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) snake_case : Dict = 1 snake_case : Optional[Any] = self.unet(UpperCAmelCase__ , UpperCAmelCase__ ).sample snake_case : List[Any] = self.scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample snake_case : List[Any] = scheduler_output - scheduler_output + torch.ones_like(UpperCAmelCase__ ) return result
84
1
import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() _a : Dict = logging.get_logger(__name__) _a : Optional[int] = 'https://openaipublic.azureedge.net/jukebox/models/' _a : int = { 'jukebox-1b-lyrics': [ '5b/vqvae.pth.tar', '5b/prior_level_0.pth.tar', '5b/prior_level_1.pth.tar', '1b_lyrics/prior_level_2.pth.tar', ], 'jukebox-5b-lyrics': [ '5b/vqvae.pth.tar', '5b/prior_level_0.pth.tar', '5b/prior_level_1.pth.tar', '5b_lyrics/prior_level_2.pth.tar', ], } def a_ ( __magic_name__ ) -> int: """simple docstring""" if key.endswith('''.model.1.bias''' ) and len(key.split('''.''' ) ) > 10: snake_case : str = key.replace('''.model.1.bias''' , '''.conv1d_1.bias''' ) elif key.endswith('''.model.1.weight''' ) and len(key.split('''.''' ) ) > 10: snake_case : Any = key.replace('''.model.1.weight''' , '''.conv1d_1.weight''' ) elif key.endswith('''.model.3.bias''' ) and len(key.split('''.''' ) ) > 10: snake_case : Union[str, Any] = key.replace('''.model.3.bias''' , '''.conv1d_2.bias''' ) elif key.endswith('''.model.3.weight''' ) and len(key.split('''.''' ) ) > 10: snake_case : Union[str, Any] = key.replace('''.model.3.weight''' , '''.conv1d_2.weight''' ) if "conditioner_blocks.0." in key: snake_case : int = key.replace('''conditioner_blocks.0''' , '''conditioner_blocks''' ) if "prime_prior" in key: snake_case : str = key.replace('''prime_prior''' , '''encoder''' ) if ".emb." in key and "total" not in key and "absolute" not in key and "relative" not in key: snake_case : str = key.replace('''.emb.''' , '''.''' ) if key.endswith('''k''' ): # replace vqvae.X.k with vqvae.X.codebook return key.replace('''.k''' , '''.codebook''' ) if "y_emb." in key: return key.replace('''y_emb.''' , '''metadata_embedding.''' ) if "x_emb.emb." in key: snake_case : Optional[Any] = key.replace('''0.x_emb.emb''' , '''embed_tokens''' ) if "prime_state_ln" in key: return key.replace('''prime_state_ln''' , '''encoder.final_layer_norm''' ) if ".ln" in key: return key.replace('''.ln''' , '''.layer_norm''' ) if "_ln" in key: return key.replace('''_ln''' , '''_layer_norm''' ) if "prime_state_proj" in key: return key.replace('''prime_state_proj''' , '''encoder.proj_in''' ) if "prime_x_out" in key: return key.replace('''prime_x_out''' , '''encoder.lm_head''' ) if "prior.x_out" in key: return key.replace('''x_out''' , '''fc_proj_out''' ) if "x_emb" in key: return key.replace('''x_emb''' , '''embed_tokens''' ) return key def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> int: """simple docstring""" snake_case : Union[str, Any] = {} import re snake_case : List[Any] = re.compile(R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)''' ) snake_case : int = re.compile( R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) snake_case : Optional[Any] = re.compile(R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)''' ) snake_case : Dict = re.compile(R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)''' ) snake_case : List[str] = re.compile( R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) snake_case : str = re.compile(R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)''' ) snake_case : List[Any] = re.compile(R'''conditioner_blocks.(\d*).cond.model.(\d*).(\d).(bias|weight)''' ) snake_case : Any = re.compile( R'''conditioner_blocks.(\d*).cond.model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) snake_case : int = re.compile(R'''conditioner_blocks.(\d*).cond.model.(\d*).(bias|weight)''' ) for original_key, value in state_dict.items(): # rename vqvae.encoder keys if re_encoder_block_conv_in.fullmatch(__magic_name__ ): snake_case : Union[str, Any] = re_encoder_block_conv_in.match(__magic_name__ ) snake_case : List[Any] = regex_match.groups() snake_case : Union[str, Any] = int(groups[2] ) * 2 + int(groups[3] ) snake_case : List[str] = F"encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.{groups[-1]}" snake_case : Any = re_encoder_block_conv_in.sub(__magic_name__ , __magic_name__ ) elif re_encoder_block_resnet.fullmatch(__magic_name__ ): snake_case : Tuple = re_encoder_block_resnet.match(__magic_name__ ) snake_case : int = regex_match.groups() snake_case : Union[str, Any] = int(groups[2] ) * 2 + int(groups[3] ) snake_case : Optional[int] = {'''1''': 1, '''3''': 2}[groups[-2]] snake_case : Any = F"encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}." snake_case : List[Any] = F"resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}" snake_case : Any = prefix + resnet_block snake_case : str = re_encoder_block_resnet.sub(__magic_name__ , __magic_name__ ) elif re_encoder_block_proj_out.fullmatch(__magic_name__ ): snake_case : Dict = re_encoder_block_proj_out.match(__magic_name__ ) snake_case : Dict = regex_match.groups() snake_case : Optional[Any] = F"encoders.{groups[0]}.level_blocks.{groups[1]}.proj_out.{groups[-1]}" snake_case : Union[str, Any] = re_encoder_block_proj_out.sub(__magic_name__ , __magic_name__ ) # rename vqvae.decoder keys elif re_decoder_block_conv_out.fullmatch(__magic_name__ ): snake_case : List[Any] = re_decoder_block_conv_out.match(__magic_name__ ) snake_case : List[str] = regex_match.groups() snake_case : Optional[int] = int(groups[2] ) * 2 + int(groups[3] ) - 2 snake_case : Any = F"decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.{groups[-1]}" snake_case : Union[str, Any] = re_decoder_block_conv_out.sub(__magic_name__ , __magic_name__ ) elif re_decoder_block_resnet.fullmatch(__magic_name__ ): snake_case : str = re_decoder_block_resnet.match(__magic_name__ ) snake_case : Tuple = regex_match.groups() snake_case : Optional[Any] = int(groups[2] ) * 2 + int(groups[3] ) - 2 snake_case : Union[str, Any] = {'''1''': 1, '''3''': 2}[groups[-2]] snake_case : List[str] = F"decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}." snake_case : str = F"resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}" snake_case : Any = prefix + resnet_block snake_case : List[Any] = re_decoder_block_resnet.sub(__magic_name__ , __magic_name__ ) elif re_decoder_block_proj_in.fullmatch(__magic_name__ ): snake_case : Optional[Any] = re_decoder_block_proj_in.match(__magic_name__ ) snake_case : Any = regex_match.groups() snake_case : Optional[Any] = F"decoders.{groups[0]}.level_blocks.{groups[1]}.proj_in.{groups[-1]}" snake_case : Optional[int] = re_decoder_block_proj_in.sub(__magic_name__ , __magic_name__ ) # rename prior cond.model to upsampler.upsample_block and resnet elif re_prior_cond_conv_out.fullmatch(__magic_name__ ): snake_case : int = re_prior_cond_conv_out.match(__magic_name__ ) snake_case : Optional[Any] = regex_match.groups() snake_case : Dict = int(groups[1] ) * 2 + int(groups[2] ) - 2 snake_case : Dict = F"conditioner_blocks.upsampler.upsample_block.{block_index}.{groups[-1]}" snake_case : str = re_prior_cond_conv_out.sub(__magic_name__ , __magic_name__ ) elif re_prior_cond_resnet.fullmatch(__magic_name__ ): snake_case : List[str] = re_prior_cond_resnet.match(__magic_name__ ) snake_case : Optional[Any] = regex_match.groups() snake_case : Any = int(groups[1] ) * 2 + int(groups[2] ) - 2 snake_case : Dict = {'''1''': 1, '''3''': 2}[groups[-2]] snake_case : Union[str, Any] = F"conditioner_blocks.upsampler.upsample_block.{block_index}." snake_case : Union[str, Any] = F"resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}" snake_case : List[str] = prefix + resnet_block snake_case : Dict = re_prior_cond_resnet.sub(__magic_name__ , __magic_name__ ) elif re_prior_cond_proj_in.fullmatch(__magic_name__ ): snake_case : Dict = re_prior_cond_proj_in.match(__magic_name__ ) snake_case : int = regex_match.groups() snake_case : Union[str, Any] = F"conditioner_blocks.upsampler.proj_in.{groups[-1]}" snake_case : int = re_prior_cond_proj_in.sub(__magic_name__ , __magic_name__ ) # keep original key else: snake_case : Optional[int] = original_key snake_case : Union[str, Any] = replace_key(__magic_name__ ) if F"{key_prefix}.{key}" not in model_state_dict or key is None: print(F"failed converting {original_key} to {key}, does not match" ) # handle missmatched shape elif value.shape != model_state_dict[F"{key_prefix}.{key}"].shape: snake_case : List[str] = model_state_dict[F"{key_prefix}.{key}"] print(F"{original_key}-> {key} : \nshape {val.shape} and { value.shape}, do not match" ) snake_case : List[str] = original_key snake_case : Optional[Any] = original_key snake_case : Optional[Any] = value return new_dict @torch.no_grad() def a_ ( __magic_name__=None , __magic_name__=None ) -> str: """simple docstring""" for file in MODEL_MAPPING[model_name]: if not os.path.isfile(F"{pytorch_dump_folder_path}/{file.split('/' )[-1]}" ): snake_case : Dict = requests.get(F"{PREFIX}{file}" , allow_redirects=__magic_name__ ) os.makedirs(F"{pytorch_dump_folder_path}/" , exist_ok=__magic_name__ ) open(F"{pytorch_dump_folder_path}/{file.split('/' )[-1]}" , '''wb''' ).write(r.content ) snake_case : str = MODEL_MAPPING[model_name.split('''/''' )[-1]] snake_case : Dict = JukeboxConfig.from_pretrained(__magic_name__ ) snake_case : Optional[int] = JukeboxModel(__magic_name__ ) snake_case : str = [] snake_case : str = {} for i, dict_name in enumerate(__magic_name__ ): snake_case : List[str] = torch.load(F"{pytorch_dump_folder_path}/{dict_name.split('/' )[-1]}" )['''model'''] snake_case : Union[str, Any] = {} for k in old_dic.keys(): if k.endswith('''.b''' ): snake_case : Union[str, Any] = old_dic[k] elif k.endswith('''.w''' ): snake_case : str = old_dic[k] elif "level_2" not in dict_name and "cond.model." in k: snake_case : str = old_dic[k] else: snake_case : Any = old_dic[k] snake_case : List[str] = '''vqvae''' if i == 0 else F"priors.{3 - i}" snake_case : str = fix_jukebox_keys(__magic_name__ , model.state_dict() , __magic_name__ , __magic_name__ ) weight_dict.append(__magic_name__ ) snake_case : Optional[Any] = weight_dict.pop(0 ) model.vqvae.load_state_dict(__magic_name__ ) for i in range(len(__magic_name__ ) ): model.priors[i].load_state_dict(weight_dict[2 - i] ) Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) with open(F"{pytorch_dump_folder_path}/mapping.json" , '''w''' ) as txtfile: json.dump(__magic_name__ , __magic_name__ ) print(F"Saving model {model_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(__magic_name__ ) return weight_dict if __name__ == "__main__": _a : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='jukebox-5b-lyrics', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='jukebox-5b-lyrics-converted', type=str, help='Path to the output PyTorch model directory.', ) _a : Optional[Any] = parser.parse_args() convert_openai_checkpoint(args.model_name, args.pytorch_dump_folder_path)
84
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class a_ ( a ): A__ : List[str] = ['image_processor', 'tokenizer'] A__ : Any = 'CLIPImageProcessor' A__ : Optional[int] = ('CLIPTokenizer', 'CLIPTokenizerFast') def __init__( self : Union[str, Any] , UpperCAmelCase__ : str=None , UpperCAmelCase__ : Union[str, Any]=None , **UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : Optional[int] = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , UpperCAmelCase__ , ) snake_case : List[Any] = kwargs.pop('''feature_extractor''' ) snake_case : Optional[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(UpperCAmelCase__ , UpperCAmelCase__ ) def __call__( self : Any , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : int=None , **UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: snake_case : int = self.tokenizer(UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , **UpperCAmelCase__ ) if images is not None: snake_case : Dict = self.image_processor(UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , **UpperCAmelCase__ ) if text is not None and images is not None: snake_case : Tuple = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**UpperCAmelCase__ ) , tensor_type=UpperCAmelCase__ ) def lowerCAmelCase( self : List[str] , *UpperCAmelCase__ : Dict , **UpperCAmelCase__ : int ): """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : str ): """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : int = self.tokenizer.model_input_names snake_case : Tuple = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , UpperCAmelCase__ , ) return self.image_processor_class @property def lowerCAmelCase( self : Any ): """simple docstring""" warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , UpperCAmelCase__ , ) return self.image_processor
84
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer _a : str = logging.get_logger(__name__) _a : Optional[int] = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} _a : Optional[Any] = { 'vocab_file': { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json' ), }, } _a : Union[str, Any] = { 'yjernite/retribert-base-uncased': 512, } _a : Tuple = { 'yjernite/retribert-base-uncased': {'do_lower_case': True}, } class a_ ( a ): A__ : List[str] = VOCAB_FILES_NAMES A__ : Any = PRETRAINED_VOCAB_FILES_MAP A__ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Any = PRETRAINED_INIT_CONFIGURATION A__ : Optional[Any] = RetriBertTokenizer A__ : Any = ['input_ids', 'attention_mask'] def __init__( self : Optional[int] , UpperCAmelCase__ : Any=None , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Dict="[UNK]" , UpperCAmelCase__ : str="[SEP]" , UpperCAmelCase__ : Union[str, Any]="[PAD]" , UpperCAmelCase__ : Dict="[CLS]" , UpperCAmelCase__ : Optional[Any]="[MASK]" , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Optional[int]=None , **UpperCAmelCase__ : Dict , ): """simple docstring""" super().__init__( UpperCAmelCase__ , tokenizer_file=UpperCAmelCase__ , do_lower_case=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , tokenize_chinese_chars=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ , **UpperCAmelCase__ , ) snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , UpperCAmelCase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' , UpperCAmelCase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , UpperCAmelCase__ ) != tokenize_chinese_chars ): snake_case : int = getattr(UpperCAmelCase__ , normalizer_state.pop('''type''' ) ) snake_case : List[Any] = do_lower_case snake_case : Union[str, Any] = strip_accents snake_case : int = tokenize_chinese_chars snake_case : int = normalizer_class(**UpperCAmelCase__ ) snake_case : Union[str, Any] = do_lower_case def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Union[str, Any]=None ): """simple docstring""" snake_case : str = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ): """simple docstring""" snake_case : List[Any] = [self.sep_token_id] snake_case : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCAmelCase( self : str , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ): """simple docstring""" snake_case : Tuple = self._tokenizer.model.save(UpperCAmelCase__ , name=UpperCAmelCase__ ) return tuple(UpperCAmelCase__ )
84
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets _a : Optional[Any] = '\\n@inproceedings{pillutla-etal:mauve:neurips2021,\n title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers},\n author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid},\n booktitle = {NeurIPS},\n year = {2021}\n}\n\n' _a : str = '\\nMAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure.\n\nMAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences.\n\nFor details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021).\n\nThis metrics is a wrapper around the official implementation of MAUVE:\nhttps://github.com/krishnap25/mauve\n' _a : List[Any] = '\nCalculates MAUVE scores between two lists of generated text and reference text.\nArgs:\n predictions: list of generated text to score. Each predictions\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\nOptional Args:\n num_buckets: the size of the histogram to quantize P and Q. Options: \'auto\' (default) or an integer\n pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1\n kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9\n kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5\n kmeans_max_iter: maximum number of k-means iterations. Default 500\n featurize_model_name: name of the model from which features are obtained. Default \'gpt2-large\' Use one of [\'gpt2\', \'gpt2-medium\', \'gpt2-large\', \'gpt2-xl\'].\n device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU\n max_text_length: maximum number of tokens to consider. Default 1024\n divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25\n mauve_scaling_factor: "c" from the paper. Default 5.\n verbose: If True (default), print running time updates\n seed: random seed to initialize k-means cluster assignments.\nReturns:\n mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer,\n frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer,\n divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve,\n p_hist: a discrete distribution, which is a quantized version of the text distribution p_text,\n q_hist: same as above, but with q_text.\nExamples:\n\n >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest\n >>> import datasets\n >>> mauve = datasets.load_metric(\'mauve\')\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP\n >>> print(out.mauve) # doctest: +SKIP\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): def lowerCAmelCase( self : Any ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://github.com/krishnap25/mauve''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/krishnap25/mauve'''] , reference_urls=[ '''https://arxiv.org/abs/2102.01454''', '''https://github.com/krishnap25/mauve''', ] , ) def lowerCAmelCase( self : int , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : List[str]="auto" , UpperCAmelCase__ : Tuple=-1 , UpperCAmelCase__ : Optional[int]=0.9 , UpperCAmelCase__ : List[Any]=5 , UpperCAmelCase__ : List[Any]=500 , UpperCAmelCase__ : Union[str, Any]="gpt2-large" , UpperCAmelCase__ : Optional[Any]=-1 , UpperCAmelCase__ : int=1_024 , UpperCAmelCase__ : List[Any]=25 , UpperCAmelCase__ : Union[str, Any]=5 , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : List[Any]=25 , ): """simple docstring""" snake_case : List[str] = compute_mauve( p_text=UpperCAmelCase__ , q_text=UpperCAmelCase__ , p_features=UpperCAmelCase__ , q_features=UpperCAmelCase__ , p_tokens=UpperCAmelCase__ , q_tokens=UpperCAmelCase__ , num_buckets=UpperCAmelCase__ , pca_max_data=UpperCAmelCase__ , kmeans_explained_var=UpperCAmelCase__ , kmeans_num_redo=UpperCAmelCase__ , kmeans_max_iter=UpperCAmelCase__ , featurize_model_name=UpperCAmelCase__ , device_id=UpperCAmelCase__ , max_text_length=UpperCAmelCase__ , divergence_curve_discretization_size=UpperCAmelCase__ , mauve_scaling_factor=UpperCAmelCase__ , verbose=UpperCAmelCase__ , seed=UpperCAmelCase__ , ) return out
84
1
from typing import Optional, Tuple import jax import jax.numpy as jnp from flax import linen as nn from flax.core.frozen_dict import FrozenDict from transformers import CLIPConfig, FlaxPreTrainedModel from transformers.models.clip.modeling_flax_clip import FlaxCLIPVisionModule def a_ ( __magic_name__ , __magic_name__ , __magic_name__=1e-12 ) -> Optional[int]: """simple docstring""" snake_case : Optional[Any] = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(__magic_name__ , axis=1 ) , a_min=__magic_name__ ) ).T snake_case : Dict = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(__magic_name__ , axis=1 ) , a_min=__magic_name__ ) ).T return jnp.matmul(__magic_name__ , norm_emb_a.T ) class a_ ( nn.Module ): A__ : CLIPConfig A__ : jnp.dtype = jnp.floataa def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Optional[int] = FlaxCLIPVisionModule(self.config.vision_config ) snake_case : Optional[int] = nn.Dense(self.config.projection_dim , use_bias=UpperCAmelCase__ , dtype=self.dtype ) snake_case : Tuple = self.param('''concept_embeds''' , jax.nn.initializers.ones , (17, self.config.projection_dim) ) snake_case : Union[str, Any] = self.param( '''special_care_embeds''' , jax.nn.initializers.ones , (3, self.config.projection_dim) ) snake_case : Union[str, Any] = self.param('''concept_embeds_weights''' , jax.nn.initializers.ones , (17,) ) snake_case : List[str] = self.param('''special_care_embeds_weights''' , jax.nn.initializers.ones , (3,) ) def __call__( self : str , UpperCAmelCase__ : List[Any] ): """simple docstring""" snake_case : Optional[int] = self.vision_model(UpperCAmelCase__ )[1] snake_case : List[Any] = self.visual_projection(UpperCAmelCase__ ) snake_case : Any = jax_cosine_distance(UpperCAmelCase__ , self.special_care_embeds ) snake_case : Optional[int] = jax_cosine_distance(UpperCAmelCase__ , self.concept_embeds ) # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign image inputs snake_case : Union[str, Any] = 0.0 snake_case : Optional[int] = special_cos_dist - self.special_care_embeds_weights[None, :] + adjustment snake_case : Optional[int] = jnp.round(UpperCAmelCase__ , 3 ) snake_case : str = jnp.any(special_scores > 0 , axis=1 , keepdims=UpperCAmelCase__ ) # Use a lower threshold if an image has any special care concept snake_case : List[str] = is_special_care * 0.01 snake_case : str = cos_dist - self.concept_embeds_weights[None, :] + special_adjustment snake_case : Any = jnp.round(UpperCAmelCase__ , 3 ) snake_case : List[Any] = jnp.any(concept_scores > 0 , axis=1 ) return has_nsfw_concepts class a_ ( a ): A__ : Tuple = CLIPConfig A__ : str = 'clip_input' A__ : Tuple = FlaxStableDiffusionSafetyCheckerModule def __init__( self : Optional[int] , UpperCAmelCase__ : CLIPConfig , UpperCAmelCase__ : Optional[Tuple] = None , UpperCAmelCase__ : int = 0 , UpperCAmelCase__ : jnp.dtype = jnp.floataa , UpperCAmelCase__ : bool = True , **UpperCAmelCase__ : Optional[int] , ): """simple docstring""" if input_shape is None: snake_case : Optional[Any] = (1, 224, 224, 3) snake_case : str = self.module_class(config=UpperCAmelCase__ , dtype=UpperCAmelCase__ , **UpperCAmelCase__ ) super().__init__(UpperCAmelCase__ , UpperCAmelCase__ , input_shape=UpperCAmelCase__ , seed=UpperCAmelCase__ , dtype=UpperCAmelCase__ , _do_init=_do_init ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : jax.random.KeyArray , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : FrozenDict = None ): """simple docstring""" # init input tensor snake_case : Any = jax.random.normal(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case , snake_case : List[str] = jax.random.split(UpperCAmelCase__ ) snake_case : Optional[Any] = {'''params''': params_rng, '''dropout''': dropout_rng} snake_case : Any = self.module.init(UpperCAmelCase__ , UpperCAmelCase__ )['''params'''] return random_params def __call__( self : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : dict = None , ): """simple docstring""" snake_case : Union[str, Any] = jnp.transpose(UpperCAmelCase__ , (0, 2, 3, 1) ) return self.module.apply( {'''params''': params or self.params} , jnp.array(UpperCAmelCase__ , dtype=jnp.floataa ) , rngs={} , )
84
import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" if "cls_token" in name: snake_case : Tuple = name.replace('''cls_token''' , '''vit.embeddings.cls_token''' ) if "mask_token" in name: snake_case : Optional[int] = name.replace('''mask_token''' , '''decoder.mask_token''' ) if "decoder_pos_embed" in name: snake_case : List[str] = name.replace('''decoder_pos_embed''' , '''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: snake_case : List[str] = name.replace('''pos_embed''' , '''vit.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: snake_case : List[Any] = name.replace('''patch_embed.proj''' , '''vit.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: snake_case : int = name.replace('''patch_embed.norm''' , '''vit.embeddings.norm''' ) if "decoder_blocks" in name: snake_case : int = name.replace('''decoder_blocks''' , '''decoder.decoder_layers''' ) if "blocks" in name: snake_case : Optional[Any] = name.replace('''blocks''' , '''vit.encoder.layer''' ) if "attn.proj" in name: snake_case : str = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: snake_case : Any = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: snake_case : List[str] = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: snake_case : Tuple = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: snake_case : Tuple = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: snake_case : Tuple = name.replace('''mlp.fc2''' , '''output.dense''' ) if "decoder_embed" in name: snake_case : Dict = name.replace('''decoder_embed''' , '''decoder.decoder_embed''' ) if "decoder_norm" in name: snake_case : Dict = name.replace('''decoder_norm''' , '''decoder.decoder_norm''' ) if "decoder_pred" in name: snake_case : Dict = name.replace('''decoder_pred''' , '''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name: snake_case : Optional[int] = name.replace('''norm.weight''' , '''vit.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name: snake_case : List[str] = name.replace('''norm.bias''' , '''vit.layernorm.bias''' ) return name def a_ ( __magic_name__ , __magic_name__ ) -> str: """simple docstring""" for key in orig_state_dict.copy().keys(): snake_case : Union[str, Any] = orig_state_dict.pop(__magic_name__ ) if "qkv" in key: snake_case : Optional[int] = key.split('''.''' ) snake_case : int = int(key_split[1] ) if "decoder_blocks" in key: snake_case : List[str] = config.decoder_hidden_size snake_case : List[Any] = '''decoder.decoder_layers.''' if "weight" in key: snake_case : str = val[:dim, :] snake_case : Optional[Any] = val[dim : dim * 2, :] snake_case : Any = val[-dim:, :] elif "bias" in key: snake_case : Optional[Any] = val[:dim] snake_case : List[Any] = val[dim : dim * 2] snake_case : List[Any] = val[-dim:] else: snake_case : Optional[int] = config.hidden_size snake_case : Tuple = '''vit.encoder.layer.''' if "weight" in key: snake_case : Optional[Any] = val[:dim, :] snake_case : str = val[dim : dim * 2, :] snake_case : Union[str, Any] = val[-dim:, :] elif "bias" in key: snake_case : Tuple = val[:dim] snake_case : int = val[dim : dim * 2] snake_case : Optional[Any] = val[-dim:] else: snake_case : Optional[Any] = val return orig_state_dict def a_ ( __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" snake_case : List[str] = ViTMAEConfig() if "large" in checkpoint_url: snake_case : str = 1_024 snake_case : Tuple = 4_096 snake_case : Optional[Any] = 24 snake_case : List[Any] = 16 elif "huge" in checkpoint_url: snake_case : Tuple = 14 snake_case : int = 1_280 snake_case : Dict = 5_120 snake_case : Tuple = 32 snake_case : Optional[Any] = 16 snake_case : Optional[Any] = ViTMAEForPreTraining(__magic_name__ ) snake_case : Optional[Any] = torch.hub.load_state_dict_from_url(__magic_name__ , map_location='''cpu''' )['''model'''] snake_case : int = ViTMAEImageProcessor(size=config.image_size ) snake_case : Dict = convert_state_dict(__magic_name__ , __magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() snake_case : Tuple = '''https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg''' snake_case : List[Any] = Image.open(requests.get(__magic_name__ , stream=__magic_name__ ).raw ) snake_case : Dict = ViTMAEImageProcessor(size=config.image_size ) snake_case : str = image_processor(images=__magic_name__ , return_tensors='''pt''' ) # forward pass torch.manual_seed(2 ) snake_case : Union[str, Any] = model(**__magic_name__ ) snake_case : Optional[Any] = outputs.logits if "large" in checkpoint_url: snake_case : Any = torch.tensor( [[-0.7309, -0.7128, -1.0169], [-1.0161, -0.9058, -1.1878], [-1.0478, -0.9411, -1.1911]] ) elif "huge" in checkpoint_url: snake_case : List[Any] = torch.tensor( [[-1.1599, -0.9199, -1.2221], [-1.1952, -0.9269, -1.2307], [-1.2143, -0.9337, -1.2262]] ) else: snake_case : Dict = torch.tensor( [[-0.9192, -0.8481, -1.1259], [-1.1349, -1.0034, -1.2599], [-1.1757, -1.0429, -1.2726]] ) # verify logits assert torch.allclose(logits[0, :3, :3] , __magic_name__ , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(__magic_name__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__magic_name__ ) if __name__ == "__main__": _a : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth', type=str, help='URL of the checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _a : str = parser.parse_args() convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
84
1
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class a_ ( a ): def __init__( self : str , UpperCAmelCase__ : pyspark.sql.DataFrame , UpperCAmelCase__ : Optional[NamedSplit] = None , UpperCAmelCase__ : Optional[Features] = None , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : str = None , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : str = None , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : str = "arrow" , **UpperCAmelCase__ : Union[str, Any] , ): """simple docstring""" super().__init__( split=UpperCAmelCase__ , features=UpperCAmelCase__ , cache_dir=UpperCAmelCase__ , keep_in_memory=UpperCAmelCase__ , streaming=UpperCAmelCase__ , **UpperCAmelCase__ , ) snake_case : str = load_from_cache_file snake_case : int = file_format snake_case : Optional[Any] = Spark( df=UpperCAmelCase__ , features=UpperCAmelCase__ , cache_dir=UpperCAmelCase__ , working_dir=UpperCAmelCase__ , **UpperCAmelCase__ , ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) snake_case : Optional[Any] = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=UpperCAmelCase__ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
84
import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _a : Optional[Any] = 16 _a : Union[str, Any] = 32 def a_ ( __magic_name__ , __magic_name__ = 16 ) -> Dict: """simple docstring""" snake_case : Tuple = AutoTokenizer.from_pretrained('''bert-base-cased''' ) snake_case : Any = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(__magic_name__ ): # max_length=None => use the model max length (it's actually the default) snake_case : Union[str, Any] = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=__magic_name__ , max_length=__magic_name__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): snake_case : Union[str, Any] = datasets.map( __magic_name__ , batched=__magic_name__ , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library snake_case : Optional[Any] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(__magic_name__ ): # On TPU it's best to pad everything to the same length or training will be very slow. snake_case : str = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": snake_case : Tuple = 16 elif accelerator.mixed_precision != "no": snake_case : Dict = 8 else: snake_case : Union[str, Any] = None return tokenizer.pad( __magic_name__ , padding='''longest''' , max_length=__magic_name__ , pad_to_multiple_of=__magic_name__ , return_tensors='''pt''' , ) # Instantiate dataloaders. snake_case : str = DataLoader( tokenized_datasets['''train'''] , shuffle=__magic_name__ , collate_fn=__magic_name__ , batch_size=__magic_name__ ) snake_case : List[str] = DataLoader( tokenized_datasets['''validation'''] , shuffle=__magic_name__ , collate_fn=__magic_name__ , batch_size=__magic_name__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders _a : Optional[int] = mocked_dataloaders # noqa: F811 def a_ ( __magic_name__ , __magic_name__ ) -> Optional[Any]: """simple docstring""" if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , __magic_name__ ) == "1": snake_case : Optional[int] = 2 # Initialize accelerator snake_case : int = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs snake_case : Dict = config['''lr'''] snake_case : Any = int(config['''num_epochs'''] ) snake_case : List[str] = int(config['''seed'''] ) snake_case : List[Any] = int(config['''batch_size'''] ) snake_case : Tuple = evaluate.load('''glue''' , '''mrpc''' ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=__magic_name__ ) def inner_training_loop(__magic_name__ ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(__magic_name__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) snake_case : str = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=__magic_name__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). snake_case : Optional[int] = model.to(accelerator.device ) # Instantiate optimizer snake_case : Optional[int] = AdamW(params=model.parameters() , lr=__magic_name__ ) snake_case , snake_case : List[Any] = get_dataloaders(__magic_name__ , __magic_name__ ) # Instantiate scheduler snake_case : int = get_linear_schedule_with_warmup( optimizer=__magic_name__ , num_warmup_steps=100 , num_training_steps=(len(__magic_name__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. snake_case , snake_case , snake_case , snake_case , snake_case : Tuple = accelerator.prepare( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # Now we train the model for epoch in range(__magic_name__ ): model.train() for step, batch in enumerate(__magic_name__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) snake_case : int = model(**__magic_name__ ) snake_case : Optional[int] = outputs.loss accelerator.backward(__magic_name__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__magic_name__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): snake_case : List[str] = model(**__magic_name__ ) snake_case : List[Any] = outputs.logits.argmax(dim=-1 ) snake_case , snake_case : Dict = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=__magic_name__ , references=__magic_name__ , ) snake_case : Tuple = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"epoch {epoch}:" , __magic_name__ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def a_ ( ) -> Union[str, Any]: """simple docstring""" snake_case : int = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=__magic_name__ , default=__magic_name__ , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) snake_case : Optional[Any] = parser.parse_args() snake_case : Optional[int] = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(__magic_name__ , __magic_name__ ) if __name__ == "__main__": main()
84
1
from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers _a : List[str] = [ 'python', 'tqdm', 'regex', 'requests', 'packaging', 'filelock', 'numpy', 'tokenizers', 'huggingface-hub', 'safetensors', 'accelerate', 'pyyaml', ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f"can't find {pkg} in {deps.keys()}, check dependency_versions_table.py") def a_ ( __magic_name__ , __magic_name__=None ) -> List[str]: """simple docstring""" require_version(deps[pkg] , __magic_name__ )
84
from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract _a : Dict = logging.get_logger(__name__) def a_ ( __magic_name__ , __magic_name__ , __magic_name__ ) -> List[Any]: """simple docstring""" return [ int(1_000 * (box[0] / width) ), int(1_000 * (box[1] / height) ), int(1_000 * (box[2] / width) ), int(1_000 * (box[3] / height) ), ] def a_ ( __magic_name__ , __magic_name__ , __magic_name__ = None ) -> str: """simple docstring""" snake_case : Any = tesseract_config if tesseract_config is not None else '''''' # apply OCR snake_case : str = to_pil_image(__magic_name__ ) snake_case , snake_case : Union[str, Any] = pil_image.size snake_case : List[Any] = pytesseract.image_to_data(__magic_name__ , lang=__magic_name__ , output_type='''dict''' , config=__magic_name__ ) snake_case , snake_case , snake_case , snake_case , snake_case : Optional[Any] = data['''text'''], data['''left'''], data['''top'''], data['''width'''], data['''height'''] # filter empty words and corresponding coordinates snake_case : Union[str, Any] = [idx for idx, word in enumerate(__magic_name__ ) if not word.strip()] snake_case : Union[str, Any] = [word for idx, word in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : Optional[Any] = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format snake_case : List[Any] = [] for x, y, w, h in zip(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): snake_case : Optional[int] = [x, y, x + w, y + h] actual_boxes.append(__magic_name__ ) # finally, normalize the bounding boxes snake_case : List[Any] = [] for box in actual_boxes: normalized_boxes.append(normalize_box(__magic_name__ , __magic_name__ , __magic_name__ ) ) assert len(__magic_name__ ) == len(__magic_name__ ), "Not as many words as there are bounding boxes" return words, normalized_boxes class a_ ( a ): A__ : int = ['pixel_values'] def __init__( self : Optional[int] , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : PILImageResampling = PILImageResampling.BILINEAR , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[str] = "" , **UpperCAmelCase__ : int , ): """simple docstring""" super().__init__(**UpperCAmelCase__ ) snake_case : Any = size if size is not None else {'''height''': 224, '''width''': 224} snake_case : Tuple = get_size_dict(UpperCAmelCase__ ) snake_case : Dict = do_resize snake_case : str = size snake_case : Optional[int] = resample snake_case : Union[str, Any] = apply_ocr snake_case : int = ocr_lang snake_case : Union[str, Any] = tesseract_config def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : np.ndarray , UpperCAmelCase__ : Dict[str, int] , UpperCAmelCase__ : PILImageResampling = PILImageResampling.BILINEAR , UpperCAmelCase__ : Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase__ : Any , ): """simple docstring""" snake_case : Dict = get_size_dict(UpperCAmelCase__ ) if "height" not in size or "width" not in size: raise ValueError(F"The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}" ) snake_case : Tuple = (size['''height'''], size['''width''']) return resize(UpperCAmelCase__ , size=UpperCAmelCase__ , resample=UpperCAmelCase__ , data_format=UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : ImageInput , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : PILImageResampling = None , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[Union[str, TensorType]] = None , UpperCAmelCase__ : ChannelDimension = ChannelDimension.FIRST , **UpperCAmelCase__ : List[Any] , ): """simple docstring""" snake_case : Tuple = do_resize if do_resize is not None else self.do_resize snake_case : List[Any] = size if size is not None else self.size snake_case : Tuple = get_size_dict(UpperCAmelCase__ ) snake_case : str = resample if resample is not None else self.resample snake_case : Optional[int] = apply_ocr if apply_ocr is not None else self.apply_ocr snake_case : Any = ocr_lang if ocr_lang is not None else self.ocr_lang snake_case : Optional[int] = tesseract_config if tesseract_config is not None else self.tesseract_config snake_case : List[str] = make_list_of_images(UpperCAmelCase__ ) if not valid_images(UpperCAmelCase__ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) # All transformations expect numpy arrays. snake_case : Any = [to_numpy_array(UpperCAmelCase__ ) for image in images] if apply_ocr: requires_backends(self , '''pytesseract''' ) snake_case : Optional[int] = [] snake_case : Union[str, Any] = [] for image in images: snake_case , snake_case : List[Any] = apply_tesseract(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) words_batch.append(UpperCAmelCase__ ) boxes_batch.append(UpperCAmelCase__ ) if do_resize: snake_case : Any = [self.resize(image=UpperCAmelCase__ , size=UpperCAmelCase__ , resample=UpperCAmelCase__ ) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) snake_case : int = [flip_channel_order(UpperCAmelCase__ ) for image in images] snake_case : Optional[Any] = [to_channel_dimension_format(UpperCAmelCase__ , UpperCAmelCase__ ) for image in images] snake_case : List[Any] = BatchFeature(data={'''pixel_values''': images} , tensor_type=UpperCAmelCase__ ) if apply_ocr: snake_case : Dict = words_batch snake_case : Dict = boxes_batch return data
84
1
from __future__ import annotations import os from typing import Any import requests _a : Dict = 'https://api.github.com' # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user _a : str = BASE_URL + '/user' # https://github.com/settings/tokens _a : Union[str, Any] = os.environ.get('USER_TOKEN', '') def a_ ( __magic_name__ ) -> dict[Any, Any]: """simple docstring""" snake_case : Union[str, Any] = { '''Authorization''': F"token {auth_token}", '''Accept''': '''application/vnd.github.v3+json''', } return requests.get(__magic_name__ , headers=__magic_name__ ).json() if __name__ == "__main__": # pragma: no cover if USER_TOKEN: for key, value in fetch_github_info(USER_TOKEN).items(): print(f"{key}: {value}") else: raise ValueError('\'USER_TOKEN\' field cannot be empty.')
84
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class a_ : def __init__( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Union[str, Any]=13 , UpperCAmelCase__ : Union[str, Any]=2 , UpperCAmelCase__ : List[Any]=24 , UpperCAmelCase__ : Union[str, Any]=16 , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : str=True , UpperCAmelCase__ : int=32 , UpperCAmelCase__ : Tuple=5 , UpperCAmelCase__ : Dict=4 , UpperCAmelCase__ : Optional[int]=37 , UpperCAmelCase__ : Optional[int]="gelu" , UpperCAmelCase__ : Dict=0.1 , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : Optional[int]=10 , UpperCAmelCase__ : Any=0.02 , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : List[Any]=2 , UpperCAmelCase__ : Optional[Any]=2 , ): """simple docstring""" snake_case : Tuple = parent snake_case : Dict = batch_size snake_case : str = patch_size snake_case : Union[str, Any] = max_length snake_case : str = num_mel_bins snake_case : Any = is_training snake_case : Union[str, Any] = use_labels snake_case : Tuple = hidden_size snake_case : Dict = num_hidden_layers snake_case : Any = num_attention_heads snake_case : Any = intermediate_size snake_case : List[Any] = hidden_act snake_case : str = hidden_dropout_prob snake_case : str = attention_probs_dropout_prob snake_case : str = type_sequence_label_size snake_case : Optional[int] = initializer_range snake_case : str = scope snake_case : int = frequency_stride snake_case : Union[str, Any] = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) snake_case : Optional[int] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 snake_case : Any = (self.max_length - self.patch_size) // self.time_stride + 1 snake_case : Union[str, Any] = frequency_out_dimension * time_out_dimension snake_case : Union[str, Any] = num_patches + 2 def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Optional[int] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) snake_case : str = None if self.use_labels: snake_case : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case : List[str] = self.get_config() return config, input_values, labels def lowerCAmelCase( self : Any ): """simple docstring""" return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCAmelCase__ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : str = ASTModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : Any = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : int = self.prepare_config_and_inputs() ( ( snake_case ) , ( snake_case ) , ( snake_case ) , ) : int = config_and_inputs snake_case : Tuple = {'''input_values''': input_values} return config, inputs_dict @require_torch class a_ ( a , a , unittest.TestCase ): A__ : List[Any] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) A__ : int = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) A__ : Optional[Any] = False A__ : Dict = False A__ : int = False A__ : Optional[int] = False def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : int ): """simple docstring""" if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[int] = ASTModelTester(self ) snake_case : Optional[int] = ConfigTester(self , config_class=UpperCAmelCase__ , has_text_modality=UpperCAmelCase__ , hidden_size=37 ) def lowerCAmelCase( self : List[str] ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def lowerCAmelCase( self : Tuple ): """simple docstring""" pass def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Optional[Any] = model_class(UpperCAmelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase__ , nn.Linear ) ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Any = model_class(UpperCAmelCase__ ) snake_case : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case : str = [*signature.parameters.keys()] snake_case : List[str] = ['''input_values'''] self.assertListEqual(arg_names[:1] , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case : List[str] = ASTModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) def a_ ( ) -> Dict: """simple docstring""" snake_case : Dict = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) snake_case , snake_case : int = torchaudio.load(__magic_name__ ) return audio, sampling_rate @require_torch @require_torchaudio class a_ ( unittest.TestCase ): @cached_property def lowerCAmelCase( self : Any ): """simple docstring""" return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : List[str] = self.default_feature_extractor snake_case : str = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(UpperCAmelCase__ ) snake_case : str = self.default_feature_extractor snake_case , snake_case : int = prepare_audio() snake_case : Optional[int] = audio.squeeze().numpy() snake_case : Optional[Any] = feature_extractor(UpperCAmelCase__ , sampling_rate=UpperCAmelCase__ , return_tensors='''pt''' ).to(UpperCAmelCase__ ) # forward pass with torch.no_grad(): snake_case : Union[str, Any] = model(**UpperCAmelCase__ ) # verify the logits snake_case : Any = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase__ ) snake_case : str = torch.tensor([-0.8760, -7.0042, -8.6602] ).to(UpperCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase__ , atol=1e-4 ) )
84
1
def a_ ( __magic_name__ ) -> Any: """simple docstring""" snake_case : Union[str, Any] = [0] * len(__magic_name__ ) snake_case : Optional[int] = [] snake_case : List[Any] = [] snake_case : Optional[int] = 0 for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(__magic_name__ ) ): if indegree[i] == 0: queue.append(__magic_name__ ) while queue: snake_case : Dict = queue.pop(0 ) cnt += 1 topo.append(__magic_name__ ) for x in graph[vertex]: indegree[x] -= 1 if indegree[x] == 0: queue.append(__magic_name__ ) if cnt != len(__magic_name__ ): print('''Cycle exists''' ) else: print(__magic_name__ ) # Adjacency List of Graph _a : Union[str, Any] = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []} topological_sort(graph)
84
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _a : Union[str, Any] = logging.getLogger(__name__) def a_ ( __magic_name__ , __magic_name__ ) -> Optional[int]: """simple docstring""" return (preds == labels).mean() @dataclass class a_ : A__ : str = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) @dataclass class a_ : A__ : str = field(metadata={'help': 'The name of the task to train on: ' + ', '.join(processors.keys() )} ) A__ : str = field(metadata={'help': 'Should contain the data files for the task.'} ) A__ : int = field( default=128 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) A__ : bool = field( default=a , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) def a_ ( ) -> Dict: """simple docstring""" snake_case : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) snake_case , snake_case , snake_case : Tuple = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"Output directory ({training_args.output_dir}) already exists and is not empty. Use" ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , __magic_name__ ) # Set seed set_seed(training_args.seed ) try: snake_case : int = processors[data_args.task_name]() snake_case : List[str] = processor.get_labels() snake_case : str = len(__magic_name__ ) except KeyError: raise ValueError('''Task not found: %s''' % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=__magic_name__ , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) snake_case : str = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) snake_case : Any = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=__magic_name__ , cache_dir=model_args.cache_dir , ) # Get datasets snake_case : Optional[int] = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=__magic_name__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) snake_case : Any = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=__magic_name__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(__magic_name__ ) -> Dict: snake_case : str = np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(__magic_name__ , p.label_ids )} # Data collator snake_case : Dict = DataCollatorWithPadding(__magic_name__ , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer snake_case : List[Any] = Trainer( model=__magic_name__ , args=__magic_name__ , train_dataset=__magic_name__ , eval_dataset=__magic_name__ , compute_metrics=__magic_name__ , data_collator=__magic_name__ , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation snake_case : Optional[int] = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) snake_case : Optional[Any] = trainer.evaluate() snake_case : Union[str, Any] = os.path.join(training_args.output_dir , '''eval_results.txt''' ) if trainer.is_world_master(): with open(__magic_name__ , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(''' %s = %s''' , __magic_name__ , __magic_name__ ) writer.write('''%s = %s\n''' % (key, value) ) results.update(__magic_name__ ) return results def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" main() if __name__ == "__main__": main()
84
1
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableDiffusionUpscalePipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class a_ ( unittest.TestCase ): def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : List[Any] = 1 snake_case : Optional[int] = 3 snake_case : Tuple = (32, 32) snake_case : List[str] = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(UpperCAmelCase__ ) return image @property def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" torch.manual_seed(0 ) snake_case : Tuple = UNetaDConditionModel( block_out_channels=(32, 32, 64) , layers_per_block=2 , sample_size=32 , in_channels=7 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , attention_head_dim=8 , use_linear_projection=UpperCAmelCase__ , only_cross_attention=(True, True, False) , num_class_embeds=100 , ) return model @property def lowerCAmelCase( self : Tuple ): """simple docstring""" torch.manual_seed(0 ) snake_case : Optional[Any] = AutoencoderKL( block_out_channels=[32, 32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) return model @property def lowerCAmelCase( self : int ): """simple docstring""" torch.manual_seed(0 ) snake_case : Tuple = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act='''gelu''' , projection_dim=512 , ) return CLIPTextModel(UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[str] = '''cpu''' # ensure determinism for the device-dependent torch.Generator snake_case : int = self.dummy_cond_unet_upscale snake_case : List[Any] = DDPMScheduler() snake_case : str = DDIMScheduler(prediction_type='''v_prediction''' ) snake_case : Any = self.dummy_vae snake_case : Union[str, Any] = self.dummy_text_encoder snake_case : Dict = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) snake_case : Union[str, Any] = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] snake_case : Optional[Any] = Image.fromarray(np.uinta(UpperCAmelCase__ ) ).convert('''RGB''' ).resize((64, 64) ) # make sure here that pndm scheduler skips prk snake_case : Optional[int] = StableDiffusionUpscalePipeline( unet=UpperCAmelCase__ , low_res_scheduler=UpperCAmelCase__ , scheduler=UpperCAmelCase__ , vae=UpperCAmelCase__ , text_encoder=UpperCAmelCase__ , tokenizer=UpperCAmelCase__ , max_noise_level=350 , ) snake_case : Any = sd_pipe.to(UpperCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) snake_case : List[str] = '''A painting of a squirrel eating a burger''' snake_case : Tuple = torch.Generator(device=UpperCAmelCase__ ).manual_seed(0 ) snake_case : List[str] = sd_pipe( [prompt] , image=UpperCAmelCase__ , generator=UpperCAmelCase__ , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='''np''' , ) snake_case : List[str] = output.images snake_case : Tuple = torch.Generator(device=UpperCAmelCase__ ).manual_seed(0 ) snake_case : Any = sd_pipe( [prompt] , image=UpperCAmelCase__ , generator=UpperCAmelCase__ , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='''np''' , return_dict=UpperCAmelCase__ , )[0] snake_case : str = image[0, -3:, -3:, -1] snake_case : Tuple = image_from_tuple[0, -3:, -3:, -1] snake_case : List[Any] = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) snake_case : Optional[Any] = np.array([0.3113, 0.3910, 0.4272, 0.4859, 0.5061, 0.4652, 0.5362, 0.5715, 0.5661] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : Optional[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator snake_case : Optional[int] = self.dummy_cond_unet_upscale snake_case : str = DDPMScheduler() snake_case : int = DDIMScheduler(prediction_type='''v_prediction''' ) snake_case : List[Any] = self.dummy_vae snake_case : Optional[Any] = self.dummy_text_encoder snake_case : Dict = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) snake_case : List[Any] = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] snake_case : int = Image.fromarray(np.uinta(UpperCAmelCase__ ) ).convert('''RGB''' ).resize((64, 64) ) # make sure here that pndm scheduler skips prk snake_case : Dict = StableDiffusionUpscalePipeline( unet=UpperCAmelCase__ , low_res_scheduler=UpperCAmelCase__ , scheduler=UpperCAmelCase__ , vae=UpperCAmelCase__ , text_encoder=UpperCAmelCase__ , tokenizer=UpperCAmelCase__ , max_noise_level=350 , ) snake_case : List[str] = sd_pipe.to(UpperCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) snake_case : str = '''A painting of a squirrel eating a burger''' snake_case : Any = sd_pipe( 2 * [prompt] , image=2 * [low_res_image] , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='''np''' , ) snake_case : Dict = output.images assert image.shape[0] == 2 snake_case : Optional[Any] = torch.Generator(device=UpperCAmelCase__ ).manual_seed(0 ) snake_case : Optional[int] = sd_pipe( [prompt] , image=UpperCAmelCase__ , generator=UpperCAmelCase__ , num_images_per_prompt=2 , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='''np''' , ) snake_case : Optional[Any] = output.images assert image.shape[0] == 2 @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Any = self.dummy_cond_unet_upscale snake_case : int = DDPMScheduler() snake_case : List[Any] = DDIMScheduler(prediction_type='''v_prediction''' ) snake_case : Tuple = self.dummy_vae snake_case : List[Any] = self.dummy_text_encoder snake_case : List[Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) snake_case : int = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] snake_case : List[str] = Image.fromarray(np.uinta(UpperCAmelCase__ ) ).convert('''RGB''' ).resize((64, 64) ) # put models in fp16, except vae as it overflows in fp16 snake_case : List[str] = unet.half() snake_case : Optional[int] = text_encoder.half() # make sure here that pndm scheduler skips prk snake_case : Union[str, Any] = StableDiffusionUpscalePipeline( unet=UpperCAmelCase__ , low_res_scheduler=UpperCAmelCase__ , scheduler=UpperCAmelCase__ , vae=UpperCAmelCase__ , text_encoder=UpperCAmelCase__ , tokenizer=UpperCAmelCase__ , max_noise_level=350 , ) snake_case : Tuple = sd_pipe.to(UpperCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) snake_case : Union[str, Any] = '''A painting of a squirrel eating a burger''' snake_case : int = torch.manual_seed(0 ) snake_case : Any = sd_pipe( [prompt] , image=UpperCAmelCase__ , generator=UpperCAmelCase__ , num_inference_steps=2 , output_type='''np''' , ).images snake_case : Tuple = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) @slow @require_torch_gpu class a_ ( unittest.TestCase ): def lowerCAmelCase( self : List[Any] ): """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : List[str] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-upscale/low_res_cat.png''' ) snake_case : Union[str, Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale''' '''/upsampled_cat.npy''' ) snake_case : List[Any] = '''stabilityai/stable-diffusion-x4-upscaler''' snake_case : Optional[int] = StableDiffusionUpscalePipeline.from_pretrained(UpperCAmelCase__ ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) pipe.enable_attention_slicing() snake_case : Any = '''a cat sitting on a park bench''' snake_case : Tuple = torch.manual_seed(0 ) snake_case : Any = pipe( prompt=UpperCAmelCase__ , image=UpperCAmelCase__ , generator=UpperCAmelCase__ , output_type='''np''' , ) snake_case : Optional[Any] = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1e-3 def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : List[str] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-upscale/low_res_cat.png''' ) snake_case : Optional[int] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale''' '''/upsampled_cat_fp16.npy''' ) snake_case : Any = '''stabilityai/stable-diffusion-x4-upscaler''' snake_case : int = StableDiffusionUpscalePipeline.from_pretrained( UpperCAmelCase__ , torch_dtype=torch.floataa , ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) pipe.enable_attention_slicing() snake_case : Tuple = '''a cat sitting on a park bench''' snake_case : int = torch.manual_seed(0 ) snake_case : List[Any] = pipe( prompt=UpperCAmelCase__ , image=UpperCAmelCase__ , generator=UpperCAmelCase__ , output_type='''np''' , ) snake_case : Optional[Any] = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5e-1 def lowerCAmelCase( self : Any ): """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() snake_case : Optional[int] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-upscale/low_res_cat.png''' ) snake_case : Dict = '''stabilityai/stable-diffusion-x4-upscaler''' snake_case : Optional[Any] = StableDiffusionUpscalePipeline.from_pretrained( UpperCAmelCase__ , torch_dtype=torch.floataa , ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() snake_case : List[str] = '''a cat sitting on a park bench''' snake_case : Optional[Any] = torch.manual_seed(0 ) snake_case : List[str] = pipe( prompt=UpperCAmelCase__ , image=UpperCAmelCase__ , generator=UpperCAmelCase__ , num_inference_steps=5 , output_type='''np''' , ) snake_case : Tuple = torch.cuda.max_memory_allocated() # make sure that less than 2.9 GB is allocated assert mem_bytes < 2.9 * 10**9
84
import re def a_ ( __magic_name__ ) -> bool: """simple docstring""" snake_case : List[str] = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(__magic_name__ , __magic_name__ ) ) if __name__ == "__main__": _a : Any = '0094702343221' print(is_sri_lankan_phone_number(phone))
84
1
_a : List[str] = 0 # The first color of the flag. _a : Any = 1 # The second color of the flag. _a : Union[str, Any] = 2 # The third color of the flag. _a : Union[str, Any] = (red, white, blue) def a_ ( __magic_name__ ) -> list: """simple docstring""" if not sequence: return [] if len(__magic_name__ ) == 1: return list(__magic_name__ ) snake_case : Tuple = 0 snake_case : Any = len(__magic_name__ ) - 1 snake_case : str = 0 while mid <= high: if sequence[mid] == colors[0]: snake_case , snake_case : Optional[Any] = sequence[mid], sequence[low] low += 1 mid += 1 elif sequence[mid] == colors[1]: mid += 1 elif sequence[mid] == colors[2]: snake_case , snake_case : List[str] = sequence[high], sequence[mid] high -= 1 else: snake_case : int = F"The elements inside the sequence must contains only {colors} values" raise ValueError(__magic_name__ ) return sequence if __name__ == "__main__": import doctest doctest.testmod() _a : Union[str, Any] = input('Enter numbers separated by commas:\n').strip() _a : Optional[Any] = [int(item.strip()) for item in user_input.split(',')] print(f"{dutch_national_flag_sort(unsorted)}")
84
import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class a_ ( unittest.TestCase ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[int]=7 , UpperCAmelCase__ : Union[str, Any]=3 , UpperCAmelCase__ : int=18 , UpperCAmelCase__ : Optional[int]=30 , UpperCAmelCase__ : Optional[int]=400 , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : int=True , ): """simple docstring""" snake_case : int = size if size is not None else {'''height''': 18, '''width''': 18} snake_case : Optional[Any] = parent snake_case : Any = batch_size snake_case : Any = num_channels snake_case : Union[str, Any] = image_size snake_case : Dict = min_resolution snake_case : Dict = max_resolution snake_case : int = do_resize snake_case : List[str] = size snake_case : List[Any] = apply_ocr def lowerCAmelCase( self : int ): """simple docstring""" return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class a_ ( a , unittest.TestCase ): A__ : List[Any] = LayoutLMvaImageProcessor if is_pytesseract_available() else None def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : Optional[Any] = LayoutLMvaImageProcessingTester(self ) @property def lowerCAmelCase( self : Dict ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[str] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase__ , '''do_resize''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''size''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''apply_ocr''' ) ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 18, '''width''': 18} ) snake_case : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) def lowerCAmelCase( self : str ): """simple docstring""" pass def lowerCAmelCase( self : Dict ): """simple docstring""" # Initialize image_processing snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , Image.Image ) # Test not batched input snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) self.assertIsInstance(encoding.words , UpperCAmelCase__ ) self.assertIsInstance(encoding.boxes , UpperCAmelCase__ ) # Test batched snake_case : Dict = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" # Initialize image_processing snake_case : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , numpify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , np.ndarray ) # Test not batched input snake_case : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched snake_case : List[Any] = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" # Initialize image_processing snake_case : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , torchify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , torch.Tensor ) # Test not batched input snake_case : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched snake_case : Tuple = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" # with apply_OCR = True snake_case : int = LayoutLMvaImageProcessor() from datasets import load_dataset snake_case : List[Any] = load_dataset('''hf-internal-testing/fixtures_docvqa''' , split='''test''' ) snake_case : List[Any] = Image.open(ds[0]['''file'''] ).convert('''RGB''' ) snake_case : Any = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 snake_case : Optional[Any] = [['''11:14''', '''to''', '''11:39''', '''a.m''', '''11:39''', '''to''', '''11:44''', '''a.m.''', '''11:44''', '''a.m.''', '''to''', '''12:25''', '''p.m.''', '''12:25''', '''to''', '''12:58''', '''p.m.''', '''12:58''', '''to''', '''4:00''', '''p.m.''', '''2:00''', '''to''', '''5:00''', '''p.m.''', '''Coffee''', '''Break''', '''Coffee''', '''will''', '''be''', '''served''', '''for''', '''men''', '''and''', '''women''', '''in''', '''the''', '''lobby''', '''adjacent''', '''to''', '''exhibit''', '''area.''', '''Please''', '''move''', '''into''', '''exhibit''', '''area.''', '''(Exhibits''', '''Open)''', '''TRRF''', '''GENERAL''', '''SESSION''', '''(PART''', '''|)''', '''Presiding:''', '''Lee''', '''A.''', '''Waller''', '''TRRF''', '''Vice''', '''President''', '''“Introductory''', '''Remarks”''', '''Lee''', '''A.''', '''Waller,''', '''TRRF''', '''Vice''', '''Presi-''', '''dent''', '''Individual''', '''Interviews''', '''with''', '''TRRF''', '''Public''', '''Board''', '''Members''', '''and''', '''Sci-''', '''entific''', '''Advisory''', '''Council''', '''Mem-''', '''bers''', '''Conducted''', '''by''', '''TRRF''', '''Treasurer''', '''Philip''', '''G.''', '''Kuehn''', '''to''', '''get''', '''answers''', '''which''', '''the''', '''public''', '''refrigerated''', '''warehousing''', '''industry''', '''is''', '''looking''', '''for.''', '''Plus''', '''questions''', '''from''', '''the''', '''floor.''', '''Dr.''', '''Emil''', '''M.''', '''Mrak,''', '''University''', '''of''', '''Cal-''', '''ifornia,''', '''Chairman,''', '''TRRF''', '''Board;''', '''Sam''', '''R.''', '''Cecil,''', '''University''', '''of''', '''Georgia''', '''College''', '''of''', '''Agriculture;''', '''Dr.''', '''Stanley''', '''Charm,''', '''Tufts''', '''University''', '''School''', '''of''', '''Medicine;''', '''Dr.''', '''Robert''', '''H.''', '''Cotton,''', '''ITT''', '''Continental''', '''Baking''', '''Company;''', '''Dr.''', '''Owen''', '''Fennema,''', '''University''', '''of''', '''Wis-''', '''consin;''', '''Dr.''', '''Robert''', '''E.''', '''Hardenburg,''', '''USDA.''', '''Questions''', '''and''', '''Answers''', '''Exhibits''', '''Open''', '''Capt.''', '''Jack''', '''Stoney''', '''Room''', '''TRRF''', '''Scientific''', '''Advisory''', '''Council''', '''Meeting''', '''Ballroom''', '''Foyer''']] # noqa: E231 snake_case : Union[str, Any] = [[[141, 57, 214, 69], [228, 58, 252, 69], [141, 75, 216, 88], [230, 79, 280, 88], [142, 260, 218, 273], [230, 261, 255, 273], [143, 279, 218, 290], [231, 282, 290, 291], [143, 342, 218, 354], [231, 345, 289, 355], [202, 362, 227, 373], [143, 379, 220, 392], [231, 382, 291, 394], [144, 714, 220, 726], [231, 715, 256, 726], [144, 732, 220, 745], [232, 736, 291, 747], [144, 769, 218, 782], [231, 770, 256, 782], [141, 788, 202, 801], [215, 791, 274, 804], [143, 826, 204, 838], [215, 826, 240, 838], [142, 844, 202, 857], [215, 847, 274, 859], [334, 57, 427, 69], [440, 57, 522, 69], [369, 75, 461, 88], [469, 75, 516, 88], [528, 76, 562, 88], [570, 76, 667, 88], [675, 75, 711, 87], [721, 79, 778, 88], [789, 75, 840, 88], [369, 97, 470, 107], [484, 94, 507, 106], [518, 94, 562, 107], [576, 94, 655, 110], [668, 94, 792, 109], [804, 95, 829, 107], [369, 113, 465, 125], [477, 116, 547, 125], [562, 113, 658, 125], [671, 116, 748, 125], [761, 113, 811, 125], [369, 131, 465, 143], [477, 133, 548, 143], [563, 130, 698, 145], [710, 130, 802, 146], [336, 171, 412, 183], [423, 171, 572, 183], [582, 170, 716, 184], [728, 171, 817, 187], [829, 171, 844, 186], [338, 197, 482, 212], [507, 196, 557, 209], [569, 196, 595, 208], [610, 196, 702, 209], [505, 214, 583, 226], [595, 214, 656, 227], [670, 215, 807, 227], [335, 259, 543, 274], [556, 259, 708, 272], [372, 279, 422, 291], [435, 279, 460, 291], [474, 279, 574, 292], [587, 278, 664, 291], [676, 278, 738, 291], [751, 279, 834, 291], [372, 298, 434, 310], [335, 341, 483, 354], [497, 341, 655, 354], [667, 341, 728, 354], [740, 341, 825, 354], [335, 360, 430, 372], [442, 360, 534, 372], [545, 359, 687, 372], [697, 360, 754, 372], [765, 360, 823, 373], [334, 378, 428, 391], [440, 378, 577, 394], [590, 378, 705, 391], [720, 378, 801, 391], [334, 397, 400, 409], [370, 416, 529, 429], [544, 416, 576, 432], [587, 416, 665, 428], [677, 416, 814, 429], [372, 435, 452, 450], [465, 434, 495, 447], [511, 434, 600, 447], [611, 436, 637, 447], [649, 436, 694, 451], [705, 438, 824, 447], [369, 453, 452, 466], [464, 454, 509, 466], [522, 453, 611, 469], [625, 453, 792, 469], [370, 472, 556, 488], [570, 472, 684, 487], [697, 472, 718, 485], [732, 472, 835, 488], [369, 490, 411, 503], [425, 490, 484, 503], [496, 490, 635, 506], [645, 490, 707, 503], [718, 491, 761, 503], [771, 490, 840, 503], [336, 510, 374, 521], [388, 510, 447, 522], [460, 510, 489, 521], [503, 510, 580, 522], [592, 509, 736, 525], [745, 509, 770, 522], [781, 509, 840, 522], [338, 528, 434, 541], [448, 528, 596, 541], [609, 527, 687, 540], [700, 528, 792, 541], [336, 546, 397, 559], [407, 546, 431, 559], [443, 546, 525, 560], [537, 546, 680, 562], [688, 546, 714, 559], [722, 546, 837, 562], [336, 565, 449, 581], [461, 565, 485, 577], [497, 565, 665, 581], [681, 565, 718, 577], [732, 565, 837, 580], [337, 584, 438, 597], [452, 583, 521, 596], [535, 584, 677, 599], [690, 583, 787, 596], [801, 583, 825, 596], [338, 602, 478, 615], [492, 602, 530, 614], [543, 602, 638, 615], [650, 602, 676, 614], [688, 602, 788, 615], [802, 602, 843, 614], [337, 621, 502, 633], [516, 621, 615, 637], [629, 621, 774, 636], [789, 621, 827, 633], [337, 639, 418, 652], [432, 640, 571, 653], [587, 639, 731, 655], [743, 639, 769, 652], [780, 639, 841, 652], [338, 658, 440, 673], [455, 658, 491, 670], [508, 658, 602, 671], [616, 658, 638, 670], [654, 658, 835, 674], [337, 677, 429, 689], [337, 714, 482, 726], [495, 714, 548, 726], [561, 714, 683, 726], [338, 770, 461, 782], [474, 769, 554, 785], [489, 788, 562, 803], [576, 788, 643, 801], [656, 787, 751, 804], [764, 788, 844, 801], [334, 825, 421, 838], [430, 824, 574, 838], [584, 824, 723, 841], [335, 844, 450, 857], [464, 843, 583, 860], [628, 862, 755, 875], [769, 861, 848, 878]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , UpperCAmelCase__ ) self.assertListEqual(encoding.boxes , UpperCAmelCase__ ) # with apply_OCR = False snake_case : str = LayoutLMvaImageProcessor(apply_ocr=UpperCAmelCase__ ) snake_case : Optional[Any] = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) )
84
1
def a_ ( __magic_name__ ) -> Any: """simple docstring""" snake_case : Dict = [0] * len(__magic_name__ ) snake_case : List[Any] = [] snake_case : Optional[int] = [1] * len(__magic_name__ ) for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(__magic_name__ ) ): if indegree[i] == 0: queue.append(__magic_name__ ) while queue: snake_case : Dict = queue.pop(0 ) for x in graph[vertex]: indegree[x] -= 1 if long_dist[vertex] + 1 > long_dist[x]: snake_case : List[Any] = long_dist[vertex] + 1 if indegree[x] == 0: queue.append(__magic_name__ ) print(max(__magic_name__ ) ) # Adjacency list of Graph _a : Optional[int] = {0: [2, 3, 4], 1: [2, 7], 2: [5], 3: [5, 7], 4: [7], 5: [6], 6: [7], 7: []} longest_distance(graph)
84
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _a : Tuple = logging.get_logger(__name__) # pylint: disable=invalid-name _a : Dict = '\n Examples:\n ```py\n >>> import torch\n >>> import numpy as np\n\n >>> from diffusers import KandinskyV22PriorPipeline, KandinskyV22ControlnetPipeline\n >>> from transformers import pipeline\n >>> from diffusers.utils import load_image\n\n\n >>> def make_hint(image, depth_estimator):\n ... image = depth_estimator(image)["depth"]\n ... image = np.array(image)\n ... image = image[:, :, None]\n ... image = np.concatenate([image, image, image], axis=2)\n ... detected_map = torch.from_numpy(image).float() / 255.0\n ... hint = detected_map.permute(2, 0, 1)\n ... return hint\n\n\n >>> depth_estimator = pipeline("depth-estimation")\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16\n ... )\n >>> pipe_prior = pipe_prior.to("cuda")\n\n >>> pipe = KandinskyV22ControlnetPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-controlnet-depth", torch_dtype=torch.float16\n ... )\n >>> pipe = pipe.to("cuda")\n\n\n >>> img = load_image(\n ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main"\n ... "/kandinsky/cat.png"\n ... ).resize((768, 768))\n\n >>> hint = make_hint(img, depth_estimator).unsqueeze(0).half().to("cuda")\n\n >>> prompt = "A robot, 4k photo"\n >>> negative_prior_prompt = "lowres, text, error, cropped, worst quality, low quality, jpeg artifacts, ugly, duplicate, morbid, mutilated, out of frame, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, mutation, deformed, blurry, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, username, watermark, signature"\n\n >>> generator = torch.Generator(device="cuda").manual_seed(43)\n\n >>> image_emb, zero_image_emb = pipe_prior(\n ... prompt=prompt, negative_prompt=negative_prior_prompt, generator=generator\n ... ).to_tuple()\n\n >>> images = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... hint=hint,\n ... num_inference_steps=50,\n ... generator=generator,\n ... height=768,\n ... width=768,\n ... ).images\n\n >>> images[0].save("robot_cat.png")\n ```\n' def a_ ( __magic_name__ , __magic_name__ , __magic_name__=8 ) -> str: """simple docstring""" snake_case : List[Any] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 snake_case : Tuple = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class a_ ( a ): def __init__( self : Optional[int] , UpperCAmelCase__ : UNetaDConditionModel , UpperCAmelCase__ : DDPMScheduler , UpperCAmelCase__ : VQModel , ): """simple docstring""" super().__init__() self.register_modules( unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ , movq=UpperCAmelCase__ , ) snake_case : List[Any] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def lowerCAmelCase( self : int , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Any ): """simple docstring""" if latents is None: snake_case : int = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=UpperCAmelCase__ , dtype=UpperCAmelCase__ ) else: if latents.shape != shape: raise ValueError(F"Unexpected latents shape, got {latents.shape}, expected {shape}" ) snake_case : Optional[Any] = latents.to(UpperCAmelCase__ ) snake_case : List[Any] = latents * scheduler.init_noise_sigma return latents def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Optional[int]=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) snake_case : Union[str, Any] = torch.device(F"cuda:{gpu_id}" ) snake_case : Dict = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : Any=0 ): """simple docstring""" if is_accelerate_available() and is_accelerate_version('''>=''' , '''0.17.0.dev0''' ): from accelerate import cpu_offload_with_hook else: raise ImportError('''`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.''' ) snake_case : Optional[int] = torch.device(F"cuda:{gpu_id}" ) if self.device.type != "cpu": self.to('''cpu''' , silence_dtype_warnings=UpperCAmelCase__ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) snake_case : List[str] = None for cpu_offloaded_model in [self.unet, self.movq]: snake_case , snake_case : Optional[int] = cpu_offload_with_hook(UpperCAmelCase__ , UpperCAmelCase__ , prev_module_hook=UpperCAmelCase__ ) # We'll offload the last model manually. snake_case : Tuple = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" if not hasattr(self.unet , '''_hf_hook''' ): return self.device for module in self.unet.modules(): if ( hasattr(UpperCAmelCase__ , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(UpperCAmelCase__ ) def __call__( self : List[str] , UpperCAmelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase__ : torch.FloatTensor , UpperCAmelCase__ : int = 512 , UpperCAmelCase__ : int = 512 , UpperCAmelCase__ : int = 100 , UpperCAmelCase__ : float = 4.0 , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase__ : Optional[torch.FloatTensor] = None , UpperCAmelCase__ : Optional[str] = "pil" , UpperCAmelCase__ : bool = True , ): """simple docstring""" snake_case : Optional[int] = self._execution_device snake_case : Union[str, Any] = guidance_scale > 1.0 if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Any = torch.cat(UpperCAmelCase__ , dim=0 ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Union[str, Any] = torch.cat(UpperCAmelCase__ , dim=0 ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : int = torch.cat(UpperCAmelCase__ , dim=0 ) snake_case : List[Any] = image_embeds.shape[0] * num_images_per_prompt if do_classifier_free_guidance: snake_case : Dict = image_embeds.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Optional[Any] = negative_image_embeds.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Tuple = hint.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Optional[Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=UpperCAmelCase__ ) snake_case : List[str] = torch.cat([hint, hint] , dim=0 ).to(dtype=self.unet.dtype , device=UpperCAmelCase__ ) self.scheduler.set_timesteps(UpperCAmelCase__ , device=UpperCAmelCase__ ) snake_case : str = self.scheduler.timesteps snake_case : Optional[Any] = self.movq.config.latent_channels snake_case , snake_case : Optional[Any] = downscale_height_and_width(UpperCAmelCase__ , UpperCAmelCase__ , self.movq_scale_factor ) # create initial latent snake_case : Dict = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , self.scheduler , ) for i, t in enumerate(self.progress_bar(UpperCAmelCase__ ) ): # expand the latents if we are doing classifier free guidance snake_case : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents snake_case : Optional[int] = {'''image_embeds''': image_embeds, '''hint''': hint} snake_case : Any = self.unet( sample=UpperCAmelCase__ , timestep=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , added_cond_kwargs=UpperCAmelCase__ , return_dict=UpperCAmelCase__ , )[0] if do_classifier_free_guidance: snake_case , snake_case : Dict = noise_pred.split(latents.shape[1] , dim=1 ) snake_case , snake_case : Any = noise_pred.chunk(2 ) snake_case , snake_case : Dict = variance_pred.chunk(2 ) snake_case : str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) snake_case : List[str] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , '''variance_type''' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): snake_case , snake_case : Tuple = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 snake_case : List[Any] = self.scheduler.step( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , generator=UpperCAmelCase__ , )[0] # post-processing snake_case : List[Any] = self.movq.decode(UpperCAmelCase__ , force_not_quantize=UpperCAmelCase__ )['''sample'''] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}" ) if output_type in ["np", "pil"]: snake_case : Optional[Any] = image * 0.5 + 0.5 snake_case : int = image.clamp(0 , 1 ) snake_case : List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": snake_case : str = self.numpy_to_pil(UpperCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase__ )
84
1
def a_ ( __magic_name__ = 1_000 ) -> int: """simple docstring""" snake_case : Dict = -1 snake_case : Tuple = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c snake_case : Union[str, Any] = (n * n - 2 * a * n) // (2 * n - 2 * a) snake_case : str = n - a - b if c * c == (a * a + b * b): snake_case : Tuple = a * b * c if candidate >= product: snake_case : Optional[int] = candidate return product if __name__ == "__main__": print(f"{solution() = }")
84
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _a : List[Any] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class a_ ( a , unittest.TestCase ): A__ : Dict = ReformerTokenizer A__ : Optional[int] = ReformerTokenizerFast A__ : str = True A__ : Tuple = False A__ : str = True def lowerCAmelCase( self : List[Any] ): """simple docstring""" super().setUp() snake_case : str = ReformerTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : int = '''<s>''' snake_case : List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase__ ) , UpperCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase__ ) , UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Any = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(UpperCAmelCase__ ) , 1_000 ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def lowerCAmelCase( self : Dict ): """simple docstring""" if not self.test_rust_tokenizer: return snake_case : Any = self.get_tokenizer() snake_case : str = self.get_rust_tokenizer() snake_case : Tuple = '''I was born in 92000, and this is falsé.''' snake_case : str = tokenizer.tokenize(UpperCAmelCase__ ) snake_case : int = rust_tokenizer.tokenize(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : Union[str, Any] = tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) snake_case : List[str] = rust_tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : List[str] = self.get_rust_tokenizer() snake_case : Optional[int] = tokenizer.encode(UpperCAmelCase__ ) snake_case : Optional[Any] = rust_tokenizer.encode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : List[Any]=15 ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): snake_case : str = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) # Simple input snake_case : Union[str, Any] = '''This is a simple input''' snake_case : List[str] = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case : int = ('''This is a simple input''', '''This is a pair''') snake_case : int = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Simple input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Simple input self.assertRaises( UpperCAmelCase__ , tokenizer_r.batch_encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' , ) # Pair input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Pair input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Pair input self.assertRaises( UpperCAmelCase__ , tokenizer_r.batch_encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' , ) def lowerCAmelCase( self : str ): """simple docstring""" pass def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Union[str, Any] = ReformerTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) snake_case : List[str] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(UpperCAmelCase__ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , [285, 46, 10, 170, 382] , ) snake_case : int = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) snake_case : int = tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) snake_case : List[Any] = tokenizer.convert_ids_to_tokens(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def lowerCAmelCase( self : Tuple ): """simple docstring""" return ReformerTokenizer.from_pretrained('''google/reformer-crime-and-punishment''' ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : Any = '''Hello World!''' snake_case : Optional[Any] = [126, 32, 262, 152, 38, 72, 287] self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @slow def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[Any] = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) snake_case : Dict = [ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @require_torch @slow def lowerCAmelCase( self : List[Any] ): """simple docstring""" import torch from transformers import ReformerConfig, ReformerModel # Build sequence snake_case : Any = list(self.big_tokenizer.get_vocab().keys() )[:10] snake_case : Union[str, Any] = ''' '''.join(UpperCAmelCase__ ) snake_case : Optional[int] = self.big_tokenizer.encode_plus(UpperCAmelCase__ , return_tensors='''pt''' ) snake_case : List[str] = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors='''pt''' ) snake_case : Optional[Any] = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) snake_case : Tuple = encoded_sequence['''input_ids'''].shape snake_case : List[Any] = ReformerModel(UpperCAmelCase__ ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**UpperCAmelCase__ ) model(**UpperCAmelCase__ ) @slow def lowerCAmelCase( self : Optional[int] ): """simple docstring""" # fmt: off snake_case : Tuple = {'''input_ids''': [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 snake_case : Tuple = [ '''This is a very simple sentence.''', '''The quick brown fox jumps over the lazy dog.''', ] self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase__ , model_name='''google/reformer-crime-and-punishment''' , revision='''0e6c3decb8211d49bf881013425dc8b0448b3f5a''' , padding=UpperCAmelCase__ , sequences=UpperCAmelCase__ , )
84
1
import os import sys _a : Any = os.path.join(os.path.dirname(__file__), 'src') sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) _a : Tuple = [ 'torch', 'numpy', 'tokenizers', 'filelock', 'requests', 'tqdm', 'regex', 'sentencepiece', 'sacremoses', 'importlib_metadata', 'huggingface_hub', ] @add_start_docstrings(AutoConfig.__doc__ ) def a_ ( *__magic_name__ , **__magic_name__ ) -> int: """simple docstring""" return AutoConfig.from_pretrained(*__magic_name__ , **__magic_name__ ) @add_start_docstrings(AutoTokenizer.__doc__ ) def a_ ( *__magic_name__ , **__magic_name__ ) -> Any: """simple docstring""" return AutoTokenizer.from_pretrained(*__magic_name__ , **__magic_name__ ) @add_start_docstrings(AutoModel.__doc__ ) def a_ ( *__magic_name__ , **__magic_name__ ) -> Any: """simple docstring""" return AutoModel.from_pretrained(*__magic_name__ , **__magic_name__ ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def a_ ( *__magic_name__ , **__magic_name__ ) -> int: """simple docstring""" return AutoModelForCausalLM.from_pretrained(*__magic_name__ , **__magic_name__ ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def a_ ( *__magic_name__ , **__magic_name__ ) -> Optional[int]: """simple docstring""" return AutoModelForMaskedLM.from_pretrained(*__magic_name__ , **__magic_name__ ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def a_ ( *__magic_name__ , **__magic_name__ ) -> Any: """simple docstring""" return AutoModelForSequenceClassification.from_pretrained(*__magic_name__ , **__magic_name__ ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def a_ ( *__magic_name__ , **__magic_name__ ) -> Tuple: """simple docstring""" return AutoModelForQuestionAnswering.from_pretrained(*__magic_name__ , **__magic_name__ )
84
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def a_ ( __magic_name__ ) -> Tuple: """simple docstring""" snake_case , snake_case : Any = image.size snake_case , snake_case : List[str] = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 snake_case : List[str] = image.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) snake_case : Dict = np.array(__magic_name__ ).astype(np.floataa ) / 255.0 snake_case : Optional[Any] = image[None].transpose(0 , 3 , 1 , 2 ) snake_case : Tuple = torch.from_numpy(__magic_name__ ) return 2.0 * image - 1.0 class a_ ( a ): def __init__( self : Optional[Any] , UpperCAmelCase__ : VQModel , UpperCAmelCase__ : UNetaDModel , UpperCAmelCase__ : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): """simple docstring""" super().__init__() self.register_modules(vqvae=UpperCAmelCase__ , unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ ) @torch.no_grad() def __call__( self : Any , UpperCAmelCase__ : Union[torch.Tensor, PIL.Image.Image] = None , UpperCAmelCase__ : Optional[int] = 1 , UpperCAmelCase__ : Optional[int] = 100 , UpperCAmelCase__ : Optional[float] = 0.0 , UpperCAmelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase__ : Optional[str] = "pil" , UpperCAmelCase__ : bool = True , ): """simple docstring""" if isinstance(UpperCAmelCase__ , PIL.Image.Image ): snake_case : Optional[int] = 1 elif isinstance(UpperCAmelCase__ , torch.Tensor ): snake_case : Any = image.shape[0] else: raise ValueError(F"`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(UpperCAmelCase__ )}" ) if isinstance(UpperCAmelCase__ , PIL.Image.Image ): snake_case : Optional[Any] = preprocess(UpperCAmelCase__ ) snake_case , snake_case : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image snake_case : List[Any] = (batch_size, self.unet.config.in_channels // 2, height, width) snake_case : str = next(self.unet.parameters() ).dtype snake_case : Dict = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=self.device , dtype=UpperCAmelCase__ ) snake_case : Any = image.to(device=self.device , dtype=UpperCAmelCase__ ) # set timesteps and move to the correct device self.scheduler.set_timesteps(UpperCAmelCase__ , device=self.device ) snake_case : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler snake_case : Union[str, Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] snake_case : Any = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) snake_case : Optional[Any] = {} if accepts_eta: snake_case : Dict = eta for t in self.progress_bar(UpperCAmelCase__ ): # concat latents and low resolution image in the channel dimension. snake_case : Optional[int] = torch.cat([latents, image] , dim=1 ) snake_case : str = self.scheduler.scale_model_input(UpperCAmelCase__ , UpperCAmelCase__ ) # predict the noise residual snake_case : int = self.unet(UpperCAmelCase__ , UpperCAmelCase__ ).sample # compute the previous noisy sample x_t -> x_t-1 snake_case : Any = self.scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample # decode the image latents with the VQVAE snake_case : Optional[int] = self.vqvae.decode(UpperCAmelCase__ ).sample snake_case : int = torch.clamp(UpperCAmelCase__ , -1.0 , 1.0 ) snake_case : Dict = image / 2 + 0.5 snake_case : int = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case : Any = self.numpy_to_pil(UpperCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase__ )
84
1
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class a_ ( a ): def __init__( self : Any , UpperCAmelCase__ : Any , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Dict = params snake_case : int = np.array(UpperCAmelCase__ ) snake_case : Optional[Any] = np.array([len(UpperCAmelCase__ ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : Tuple , UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" return (self.token_ids[index], self.lengths[index]) def __len__( self : Optional[Any] ): """simple docstring""" return len(self.lengths ) def lowerCAmelCase( self : int ): """simple docstring""" assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : Optional[int] = self.params.max_model_input_size snake_case : List[Any] = self.lengths > max_len logger.info(F"Splitting {sum(UpperCAmelCase__ )} too long sequences." ) def divide_chunks(UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[int] ): return [l[i : i + n] for i in range(0 , len(UpperCAmelCase__ ) , UpperCAmelCase__ )] snake_case : Optional[int] = [] snake_case : str = [] if self.params.mlm: snake_case , snake_case : Optional[Any] = self.params.special_tok_ids['''cls_token'''], self.params.special_tok_ids['''sep_token'''] else: snake_case , snake_case : Optional[Any] = self.params.special_tok_ids['''bos_token'''], self.params.special_tok_ids['''eos_token'''] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: snake_case : Union[str, Any] = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: snake_case : int = np.insert(UpperCAmelCase__ , 0 , UpperCAmelCase__ ) if sub_s[-1] != sep_id: snake_case : Dict = np.insert(UpperCAmelCase__ , len(UpperCAmelCase__ ) , UpperCAmelCase__ ) assert len(UpperCAmelCase__ ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(UpperCAmelCase__ ) new_tok_ids.extend(UpperCAmelCase__ ) new_lengths.extend([len(UpperCAmelCase__ ) for l in sub_seqs] ) snake_case : Optional[Any] = np.array(UpperCAmelCase__ ) snake_case : Dict = np.array(UpperCAmelCase__ ) def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : str = len(self ) snake_case : List[Any] = self.lengths > 11 snake_case : Any = self.token_ids[indices] snake_case : Optional[Any] = self.lengths[indices] snake_case : Dict = len(self ) logger.info(F"Remove {init_size - new_size} too short (<=11 tokens) sequences." ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" if "unk_token" not in self.params.special_tok_ids: return else: snake_case : Optional[Any] = self.params.special_tok_ids['''unk_token'''] snake_case : str = len(self ) snake_case : List[str] = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) snake_case : int = (unk_occs / self.lengths) < 0.5 snake_case : str = self.token_ids[indices] snake_case : List[Any] = self.lengths[indices] snake_case : List[Any] = len(self ) logger.info(F"Remove {init_size - new_size} sequences with a high level of unknown tokens (50%)." ) def lowerCAmelCase( self : Tuple ): """simple docstring""" if not self.params.is_master: return logger.info(F"{len(self )} sequences" ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : Any ): """simple docstring""" snake_case : Union[str, Any] = [t[0] for t in batch] snake_case : Any = [t[1] for t in batch] assert len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) # Max for paddings snake_case : Any = max(UpperCAmelCase__ ) # Pad token ids if self.params.mlm: snake_case : List[str] = self.params.special_tok_ids['''pad_token'''] else: snake_case : Optional[int] = self.params.special_tok_ids['''unk_token'''] snake_case : Optional[int] = [list(t.astype(UpperCAmelCase__ ) ) + [pad_idx] * (max_seq_len_ - len(UpperCAmelCase__ )) for t in token_ids] assert len(tk_ ) == len(UpperCAmelCase__ ) assert all(len(UpperCAmelCase__ ) == max_seq_len_ for t in tk_ ) snake_case : Dict = torch.tensor(tk_ ) # (bs, max_seq_len_) snake_case : List[str] = torch.tensor(UpperCAmelCase__ ) # (bs) return tk_t, lg_t
84
import os import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers.models.realm.configuration_realm import RealmConfig from transformers.models.realm.retrieval_realm import _REALM_BLOCK_RECORDS_FILENAME, RealmRetriever from transformers.models.realm.tokenization_realm import VOCAB_FILES_NAMES, RealmTokenizer class a_ ( a ): def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[Any] = tempfile.mkdtemp() snake_case : Dict = 5 # Realm tok snake_case : str = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''test''', '''question''', '''this''', '''is''', '''the''', '''first''', '''second''', '''third''', '''fourth''', '''fifth''', '''record''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] snake_case : Tuple = os.path.join(self.tmpdirname , '''realm_tokenizer''' ) os.makedirs(UpperCAmelCase__ , exist_ok=UpperCAmelCase__ ) snake_case : Any = os.path.join(UpperCAmelCase__ , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) snake_case : Tuple = os.path.join(self.tmpdirname , '''realm_block_records''' ) os.makedirs(UpperCAmelCase__ , exist_ok=UpperCAmelCase__ ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" return RealmTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''realm_tokenizer''' ) ) def lowerCAmelCase( self : Any ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Any = RealmConfig(num_block_records=self.num_block_records ) return config def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Optional[int] = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''question''': ['''foo''', '''bar'''], '''answers''': [['''Foo''', '''Bar'''], ['''Bar''']], } ) return dataset def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Dict = np.array( [ b'''This is the first record''', b'''This is the second record''', b'''This is the third record''', b'''This is the fourth record''', b'''This is the fifth record''', b'''This is a longer longer longer record''', ] , dtype=UpperCAmelCase__ , ) return block_records def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Tuple = RealmRetriever( block_records=self.get_dummy_block_records() , tokenizer=self.get_tokenizer() , ) return retriever def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[str] = self.get_config() snake_case : Optional[Any] = self.get_dummy_retriever() snake_case : Optional[int] = retriever.tokenizer snake_case : Dict = np.array([0, 3] , dtype='''long''' ) snake_case : Optional[int] = tokenizer(['''Test question'''] ).input_ids snake_case : Union[str, Any] = tokenizer( ['''the fourth'''] , add_special_tokens=UpperCAmelCase__ , return_token_type_ids=UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , ).input_ids snake_case : Optional[Any] = config.reader_seq_len snake_case , snake_case , snake_case , snake_case : List[str] = retriever( UpperCAmelCase__ , UpperCAmelCase__ , answer_ids=UpperCAmelCase__ , max_length=UpperCAmelCase__ , return_tensors='''np''' ) self.assertEqual(len(UpperCAmelCase__ ) , 2 ) self.assertEqual(len(UpperCAmelCase__ ) , 2 ) self.assertEqual(len(UpperCAmelCase__ ) , 2 ) self.assertEqual(concat_inputs.input_ids.shape , (2, 10) ) self.assertEqual(concat_inputs.attention_mask.shape , (2, 10) ) self.assertEqual(concat_inputs.token_type_ids.shape , (2, 10) ) self.assertEqual(concat_inputs.special_tokens_mask.shape , (2, 10) ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[0] ) , ['''[CLS]''', '''test''', '''question''', '''[SEP]''', '''this''', '''is''', '''the''', '''first''', '''record''', '''[SEP]'''] , ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[1] ) , ['''[CLS]''', '''test''', '''question''', '''[SEP]''', '''this''', '''is''', '''the''', '''fourth''', '''record''', '''[SEP]'''] , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[Any] = self.get_config() snake_case : Optional[int] = self.get_dummy_retriever() snake_case : List[str] = retriever.tokenizer snake_case : Optional[Any] = np.array([0, 3, 5] , dtype='''long''' ) snake_case : Optional[int] = tokenizer(['''Test question'''] ).input_ids snake_case : Any = tokenizer( ['''the fourth''', '''longer longer'''] , add_special_tokens=UpperCAmelCase__ , return_token_type_ids=UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , ).input_ids snake_case : List[Any] = config.reader_seq_len snake_case , snake_case , snake_case , snake_case : Union[str, Any] = retriever( UpperCAmelCase__ , UpperCAmelCase__ , answer_ids=UpperCAmelCase__ , max_length=UpperCAmelCase__ , return_tensors='''np''' ) self.assertEqual([False, True, True] , UpperCAmelCase__ ) self.assertEqual([[-1, -1, -1], [6, -1, -1], [6, 7, 8]] , UpperCAmelCase__ ) self.assertEqual([[-1, -1, -1], [7, -1, -1], [7, 8, 9]] , UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : int = self.get_dummy_retriever() retriever.save_pretrained(os.path.join(self.tmpdirname , '''realm_block_records''' ) ) # Test local path snake_case : Optional[Any] = retriever.from_pretrained(os.path.join(self.tmpdirname , '''realm_block_records''' ) ) self.assertEqual(retriever.block_records[0] , b'''This is the first record''' ) # Test mocked remote path with patch('''transformers.models.realm.retrieval_realm.hf_hub_download''' ) as mock_hf_hub_download: snake_case : Any = os.path.join( os.path.join(self.tmpdirname , '''realm_block_records''' ) , _REALM_BLOCK_RECORDS_FILENAME ) snake_case : Any = RealmRetriever.from_pretrained('''google/realm-cc-news-pretrained-openqa''' ) self.assertEqual(retriever.block_records[0] , b'''This is the first record''' )
84
1
def a_ ( __magic_name__ ) -> bool: """simple docstring""" if p < 2: raise ValueError('''p should not be less than 2!''' ) elif p == 2: return True snake_case : int = 4 snake_case : Optional[Any] = (1 << p) - 1 for _ in range(p - 2 ): snake_case : Optional[Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
84
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _a : str = {'configuration_encoder_decoder': ['EncoderDecoderConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = ['EncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Dict = ['TFEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = ['FlaxEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys _a : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
84
1
import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" if "cls_token" in name: snake_case : Tuple = name.replace('''cls_token''' , '''vit.embeddings.cls_token''' ) if "mask_token" in name: snake_case : Optional[int] = name.replace('''mask_token''' , '''decoder.mask_token''' ) if "decoder_pos_embed" in name: snake_case : List[str] = name.replace('''decoder_pos_embed''' , '''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: snake_case : List[str] = name.replace('''pos_embed''' , '''vit.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: snake_case : List[Any] = name.replace('''patch_embed.proj''' , '''vit.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: snake_case : int = name.replace('''patch_embed.norm''' , '''vit.embeddings.norm''' ) if "decoder_blocks" in name: snake_case : int = name.replace('''decoder_blocks''' , '''decoder.decoder_layers''' ) if "blocks" in name: snake_case : Optional[Any] = name.replace('''blocks''' , '''vit.encoder.layer''' ) if "attn.proj" in name: snake_case : str = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: snake_case : Any = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: snake_case : List[str] = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: snake_case : Tuple = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: snake_case : Tuple = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: snake_case : Tuple = name.replace('''mlp.fc2''' , '''output.dense''' ) if "decoder_embed" in name: snake_case : Dict = name.replace('''decoder_embed''' , '''decoder.decoder_embed''' ) if "decoder_norm" in name: snake_case : Dict = name.replace('''decoder_norm''' , '''decoder.decoder_norm''' ) if "decoder_pred" in name: snake_case : Dict = name.replace('''decoder_pred''' , '''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name: snake_case : Optional[int] = name.replace('''norm.weight''' , '''vit.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name: snake_case : List[str] = name.replace('''norm.bias''' , '''vit.layernorm.bias''' ) return name def a_ ( __magic_name__ , __magic_name__ ) -> str: """simple docstring""" for key in orig_state_dict.copy().keys(): snake_case : Union[str, Any] = orig_state_dict.pop(__magic_name__ ) if "qkv" in key: snake_case : Optional[int] = key.split('''.''' ) snake_case : int = int(key_split[1] ) if "decoder_blocks" in key: snake_case : List[str] = config.decoder_hidden_size snake_case : List[Any] = '''decoder.decoder_layers.''' if "weight" in key: snake_case : str = val[:dim, :] snake_case : Optional[Any] = val[dim : dim * 2, :] snake_case : Any = val[-dim:, :] elif "bias" in key: snake_case : Optional[Any] = val[:dim] snake_case : List[Any] = val[dim : dim * 2] snake_case : List[Any] = val[-dim:] else: snake_case : Optional[int] = config.hidden_size snake_case : Tuple = '''vit.encoder.layer.''' if "weight" in key: snake_case : Optional[Any] = val[:dim, :] snake_case : str = val[dim : dim * 2, :] snake_case : Union[str, Any] = val[-dim:, :] elif "bias" in key: snake_case : Tuple = val[:dim] snake_case : int = val[dim : dim * 2] snake_case : Optional[Any] = val[-dim:] else: snake_case : Optional[Any] = val return orig_state_dict def a_ ( __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" snake_case : List[str] = ViTMAEConfig() if "large" in checkpoint_url: snake_case : str = 1_024 snake_case : Tuple = 4_096 snake_case : Optional[Any] = 24 snake_case : List[Any] = 16 elif "huge" in checkpoint_url: snake_case : Tuple = 14 snake_case : int = 1_280 snake_case : Dict = 5_120 snake_case : Tuple = 32 snake_case : Optional[Any] = 16 snake_case : Optional[Any] = ViTMAEForPreTraining(__magic_name__ ) snake_case : Optional[Any] = torch.hub.load_state_dict_from_url(__magic_name__ , map_location='''cpu''' )['''model'''] snake_case : int = ViTMAEImageProcessor(size=config.image_size ) snake_case : Dict = convert_state_dict(__magic_name__ , __magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() snake_case : Tuple = '''https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg''' snake_case : List[Any] = Image.open(requests.get(__magic_name__ , stream=__magic_name__ ).raw ) snake_case : Dict = ViTMAEImageProcessor(size=config.image_size ) snake_case : str = image_processor(images=__magic_name__ , return_tensors='''pt''' ) # forward pass torch.manual_seed(2 ) snake_case : Union[str, Any] = model(**__magic_name__ ) snake_case : Optional[Any] = outputs.logits if "large" in checkpoint_url: snake_case : Any = torch.tensor( [[-0.7309, -0.7128, -1.0169], [-1.0161, -0.9058, -1.1878], [-1.0478, -0.9411, -1.1911]] ) elif "huge" in checkpoint_url: snake_case : List[Any] = torch.tensor( [[-1.1599, -0.9199, -1.2221], [-1.1952, -0.9269, -1.2307], [-1.2143, -0.9337, -1.2262]] ) else: snake_case : Dict = torch.tensor( [[-0.9192, -0.8481, -1.1259], [-1.1349, -1.0034, -1.2599], [-1.1757, -1.0429, -1.2726]] ) # verify logits assert torch.allclose(logits[0, :3, :3] , __magic_name__ , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(__magic_name__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__magic_name__ ) if __name__ == "__main__": _a : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth', type=str, help='URL of the checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _a : str = parser.parse_args() convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
84
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer _a : str = logging.get_logger(__name__) _a : Optional[int] = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} _a : Optional[Any] = { 'vocab_file': { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json' ), }, } _a : Union[str, Any] = { 'yjernite/retribert-base-uncased': 512, } _a : Tuple = { 'yjernite/retribert-base-uncased': {'do_lower_case': True}, } class a_ ( a ): A__ : List[str] = VOCAB_FILES_NAMES A__ : Any = PRETRAINED_VOCAB_FILES_MAP A__ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Any = PRETRAINED_INIT_CONFIGURATION A__ : Optional[Any] = RetriBertTokenizer A__ : Any = ['input_ids', 'attention_mask'] def __init__( self : Optional[int] , UpperCAmelCase__ : Any=None , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Dict="[UNK]" , UpperCAmelCase__ : str="[SEP]" , UpperCAmelCase__ : Union[str, Any]="[PAD]" , UpperCAmelCase__ : Dict="[CLS]" , UpperCAmelCase__ : Optional[Any]="[MASK]" , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Optional[int]=None , **UpperCAmelCase__ : Dict , ): """simple docstring""" super().__init__( UpperCAmelCase__ , tokenizer_file=UpperCAmelCase__ , do_lower_case=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , tokenize_chinese_chars=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ , **UpperCAmelCase__ , ) snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , UpperCAmelCase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' , UpperCAmelCase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , UpperCAmelCase__ ) != tokenize_chinese_chars ): snake_case : int = getattr(UpperCAmelCase__ , normalizer_state.pop('''type''' ) ) snake_case : List[Any] = do_lower_case snake_case : Union[str, Any] = strip_accents snake_case : int = tokenize_chinese_chars snake_case : int = normalizer_class(**UpperCAmelCase__ ) snake_case : Union[str, Any] = do_lower_case def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Union[str, Any]=None ): """simple docstring""" snake_case : str = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ): """simple docstring""" snake_case : List[Any] = [self.sep_token_id] snake_case : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCAmelCase( self : str , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ): """simple docstring""" snake_case : Tuple = self._tokenizer.model.save(UpperCAmelCase__ , name=UpperCAmelCase__ ) return tuple(UpperCAmelCase__ )
84
1
import argparse import copy def a_ ( __magic_name__ ) -> Optional[Any]: """simple docstring""" snake_case : List[str] = {} with open(__magic_name__ ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: snake_case : Optional[Any] = [] _list.append([line.split()[1], line.split()[2]] ) snake_case : int = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: snake_case : Any = [] _list.append([line.split()[0], line.split()[2]] ) snake_case : str = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def a_ ( __magic_name__ , __magic_name__ ) -> Optional[Any]: """simple docstring""" with open(__magic_name__ ) as f: snake_case : Optional[int] = f.read(1 ) snake_case : List[Any] = start_node snake_case : Any = [] snake_case : List[Any] = start_node snake_case : List[Any] = 0 while visiting not in first_solution: snake_case : Union[str, Any] = 10_000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(__magic_name__ ) and k[0] not in first_solution: snake_case : Union[str, Any] = k[1] snake_case : List[str] = k[0] first_solution.append(__magic_name__ ) snake_case : str = distance_of_first_solution + int(__magic_name__ ) snake_case : List[Any] = best_node first_solution.append(__magic_name__ ) snake_case : int = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 snake_case : Union[str, Any] = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 10_000 ) return first_solution, distance_of_first_solution def a_ ( __magic_name__ , __magic_name__ ) -> Dict: """simple docstring""" snake_case : Optional[int] = [] for n in solution[1:-1]: snake_case : Any = solution.index(__magic_name__ ) for kn in solution[1:-1]: snake_case : List[Any] = solution.index(__magic_name__ ) if n == kn: continue snake_case : Tuple = copy.deepcopy(__magic_name__ ) snake_case : List[str] = kn snake_case : str = n snake_case : Dict = 0 for k in _tmp[:-1]: snake_case : str = _tmp[_tmp.index(__magic_name__ ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: snake_case : Dict = distance + int(i[1] ) _tmp.append(__magic_name__ ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) snake_case : Tuple = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda __magic_name__ : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" snake_case : int = 1 snake_case : Dict = first_solution snake_case : Union[str, Any] = [] snake_case : List[str] = distance_of_first_solution snake_case : Any = solution while count <= iters: snake_case : Union[str, Any] = find_neighborhood(__magic_name__ , __magic_name__ ) snake_case : int = 0 snake_case : List[str] = neighborhood[index_of_best_solution] snake_case : List[str] = len(__magic_name__ ) - 1 snake_case : List[str] = False while not found: snake_case : Dict = 0 while i < len(__magic_name__ ): if best_solution[i] != solution[i]: snake_case : Optional[Any] = best_solution[i] snake_case : Optional[Any] = solution[i] break snake_case : str = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) snake_case : List[str] = True snake_case : str = best_solution[:-1] snake_case : str = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: snake_case : List[Any] = cost snake_case : List[Any] = solution else: snake_case : Tuple = index_of_best_solution + 1 snake_case : int = neighborhood[index_of_best_solution] if len(__magic_name__ ) >= size: tabu_list.pop(0 ) snake_case : List[Any] = count + 1 return best_solution_ever, best_cost def a_ ( __magic_name__=None ) -> List[Any]: """simple docstring""" snake_case : Tuple = generate_neighbours(args.File ) snake_case , snake_case : List[Any] = generate_first_solution( args.File , __magic_name__ ) snake_case , snake_case : List[Any] = tabu_search( __magic_name__ , __magic_name__ , __magic_name__ , args.Iterations , args.Size , ) print(F"Best solution: {best_sol}, with total distance: {best_cost}." ) if __name__ == "__main__": _a : str = argparse.ArgumentParser(description='Tabu Search') parser.add_argument( '-f', '--File', type=str, help='Path to the file containing the data', required=True, ) parser.add_argument( '-i', '--Iterations', type=int, help='How many iterations the algorithm should perform', required=True, ) parser.add_argument( '-s', '--Size', type=int, help='Size of the tabu list', required=True ) # Pass the arguments to main method main(parser.parse_args())
84
import string import numpy def a_ ( __magic_name__ , __magic_name__ ) -> int: """simple docstring""" return b if a == 0 else greatest_common_divisor(b % a , __magic_name__ ) class a_ : A__ : List[Any] = string.ascii_uppercase + string.digits # This cipher takes alphanumerics into account # i.e. a total of 36 characters # take x and return x % len(key_string) A__ : List[str] = numpy.vectorize(lambda a : x % 36 ) A__ : Dict = numpy.vectorize(a ) def __init__( self : List[str] , UpperCAmelCase__ : numpy.ndarray ): """simple docstring""" snake_case : int = self.modulus(UpperCAmelCase__ ) # mod36 calc's on the encrypt key self.check_determinant() # validate the determinant of the encryption key snake_case : List[str] = encrypt_key.shape[0] def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : str ): """simple docstring""" return self.key_string.index(UpperCAmelCase__ ) def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : int ): """simple docstring""" return self.key_string[round(UpperCAmelCase__ )] def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[Any] = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: snake_case : Tuple = det % len(self.key_string ) snake_case : Tuple = len(self.key_string ) if greatest_common_divisor(UpperCAmelCase__ , len(self.key_string ) ) != 1: snake_case : List[Any] = ( F"determinant modular {req_l} of encryption key({det}) " F"is not co prime w.r.t {req_l}.\nTry another key." ) raise ValueError(UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Optional[int] = [char for char in text.upper() if char in self.key_string] snake_case : Optional[int] = chars[-1] while len(UpperCAmelCase__ ) % self.break_key != 0: chars.append(UpperCAmelCase__ ) return "".join(UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Optional[int] = self.process_text(text.upper() ) snake_case : Optional[int] = '''''' for i in range(0 , len(UpperCAmelCase__ ) - self.break_key + 1 , self.break_key ): snake_case : int = text[i : i + self.break_key] snake_case : int = [self.replace_letters(UpperCAmelCase__ ) for char in batch] snake_case : Tuple = numpy.array([vec] ).T snake_case : Optional[Any] = self.modulus(self.encrypt_key.dot(UpperCAmelCase__ ) ).T.tolist()[ 0 ] snake_case : Dict = ''''''.join( self.replace_digits(UpperCAmelCase__ ) for num in batch_encrypted ) encrypted += encrypted_batch return encrypted def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Optional[int] = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: snake_case : int = det % len(self.key_string ) snake_case : Dict = None for i in range(len(self.key_string ) ): if (det * i) % len(self.key_string ) == 1: snake_case : Any = i break snake_case : Any = ( det_inv * numpy.linalg.det(self.encrypt_key ) * numpy.linalg.inv(self.encrypt_key ) ) return self.to_int(self.modulus(UpperCAmelCase__ ) ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Any = self.make_decrypt_key() snake_case : Optional[Any] = self.process_text(text.upper() ) snake_case : int = '''''' for i in range(0 , len(UpperCAmelCase__ ) - self.break_key + 1 , self.break_key ): snake_case : Any = text[i : i + self.break_key] snake_case : int = [self.replace_letters(UpperCAmelCase__ ) for char in batch] snake_case : List[str] = numpy.array([vec] ).T snake_case : Optional[Any] = self.modulus(decrypt_key.dot(UpperCAmelCase__ ) ).T.tolist()[0] snake_case : int = ''''''.join( self.replace_digits(UpperCAmelCase__ ) for num in batch_decrypted ) decrypted += decrypted_batch return decrypted def a_ ( ) -> None: """simple docstring""" snake_case : Any = int(input('''Enter the order of the encryption key: ''' ) ) snake_case : List[Any] = [] print('''Enter each row of the encryption key with space separated integers''' ) for _ in range(__magic_name__ ): snake_case : Optional[Any] = [int(__magic_name__ ) for x in input().split()] hill_matrix.append(__magic_name__ ) snake_case : List[str] = HillCipher(numpy.array(__magic_name__ ) ) print('''Would you like to encrypt or decrypt some text? (1 or 2)''' ) snake_case : int = input('''\n1. Encrypt\n2. Decrypt\n''' ) if option == "1": snake_case : List[Any] = input('''What text would you like to encrypt?: ''' ) print('''Your encrypted text is:''' ) print(hc.encrypt(__magic_name__ ) ) elif option == "2": snake_case : int = input('''What text would you like to decrypt?: ''' ) print('''Your decrypted text is:''' ) print(hc.decrypt(__magic_name__ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
84
1
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: _a : Any = None _a : Union[str, Any] = logging.get_logger(__name__) _a : Dict = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} _a : int = { 'vocab_file': { 'facebook/mbart-large-en-ro': ( 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model' ), 'facebook/mbart-large-cc25': ( 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model' ), }, 'tokenizer_file': { 'facebook/mbart-large-en-ro': 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json', 'facebook/mbart-large-cc25': 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json', }, } _a : List[Any] = { 'facebook/mbart-large-en-ro': 1_024, 'facebook/mbart-large-cc25': 1_024, } # fmt: off _a : Dict = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN'] class a_ ( a ): A__ : Optional[int] = VOCAB_FILES_NAMES A__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : str = PRETRAINED_VOCAB_FILES_MAP A__ : int = ['input_ids', 'attention_mask'] A__ : Union[str, Any] = MBartTokenizer A__ : List[int] = [] A__ : List[int] = [] def __init__( self : Any , UpperCAmelCase__ : str=None , UpperCAmelCase__ : Any=None , UpperCAmelCase__ : Union[str, Any]="<s>" , UpperCAmelCase__ : str="</s>" , UpperCAmelCase__ : int="</s>" , UpperCAmelCase__ : Any="<s>" , UpperCAmelCase__ : str="<unk>" , UpperCAmelCase__ : str="<pad>" , UpperCAmelCase__ : str="<mask>" , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : List[str]=None , **UpperCAmelCase__ : str , ): """simple docstring""" # Mask token behave like a normal word, i.e. include the space before it snake_case : List[str] = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else mask_token super().__init__( vocab_file=UpperCAmelCase__ , tokenizer_file=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , src_lang=UpperCAmelCase__ , tgt_lang=UpperCAmelCase__ , additional_special_tokens=UpperCAmelCase__ , **UpperCAmelCase__ , ) snake_case : List[str] = vocab_file snake_case : Tuple = False if not self.vocab_file else True snake_case : Optional[Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'''additional_special_tokens''': _additional_special_tokens} ) snake_case : Tuple = { lang_code: self.convert_tokens_to_ids(UpperCAmelCase__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } snake_case : Union[str, Any] = src_lang if src_lang is not None else '''en_XX''' snake_case : Optional[int] = self.convert_tokens_to_ids(self._src_lang ) snake_case : int = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" return self._src_lang @src_lang.setter def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Optional[int] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ): """simple docstring""" if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ): """simple docstring""" snake_case : str = [self.sep_token_id] snake_case : Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] , UpperCAmelCase__ : Optional[str] , **UpperCAmelCase__ : List[Any] ): """simple docstring""" if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) snake_case : str = src_lang snake_case : Optional[int] = self(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , **UpperCAmelCase__ ) snake_case : str = self.convert_tokens_to_ids(UpperCAmelCase__ ) snake_case : str = tgt_lang_id return inputs def lowerCAmelCase( self : Any , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : str = "en_XX" , UpperCAmelCase__ : Optional[List[str]] = None , UpperCAmelCase__ : str = "ro_RO" , **UpperCAmelCase__ : Any , ): """simple docstring""" snake_case : List[str] = src_lang snake_case : Optional[int] = tgt_lang return super().prepare_seqaseq_batch(UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Dict ): """simple docstring""" return self.set_src_lang_special_tokens(self.src_lang ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" return self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : Tuple ): """simple docstring""" snake_case : Tuple = self.convert_tokens_to_ids(UpperCAmelCase__ ) snake_case : Union[str, Any] = [] snake_case : Optional[int] = [self.eos_token_id, self.cur_lang_code] snake_case : Optional[Any] = self.convert_ids_to_tokens(self.prefix_tokens ) snake_case : Tuple = self.convert_ids_to_tokens(self.suffix_tokens ) snake_case : Dict = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : str ): """simple docstring""" snake_case : str = self.convert_tokens_to_ids(UpperCAmelCase__ ) snake_case : Optional[Any] = [] snake_case : List[str] = [self.eos_token_id, self.cur_lang_code] snake_case : Any = self.convert_ids_to_tokens(self.prefix_tokens ) snake_case : Any = self.convert_ids_to_tokens(self.suffix_tokens ) snake_case : Optional[Any] = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(UpperCAmelCase__ ): logger.error(F"Vocabulary path ({save_directory}) should be a directory." ) return snake_case : Any = os.path.join( UpperCAmelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase__ ): copyfile(self.vocab_file , UpperCAmelCase__ ) return (out_vocab_file,)
84
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class a_ ( a ): A__ : List[Any] = 'Salesforce/blip-image-captioning-base' A__ : Dict = ( 'This is a tool that generates a description of an image. It takes an input named `image` which should be the ' 'image to caption, and returns a text that contains the description in English.' ) A__ : str = 'image_captioner' A__ : Dict = AutoModelForVisionaSeq A__ : Optional[Any] = ['image'] A__ : List[str] = ['text'] def __init__( self : List[str] , *UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" requires_backends(self , ['''vision'''] ) super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : "Image" ): """simple docstring""" return self.pre_processor(images=UpperCAmelCase__ , return_tensors='''pt''' ) def lowerCAmelCase( self : Any , UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" return self.model.generate(**UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : List[Any] ): """simple docstring""" return self.pre_processor.batch_decode(UpperCAmelCase__ , skip_special_tokens=UpperCAmelCase__ )[0].strip()
84
1
def a_ ( __magic_name__ ) -> bool: """simple docstring""" snake_case : List[str] = [int(__magic_name__ ) for i in ip_va_address.split('''.''' ) if i.isdigit()] return len(__magic_name__ ) == 4 and all(0 <= int(__magic_name__ ) <= 254 for octet in octets ) if __name__ == "__main__": _a : List[Any] = input().strip() _a : Any = 'valid' if is_ip_va_address_valid(ip) else 'invalid' print(f"{ip} is a {valid_or_invalid} IP v4 address.")
84
def a_ ( __magic_name__ ) -> bool: """simple docstring""" if p < 2: raise ValueError('''p should not be less than 2!''' ) elif p == 2: return True snake_case : int = 4 snake_case : Optional[Any] = (1 << p) - 1 for _ in range(p - 2 ): snake_case : Optional[Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
84
1
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model _a : List[str] = '0.12' # assumed parallelism: 8 if is_torch_available(): import torch def a_ ( __magic_name__ , __magic_name__ , __magic_name__=None ) -> int: """simple docstring""" if rng is None: snake_case : Union[str, Any] = random.Random() snake_case : str = 1 for dim in shape: total_dims *= dim snake_case : Optional[Any] = [] for _ in range(__magic_name__ ): values.append(rng.randint(0 , vocab_size - 1 ) ) snake_case : List[str] = np.array(__magic_name__ , dtype=jnp.intaa ).reshape(__magic_name__ ) return output def a_ ( __magic_name__ , __magic_name__=None ) -> str: """simple docstring""" snake_case : Union[str, Any] = ids_tensor(__magic_name__ , vocab_size=2 , rng=__magic_name__ ) # make sure that at least one token is attended to for each batch snake_case : str = 1 return attn_mask @require_flax class a_ : A__ : str = None A__ : Union[str, Any] = () def lowerCAmelCase( self : int ): """simple docstring""" snake_case , snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 snake_case : Tuple = 2 snake_case : str = inputs['''input_ids'''].shape[-1] // 2 snake_case : Optional[int] = inputs['''input_ids'''][:max_batch_size, :sequence_length] snake_case : int = jnp.ones_like(UpperCAmelCase__ ) snake_case : List[Any] = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens snake_case : Dict = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` snake_case : Optional[Any] = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case , snake_case , snake_case , snake_case : Optional[Any] = self._get_input_ids_and_config() snake_case : int = False snake_case : Optional[Any] = max_length snake_case : str = 0 for model_class in self.all_generative_model_classes: snake_case : List[str] = model_class(UpperCAmelCase__ ) snake_case : int = model_class.__name__[4:] # Skip the "Flax" at the beginning snake_case : Optional[int] = getattr(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : Dict = pt_model_class(UpperCAmelCase__ ).eval() snake_case : Union[str, Any] = load_flax_weights_in_pytorch_model(UpperCAmelCase__ , flax_model.params ) snake_case : Optional[int] = flax_model.generate(UpperCAmelCase__ ).sequences snake_case : Tuple = pt_model.generate(torch.tensor(UpperCAmelCase__ , dtype=torch.long ) ) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: snake_case : List[Any] = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() , flax_generation_outputs.tolist() ) def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case , snake_case , snake_case , snake_case : List[Any] = self._get_input_ids_and_config() snake_case : Dict = False snake_case : List[str] = max_length for model_class in self.all_generative_model_classes: snake_case : List[str] = model_class(UpperCAmelCase__ ) snake_case : int = model.generate(UpperCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase__ ) snake_case : Union[str, Any] = jit(model.generate ) snake_case : Union[str, Any] = jit_generate(UpperCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def lowerCAmelCase( self : int ): """simple docstring""" snake_case , snake_case , snake_case , snake_case : Optional[int] = self._get_input_ids_and_config() snake_case : List[Any] = True snake_case : List[str] = max_length for model_class in self.all_generative_model_classes: snake_case : Union[str, Any] = model_class(UpperCAmelCase__ ) snake_case : Tuple = model.generate(UpperCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase__ ) snake_case : Dict = jit(model.generate ) snake_case : int = jit_generate(UpperCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case , snake_case , snake_case , snake_case : Union[str, Any] = self._get_input_ids_and_config() snake_case : str = False snake_case : Any = max_length snake_case : str = 2 for model_class in self.all_generative_model_classes: snake_case : List[str] = model_class(UpperCAmelCase__ ) snake_case : Union[str, Any] = model.generate(UpperCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase__ ) snake_case : Tuple = jit(model.generate ) snake_case : int = jit_generate(UpperCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def lowerCAmelCase( self : Any ): """simple docstring""" snake_case , snake_case , snake_case , snake_case : Dict = self._get_input_ids_and_config() snake_case : Dict = False snake_case : int = max_length snake_case : str = 2 snake_case : str = 2 for model_class in self.all_generative_model_classes: snake_case : List[str] = model_class(UpperCAmelCase__ ) snake_case : str = model.generate(UpperCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[0] , input_ids.shape[0] * config.num_return_sequences ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case , snake_case , snake_case : Optional[Any] = self._get_input_ids_and_config() snake_case : Union[str, Any] = True snake_case : Optional[int] = max_length snake_case : Union[str, Any] = 0.8 snake_case : Optional[Any] = 10 snake_case : Tuple = 0.3 snake_case : int = 1 snake_case : str = 8 snake_case : Dict = 9 for model_class in self.all_generative_model_classes: snake_case : Tuple = model_class(UpperCAmelCase__ ) snake_case : int = model.generate(UpperCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase__ ) snake_case : int = jit(model.generate ) snake_case : str = jit_generate(UpperCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case , snake_case , snake_case , snake_case : int = self._get_input_ids_and_config() snake_case : str = max_length snake_case : int = 1 snake_case : Union[str, Any] = 8 snake_case : Optional[Any] = 9 for model_class in self.all_generative_model_classes: snake_case : Union[str, Any] = model_class(UpperCAmelCase__ ) snake_case : Tuple = model.generate(UpperCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase__ ) snake_case : int = jit(model.generate ) snake_case : Dict = jit_generate(UpperCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def lowerCAmelCase( self : str ): """simple docstring""" snake_case , snake_case , snake_case , snake_case : Any = self._get_input_ids_and_config() snake_case : List[Any] = max_length snake_case : List[str] = 2 snake_case : List[Any] = 1 snake_case : str = 8 snake_case : Optional[Any] = 9 for model_class in self.all_generative_model_classes: snake_case : List[Any] = model_class(UpperCAmelCase__ ) snake_case : Any = model.generate(UpperCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase__ ) snake_case : Tuple = jit(model.generate ) snake_case : Optional[int] = jit_generate(UpperCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case , snake_case , snake_case , snake_case : List[Any] = self._get_input_ids_and_config() # pad attention mask on the left snake_case : int = attention_mask.at[(0, 0)].set(0 ) snake_case : Any = False snake_case : List[Any] = max_length for model_class in self.all_generative_model_classes: snake_case : Optional[Any] = model_class(UpperCAmelCase__ ) snake_case : Union[str, Any] = model.generate(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase__ ) snake_case : str = jit(model.generate ) snake_case : Optional[Any] = jit_generate(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def lowerCAmelCase( self : Any ): """simple docstring""" snake_case , snake_case , snake_case , snake_case : Dict = self._get_input_ids_and_config() # pad attention mask on the left snake_case : Any = attention_mask.at[(0, 0)].set(0 ) snake_case : Dict = True snake_case : Any = max_length for model_class in self.all_generative_model_classes: snake_case : Optional[Any] = model_class(UpperCAmelCase__ ) snake_case : str = model.generate(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase__ ) snake_case : Tuple = jit(model.generate ) snake_case : Optional[int] = jit_generate(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def lowerCAmelCase( self : str ): """simple docstring""" snake_case , snake_case , snake_case , snake_case : Any = self._get_input_ids_and_config() # pad attention mask on the left snake_case : List[Any] = attention_mask.at[(0, 0)].set(0 ) snake_case : Any = 2 snake_case : List[str] = max_length for model_class in self.all_generative_model_classes: snake_case : Tuple = model_class(UpperCAmelCase__ ) snake_case : List[str] = model.generate(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] , UpperCAmelCase__ ) snake_case : Tuple = jit(model.generate ) snake_case : Optional[Any] = jit_generate(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) @require_flax class a_ ( unittest.TestCase ): def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : List[str] = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-bert''' ) snake_case : int = FlaxAutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) snake_case : Any = '''Hello world''' snake_case : Optional[Any] = tokenizer(UpperCAmelCase__ , return_tensors='''np''' ).input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(UpperCAmelCase__ , '''do_samples''' ): model.generate(UpperCAmelCase__ , do_samples=UpperCAmelCase__ ) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(UpperCAmelCase__ , '''foo''' ): snake_case : Any = {'''foo''': '''bar'''} model.generate(UpperCAmelCase__ , **UpperCAmelCase__ )
84
from sklearn.metrics import fa_score import datasets _a : List[str] = '\nThe F1 score is the harmonic mean of the precision and recall. It can be computed with the equation:\nF1 = 2 * (precision * recall) / (precision + recall)\n' _a : Dict = '\nArgs:\n predictions (`list` of `int`): Predicted labels.\n references (`list` of `int`): Ground truth labels.\n labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None.\n pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1.\n average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`.\n\n - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary.\n - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives.\n - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.\n - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall.\n - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification).\n sample_weight (`list` of `float`): Sample weights Defaults to None.\n\nReturns:\n f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better.\n\nExamples:\n\n Example 1-A simple binary example\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0])\n >>> print(results)\n {\'f1\': 0.5}\n\n Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0)\n >>> print(round(results[\'f1\'], 2))\n 0.67\n\n Example 3-The same simple binary example as in Example 1, but with `sample_weight` included.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3])\n >>> print(round(results[\'f1\'], 2))\n 0.35\n\n Example 4-A multiclass example, with different values for the `average` input.\n >>> predictions = [0, 2, 1, 0, 0, 1]\n >>> references = [0, 1, 2, 0, 1, 2]\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro")\n >>> print(round(results[\'f1\'], 2))\n 0.33\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=None)\n >>> print(results)\n {\'f1\': array([0.8, 0. , 0. ])}\n' _a : List[Any] = '\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P.\n and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and\n Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.},\n journal={Journal of Machine Learning Research},\n volume={12},\n pages={2825--2830},\n year={2011}\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): def lowerCAmelCase( self : Any ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''int32''' ) ), '''references''': datasets.Sequence(datasets.Value('''int32''' ) ), } if self.config_name == '''multilabel''' else { '''predictions''': datasets.Value('''int32''' ), '''references''': datasets.Value('''int32''' ), } ) , reference_urls=['''https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html'''] , ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Optional[Any]=1 , UpperCAmelCase__ : List[str]="binary" , UpperCAmelCase__ : str=None ): """simple docstring""" snake_case : List[Any] = fa_score( UpperCAmelCase__ , UpperCAmelCase__ , labels=UpperCAmelCase__ , pos_label=UpperCAmelCase__ , average=UpperCAmelCase__ , sample_weight=UpperCAmelCase__ ) return {"f1": float(UpperCAmelCase__ ) if score.size == 1 else score}
84
1
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def a_ ( __magic_name__ ) -> Tuple: """simple docstring""" snake_case , snake_case : Any = image.size snake_case , snake_case : List[str] = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 snake_case : List[str] = image.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) snake_case : Dict = np.array(__magic_name__ ).astype(np.floataa ) / 255.0 snake_case : Optional[Any] = image[None].transpose(0 , 3 , 1 , 2 ) snake_case : Tuple = torch.from_numpy(__magic_name__ ) return 2.0 * image - 1.0 class a_ ( a ): def __init__( self : Optional[Any] , UpperCAmelCase__ : VQModel , UpperCAmelCase__ : UNetaDModel , UpperCAmelCase__ : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): """simple docstring""" super().__init__() self.register_modules(vqvae=UpperCAmelCase__ , unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ ) @torch.no_grad() def __call__( self : Any , UpperCAmelCase__ : Union[torch.Tensor, PIL.Image.Image] = None , UpperCAmelCase__ : Optional[int] = 1 , UpperCAmelCase__ : Optional[int] = 100 , UpperCAmelCase__ : Optional[float] = 0.0 , UpperCAmelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase__ : Optional[str] = "pil" , UpperCAmelCase__ : bool = True , ): """simple docstring""" if isinstance(UpperCAmelCase__ , PIL.Image.Image ): snake_case : Optional[int] = 1 elif isinstance(UpperCAmelCase__ , torch.Tensor ): snake_case : Any = image.shape[0] else: raise ValueError(F"`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(UpperCAmelCase__ )}" ) if isinstance(UpperCAmelCase__ , PIL.Image.Image ): snake_case : Optional[Any] = preprocess(UpperCAmelCase__ ) snake_case , snake_case : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image snake_case : List[Any] = (batch_size, self.unet.config.in_channels // 2, height, width) snake_case : str = next(self.unet.parameters() ).dtype snake_case : Dict = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=self.device , dtype=UpperCAmelCase__ ) snake_case : Any = image.to(device=self.device , dtype=UpperCAmelCase__ ) # set timesteps and move to the correct device self.scheduler.set_timesteps(UpperCAmelCase__ , device=self.device ) snake_case : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler snake_case : Union[str, Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] snake_case : Any = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) snake_case : Optional[Any] = {} if accepts_eta: snake_case : Dict = eta for t in self.progress_bar(UpperCAmelCase__ ): # concat latents and low resolution image in the channel dimension. snake_case : Optional[int] = torch.cat([latents, image] , dim=1 ) snake_case : str = self.scheduler.scale_model_input(UpperCAmelCase__ , UpperCAmelCase__ ) # predict the noise residual snake_case : int = self.unet(UpperCAmelCase__ , UpperCAmelCase__ ).sample # compute the previous noisy sample x_t -> x_t-1 snake_case : Any = self.scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample # decode the image latents with the VQVAE snake_case : Optional[int] = self.vqvae.decode(UpperCAmelCase__ ).sample snake_case : int = torch.clamp(UpperCAmelCase__ , -1.0 , 1.0 ) snake_case : Dict = image / 2 + 0.5 snake_case : int = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case : Any = self.numpy_to_pil(UpperCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase__ )
84
def a_ ( __magic_name__ ) -> int: """simple docstring""" if not isinstance(__magic_name__ , __magic_name__ ): raise TypeError('''only integers accepted as input''' ) else: snake_case : str = str(abs(__magic_name__ ) ) snake_case : Optional[Any] = [list(__magic_name__ ) for char in range(len(__magic_name__ ) )] for index in range(len(__magic_name__ ) ): num_transpositions[index].pop(__magic_name__ ) return max( int(''''''.join(list(__magic_name__ ) ) ) for transposition in num_transpositions ) if __name__ == "__main__": __import__('doctest').testmod()
84
1
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class a_ ( a ): def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Optional[Any] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCAmelCase__ , '''tf_padding''' ) ) self.parent.assertTrue(hasattr(UpperCAmelCase__ , '''depth_multiplier''' ) ) class a_ : def __init__( self : Tuple , UpperCAmelCase__ : str , UpperCAmelCase__ : int=13 , UpperCAmelCase__ : Union[str, Any]=3 , UpperCAmelCase__ : Optional[Any]=32 , UpperCAmelCase__ : Optional[int]=0.25 , UpperCAmelCase__ : Dict=8 , UpperCAmelCase__ : List[Any]=8 , UpperCAmelCase__ : Optional[int]=6 , UpperCAmelCase__ : List[str]=32 , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Any=True , UpperCAmelCase__ : Any=True , UpperCAmelCase__ : Optional[int]="relu6" , UpperCAmelCase__ : Dict=1_280 , UpperCAmelCase__ : Any=0.1 , UpperCAmelCase__ : List[str]=0.02 , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Optional[int]=10 , UpperCAmelCase__ : Any=None , ): """simple docstring""" snake_case : List[str] = parent snake_case : Optional[Any] = batch_size snake_case : Tuple = num_channels snake_case : Optional[Any] = image_size snake_case : Optional[int] = depth_multiplier snake_case : Any = depth_divisible_by snake_case : Optional[int] = min_depth snake_case : List[str] = expand_ratio snake_case : Union[str, Any] = tf_padding snake_case : Optional[Any] = output_stride snake_case : Any = first_layer_is_expansion snake_case : Tuple = finegrained_output snake_case : str = hidden_act snake_case : Dict = last_hidden_size if finegrained_output else int(last_hidden_size * depth_multiplier ) snake_case : Union[str, Any] = classifier_dropout_prob snake_case : str = use_labels snake_case : Union[str, Any] = is_training snake_case : int = num_labels snake_case : Tuple = initializer_range snake_case : Dict = scope def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case : Union[str, Any] = None snake_case : List[str] = None if self.use_labels: snake_case : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) snake_case : Optional[int] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) snake_case : Tuple = self.get_config() return config, pixel_values, labels, pixel_labels def lowerCAmelCase( self : Dict ): """simple docstring""" return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , depth_divisible_by=self.depth_divisible_by , min_depth=self.min_depth , expand_ratio=self.expand_ratio , output_stride=self.output_stride , first_layer_is_expansion=self.first_layer_is_expansion , finegrained_output=self.finegrained_output , hidden_act=self.hidden_act , tf_padding=self.tf_padding , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def lowerCAmelCase( self : Any , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Any , UpperCAmelCase__ : int ): """simple docstring""" snake_case : Dict = MobileNetVaModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : str = model(UpperCAmelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) self.parent.assertEqual( result.pooler_output.shape , (self.batch_size, self.last_hidden_size) , ) def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : Tuple = self.num_labels snake_case : Any = MobileNetVaForImageClassification(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : str = model(UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Any , UpperCAmelCase__ : Any ): """simple docstring""" snake_case : Dict = self.num_labels snake_case : Dict = MobileNetVaForSemanticSegmentation(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : Dict = model(UpperCAmelCase__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) snake_case : Dict = model(UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : str = self.prepare_config_and_inputs() snake_case , snake_case , snake_case , snake_case : List[Any] = config_and_inputs snake_case : Optional[int] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class a_ ( a , a , unittest.TestCase ): A__ : Union[str, Any] = ( (MobileNetVaModel, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation) if is_torch_available() else () ) A__ : Any = ( { 'feature-extraction': MobileNetVaModel, 'image-classification': MobileNetVaForImageClassification, 'image-segmentation': MobileNetVaForSemanticSegmentation, } if is_torch_available() else {} ) A__ : Tuple = False A__ : Tuple = False A__ : Tuple = False A__ : List[str] = False def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Any = MobileNetVaModelTester(self ) snake_case : Optional[Any] = MobileNetVaConfigTester(self , config_class=UpperCAmelCase__ , has_text_modality=UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''MobileNetV2 does not use inputs_embeds''' ) def lowerCAmelCase( self : int ): """simple docstring""" pass @unittest.skip(reason='''MobileNetV2 does not support input and output embeddings''' ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" pass @unittest.skip(reason='''MobileNetV2 does not output attentions''' ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" pass def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case , snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Optional[Any] = model_class(UpperCAmelCase__ ) snake_case : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case : Optional[int] = [*signature.parameters.keys()] snake_case : List[str] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , UpperCAmelCase__ ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) def lowerCAmelCase( self : Any ): """simple docstring""" def check_hidden_states_output(UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Dict ): snake_case : Any = model_class(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() with torch.no_grad(): snake_case : List[str] = model(**self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) ) snake_case : Optional[int] = outputs.hidden_states snake_case : Any = 16 self.assertEqual(len(UpperCAmelCase__ ) , UpperCAmelCase__ ) snake_case , snake_case : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Tuple = True check_hidden_states_output(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case : List[str] = True check_hidden_states_output(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*UpperCAmelCase__ ) @slow def lowerCAmelCase( self : int ): """simple docstring""" for model_name in MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case : str = MobileNetVaModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) def a_ ( ) -> Dict: """simple docstring""" snake_case : int = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class a_ ( unittest.TestCase ): @cached_property def lowerCAmelCase( self : str ): """simple docstring""" return ( MobileNetVaImageProcessor.from_pretrained('''google/mobilenet_v2_1.0_224''' ) if is_vision_available() else None ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : Optional[int] = MobileNetVaForImageClassification.from_pretrained('''google/mobilenet_v2_1.0_224''' ).to(UpperCAmelCase__ ) snake_case : Dict = self.default_image_processor snake_case : Tuple = prepare_img() snake_case : Union[str, Any] = image_processor(images=UpperCAmelCase__ , return_tensors='''pt''' ).to(UpperCAmelCase__ ) # forward pass with torch.no_grad(): snake_case : Dict = model(**UpperCAmelCase__ ) # verify the logits snake_case : int = torch.Size((1, 1_001) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase__ ) snake_case : List[Any] = torch.tensor([0.2445, -1.1993, 0.1905] ).to(UpperCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase__ , atol=1e-4 ) ) @slow def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Tuple = MobileNetVaForSemanticSegmentation.from_pretrained('''google/deeplabv3_mobilenet_v2_1.0_513''' ) snake_case : Any = model.to(UpperCAmelCase__ ) snake_case : Optional[int] = MobileNetVaImageProcessor.from_pretrained('''google/deeplabv3_mobilenet_v2_1.0_513''' ) snake_case : str = prepare_img() snake_case : Any = image_processor(images=UpperCAmelCase__ , return_tensors='''pt''' ).to(UpperCAmelCase__ ) # forward pass with torch.no_grad(): snake_case : List[str] = model(**UpperCAmelCase__ ) snake_case : List[str] = outputs.logits # verify the logits snake_case : Union[str, Any] = torch.Size((1, 21, 65, 65) ) self.assertEqual(logits.shape , UpperCAmelCase__ ) snake_case : Union[str, Any] = torch.tensor( [ [[17.5790, 17.7581, 18.3355], [18.3257, 18.4230, 18.8973], [18.6169, 18.8650, 19.2187]], [[-2.1595, -2.0977, -2.3741], [-2.4226, -2.3028, -2.6835], [-2.7819, -2.5991, -2.7706]], [[4.2058, 4.8317, 4.7638], [4.4136, 5.0361, 4.9383], [4.5028, 4.9644, 4.8734]], ] , device=UpperCAmelCase__ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , UpperCAmelCase__ , atol=1e-4 ) )
84
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class a_ : def __init__( self : Union[str, Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any]=99 , UpperCAmelCase__ : Dict=13 , UpperCAmelCase__ : int=7 , UpperCAmelCase__ : Any=9 , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Tuple=False , UpperCAmelCase__ : Tuple=32 , UpperCAmelCase__ : Dict=5 , UpperCAmelCase__ : Optional[int]=4 , UpperCAmelCase__ : List[Any]=37 , UpperCAmelCase__ : Union[str, Any]=8 , UpperCAmelCase__ : Optional[Any]=0.1 , UpperCAmelCase__ : str=0.002 , UpperCAmelCase__ : str=1 , UpperCAmelCase__ : Any=0 , UpperCAmelCase__ : Union[str, Any]=0 , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Optional[Any]=None , ): """simple docstring""" snake_case : Union[str, Any] = parent snake_case : Union[str, Any] = batch_size snake_case : Any = encoder_seq_length snake_case : str = decoder_seq_length # For common tests snake_case : Optional[int] = self.decoder_seq_length snake_case : Optional[Any] = is_training snake_case : List[Any] = use_attention_mask snake_case : Union[str, Any] = use_labels snake_case : Any = vocab_size snake_case : Optional[int] = hidden_size snake_case : List[str] = num_hidden_layers snake_case : Union[str, Any] = num_attention_heads snake_case : Any = d_ff snake_case : Any = relative_attention_num_buckets snake_case : Optional[Any] = dropout_rate snake_case : int = initializer_factor snake_case : Optional[Any] = eos_token_id snake_case : Dict = pad_token_id snake_case : Optional[Any] = decoder_start_token_id snake_case : Union[str, Any] = None snake_case : List[str] = decoder_layers def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" return TaConfig.from_pretrained('''google/umt5-base''' ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Any , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : List[Any]=None , ): """simple docstring""" if attention_mask is None: snake_case : Union[str, Any] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: snake_case : Any = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: snake_case : List[Any] = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=UpperCAmelCase__ ) if decoder_head_mask is None: snake_case : Tuple = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase__ ) if cross_attn_head_mask is None: snake_case : Union[str, Any] = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase__ ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input snake_case : List[str] = input_ids.clamp(self.pad_token_id + 1 ) snake_case : List[str] = decoder_input_ids.clamp(self.pad_token_id + 1 ) snake_case : str = self.get_config() snake_case : Tuple = config.num_attention_heads snake_case : List[Any] = self.prepare_inputs_dict(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) return config, input_dict def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : List[str] = self.prepare_config_and_inputs() return config, inputs_dict def lowerCAmelCase( self : Dict ): """simple docstring""" return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def lowerCAmelCase( self : Tuple ): """simple docstring""" return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def lowerCAmelCase( self : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Tuple , ): """simple docstring""" snake_case : str = UMTaModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : str = model( input_ids=UpperCAmelCase__ , decoder_input_ids=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , decoder_attention_mask=UpperCAmelCase__ , ) snake_case : int = model(input_ids=UpperCAmelCase__ , decoder_input_ids=UpperCAmelCase__ ) snake_case : int = result.last_hidden_state snake_case : Dict = result.past_key_values snake_case : Dict = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(UpperCAmelCase__ ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def lowerCAmelCase( self : int , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[str] , ): """simple docstring""" snake_case : int = UMTaModel(config=UpperCAmelCase__ ).get_decoder().to(UpperCAmelCase__ ).eval() # first forward pass snake_case : List[Any] = model(UpperCAmelCase__ , use_cache=UpperCAmelCase__ ) snake_case : List[Any] = model(UpperCAmelCase__ ) snake_case : Any = model(UpperCAmelCase__ , use_cache=UpperCAmelCase__ ) self.parent.assertTrue(len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) ) self.parent.assertTrue(len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) + 1 ) snake_case , snake_case : List[str] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case : Any = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and snake_case : Union[str, Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case : Any = model(UpperCAmelCase__ )['''last_hidden_state'''] snake_case : Optional[Any] = model(UpperCAmelCase__ , past_key_values=UpperCAmelCase__ )['''last_hidden_state'''] # select random slice snake_case : Tuple = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case : Union[str, Any] = output_from_no_past[:, -1, random_slice_idx].detach() snake_case : Tuple = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1e-3 ) ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : int , ): """simple docstring""" snake_case : int = UMTaModel(config=UpperCAmelCase__ ).to(UpperCAmelCase__ ).half().eval() snake_case : str = model(**UpperCAmelCase__ )['''last_hidden_state'''] self.parent.assertFalse(torch.isnan(UpperCAmelCase__ ).any().item() ) @require_torch class a_ ( a , a , a , unittest.TestCase ): A__ : str = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) A__ : str = (UMTaForConditionalGeneration,) if is_torch_available() else () A__ : Any = ( { 'conversational': UMTaForConditionalGeneration, 'feature-extraction': UMTaModel, 'summarization': UMTaForConditionalGeneration, 'text2text-generation': UMTaForConditionalGeneration, 'translation': UMTaForConditionalGeneration, 'question-answering': UMTaForQuestionAnswering, } if is_torch_available() else {} ) A__ : Dict = True A__ : List[str] = False A__ : Optional[int] = False A__ : Optional[int] = True A__ : List[str] = True # The small UMT5 model needs higher percentages for CPU/MP tests A__ : int = [0.8, 0.9] def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Union[str, Any] = UMTaModelTester(self ) @unittest.skip('''Test has a segmentation fault on torch 1.8.0''' ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Tuple = self.model_tester.prepare_config_and_inputs() snake_case : Optional[Any] = UMTaModel(config_and_inputs[0] ).to(UpperCAmelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( UpperCAmelCase__ , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"{tmpdirname}/t5_test.onnx" , export_params=UpperCAmelCase__ , opset_version=9 , input_names=['''input_ids''', '''decoder_input_ids'''] , ) @unittest.skipIf(torch_device == '''cpu''' , '''Cant do half precision''' ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*UpperCAmelCase__ ) def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : Optional[int] = ['''encoder_attentions''', '''decoder_attentions''', '''cross_attentions'''] snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() snake_case : int = config_and_inputs[0] snake_case : Union[str, Any] = UMTaForConditionalGeneration(UpperCAmelCase__ ).eval() model.to(UpperCAmelCase__ ) snake_case : str = { '''head_mask''': torch.zeros(config.num_layers , config.num_heads , device=UpperCAmelCase__ ), '''decoder_head_mask''': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ), '''cross_attn_head_mask''': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ), } for attn_name, (name, mask) in zip(UpperCAmelCase__ , head_masking.items() ): snake_case : int = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": snake_case : List[str] = torch.ones( config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ) snake_case : Union[str, Any] = model.generate( config_and_inputs[1]['''input_ids'''] , num_beams=1 , max_length=3 , output_attentions=UpperCAmelCase__ , return_dict_in_generate=UpperCAmelCase__ , **UpperCAmelCase__ , ) # We check the state of decoder_attentions and cross_attentions just from the last step snake_case : List[str] = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('''Does not work on the tiny model as we keep hitting edge cases.''' ) def lowerCAmelCase( self : Any ): """simple docstring""" pass @require_torch @require_sentencepiece @require_tokenizers class a_ ( unittest.TestCase ): @slow @unittest.skip( '''Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged''' ) def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Optional[Any] = UMTaForConditionalGeneration.from_pretrained('''google/umt5-small''' , return_dict=UpperCAmelCase__ ).to(UpperCAmelCase__ ) snake_case : int = AutoTokenizer.from_pretrained('''google/umt5-small''' , use_fast=UpperCAmelCase__ , legacy=UpperCAmelCase__ ) snake_case : List[str] = [ '''Bonjour monsieur <extra_id_0> bien <extra_id_1>.''', '''No se como puedo <extra_id_0>.''', '''This is the reason why we <extra_id_0> them.''', '''The <extra_id_0> walks in <extra_id_1>, seats''', '''A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.''', ] snake_case : Dict = tokenizer(UpperCAmelCase__ , return_tensors='''pt''' , padding=UpperCAmelCase__ ).input_ids # fmt: off snake_case : Optional[Any] = torch.tensor( [ [ 38_530, 210_703, 256_299, 1_410, 256_298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25_922, 256_299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1_460, 339, 312, 19_014, 10_620, 758, 256_299, 2_355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256_299, 14_869, 281, 301, 256_298, 275, 119_983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256_299, 14_869, 281, 2_234, 289, 2_275, 333,61_391, 289, 256_298, 543, 256_297, 168_714, 329, 256_296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : List[Any] = model.generate(input_ids.to(UpperCAmelCase__ ) ) snake_case : int = [ '''<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>''', '''<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', ] snake_case : Tuple = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ )
84
1
# Copyright 2022 The HuggingFace Team and The OpenBMB Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _a : Union[str, Any] = { 'configuration_cpmant': ['CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'CpmAntConfig'], 'tokenization_cpmant': ['CpmAntTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = [ 'CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST', 'CpmAntForCausalLM', 'CpmAntModel', 'CpmAntPreTrainedModel', ] if TYPE_CHECKING: from .configuration_cpmant import CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP, CpmAntConfig from .tokenization_cpmant import CpmAntTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_cpmant import ( CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST, CpmAntForCausalLM, CpmAntModel, CpmAntPreTrainedModel, ) else: import sys _a : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
84
import torch from diffusers import DiffusionPipeline class a_ ( a ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[Any] ): """simple docstring""" super().__init__() self.register_modules(unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ ) def __call__( self : Optional[int] ): """simple docstring""" snake_case : Any = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) snake_case : Dict = 1 snake_case : Optional[Any] = self.unet(UpperCAmelCase__ , UpperCAmelCase__ ).sample snake_case : List[Any] = self.scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample snake_case : List[Any] = scheduler_output - scheduler_output + torch.ones_like(UpperCAmelCase__ ) return result
84
1
import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class a_ ( unittest.TestCase ): def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : Tuple = 1 snake_case : int = 3 snake_case : Dict = (32, 32) snake_case : Dict = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(UpperCAmelCase__ ) return image @property def lowerCAmelCase( self : int ): """simple docstring""" torch.manual_seed(0 ) snake_case : int = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) return model @property def lowerCAmelCase( self : List[Any] ): """simple docstring""" torch.manual_seed(0 ) snake_case : int = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) return model @property def lowerCAmelCase( self : Tuple ): """simple docstring""" torch.manual_seed(0 ) snake_case : Union[str, Any] = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5_006 , ) return RobertaSeriesModelWithTransformation(UpperCAmelCase__ ) @property def lowerCAmelCase( self : List[str] ): """simple docstring""" def extract(*UpperCAmelCase__ : str , **UpperCAmelCase__ : Tuple ): class a_ : def __init__( self : Optional[Any] ): """simple docstring""" snake_case : Tuple = torch.ones([0] ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : List[Any] ): """simple docstring""" self.pixel_values.to(UpperCAmelCase__ ) return self return Out() return extract def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator snake_case : Any = self.dummy_cond_unet snake_case : Tuple = PNDMScheduler(skip_prk_steps=UpperCAmelCase__ ) snake_case : List[str] = self.dummy_vae snake_case : Optional[int] = self.dummy_text_encoder snake_case : Dict = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) snake_case : Any = 77 snake_case : Union[str, Any] = self.dummy_image.to(UpperCAmelCase__ ) snake_case : Union[str, Any] = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk snake_case : Dict = AltDiffusionImgaImgPipeline( unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ , vae=UpperCAmelCase__ , text_encoder=UpperCAmelCase__ , tokenizer=UpperCAmelCase__ , safety_checker=UpperCAmelCase__ , feature_extractor=self.dummy_extractor , ) snake_case : Tuple = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=UpperCAmelCase__ ) snake_case : Optional[Any] = alt_pipe.to(UpperCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) snake_case : Optional[int] = '''A painting of a squirrel eating a burger''' snake_case : str = torch.Generator(device=UpperCAmelCase__ ).manual_seed(0 ) snake_case : List[str] = alt_pipe( [prompt] , generator=UpperCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , image=UpperCAmelCase__ , ) snake_case : Any = output.images snake_case : Dict = torch.Generator(device=UpperCAmelCase__ ).manual_seed(0 ) snake_case : Optional[Any] = alt_pipe( [prompt] , generator=UpperCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , image=UpperCAmelCase__ , return_dict=UpperCAmelCase__ , )[0] snake_case : Dict = image[0, -3:, -3:, -1] snake_case : Union[str, Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) snake_case : Any = np.array([0.4427, 0.3731, 0.4249, 0.4941, 0.4546, 0.4148, 0.4193, 0.4666, 0.4499] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5e-3 @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : Optional[Any] = self.dummy_cond_unet snake_case : Union[str, Any] = PNDMScheduler(skip_prk_steps=UpperCAmelCase__ ) snake_case : Optional[int] = self.dummy_vae snake_case : Any = self.dummy_text_encoder snake_case : Optional[Any] = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) snake_case : List[str] = 77 snake_case : str = self.dummy_image.to(UpperCAmelCase__ ) # put models in fp16 snake_case : List[Any] = unet.half() snake_case : Union[str, Any] = vae.half() snake_case : Tuple = bert.half() # make sure here that pndm scheduler skips prk snake_case : List[Any] = AltDiffusionImgaImgPipeline( unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ , vae=UpperCAmelCase__ , text_encoder=UpperCAmelCase__ , tokenizer=UpperCAmelCase__ , safety_checker=UpperCAmelCase__ , feature_extractor=self.dummy_extractor , ) snake_case : Optional[Any] = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=UpperCAmelCase__ ) snake_case : List[Any] = alt_pipe.to(UpperCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) snake_case : Any = '''A painting of a squirrel eating a burger''' snake_case : Any = torch.manual_seed(0 ) snake_case : int = alt_pipe( [prompt] , generator=UpperCAmelCase__ , num_inference_steps=2 , output_type='''np''' , image=UpperCAmelCase__ , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[int] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) # resize to resolution that is divisible by 8 but not 16 or 32 snake_case : Optional[int] = init_image.resize((760, 504) ) snake_case : int = '''BAAI/AltDiffusion''' snake_case : Optional[int] = AltDiffusionImgaImgPipeline.from_pretrained( UpperCAmelCase__ , safety_checker=UpperCAmelCase__ , ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) pipe.enable_attention_slicing() snake_case : Optional[int] = '''A fantasy landscape, trending on artstation''' snake_case : int = torch.manual_seed(0 ) snake_case : Dict = pipe( prompt=UpperCAmelCase__ , image=UpperCAmelCase__ , strength=0.75 , guidance_scale=7.5 , generator=UpperCAmelCase__ , output_type='''np''' , ) snake_case : int = output.images[0] snake_case : str = image[255:258, 383:386, -1] assert image.shape == (504, 760, 3) snake_case : int = np.array([0.9358, 0.9397, 0.9599, 0.9901, 1.0000, 1.0000, 0.9882, 1.0000, 1.0000] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class a_ ( unittest.TestCase ): def lowerCAmelCase( self : List[str] ): """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Union[str, Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) snake_case : Tuple = init_image.resize((768, 512) ) snake_case : List[Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy''' ) snake_case : Union[str, Any] = '''BAAI/AltDiffusion''' snake_case : List[Any] = AltDiffusionImgaImgPipeline.from_pretrained( UpperCAmelCase__ , safety_checker=UpperCAmelCase__ , ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) pipe.enable_attention_slicing() snake_case : Union[str, Any] = '''A fantasy landscape, trending on artstation''' snake_case : int = torch.manual_seed(0 ) snake_case : Dict = pipe( prompt=UpperCAmelCase__ , image=UpperCAmelCase__ , strength=0.75 , guidance_scale=7.5 , generator=UpperCAmelCase__ , output_type='''np''' , ) snake_case : Optional[Any] = output.images[0] assert image.shape == (512, 768, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1e-2
84
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class a_ ( a ): A__ : List[str] = ['image_processor', 'tokenizer'] A__ : Any = 'CLIPImageProcessor' A__ : Optional[int] = ('CLIPTokenizer', 'CLIPTokenizerFast') def __init__( self : Union[str, Any] , UpperCAmelCase__ : str=None , UpperCAmelCase__ : Union[str, Any]=None , **UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : Optional[int] = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , UpperCAmelCase__ , ) snake_case : List[Any] = kwargs.pop('''feature_extractor''' ) snake_case : Optional[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(UpperCAmelCase__ , UpperCAmelCase__ ) def __call__( self : Any , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : int=None , **UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: snake_case : int = self.tokenizer(UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , **UpperCAmelCase__ ) if images is not None: snake_case : Dict = self.image_processor(UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , **UpperCAmelCase__ ) if text is not None and images is not None: snake_case : Tuple = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**UpperCAmelCase__ ) , tensor_type=UpperCAmelCase__ ) def lowerCAmelCase( self : List[str] , *UpperCAmelCase__ : Dict , **UpperCAmelCase__ : int ): """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : str ): """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : int = self.tokenizer.model_input_names snake_case : Tuple = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , UpperCAmelCase__ , ) return self.image_processor_class @property def lowerCAmelCase( self : Any ): """simple docstring""" warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , UpperCAmelCase__ , ) return self.image_processor
84
1
import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() _a : Any = logging.get_logger(__name__) def a_ ( __magic_name__ ) -> Union[str, Any]: """simple docstring""" snake_case : int = torch.load(__magic_name__ , map_location='''cpu''' ) if "model" in sd.keys(): snake_case : Union[str, Any] = torch.load(__magic_name__ , map_location='''cpu''' )['''model'''] # pop unnecessary weights snake_case : List[str] = [ '''decoder.version''', '''decoder.output_projection.weight''', ] for key in keys_to_delete: if key in sd: sd.pop(__magic_name__ ) snake_case : Any = { '''decoder.project_in_dim.weight''': '''decoder.project_in.weight''', '''decoder.project_out_dim.weight''': '''decoder.project_out.weight''', '''decoder.layer_norm.weight''': '''decoder.final_layer_norm.weight''', '''decoder.layer_norm.bias''': '''decoder.final_layer_norm.bias''', } for old_key, new_key in keys_to_rename.items(): if old_key in sd: snake_case : Dict = sd.pop(__magic_name__ ) snake_case : Optional[Any] = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: snake_case : Tuple = sd[key] # We split QKV in separate Q,K,V snake_case : Optional[Any] = key.replace('''.qkv_proj.''' , '''.q_proj.''' ) snake_case : Optional[Any] = key.replace('''.qkv_proj.''' , '''.k_proj.''' ) snake_case : Any = key.replace('''.qkv_proj.''' , '''.v_proj.''' ) snake_case : Dict = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 snake_case , snake_case , snake_case : Union[str, Any] = torch.split(__magic_name__ , depth // 3 , dim=0 ) snake_case : Optional[int] = q snake_case : str = k snake_case : List[str] = v del sd[key] return sd @torch.no_grad() def a_ ( __magic_name__ , __magic_name__ , __magic_name__=None ) -> Dict: """simple docstring""" snake_case : Any = load_checkpoint(__magic_name__ ) if config is not None: snake_case : str = OPTConfig.from_pretrained(__magic_name__ ) else: snake_case : Optional[int] = OPTConfig() snake_case : Dict = OPTModel(__magic_name__ ).half().eval() model.load_state_dict(__magic_name__ ) # Check results Path(__magic_name__ ).mkdir(exist_ok=__magic_name__ ) model.save_pretrained(__magic_name__ ) if __name__ == "__main__": _a : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--fairseq_path', type=str, help=( 'path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:' ' https://huggingface.co/models?other=opt_metasq' ), ) parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--hf_config', default=None, type=str, help='Define HF config.') _a : List[Any] = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
84
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets _a : Optional[Any] = '\\n@inproceedings{pillutla-etal:mauve:neurips2021,\n title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers},\n author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid},\n booktitle = {NeurIPS},\n year = {2021}\n}\n\n' _a : str = '\\nMAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure.\n\nMAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences.\n\nFor details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021).\n\nThis metrics is a wrapper around the official implementation of MAUVE:\nhttps://github.com/krishnap25/mauve\n' _a : List[Any] = '\nCalculates MAUVE scores between two lists of generated text and reference text.\nArgs:\n predictions: list of generated text to score. Each predictions\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\nOptional Args:\n num_buckets: the size of the histogram to quantize P and Q. Options: \'auto\' (default) or an integer\n pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1\n kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9\n kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5\n kmeans_max_iter: maximum number of k-means iterations. Default 500\n featurize_model_name: name of the model from which features are obtained. Default \'gpt2-large\' Use one of [\'gpt2\', \'gpt2-medium\', \'gpt2-large\', \'gpt2-xl\'].\n device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU\n max_text_length: maximum number of tokens to consider. Default 1024\n divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25\n mauve_scaling_factor: "c" from the paper. Default 5.\n verbose: If True (default), print running time updates\n seed: random seed to initialize k-means cluster assignments.\nReturns:\n mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer,\n frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer,\n divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve,\n p_hist: a discrete distribution, which is a quantized version of the text distribution p_text,\n q_hist: same as above, but with q_text.\nExamples:\n\n >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest\n >>> import datasets\n >>> mauve = datasets.load_metric(\'mauve\')\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP\n >>> print(out.mauve) # doctest: +SKIP\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): def lowerCAmelCase( self : Any ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://github.com/krishnap25/mauve''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/krishnap25/mauve'''] , reference_urls=[ '''https://arxiv.org/abs/2102.01454''', '''https://github.com/krishnap25/mauve''', ] , ) def lowerCAmelCase( self : int , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : List[str]="auto" , UpperCAmelCase__ : Tuple=-1 , UpperCAmelCase__ : Optional[int]=0.9 , UpperCAmelCase__ : List[Any]=5 , UpperCAmelCase__ : List[Any]=500 , UpperCAmelCase__ : Union[str, Any]="gpt2-large" , UpperCAmelCase__ : Optional[Any]=-1 , UpperCAmelCase__ : int=1_024 , UpperCAmelCase__ : List[Any]=25 , UpperCAmelCase__ : Union[str, Any]=5 , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : List[Any]=25 , ): """simple docstring""" snake_case : List[str] = compute_mauve( p_text=UpperCAmelCase__ , q_text=UpperCAmelCase__ , p_features=UpperCAmelCase__ , q_features=UpperCAmelCase__ , p_tokens=UpperCAmelCase__ , q_tokens=UpperCAmelCase__ , num_buckets=UpperCAmelCase__ , pca_max_data=UpperCAmelCase__ , kmeans_explained_var=UpperCAmelCase__ , kmeans_num_redo=UpperCAmelCase__ , kmeans_max_iter=UpperCAmelCase__ , featurize_model_name=UpperCAmelCase__ , device_id=UpperCAmelCase__ , max_text_length=UpperCAmelCase__ , divergence_curve_discretization_size=UpperCAmelCase__ , mauve_scaling_factor=UpperCAmelCase__ , verbose=UpperCAmelCase__ , seed=UpperCAmelCase__ , ) return out
84
1
import json import os import re import unicodedata from json.encoder import INFINITY from typing import Any, Dict, List, Optional, Tuple, Union import numpy as np import regex from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_flax_available, is_tf_available, is_torch_available, logging from ...utils.generic import _is_jax, _is_numpy _a : Union[str, Any] = logging.get_logger(__name__) _a : Tuple = { 'artists_file': 'artists.json', 'lyrics_file': 'lyrics.json', 'genres_file': 'genres.json', } _a : str = { 'artists_file': { 'jukebox': 'https://huggingface.co/ArthurZ/jukebox/blob/main/artists.json', }, 'genres_file': { 'jukebox': 'https://huggingface.co/ArthurZ/jukebox/blob/main/genres.json', }, 'lyrics_file': { 'jukebox': 'https://huggingface.co/ArthurZ/jukebox/blob/main/lyrics.json', }, } _a : Union[str, Any] = { 'jukebox': 512, } class a_ ( a ): A__ : str = VOCAB_FILES_NAMES A__ : Optional[int] = PRETRAINED_VOCAB_FILES_MAP A__ : Optional[int] = PRETRAINED_LYRIC_TOKENS_SIZES A__ : int = ['input_ids', 'attention_mask'] def __init__( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : int=["v3", "v2", "v2"] , UpperCAmelCase__ : Tuple=512 , UpperCAmelCase__ : Any=5 , UpperCAmelCase__ : Optional[Any]="<|endoftext|>" , **UpperCAmelCase__ : Dict , ): """simple docstring""" snake_case : List[str] = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else unk_token super().__init__( unk_token=UpperCAmelCase__ , n_genres=UpperCAmelCase__ , version=UpperCAmelCase__ , max_n_lyric_tokens=UpperCAmelCase__ , **UpperCAmelCase__ , ) snake_case : Union[str, Any] = version snake_case : Union[str, Any] = max_n_lyric_tokens snake_case : List[str] = n_genres with open(UpperCAmelCase__ , encoding='''utf-8''' ) as vocab_handle: snake_case : Tuple = json.load(UpperCAmelCase__ ) with open(UpperCAmelCase__ , encoding='''utf-8''' ) as vocab_handle: snake_case : Union[str, Any] = json.load(UpperCAmelCase__ ) with open(UpperCAmelCase__ , encoding='''utf-8''' ) as vocab_handle: snake_case : Union[str, Any] = json.load(UpperCAmelCase__ ) snake_case : Optional[Any] = r'''[^A-Za-z0-9.,:;!?\-\'\"()\[\] \t\n]+''' # In v2, we had a n_vocab=80 and in v3 we missed + and so n_vocab=79 of characters. if len(self.lyrics_encoder ) == 79: snake_case : Optional[Any] = oov.replace(r'''\-\'''' , r'''\-+\'''' ) snake_case : List[str] = regex.compile(UpperCAmelCase__ ) snake_case : Tuple = {v: k for k, v in self.artists_encoder.items()} snake_case : Union[str, Any] = {v: k for k, v in self.genres_encoder.items()} snake_case : Tuple = {v: k for k, v in self.lyrics_encoder.items()} @property def lowerCAmelCase( self : str ): """simple docstring""" return len(self.artists_encoder ) + len(self.genres_encoder ) + len(self.lyrics_encoder ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" return dict(self.artists_encoder , self.genres_encoder , self.lyrics_encoder ) def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : Any , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : str = [self.artists_encoder.get(UpperCAmelCase__ , 0 ) for artist in list_artists] for genres in range(len(UpperCAmelCase__ ) ): snake_case : Optional[int] = [self.genres_encoder.get(UpperCAmelCase__ , 0 ) for genre in list_genres[genres]] snake_case : Optional[int] = list_genres[genres] + [-1] * (self.n_genres - len(list_genres[genres] )) snake_case : List[Any] = [[self.lyrics_encoder.get(UpperCAmelCase__ , 0 ) for character in list_lyrics[0]], [], []] return artists_id, list_genres, lyric_ids def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : str ): """simple docstring""" return list(UpperCAmelCase__ ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : Optional[Any] ): """simple docstring""" snake_case , snake_case , snake_case : Tuple = self.prepare_for_tokenization(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : Optional[int] = self._tokenize(UpperCAmelCase__ ) return artist, genre, lyrics def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : str , UpperCAmelCase__ : str , UpperCAmelCase__ : bool = False ): """simple docstring""" for idx in range(len(self.version ) ): if self.version[idx] == "v3": snake_case : List[str] = artists[idx].lower() snake_case : Optional[Any] = [genres[idx].lower()] else: snake_case : List[Any] = self._normalize(artists[idx] ) + '''.v2''' snake_case : int = [ self._normalize(UpperCAmelCase__ ) + '''.v2''' for genre in genres[idx].split('''_''' ) ] # split is for the full dictionary with combined genres if self.version[0] == "v2": snake_case : Any = regex.compile(r'''[^A-Za-z0-9.,:;!?\-\'\"()\[\] \t\n]+''' ) snake_case : int = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789.,:;!?-+\'\"()[] \t\n''' snake_case : Dict = {vocab[index]: index + 1 for index in range(len(UpperCAmelCase__ ) )} snake_case : List[str] = 0 snake_case : Tuple = len(UpperCAmelCase__ ) + 1 snake_case : Optional[int] = self.vocab snake_case : Optional[int] = {v: k for k, v in self.vocab.items()} snake_case : int = '''''' else: snake_case : Dict = regex.compile(r'''[^A-Za-z0-9.,:;!?\-+\'\"()\[\] \t\n]+''' ) snake_case : int = self._run_strip_accents(UpperCAmelCase__ ) snake_case : List[str] = lyrics.replace('''\\''' , '''\n''' ) snake_case : List[str] = self.out_of_vocab.sub('''''' , UpperCAmelCase__ ), [], [] return artists, genres, lyrics def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Optional[Any] = unicodedata.normalize('''NFD''' , UpperCAmelCase__ ) snake_case : Any = [] for char in text: snake_case : int = unicodedata.category(UpperCAmelCase__ ) if cat == "Mn": continue output.append(UpperCAmelCase__ ) return "".join(UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Optional[Any] = ( [chr(UpperCAmelCase__ ) for i in range(ord('''a''' ) , ord('''z''' ) + 1 )] + [chr(UpperCAmelCase__ ) for i in range(ord('''A''' ) , ord('''Z''' ) + 1 )] + [chr(UpperCAmelCase__ ) for i in range(ord('''0''' ) , ord('''9''' ) + 1 )] + ['''.'''] ) snake_case : Optional[Any] = frozenset(UpperCAmelCase__ ) snake_case : Dict = re.compile(r'''_+''' ) snake_case : int = ''''''.join([c if c in accepted else '''_''' for c in text.lower()] ) snake_case : List[Any] = pattern.sub('''_''' , UpperCAmelCase__ ).strip('''_''' ) return text def lowerCAmelCase( self : Dict , UpperCAmelCase__ : List[str] ): """simple docstring""" return " ".join(UpperCAmelCase__ ) def lowerCAmelCase( self : Any , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[Union[str, TensorType]] = None , UpperCAmelCase__ : bool = False ): """simple docstring""" # Convert to TensorType if not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : List[str] = TensorType(UpperCAmelCase__ ) # Get a function reference for the correct framework if tensor_type == TensorType.TENSORFLOW: if not is_tf_available(): raise ImportError( '''Unable to convert output to TensorFlow tensors format, TensorFlow is not installed.''' ) import tensorflow as tf snake_case : Dict = tf.constant snake_case : List[str] = tf.is_tensor elif tensor_type == TensorType.PYTORCH: if not is_torch_available(): raise ImportError('''Unable to convert output to PyTorch tensors format, PyTorch is not installed.''' ) import torch snake_case : str = torch.tensor snake_case : List[Any] = torch.is_tensor elif tensor_type == TensorType.JAX: if not is_flax_available(): raise ImportError('''Unable to convert output to JAX tensors format, JAX is not installed.''' ) import jax.numpy as jnp # noqa: F811 snake_case : Optional[int] = jnp.array snake_case : Optional[Any] = _is_jax else: snake_case : int = np.asarray snake_case : Union[str, Any] = _is_numpy # Do the tensor conversion in batch try: if prepend_batch_axis: snake_case : Optional[int] = [inputs] if not is_tensor(UpperCAmelCase__ ): snake_case : Union[str, Any] = as_tensor(UpperCAmelCase__ ) except: # noqa E722 raise ValueError( '''Unable to create tensor, you should probably activate truncation and/or padding ''' '''with \'padding=True\' \'truncation=True\' to have batched tensors with the same length.''' ) return inputs def __call__( self : Optional[int] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Any , UpperCAmelCase__ : Optional[Any]="" , UpperCAmelCase__ : int="pt" ): """simple docstring""" snake_case : Optional[Any] = [0, 0, 0] snake_case : Optional[int] = [artist] * len(self.version ) snake_case : Any = [genres] * len(self.version ) snake_case , snake_case , snake_case : Dict = self.tokenize(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) snake_case , snake_case , snake_case : List[Any] = self._convert_token_to_id(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : Union[str, Any] = [-INFINITY] * len(full_tokens[-1] ) snake_case : Optional[Any] = [ self.convert_to_tensors( [input_ids + [artists_id[i]] + genres_ids[i] + full_tokens[i]] , tensor_type=UpperCAmelCase__ ) for i in range(len(self.version ) ) ] return BatchEncoding({'''input_ids''': input_ids, '''attention_masks''': attention_masks} ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ): """simple docstring""" if not os.path.isdir(UpperCAmelCase__ ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return snake_case : Tuple = os.path.join( UpperCAmelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''artists_file'''] ) with open(UpperCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.artists_encoder , ensure_ascii=UpperCAmelCase__ ) ) snake_case : Any = os.path.join( UpperCAmelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''genres_file'''] ) with open(UpperCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.genres_encoder , ensure_ascii=UpperCAmelCase__ ) ) snake_case : Optional[Any] = os.path.join( UpperCAmelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''lyrics_file'''] ) with open(UpperCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.lyrics_encoder , ensure_ascii=UpperCAmelCase__ ) ) return (artists_file, genres_file, lyrics_file) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[str] ): """simple docstring""" snake_case : List[str] = self.artists_decoder.get(UpperCAmelCase__ ) snake_case : int = [self.genres_decoder.get(UpperCAmelCase__ ) for genre in genres_index] snake_case : int = [self.lyrics_decoder.get(UpperCAmelCase__ ) for character in lyric_index] return artist, genres, lyrics
84
import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" if "cls_token" in name: snake_case : Tuple = name.replace('''cls_token''' , '''vit.embeddings.cls_token''' ) if "mask_token" in name: snake_case : Optional[int] = name.replace('''mask_token''' , '''decoder.mask_token''' ) if "decoder_pos_embed" in name: snake_case : List[str] = name.replace('''decoder_pos_embed''' , '''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: snake_case : List[str] = name.replace('''pos_embed''' , '''vit.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: snake_case : List[Any] = name.replace('''patch_embed.proj''' , '''vit.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: snake_case : int = name.replace('''patch_embed.norm''' , '''vit.embeddings.norm''' ) if "decoder_blocks" in name: snake_case : int = name.replace('''decoder_blocks''' , '''decoder.decoder_layers''' ) if "blocks" in name: snake_case : Optional[Any] = name.replace('''blocks''' , '''vit.encoder.layer''' ) if "attn.proj" in name: snake_case : str = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: snake_case : Any = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: snake_case : List[str] = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: snake_case : Tuple = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: snake_case : Tuple = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: snake_case : Tuple = name.replace('''mlp.fc2''' , '''output.dense''' ) if "decoder_embed" in name: snake_case : Dict = name.replace('''decoder_embed''' , '''decoder.decoder_embed''' ) if "decoder_norm" in name: snake_case : Dict = name.replace('''decoder_norm''' , '''decoder.decoder_norm''' ) if "decoder_pred" in name: snake_case : Dict = name.replace('''decoder_pred''' , '''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name: snake_case : Optional[int] = name.replace('''norm.weight''' , '''vit.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name: snake_case : List[str] = name.replace('''norm.bias''' , '''vit.layernorm.bias''' ) return name def a_ ( __magic_name__ , __magic_name__ ) -> str: """simple docstring""" for key in orig_state_dict.copy().keys(): snake_case : Union[str, Any] = orig_state_dict.pop(__magic_name__ ) if "qkv" in key: snake_case : Optional[int] = key.split('''.''' ) snake_case : int = int(key_split[1] ) if "decoder_blocks" in key: snake_case : List[str] = config.decoder_hidden_size snake_case : List[Any] = '''decoder.decoder_layers.''' if "weight" in key: snake_case : str = val[:dim, :] snake_case : Optional[Any] = val[dim : dim * 2, :] snake_case : Any = val[-dim:, :] elif "bias" in key: snake_case : Optional[Any] = val[:dim] snake_case : List[Any] = val[dim : dim * 2] snake_case : List[Any] = val[-dim:] else: snake_case : Optional[int] = config.hidden_size snake_case : Tuple = '''vit.encoder.layer.''' if "weight" in key: snake_case : Optional[Any] = val[:dim, :] snake_case : str = val[dim : dim * 2, :] snake_case : Union[str, Any] = val[-dim:, :] elif "bias" in key: snake_case : Tuple = val[:dim] snake_case : int = val[dim : dim * 2] snake_case : Optional[Any] = val[-dim:] else: snake_case : Optional[Any] = val return orig_state_dict def a_ ( __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" snake_case : List[str] = ViTMAEConfig() if "large" in checkpoint_url: snake_case : str = 1_024 snake_case : Tuple = 4_096 snake_case : Optional[Any] = 24 snake_case : List[Any] = 16 elif "huge" in checkpoint_url: snake_case : Tuple = 14 snake_case : int = 1_280 snake_case : Dict = 5_120 snake_case : Tuple = 32 snake_case : Optional[Any] = 16 snake_case : Optional[Any] = ViTMAEForPreTraining(__magic_name__ ) snake_case : Optional[Any] = torch.hub.load_state_dict_from_url(__magic_name__ , map_location='''cpu''' )['''model'''] snake_case : int = ViTMAEImageProcessor(size=config.image_size ) snake_case : Dict = convert_state_dict(__magic_name__ , __magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() snake_case : Tuple = '''https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg''' snake_case : List[Any] = Image.open(requests.get(__magic_name__ , stream=__magic_name__ ).raw ) snake_case : Dict = ViTMAEImageProcessor(size=config.image_size ) snake_case : str = image_processor(images=__magic_name__ , return_tensors='''pt''' ) # forward pass torch.manual_seed(2 ) snake_case : Union[str, Any] = model(**__magic_name__ ) snake_case : Optional[Any] = outputs.logits if "large" in checkpoint_url: snake_case : Any = torch.tensor( [[-0.7309, -0.7128, -1.0169], [-1.0161, -0.9058, -1.1878], [-1.0478, -0.9411, -1.1911]] ) elif "huge" in checkpoint_url: snake_case : List[Any] = torch.tensor( [[-1.1599, -0.9199, -1.2221], [-1.1952, -0.9269, -1.2307], [-1.2143, -0.9337, -1.2262]] ) else: snake_case : Dict = torch.tensor( [[-0.9192, -0.8481, -1.1259], [-1.1349, -1.0034, -1.2599], [-1.1757, -1.0429, -1.2726]] ) # verify logits assert torch.allclose(logits[0, :3, :3] , __magic_name__ , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(__magic_name__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__magic_name__ ) if __name__ == "__main__": _a : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth', type=str, help='URL of the checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _a : str = parser.parse_args() convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
84
1
from __future__ import annotations from typing import Any def a_ ( __magic_name__ ) -> None: """simple docstring""" create_state_space_tree(__magic_name__ , [] , 0 ) def a_ ( __magic_name__ , __magic_name__ , __magic_name__ ) -> None: """simple docstring""" if index == len(__magic_name__ ): print(__magic_name__ ) return create_state_space_tree(__magic_name__ , __magic_name__ , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__magic_name__ , __magic_name__ , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _a : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(['A', 'B', 'C']) generate_all_subsequences(seq)
84
import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _a : Optional[Any] = 16 _a : Union[str, Any] = 32 def a_ ( __magic_name__ , __magic_name__ = 16 ) -> Dict: """simple docstring""" snake_case : Tuple = AutoTokenizer.from_pretrained('''bert-base-cased''' ) snake_case : Any = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(__magic_name__ ): # max_length=None => use the model max length (it's actually the default) snake_case : Union[str, Any] = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=__magic_name__ , max_length=__magic_name__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): snake_case : Union[str, Any] = datasets.map( __magic_name__ , batched=__magic_name__ , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library snake_case : Optional[Any] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(__magic_name__ ): # On TPU it's best to pad everything to the same length or training will be very slow. snake_case : str = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": snake_case : Tuple = 16 elif accelerator.mixed_precision != "no": snake_case : Dict = 8 else: snake_case : Union[str, Any] = None return tokenizer.pad( __magic_name__ , padding='''longest''' , max_length=__magic_name__ , pad_to_multiple_of=__magic_name__ , return_tensors='''pt''' , ) # Instantiate dataloaders. snake_case : str = DataLoader( tokenized_datasets['''train'''] , shuffle=__magic_name__ , collate_fn=__magic_name__ , batch_size=__magic_name__ ) snake_case : List[str] = DataLoader( tokenized_datasets['''validation'''] , shuffle=__magic_name__ , collate_fn=__magic_name__ , batch_size=__magic_name__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders _a : Optional[int] = mocked_dataloaders # noqa: F811 def a_ ( __magic_name__ , __magic_name__ ) -> Optional[Any]: """simple docstring""" if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , __magic_name__ ) == "1": snake_case : Optional[int] = 2 # Initialize accelerator snake_case : int = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs snake_case : Dict = config['''lr'''] snake_case : Any = int(config['''num_epochs'''] ) snake_case : List[str] = int(config['''seed'''] ) snake_case : List[Any] = int(config['''batch_size'''] ) snake_case : Tuple = evaluate.load('''glue''' , '''mrpc''' ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=__magic_name__ ) def inner_training_loop(__magic_name__ ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(__magic_name__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) snake_case : str = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=__magic_name__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). snake_case : Optional[int] = model.to(accelerator.device ) # Instantiate optimizer snake_case : Optional[int] = AdamW(params=model.parameters() , lr=__magic_name__ ) snake_case , snake_case : List[Any] = get_dataloaders(__magic_name__ , __magic_name__ ) # Instantiate scheduler snake_case : int = get_linear_schedule_with_warmup( optimizer=__magic_name__ , num_warmup_steps=100 , num_training_steps=(len(__magic_name__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. snake_case , snake_case , snake_case , snake_case , snake_case : Tuple = accelerator.prepare( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # Now we train the model for epoch in range(__magic_name__ ): model.train() for step, batch in enumerate(__magic_name__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) snake_case : int = model(**__magic_name__ ) snake_case : Optional[int] = outputs.loss accelerator.backward(__magic_name__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__magic_name__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): snake_case : List[str] = model(**__magic_name__ ) snake_case : List[Any] = outputs.logits.argmax(dim=-1 ) snake_case , snake_case : Dict = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=__magic_name__ , references=__magic_name__ , ) snake_case : Tuple = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"epoch {epoch}:" , __magic_name__ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def a_ ( ) -> Union[str, Any]: """simple docstring""" snake_case : int = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=__magic_name__ , default=__magic_name__ , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) snake_case : Optional[Any] = parser.parse_args() snake_case : Optional[int] = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(__magic_name__ , __magic_name__ ) if __name__ == "__main__": main()
84
1
from jiwer import compute_measures import datasets _a : int = '\\n@inproceedings{inproceedings,\n author = {Morris, Andrew and Maier, Viktoria and Green, Phil},\n year = {2004},\n month = {01},\n pages = {},\n title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.}\n}\n' _a : str = '\\nWord error rate (WER) is a common metric of the performance of an automatic speech recognition system.\n\nThe general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort.\n\nThis problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate.\n\nWord error rate can then be computed as:\n\nWER = (S + D + I) / N = (S + D + I) / (S + D + C)\n\nwhere\n\nS is the number of substitutions,\nD is the number of deletions,\nI is the number of insertions,\nC is the number of correct words,\nN is the number of words in the reference (N=S+D+C).\n\nThis value indicates the average number of errors per reference word. The lower the value, the better the\nperformance of the ASR system with a WER of 0 being a perfect score.\n' _a : Tuple = '\nCompute WER score of transcribed segments against references.\n\nArgs:\n references: List of references for each speech input.\n predictions: List of transcriptions to score.\n concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively.\n\nReturns:\n (float): the word error rate\n\nExamples:\n\n >>> predictions = ["this is the prediction", "there is an other sample"]\n >>> references = ["this is the reference", "there is another one"]\n >>> wer = datasets.load_metric("wer")\n >>> wer_score = wer.compute(predictions=predictions, references=references)\n >>> print(wer_score)\n 0.5\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): def lowerCAmelCase( self : Dict ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/jitsi/jiwer/'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/Word_error_rate''', ] , ) def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : str=None , UpperCAmelCase__ : Tuple=False ): """simple docstring""" if concatenate_texts: return compute_measures(UpperCAmelCase__ , UpperCAmelCase__ )["wer"] else: snake_case : Optional[int] = 0 snake_case : List[Any] = 0 for prediction, reference in zip(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Dict = compute_measures(UpperCAmelCase__ , UpperCAmelCase__ ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
84
from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract _a : Dict = logging.get_logger(__name__) def a_ ( __magic_name__ , __magic_name__ , __magic_name__ ) -> List[Any]: """simple docstring""" return [ int(1_000 * (box[0] / width) ), int(1_000 * (box[1] / height) ), int(1_000 * (box[2] / width) ), int(1_000 * (box[3] / height) ), ] def a_ ( __magic_name__ , __magic_name__ , __magic_name__ = None ) -> str: """simple docstring""" snake_case : Any = tesseract_config if tesseract_config is not None else '''''' # apply OCR snake_case : str = to_pil_image(__magic_name__ ) snake_case , snake_case : Union[str, Any] = pil_image.size snake_case : List[Any] = pytesseract.image_to_data(__magic_name__ , lang=__magic_name__ , output_type='''dict''' , config=__magic_name__ ) snake_case , snake_case , snake_case , snake_case , snake_case : Optional[Any] = data['''text'''], data['''left'''], data['''top'''], data['''width'''], data['''height'''] # filter empty words and corresponding coordinates snake_case : Union[str, Any] = [idx for idx, word in enumerate(__magic_name__ ) if not word.strip()] snake_case : Union[str, Any] = [word for idx, word in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : Optional[Any] = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format snake_case : List[Any] = [] for x, y, w, h in zip(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): snake_case : Optional[int] = [x, y, x + w, y + h] actual_boxes.append(__magic_name__ ) # finally, normalize the bounding boxes snake_case : List[Any] = [] for box in actual_boxes: normalized_boxes.append(normalize_box(__magic_name__ , __magic_name__ , __magic_name__ ) ) assert len(__magic_name__ ) == len(__magic_name__ ), "Not as many words as there are bounding boxes" return words, normalized_boxes class a_ ( a ): A__ : int = ['pixel_values'] def __init__( self : Optional[int] , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : PILImageResampling = PILImageResampling.BILINEAR , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[str] = "" , **UpperCAmelCase__ : int , ): """simple docstring""" super().__init__(**UpperCAmelCase__ ) snake_case : Any = size if size is not None else {'''height''': 224, '''width''': 224} snake_case : Tuple = get_size_dict(UpperCAmelCase__ ) snake_case : Dict = do_resize snake_case : str = size snake_case : Optional[int] = resample snake_case : Union[str, Any] = apply_ocr snake_case : int = ocr_lang snake_case : Union[str, Any] = tesseract_config def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : np.ndarray , UpperCAmelCase__ : Dict[str, int] , UpperCAmelCase__ : PILImageResampling = PILImageResampling.BILINEAR , UpperCAmelCase__ : Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase__ : Any , ): """simple docstring""" snake_case : Dict = get_size_dict(UpperCAmelCase__ ) if "height" not in size or "width" not in size: raise ValueError(F"The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}" ) snake_case : Tuple = (size['''height'''], size['''width''']) return resize(UpperCAmelCase__ , size=UpperCAmelCase__ , resample=UpperCAmelCase__ , data_format=UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : ImageInput , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : PILImageResampling = None , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[Union[str, TensorType]] = None , UpperCAmelCase__ : ChannelDimension = ChannelDimension.FIRST , **UpperCAmelCase__ : List[Any] , ): """simple docstring""" snake_case : Tuple = do_resize if do_resize is not None else self.do_resize snake_case : List[Any] = size if size is not None else self.size snake_case : Tuple = get_size_dict(UpperCAmelCase__ ) snake_case : str = resample if resample is not None else self.resample snake_case : Optional[int] = apply_ocr if apply_ocr is not None else self.apply_ocr snake_case : Any = ocr_lang if ocr_lang is not None else self.ocr_lang snake_case : Optional[int] = tesseract_config if tesseract_config is not None else self.tesseract_config snake_case : List[str] = make_list_of_images(UpperCAmelCase__ ) if not valid_images(UpperCAmelCase__ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) # All transformations expect numpy arrays. snake_case : Any = [to_numpy_array(UpperCAmelCase__ ) for image in images] if apply_ocr: requires_backends(self , '''pytesseract''' ) snake_case : Optional[int] = [] snake_case : Union[str, Any] = [] for image in images: snake_case , snake_case : List[Any] = apply_tesseract(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) words_batch.append(UpperCAmelCase__ ) boxes_batch.append(UpperCAmelCase__ ) if do_resize: snake_case : Any = [self.resize(image=UpperCAmelCase__ , size=UpperCAmelCase__ , resample=UpperCAmelCase__ ) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) snake_case : int = [flip_channel_order(UpperCAmelCase__ ) for image in images] snake_case : Optional[Any] = [to_channel_dimension_format(UpperCAmelCase__ , UpperCAmelCase__ ) for image in images] snake_case : List[Any] = BatchFeature(data={'''pixel_values''': images} , tensor_type=UpperCAmelCase__ ) if apply_ocr: snake_case : Dict = words_batch snake_case : Dict = boxes_batch return data
84
1
from collections import OrderedDict from typing import List, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a : Optional[Any] = logging.get_logger(__name__) _a : List[Any] = { 'google/efficientnet-b7': 'https://huggingface.co/google/efficientnet-b7/resolve/main/config.json', } class a_ ( a ): A__ : List[str] = 'efficientnet' def __init__( self : int , UpperCAmelCase__ : int = 3 , UpperCAmelCase__ : int = 600 , UpperCAmelCase__ : float = 2.0 , UpperCAmelCase__ : float = 3.1 , UpperCAmelCase__ : int = 8 , UpperCAmelCase__ : List[int] = [3, 3, 5, 3, 5, 5, 3] , UpperCAmelCase__ : List[int] = [32, 16, 24, 40, 80, 112, 192] , UpperCAmelCase__ : List[int] = [16, 24, 40, 80, 112, 192, 320] , UpperCAmelCase__ : List[int] = [] , UpperCAmelCase__ : List[int] = [1, 2, 2, 2, 1, 2, 1] , UpperCAmelCase__ : List[int] = [1, 2, 2, 3, 3, 4, 1] , UpperCAmelCase__ : List[int] = [1, 6, 6, 6, 6, 6, 6] , UpperCAmelCase__ : float = 0.25 , UpperCAmelCase__ : str = "swish" , UpperCAmelCase__ : int = 2_560 , UpperCAmelCase__ : str = "mean" , UpperCAmelCase__ : float = 0.02 , UpperCAmelCase__ : float = 0.001 , UpperCAmelCase__ : float = 0.99 , UpperCAmelCase__ : float = 0.5 , UpperCAmelCase__ : float = 0.2 , **UpperCAmelCase__ : List[str] , ): """simple docstring""" super().__init__(**UpperCAmelCase__ ) snake_case : List[str] = num_channels snake_case : List[str] = image_size snake_case : int = width_coefficient snake_case : Dict = depth_coefficient snake_case : Any = depth_divisor snake_case : List[str] = kernel_sizes snake_case : str = in_channels snake_case : Optional[Any] = out_channels snake_case : Tuple = depthwise_padding snake_case : Union[str, Any] = strides snake_case : Dict = num_block_repeats snake_case : Any = expand_ratios snake_case : Optional[int] = squeeze_expansion_ratio snake_case : Dict = hidden_act snake_case : List[str] = hidden_dim snake_case : Dict = pooling_type snake_case : List[str] = initializer_range snake_case : int = batch_norm_eps snake_case : Optional[Any] = batch_norm_momentum snake_case : Any = dropout_rate snake_case : Optional[int] = drop_connect_rate snake_case : List[Any] = sum(UpperCAmelCase__ ) * 4 class a_ ( a ): A__ : List[str] = version.parse('1.11' ) @property def lowerCAmelCase( self : str ): """simple docstring""" return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" return 1e-5
84
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class a_ : def __init__( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Union[str, Any]=13 , UpperCAmelCase__ : Union[str, Any]=2 , UpperCAmelCase__ : List[Any]=24 , UpperCAmelCase__ : Union[str, Any]=16 , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : str=True , UpperCAmelCase__ : int=32 , UpperCAmelCase__ : Tuple=5 , UpperCAmelCase__ : Dict=4 , UpperCAmelCase__ : Optional[int]=37 , UpperCAmelCase__ : Optional[int]="gelu" , UpperCAmelCase__ : Dict=0.1 , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : Optional[int]=10 , UpperCAmelCase__ : Any=0.02 , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : List[Any]=2 , UpperCAmelCase__ : Optional[Any]=2 , ): """simple docstring""" snake_case : Tuple = parent snake_case : Dict = batch_size snake_case : str = patch_size snake_case : Union[str, Any] = max_length snake_case : str = num_mel_bins snake_case : Any = is_training snake_case : Union[str, Any] = use_labels snake_case : Tuple = hidden_size snake_case : Dict = num_hidden_layers snake_case : Any = num_attention_heads snake_case : Any = intermediate_size snake_case : List[Any] = hidden_act snake_case : str = hidden_dropout_prob snake_case : str = attention_probs_dropout_prob snake_case : str = type_sequence_label_size snake_case : Optional[int] = initializer_range snake_case : str = scope snake_case : int = frequency_stride snake_case : Union[str, Any] = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) snake_case : Optional[int] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 snake_case : Any = (self.max_length - self.patch_size) // self.time_stride + 1 snake_case : Union[str, Any] = frequency_out_dimension * time_out_dimension snake_case : Union[str, Any] = num_patches + 2 def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Optional[int] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) snake_case : str = None if self.use_labels: snake_case : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case : List[str] = self.get_config() return config, input_values, labels def lowerCAmelCase( self : Any ): """simple docstring""" return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCAmelCase__ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : str = ASTModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : Any = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : int = self.prepare_config_and_inputs() ( ( snake_case ) , ( snake_case ) , ( snake_case ) , ) : int = config_and_inputs snake_case : Tuple = {'''input_values''': input_values} return config, inputs_dict @require_torch class a_ ( a , a , unittest.TestCase ): A__ : List[Any] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) A__ : int = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) A__ : Optional[Any] = False A__ : Dict = False A__ : int = False A__ : Optional[int] = False def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : int ): """simple docstring""" if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[int] = ASTModelTester(self ) snake_case : Optional[int] = ConfigTester(self , config_class=UpperCAmelCase__ , has_text_modality=UpperCAmelCase__ , hidden_size=37 ) def lowerCAmelCase( self : List[str] ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def lowerCAmelCase( self : Tuple ): """simple docstring""" pass def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Optional[Any] = model_class(UpperCAmelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase__ , nn.Linear ) ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Any = model_class(UpperCAmelCase__ ) snake_case : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case : str = [*signature.parameters.keys()] snake_case : List[str] = ['''input_values'''] self.assertListEqual(arg_names[:1] , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case : List[str] = ASTModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) def a_ ( ) -> Dict: """simple docstring""" snake_case : Dict = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) snake_case , snake_case : int = torchaudio.load(__magic_name__ ) return audio, sampling_rate @require_torch @require_torchaudio class a_ ( unittest.TestCase ): @cached_property def lowerCAmelCase( self : Any ): """simple docstring""" return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : List[str] = self.default_feature_extractor snake_case : str = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(UpperCAmelCase__ ) snake_case : str = self.default_feature_extractor snake_case , snake_case : int = prepare_audio() snake_case : Optional[int] = audio.squeeze().numpy() snake_case : Optional[Any] = feature_extractor(UpperCAmelCase__ , sampling_rate=UpperCAmelCase__ , return_tensors='''pt''' ).to(UpperCAmelCase__ ) # forward pass with torch.no_grad(): snake_case : Union[str, Any] = model(**UpperCAmelCase__ ) # verify the logits snake_case : Any = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase__ ) snake_case : str = torch.tensor([-0.8760, -7.0042, -8.6602] ).to(UpperCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase__ , atol=1e-4 ) )
84
1
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class a_ ( a ): A__ : List[str] = ['image_processor', 'tokenizer'] A__ : Any = 'CLIPImageProcessor' A__ : Optional[int] = ('CLIPTokenizer', 'CLIPTokenizerFast') def __init__( self : Union[str, Any] , UpperCAmelCase__ : str=None , UpperCAmelCase__ : Union[str, Any]=None , **UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : Optional[int] = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , UpperCAmelCase__ , ) snake_case : List[Any] = kwargs.pop('''feature_extractor''' ) snake_case : Optional[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(UpperCAmelCase__ , UpperCAmelCase__ ) def __call__( self : Any , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : int=None , **UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: snake_case : int = self.tokenizer(UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , **UpperCAmelCase__ ) if images is not None: snake_case : Dict = self.image_processor(UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , **UpperCAmelCase__ ) if text is not None and images is not None: snake_case : Tuple = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**UpperCAmelCase__ ) , tensor_type=UpperCAmelCase__ ) def lowerCAmelCase( self : List[str] , *UpperCAmelCase__ : Dict , **UpperCAmelCase__ : int ): """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : str ): """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : int = self.tokenizer.model_input_names snake_case : Tuple = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , UpperCAmelCase__ , ) return self.image_processor_class @property def lowerCAmelCase( self : Any ): """simple docstring""" warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , UpperCAmelCase__ , ) return self.image_processor
84
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _a : Union[str, Any] = logging.getLogger(__name__) def a_ ( __magic_name__ , __magic_name__ ) -> Optional[int]: """simple docstring""" return (preds == labels).mean() @dataclass class a_ : A__ : str = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) @dataclass class a_ : A__ : str = field(metadata={'help': 'The name of the task to train on: ' + ', '.join(processors.keys() )} ) A__ : str = field(metadata={'help': 'Should contain the data files for the task.'} ) A__ : int = field( default=128 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) A__ : bool = field( default=a , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) def a_ ( ) -> Dict: """simple docstring""" snake_case : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) snake_case , snake_case , snake_case : Tuple = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"Output directory ({training_args.output_dir}) already exists and is not empty. Use" ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , __magic_name__ ) # Set seed set_seed(training_args.seed ) try: snake_case : int = processors[data_args.task_name]() snake_case : List[str] = processor.get_labels() snake_case : str = len(__magic_name__ ) except KeyError: raise ValueError('''Task not found: %s''' % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=__magic_name__ , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) snake_case : str = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) snake_case : Any = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=__magic_name__ , cache_dir=model_args.cache_dir , ) # Get datasets snake_case : Optional[int] = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=__magic_name__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) snake_case : Any = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=__magic_name__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(__magic_name__ ) -> Dict: snake_case : str = np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(__magic_name__ , p.label_ids )} # Data collator snake_case : Dict = DataCollatorWithPadding(__magic_name__ , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer snake_case : List[Any] = Trainer( model=__magic_name__ , args=__magic_name__ , train_dataset=__magic_name__ , eval_dataset=__magic_name__ , compute_metrics=__magic_name__ , data_collator=__magic_name__ , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation snake_case : Optional[int] = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) snake_case : Optional[Any] = trainer.evaluate() snake_case : Union[str, Any] = os.path.join(training_args.output_dir , '''eval_results.txt''' ) if trainer.is_world_master(): with open(__magic_name__ , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(''' %s = %s''' , __magic_name__ , __magic_name__ ) writer.write('''%s = %s\n''' % (key, value) ) results.update(__magic_name__ ) return results def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" main() if __name__ == "__main__": main()
84
1
import re from filelock import FileLock try: import nltk _a : List[str] = True except (ImportError, ModuleNotFoundError): _a : Tuple = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def a_ ( __magic_name__ ) -> str: """simple docstring""" re.sub('''<n>''' , '''''' , __magic_name__ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(__magic_name__ ) )
84
import re def a_ ( __magic_name__ ) -> bool: """simple docstring""" snake_case : List[str] = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(__magic_name__ , __magic_name__ ) ) if __name__ == "__main__": _a : Any = '0094702343221' print(is_sri_lankan_phone_number(phone))
84
1
from __future__ import annotations from PIL import Image # Define glider example _a : Dict = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example _a : Dict = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def a_ ( __magic_name__ ) -> list[list[int]]: """simple docstring""" snake_case : Union[str, Any] = [] for i in range(len(__magic_name__ ) ): snake_case : str = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours snake_case : Optional[int] = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(__magic_name__ ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(__magic_name__ ) - 1: neighbour_count += cells[i + 1][j] if i < len(__magic_name__ ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. snake_case : Any = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(__magic_name__ ) return next_generation def a_ ( __magic_name__ , __magic_name__ ) -> list[Image.Image]: """simple docstring""" snake_case : str = [] for _ in range(__magic_name__ ): # Create output image snake_case : List[Any] = Image.new('''RGB''' , (len(cells[0] ), len(__magic_name__ )) ) snake_case : Any = img.load() # Save cells to image for x in range(len(__magic_name__ ) ): for y in range(len(cells[0] ) ): snake_case : Dict = 255 - cells[y][x] * 255 snake_case : int = (colour, colour, colour) # Save image images.append(__magic_name__ ) snake_case : Union[str, Any] = new_generation(__magic_name__ ) return images if __name__ == "__main__": _a : Optional[Any] = generate_images(GLIDER, 16) images[0].save('out.gif', save_all=True, append_images=images[1:])
84
import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class a_ ( unittest.TestCase ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[int]=7 , UpperCAmelCase__ : Union[str, Any]=3 , UpperCAmelCase__ : int=18 , UpperCAmelCase__ : Optional[int]=30 , UpperCAmelCase__ : Optional[int]=400 , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : int=True , ): """simple docstring""" snake_case : int = size if size is not None else {'''height''': 18, '''width''': 18} snake_case : Optional[Any] = parent snake_case : Any = batch_size snake_case : Any = num_channels snake_case : Union[str, Any] = image_size snake_case : Dict = min_resolution snake_case : Dict = max_resolution snake_case : int = do_resize snake_case : List[str] = size snake_case : List[Any] = apply_ocr def lowerCAmelCase( self : int ): """simple docstring""" return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class a_ ( a , unittest.TestCase ): A__ : List[Any] = LayoutLMvaImageProcessor if is_pytesseract_available() else None def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : Optional[Any] = LayoutLMvaImageProcessingTester(self ) @property def lowerCAmelCase( self : Dict ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[str] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase__ , '''do_resize''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''size''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''apply_ocr''' ) ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 18, '''width''': 18} ) snake_case : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) def lowerCAmelCase( self : str ): """simple docstring""" pass def lowerCAmelCase( self : Dict ): """simple docstring""" # Initialize image_processing snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , Image.Image ) # Test not batched input snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) self.assertIsInstance(encoding.words , UpperCAmelCase__ ) self.assertIsInstance(encoding.boxes , UpperCAmelCase__ ) # Test batched snake_case : Dict = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" # Initialize image_processing snake_case : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , numpify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , np.ndarray ) # Test not batched input snake_case : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched snake_case : List[Any] = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" # Initialize image_processing snake_case : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , torchify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , torch.Tensor ) # Test not batched input snake_case : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched snake_case : Tuple = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" # with apply_OCR = True snake_case : int = LayoutLMvaImageProcessor() from datasets import load_dataset snake_case : List[Any] = load_dataset('''hf-internal-testing/fixtures_docvqa''' , split='''test''' ) snake_case : List[Any] = Image.open(ds[0]['''file'''] ).convert('''RGB''' ) snake_case : Any = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 snake_case : Optional[Any] = [['''11:14''', '''to''', '''11:39''', '''a.m''', '''11:39''', '''to''', '''11:44''', '''a.m.''', '''11:44''', '''a.m.''', '''to''', '''12:25''', '''p.m.''', '''12:25''', '''to''', '''12:58''', '''p.m.''', '''12:58''', '''to''', '''4:00''', '''p.m.''', '''2:00''', '''to''', '''5:00''', '''p.m.''', '''Coffee''', '''Break''', '''Coffee''', '''will''', '''be''', '''served''', '''for''', '''men''', '''and''', '''women''', '''in''', '''the''', '''lobby''', '''adjacent''', '''to''', '''exhibit''', '''area.''', '''Please''', '''move''', '''into''', '''exhibit''', '''area.''', '''(Exhibits''', '''Open)''', '''TRRF''', '''GENERAL''', '''SESSION''', '''(PART''', '''|)''', '''Presiding:''', '''Lee''', '''A.''', '''Waller''', '''TRRF''', '''Vice''', '''President''', '''“Introductory''', '''Remarks”''', '''Lee''', '''A.''', '''Waller,''', '''TRRF''', '''Vice''', '''Presi-''', '''dent''', '''Individual''', '''Interviews''', '''with''', '''TRRF''', '''Public''', '''Board''', '''Members''', '''and''', '''Sci-''', '''entific''', '''Advisory''', '''Council''', '''Mem-''', '''bers''', '''Conducted''', '''by''', '''TRRF''', '''Treasurer''', '''Philip''', '''G.''', '''Kuehn''', '''to''', '''get''', '''answers''', '''which''', '''the''', '''public''', '''refrigerated''', '''warehousing''', '''industry''', '''is''', '''looking''', '''for.''', '''Plus''', '''questions''', '''from''', '''the''', '''floor.''', '''Dr.''', '''Emil''', '''M.''', '''Mrak,''', '''University''', '''of''', '''Cal-''', '''ifornia,''', '''Chairman,''', '''TRRF''', '''Board;''', '''Sam''', '''R.''', '''Cecil,''', '''University''', '''of''', '''Georgia''', '''College''', '''of''', '''Agriculture;''', '''Dr.''', '''Stanley''', '''Charm,''', '''Tufts''', '''University''', '''School''', '''of''', '''Medicine;''', '''Dr.''', '''Robert''', '''H.''', '''Cotton,''', '''ITT''', '''Continental''', '''Baking''', '''Company;''', '''Dr.''', '''Owen''', '''Fennema,''', '''University''', '''of''', '''Wis-''', '''consin;''', '''Dr.''', '''Robert''', '''E.''', '''Hardenburg,''', '''USDA.''', '''Questions''', '''and''', '''Answers''', '''Exhibits''', '''Open''', '''Capt.''', '''Jack''', '''Stoney''', '''Room''', '''TRRF''', '''Scientific''', '''Advisory''', '''Council''', '''Meeting''', '''Ballroom''', '''Foyer''']] # noqa: E231 snake_case : Union[str, Any] = [[[141, 57, 214, 69], [228, 58, 252, 69], [141, 75, 216, 88], [230, 79, 280, 88], [142, 260, 218, 273], [230, 261, 255, 273], [143, 279, 218, 290], [231, 282, 290, 291], [143, 342, 218, 354], [231, 345, 289, 355], [202, 362, 227, 373], [143, 379, 220, 392], [231, 382, 291, 394], [144, 714, 220, 726], [231, 715, 256, 726], [144, 732, 220, 745], [232, 736, 291, 747], [144, 769, 218, 782], [231, 770, 256, 782], [141, 788, 202, 801], [215, 791, 274, 804], [143, 826, 204, 838], [215, 826, 240, 838], [142, 844, 202, 857], [215, 847, 274, 859], [334, 57, 427, 69], [440, 57, 522, 69], [369, 75, 461, 88], [469, 75, 516, 88], [528, 76, 562, 88], [570, 76, 667, 88], [675, 75, 711, 87], [721, 79, 778, 88], [789, 75, 840, 88], [369, 97, 470, 107], [484, 94, 507, 106], [518, 94, 562, 107], [576, 94, 655, 110], [668, 94, 792, 109], [804, 95, 829, 107], [369, 113, 465, 125], [477, 116, 547, 125], [562, 113, 658, 125], [671, 116, 748, 125], [761, 113, 811, 125], [369, 131, 465, 143], [477, 133, 548, 143], [563, 130, 698, 145], [710, 130, 802, 146], [336, 171, 412, 183], [423, 171, 572, 183], [582, 170, 716, 184], [728, 171, 817, 187], [829, 171, 844, 186], [338, 197, 482, 212], [507, 196, 557, 209], [569, 196, 595, 208], [610, 196, 702, 209], [505, 214, 583, 226], [595, 214, 656, 227], [670, 215, 807, 227], [335, 259, 543, 274], [556, 259, 708, 272], [372, 279, 422, 291], [435, 279, 460, 291], [474, 279, 574, 292], [587, 278, 664, 291], [676, 278, 738, 291], [751, 279, 834, 291], [372, 298, 434, 310], [335, 341, 483, 354], [497, 341, 655, 354], [667, 341, 728, 354], [740, 341, 825, 354], [335, 360, 430, 372], [442, 360, 534, 372], [545, 359, 687, 372], [697, 360, 754, 372], [765, 360, 823, 373], [334, 378, 428, 391], [440, 378, 577, 394], [590, 378, 705, 391], [720, 378, 801, 391], [334, 397, 400, 409], [370, 416, 529, 429], [544, 416, 576, 432], [587, 416, 665, 428], [677, 416, 814, 429], [372, 435, 452, 450], [465, 434, 495, 447], [511, 434, 600, 447], [611, 436, 637, 447], [649, 436, 694, 451], [705, 438, 824, 447], [369, 453, 452, 466], [464, 454, 509, 466], [522, 453, 611, 469], [625, 453, 792, 469], [370, 472, 556, 488], [570, 472, 684, 487], [697, 472, 718, 485], [732, 472, 835, 488], [369, 490, 411, 503], [425, 490, 484, 503], [496, 490, 635, 506], [645, 490, 707, 503], [718, 491, 761, 503], [771, 490, 840, 503], [336, 510, 374, 521], [388, 510, 447, 522], [460, 510, 489, 521], [503, 510, 580, 522], [592, 509, 736, 525], [745, 509, 770, 522], [781, 509, 840, 522], [338, 528, 434, 541], [448, 528, 596, 541], [609, 527, 687, 540], [700, 528, 792, 541], [336, 546, 397, 559], [407, 546, 431, 559], [443, 546, 525, 560], [537, 546, 680, 562], [688, 546, 714, 559], [722, 546, 837, 562], [336, 565, 449, 581], [461, 565, 485, 577], [497, 565, 665, 581], [681, 565, 718, 577], [732, 565, 837, 580], [337, 584, 438, 597], [452, 583, 521, 596], [535, 584, 677, 599], [690, 583, 787, 596], [801, 583, 825, 596], [338, 602, 478, 615], [492, 602, 530, 614], [543, 602, 638, 615], [650, 602, 676, 614], [688, 602, 788, 615], [802, 602, 843, 614], [337, 621, 502, 633], [516, 621, 615, 637], [629, 621, 774, 636], [789, 621, 827, 633], [337, 639, 418, 652], [432, 640, 571, 653], [587, 639, 731, 655], [743, 639, 769, 652], [780, 639, 841, 652], [338, 658, 440, 673], [455, 658, 491, 670], [508, 658, 602, 671], [616, 658, 638, 670], [654, 658, 835, 674], [337, 677, 429, 689], [337, 714, 482, 726], [495, 714, 548, 726], [561, 714, 683, 726], [338, 770, 461, 782], [474, 769, 554, 785], [489, 788, 562, 803], [576, 788, 643, 801], [656, 787, 751, 804], [764, 788, 844, 801], [334, 825, 421, 838], [430, 824, 574, 838], [584, 824, 723, 841], [335, 844, 450, 857], [464, 843, 583, 860], [628, 862, 755, 875], [769, 861, 848, 878]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , UpperCAmelCase__ ) self.assertListEqual(encoding.boxes , UpperCAmelCase__ ) # with apply_OCR = False snake_case : str = LayoutLMvaImageProcessor(apply_ocr=UpperCAmelCase__ ) snake_case : Optional[Any] = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) )
84
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _a : Tuple = logging.get_logger(__name__) _a : int = { 'weiweishi/roc-bert-base-zh': 'https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json', } class a_ ( a ): A__ : List[str] = 'roc_bert' def __init__( self : int , UpperCAmelCase__ : List[Any]=30_522 , UpperCAmelCase__ : List[str]=768 , UpperCAmelCase__ : List[Any]=12 , UpperCAmelCase__ : Optional[int]=12 , UpperCAmelCase__ : int=3_072 , UpperCAmelCase__ : List[str]="gelu" , UpperCAmelCase__ : Any=0.1 , UpperCAmelCase__ : Optional[Any]=0.1 , UpperCAmelCase__ : Tuple=512 , UpperCAmelCase__ : int=2 , UpperCAmelCase__ : Dict=0.02 , UpperCAmelCase__ : int=1e-1_2 , UpperCAmelCase__ : Tuple=True , UpperCAmelCase__ : Any=0 , UpperCAmelCase__ : Optional[Any]="absolute" , UpperCAmelCase__ : List[Any]=None , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Tuple=True , UpperCAmelCase__ : Union[str, Any]=768 , UpperCAmelCase__ : Any=910 , UpperCAmelCase__ : str=512 , UpperCAmelCase__ : Optional[Any]=24_858 , UpperCAmelCase__ : List[str]=True , **UpperCAmelCase__ : Any , ): """simple docstring""" snake_case : str = vocab_size snake_case : Tuple = max_position_embeddings snake_case : List[Any] = hidden_size snake_case : Tuple = num_hidden_layers snake_case : str = num_attention_heads snake_case : int = intermediate_size snake_case : Union[str, Any] = hidden_act snake_case : List[str] = hidden_dropout_prob snake_case : str = attention_probs_dropout_prob snake_case : Dict = initializer_range snake_case : Optional[Any] = type_vocab_size snake_case : Optional[int] = layer_norm_eps snake_case : Dict = use_cache snake_case : str = enable_pronunciation snake_case : Union[str, Any] = enable_shape snake_case : List[str] = pronunciation_embed_dim snake_case : Union[str, Any] = pronunciation_vocab_size snake_case : Union[str, Any] = shape_embed_dim snake_case : str = shape_vocab_size snake_case : List[Any] = concat_input snake_case : Any = position_embedding_type snake_case : Dict = classifier_dropout super().__init__(pad_token_id=UpperCAmelCase__ , **UpperCAmelCase__ )
84
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _a : Tuple = logging.get_logger(__name__) # pylint: disable=invalid-name _a : Dict = '\n Examples:\n ```py\n >>> import torch\n >>> import numpy as np\n\n >>> from diffusers import KandinskyV22PriorPipeline, KandinskyV22ControlnetPipeline\n >>> from transformers import pipeline\n >>> from diffusers.utils import load_image\n\n\n >>> def make_hint(image, depth_estimator):\n ... image = depth_estimator(image)["depth"]\n ... image = np.array(image)\n ... image = image[:, :, None]\n ... image = np.concatenate([image, image, image], axis=2)\n ... detected_map = torch.from_numpy(image).float() / 255.0\n ... hint = detected_map.permute(2, 0, 1)\n ... return hint\n\n\n >>> depth_estimator = pipeline("depth-estimation")\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16\n ... )\n >>> pipe_prior = pipe_prior.to("cuda")\n\n >>> pipe = KandinskyV22ControlnetPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-controlnet-depth", torch_dtype=torch.float16\n ... )\n >>> pipe = pipe.to("cuda")\n\n\n >>> img = load_image(\n ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main"\n ... "/kandinsky/cat.png"\n ... ).resize((768, 768))\n\n >>> hint = make_hint(img, depth_estimator).unsqueeze(0).half().to("cuda")\n\n >>> prompt = "A robot, 4k photo"\n >>> negative_prior_prompt = "lowres, text, error, cropped, worst quality, low quality, jpeg artifacts, ugly, duplicate, morbid, mutilated, out of frame, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, mutation, deformed, blurry, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, username, watermark, signature"\n\n >>> generator = torch.Generator(device="cuda").manual_seed(43)\n\n >>> image_emb, zero_image_emb = pipe_prior(\n ... prompt=prompt, negative_prompt=negative_prior_prompt, generator=generator\n ... ).to_tuple()\n\n >>> images = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... hint=hint,\n ... num_inference_steps=50,\n ... generator=generator,\n ... height=768,\n ... width=768,\n ... ).images\n\n >>> images[0].save("robot_cat.png")\n ```\n' def a_ ( __magic_name__ , __magic_name__ , __magic_name__=8 ) -> str: """simple docstring""" snake_case : List[Any] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 snake_case : Tuple = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class a_ ( a ): def __init__( self : Optional[int] , UpperCAmelCase__ : UNetaDConditionModel , UpperCAmelCase__ : DDPMScheduler , UpperCAmelCase__ : VQModel , ): """simple docstring""" super().__init__() self.register_modules( unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ , movq=UpperCAmelCase__ , ) snake_case : List[Any] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def lowerCAmelCase( self : int , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Any ): """simple docstring""" if latents is None: snake_case : int = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=UpperCAmelCase__ , dtype=UpperCAmelCase__ ) else: if latents.shape != shape: raise ValueError(F"Unexpected latents shape, got {latents.shape}, expected {shape}" ) snake_case : Optional[Any] = latents.to(UpperCAmelCase__ ) snake_case : List[Any] = latents * scheduler.init_noise_sigma return latents def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Optional[int]=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) snake_case : Union[str, Any] = torch.device(F"cuda:{gpu_id}" ) snake_case : Dict = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : Any=0 ): """simple docstring""" if is_accelerate_available() and is_accelerate_version('''>=''' , '''0.17.0.dev0''' ): from accelerate import cpu_offload_with_hook else: raise ImportError('''`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.''' ) snake_case : Optional[int] = torch.device(F"cuda:{gpu_id}" ) if self.device.type != "cpu": self.to('''cpu''' , silence_dtype_warnings=UpperCAmelCase__ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) snake_case : List[str] = None for cpu_offloaded_model in [self.unet, self.movq]: snake_case , snake_case : Optional[int] = cpu_offload_with_hook(UpperCAmelCase__ , UpperCAmelCase__ , prev_module_hook=UpperCAmelCase__ ) # We'll offload the last model manually. snake_case : Tuple = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" if not hasattr(self.unet , '''_hf_hook''' ): return self.device for module in self.unet.modules(): if ( hasattr(UpperCAmelCase__ , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(UpperCAmelCase__ ) def __call__( self : List[str] , UpperCAmelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase__ : torch.FloatTensor , UpperCAmelCase__ : int = 512 , UpperCAmelCase__ : int = 512 , UpperCAmelCase__ : int = 100 , UpperCAmelCase__ : float = 4.0 , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase__ : Optional[torch.FloatTensor] = None , UpperCAmelCase__ : Optional[str] = "pil" , UpperCAmelCase__ : bool = True , ): """simple docstring""" snake_case : Optional[int] = self._execution_device snake_case : Union[str, Any] = guidance_scale > 1.0 if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Any = torch.cat(UpperCAmelCase__ , dim=0 ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Union[str, Any] = torch.cat(UpperCAmelCase__ , dim=0 ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : int = torch.cat(UpperCAmelCase__ , dim=0 ) snake_case : List[Any] = image_embeds.shape[0] * num_images_per_prompt if do_classifier_free_guidance: snake_case : Dict = image_embeds.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Optional[Any] = negative_image_embeds.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Tuple = hint.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Optional[Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=UpperCAmelCase__ ) snake_case : List[str] = torch.cat([hint, hint] , dim=0 ).to(dtype=self.unet.dtype , device=UpperCAmelCase__ ) self.scheduler.set_timesteps(UpperCAmelCase__ , device=UpperCAmelCase__ ) snake_case : str = self.scheduler.timesteps snake_case : Optional[Any] = self.movq.config.latent_channels snake_case , snake_case : Optional[Any] = downscale_height_and_width(UpperCAmelCase__ , UpperCAmelCase__ , self.movq_scale_factor ) # create initial latent snake_case : Dict = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , self.scheduler , ) for i, t in enumerate(self.progress_bar(UpperCAmelCase__ ) ): # expand the latents if we are doing classifier free guidance snake_case : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents snake_case : Optional[int] = {'''image_embeds''': image_embeds, '''hint''': hint} snake_case : Any = self.unet( sample=UpperCAmelCase__ , timestep=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , added_cond_kwargs=UpperCAmelCase__ , return_dict=UpperCAmelCase__ , )[0] if do_classifier_free_guidance: snake_case , snake_case : Dict = noise_pred.split(latents.shape[1] , dim=1 ) snake_case , snake_case : Any = noise_pred.chunk(2 ) snake_case , snake_case : Dict = variance_pred.chunk(2 ) snake_case : str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) snake_case : List[str] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , '''variance_type''' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): snake_case , snake_case : Tuple = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 snake_case : List[Any] = self.scheduler.step( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , generator=UpperCAmelCase__ , )[0] # post-processing snake_case : List[Any] = self.movq.decode(UpperCAmelCase__ , force_not_quantize=UpperCAmelCase__ )['''sample'''] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}" ) if output_type in ["np", "pil"]: snake_case : Optional[Any] = image * 0.5 + 0.5 snake_case : int = image.clamp(0 , 1 ) snake_case : List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": snake_case : str = self.numpy_to_pil(UpperCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase__ )
84
1
import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() _a : List[str] = logging.get_logger(__name__) _a : str = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn.grep_linear': 'encoder.layers.*.attention.gru_rel_pos_linear', 'self_attn.relative_attention_bias': 'encoder.layers.*.attention.rel_attn_embed', 'self_attn.grep_a': 'encoder.layers.*.attention.gru_rel_pos_const', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'ctc_proj', 'mask_emb': 'masked_spec_embed', } _a : Union[str, Any] = [ 'ctc_proj', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> List[Any]: """simple docstring""" for attribute in key.split('''.''' ): snake_case : Any = getattr(__magic_name__ , __magic_name__ ) if weight_type is not None: snake_case : int = getattr(__magic_name__ , __magic_name__ ).shape else: snake_case : List[str] = hf_pointer.shape assert hf_shape == value.shape, ( F"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" F" {value.shape} for {full_name}" ) if weight_type == "weight": snake_case : Any = value elif weight_type == "weight_g": snake_case : List[Any] = value elif weight_type == "weight_v": snake_case : List[Any] = value elif weight_type == "bias": snake_case : Optional[Any] = value else: snake_case : Any = value logger.info(F"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def a_ ( __magic_name__ , __magic_name__ ) -> int: """simple docstring""" snake_case : List[Any] = [] snake_case : Dict = fairseq_model.state_dict() snake_case : List[str] = hf_model.feature_extractor for name, value in fairseq_dict.items(): snake_case : List[Any] = False if "conv_layers" in name: load_conv_layer( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , hf_model.config.feat_extract_norm == '''group''' , ) snake_case : Any = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: snake_case : str = True if "*" in mapped_key: snake_case : str = name.split(__magic_name__ )[0].split('''.''' )[-2] snake_case : Optional[Any] = mapped_key.replace('''*''' , __magic_name__ ) if "weight_g" in name: snake_case : List[Any] = '''weight_g''' elif "weight_v" in name: snake_case : str = '''weight_v''' elif "bias" in name and "relative_attention_bias" not in name: snake_case : Union[str, Any] = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj snake_case : List[Any] = '''weight''' else: snake_case : Any = None set_recursively(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) continue if not is_used: unused_weights.append(__magic_name__ ) logger.warning(F"Unused weights: {unused_weights}" ) def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" snake_case : Any = full_name.split('''conv_layers.''' )[-1] snake_case : Tuple = name.split('''.''' ) snake_case : Tuple = int(items[0] ) snake_case : Tuple = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) snake_case : Optional[int] = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) snake_case : Union[str, Any] = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was" " found." ) snake_case : Optional[Any] = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found." ) snake_case : str = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(__magic_name__ ) @torch.no_grad() def a_ ( __magic_name__ , __magic_name__ , __magic_name__=None ) -> Any: """simple docstring""" snake_case : Dict = torch.load(__magic_name__ ) snake_case : Union[str, Any] = WavLMConfigOrig(checkpoint['''cfg'''] ) snake_case : List[str] = WavLMOrig(__magic_name__ ) model.load_state_dict(checkpoint['''model'''] ) model.eval() if config_path is not None: snake_case : str = WavLMConfig.from_pretrained(__magic_name__ ) else: snake_case : Optional[Any] = WavLMConfig() snake_case : str = WavLMModel(__magic_name__ ) recursively_load_weights(__magic_name__ , __magic_name__ ) hf_wavlm.save_pretrained(__magic_name__ ) if __name__ == "__main__": _a : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') _a : Dict = parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
84
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _a : List[Any] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class a_ ( a , unittest.TestCase ): A__ : Dict = ReformerTokenizer A__ : Optional[int] = ReformerTokenizerFast A__ : str = True A__ : Tuple = False A__ : str = True def lowerCAmelCase( self : List[Any] ): """simple docstring""" super().setUp() snake_case : str = ReformerTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : int = '''<s>''' snake_case : List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase__ ) , UpperCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase__ ) , UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Any = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(UpperCAmelCase__ ) , 1_000 ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def lowerCAmelCase( self : Dict ): """simple docstring""" if not self.test_rust_tokenizer: return snake_case : Any = self.get_tokenizer() snake_case : str = self.get_rust_tokenizer() snake_case : Tuple = '''I was born in 92000, and this is falsé.''' snake_case : str = tokenizer.tokenize(UpperCAmelCase__ ) snake_case : int = rust_tokenizer.tokenize(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : Union[str, Any] = tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) snake_case : List[str] = rust_tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : List[str] = self.get_rust_tokenizer() snake_case : Optional[int] = tokenizer.encode(UpperCAmelCase__ ) snake_case : Optional[Any] = rust_tokenizer.encode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : List[Any]=15 ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): snake_case : str = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) # Simple input snake_case : Union[str, Any] = '''This is a simple input''' snake_case : List[str] = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case : int = ('''This is a simple input''', '''This is a pair''') snake_case : int = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Simple input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Simple input self.assertRaises( UpperCAmelCase__ , tokenizer_r.batch_encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' , ) # Pair input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Pair input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Pair input self.assertRaises( UpperCAmelCase__ , tokenizer_r.batch_encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' , ) def lowerCAmelCase( self : str ): """simple docstring""" pass def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Union[str, Any] = ReformerTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) snake_case : List[str] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(UpperCAmelCase__ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , [285, 46, 10, 170, 382] , ) snake_case : int = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) snake_case : int = tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) snake_case : List[Any] = tokenizer.convert_ids_to_tokens(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def lowerCAmelCase( self : Tuple ): """simple docstring""" return ReformerTokenizer.from_pretrained('''google/reformer-crime-and-punishment''' ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : Any = '''Hello World!''' snake_case : Optional[Any] = [126, 32, 262, 152, 38, 72, 287] self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @slow def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[Any] = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) snake_case : Dict = [ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @require_torch @slow def lowerCAmelCase( self : List[Any] ): """simple docstring""" import torch from transformers import ReformerConfig, ReformerModel # Build sequence snake_case : Any = list(self.big_tokenizer.get_vocab().keys() )[:10] snake_case : Union[str, Any] = ''' '''.join(UpperCAmelCase__ ) snake_case : Optional[int] = self.big_tokenizer.encode_plus(UpperCAmelCase__ , return_tensors='''pt''' ) snake_case : List[str] = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors='''pt''' ) snake_case : Optional[Any] = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) snake_case : Tuple = encoded_sequence['''input_ids'''].shape snake_case : List[Any] = ReformerModel(UpperCAmelCase__ ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**UpperCAmelCase__ ) model(**UpperCAmelCase__ ) @slow def lowerCAmelCase( self : Optional[int] ): """simple docstring""" # fmt: off snake_case : Tuple = {'''input_ids''': [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 snake_case : Tuple = [ '''This is a very simple sentence.''', '''The quick brown fox jumps over the lazy dog.''', ] self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase__ , model_name='''google/reformer-crime-and-punishment''' , revision='''0e6c3decb8211d49bf881013425dc8b0448b3f5a''' , padding=UpperCAmelCase__ , sequences=UpperCAmelCase__ , )
84
1
import argparse import requests import torch from PIL import Image from torchvision.transforms import Compose, Normalize, Resize, ToTensor from transformers import SwinaSRConfig, SwinaSRForImageSuperResolution, SwinaSRImageProcessor def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" snake_case : List[str] = SwinaSRConfig() if "Swin2SR_ClassicalSR_X4_64" in checkpoint_url: snake_case : List[str] = 4 elif "Swin2SR_CompressedSR_X4_48" in checkpoint_url: snake_case : Any = 4 snake_case : int = 48 snake_case : Any = '''pixelshuffle_aux''' elif "Swin2SR_Lightweight_X2_64" in checkpoint_url: snake_case : int = [6, 6, 6, 6] snake_case : Any = 60 snake_case : Union[str, Any] = [6, 6, 6, 6] snake_case : Any = '''pixelshuffledirect''' elif "Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR" in checkpoint_url: snake_case : str = 4 snake_case : List[str] = '''nearest+conv''' elif "Swin2SR_Jpeg_dynamic" in checkpoint_url: snake_case : str = 1 snake_case : Tuple = 1 snake_case : List[str] = 126 snake_case : int = 7 snake_case : Union[str, Any] = 255.0 snake_case : int = '''''' return config def a_ ( __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" if "patch_embed.proj" in name and "layers" not in name: snake_case : Dict = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: snake_case : Dict = name.replace('''patch_embed.norm''' , '''embeddings.patch_embeddings.layernorm''' ) if "layers" in name: snake_case : Any = name.replace('''layers''' , '''encoder.stages''' ) if "residual_group.blocks" in name: snake_case : List[str] = name.replace('''residual_group.blocks''' , '''layers''' ) if "attn.proj" in name: snake_case : Any = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: snake_case : List[Any] = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: snake_case : List[Any] = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: snake_case : Optional[int] = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: snake_case : Optional[int] = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: snake_case : Tuple = name.replace('''mlp.fc2''' , '''output.dense''' ) if "q_bias" in name: snake_case : Tuple = name.replace('''q_bias''' , '''query.bias''' ) if "k_bias" in name: snake_case : Optional[Any] = name.replace('''k_bias''' , '''key.bias''' ) if "v_bias" in name: snake_case : Union[str, Any] = name.replace('''v_bias''' , '''value.bias''' ) if "cpb_mlp" in name: snake_case : Union[str, Any] = name.replace('''cpb_mlp''' , '''continuous_position_bias_mlp''' ) if "patch_embed.proj" in name: snake_case : Optional[Any] = name.replace('''patch_embed.proj''' , '''patch_embed.projection''' ) if name == "norm.weight": snake_case : Dict = '''layernorm.weight''' if name == "norm.bias": snake_case : List[Any] = '''layernorm.bias''' if "conv_first" in name: snake_case : Dict = name.replace('''conv_first''' , '''first_convolution''' ) if ( "upsample" in name or "conv_before_upsample" in name or "conv_bicubic" in name or "conv_up" in name or "conv_hr" in name or "conv_last" in name or "aux" in name ): # heads if "conv_last" in name: snake_case : List[Any] = name.replace('''conv_last''' , '''final_convolution''' ) if config.upsampler in ["pixelshuffle", "pixelshuffle_aux", "nearest+conv"]: if "conv_before_upsample.0" in name: snake_case : Dict = name.replace('''conv_before_upsample.0''' , '''conv_before_upsample''' ) if "upsample.0" in name: snake_case : List[str] = name.replace('''upsample.0''' , '''upsample.convolution_0''' ) if "upsample.2" in name: snake_case : Optional[Any] = name.replace('''upsample.2''' , '''upsample.convolution_1''' ) snake_case : Dict = '''upsample.''' + name elif config.upsampler == "pixelshuffledirect": snake_case : str = name.replace('''upsample.0.weight''' , '''upsample.conv.weight''' ) snake_case : List[str] = name.replace('''upsample.0.bias''' , '''upsample.conv.bias''' ) else: pass else: snake_case : int = '''swin2sr.''' + name return name def a_ ( __magic_name__ , __magic_name__ ) -> Union[str, Any]: """simple docstring""" for key in orig_state_dict.copy().keys(): snake_case : Union[str, Any] = orig_state_dict.pop(__magic_name__ ) if "qkv" in key: snake_case : str = key.split('''.''' ) snake_case : Any = int(key_split[1] ) snake_case : List[str] = int(key_split[4] ) snake_case : Tuple = config.embed_dim if "weight" in key: snake_case : Tuple = val[:dim, :] snake_case : List[str] = val[dim : dim * 2, :] snake_case : Optional[Any] = val[-dim:, :] else: snake_case : Optional[Any] = val[:dim] snake_case : Dict = val[dim : dim * 2] snake_case : Tuple = val[-dim:] pass else: snake_case : Any = val return orig_state_dict def a_ ( __magic_name__ , __magic_name__ , __magic_name__ ) -> Optional[Any]: """simple docstring""" snake_case : Optional[int] = get_config(__magic_name__ ) snake_case : Union[str, Any] = SwinaSRForImageSuperResolution(__magic_name__ ) model.eval() snake_case : int = torch.hub.load_state_dict_from_url(__magic_name__ , map_location='''cpu''' ) snake_case : Optional[int] = convert_state_dict(__magic_name__ , __magic_name__ ) snake_case , snake_case : List[Any] = model.load_state_dict(__magic_name__ , strict=__magic_name__ ) if len(__magic_name__ ) > 0: raise ValueError('''Missing keys when converting: {}'''.format(__magic_name__ ) ) for key in unexpected_keys: if not ("relative_position_index" in key or "relative_coords_table" in key or "self_mask" in key): raise ValueError(F"Unexpected key {key} in state_dict" ) # verify values snake_case : int = '''https://github.com/mv-lab/swin2sr/blob/main/testsets/real-inputs/shanghai.jpg?raw=true''' snake_case : Union[str, Any] = Image.open(requests.get(__magic_name__ , stream=__magic_name__ ).raw ).convert('''RGB''' ) snake_case : str = SwinaSRImageProcessor() # pixel_values = processor(image, return_tensors="pt").pixel_values snake_case : Union[str, Any] = 126 if '''Jpeg''' in checkpoint_url else 256 snake_case : int = Compose( [ Resize((image_size, image_size) ), ToTensor(), Normalize(mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ), ] ) snake_case : Optional[Any] = transforms(__magic_name__ ).unsqueeze(0 ) if config.num_channels == 1: snake_case : Union[str, Any] = pixel_values[:, 0, :, :].unsqueeze(1 ) snake_case : Optional[Any] = model(__magic_name__ ) # assert values if "Swin2SR_ClassicalSR_X2_64" in checkpoint_url: snake_case : Union[str, Any] = torch.Size([1, 3, 512, 512] ) snake_case : List[str] = torch.tensor( [[-0.7087, -0.7138, -0.6721], [-0.8340, -0.8095, -0.7298], [-0.9149, -0.8414, -0.7940]] ) elif "Swin2SR_ClassicalSR_X4_64" in checkpoint_url: snake_case : Union[str, Any] = torch.Size([1, 3, 1_024, 1_024] ) snake_case : Optional[int] = torch.tensor( [[-0.7775, -0.8105, -0.8933], [-0.7764, -0.8356, -0.9225], [-0.7976, -0.8686, -0.9579]] ) elif "Swin2SR_CompressedSR_X4_48" in checkpoint_url: # TODO values didn't match exactly here snake_case : Union[str, Any] = torch.Size([1, 3, 1_024, 1_024] ) snake_case : List[str] = torch.tensor( [[-0.8035, -0.7504, -0.7491], [-0.8538, -0.8124, -0.7782], [-0.8804, -0.8651, -0.8493]] ) elif "Swin2SR_Lightweight_X2_64" in checkpoint_url: snake_case : str = torch.Size([1, 3, 512, 512] ) snake_case : Tuple = torch.tensor( [[-0.7669, -0.8662, -0.8767], [-0.8810, -0.9962, -0.9820], [-0.9340, -1.0322, -1.1149]] ) elif "Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR" in checkpoint_url: snake_case : Any = torch.Size([1, 3, 1_024, 1_024] ) snake_case : str = torch.tensor( [[-0.5238, -0.5557, -0.6321], [-0.6016, -0.5903, -0.6391], [-0.6244, -0.6334, -0.6889]] ) assert ( outputs.reconstruction.shape == expected_shape ), F"Shape of reconstruction should be {expected_shape}, but is {outputs.reconstruction.shape}" assert torch.allclose(outputs.reconstruction[0, 0, :3, :3] , __magic_name__ , atol=1e-3 ) print('''Looks ok!''' ) snake_case : Union[str, Any] = { '''https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X2_64.pth''': ( '''swin2SR-classical-sr-x2-64''' ), '''https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X4_64.pth''': ( '''swin2SR-classical-sr-x4-64''' ), '''https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_CompressedSR_X4_48.pth''': ( '''swin2SR-compressed-sr-x4-48''' ), '''https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_Lightweight_X2_64.pth''': ( '''swin2SR-lightweight-x2-64''' ), '''https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR.pth''': ( '''swin2SR-realworld-sr-x4-64-bsrgan-psnr''' ), } snake_case : int = url_to_name[checkpoint_url] if pytorch_dump_folder_path is not None: print(F"Saving model {model_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(__magic_name__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) processor.save_pretrained(__magic_name__ ) if push_to_hub: model.push_to_hub(F"caidas/{model_name}" ) processor.push_to_hub(F"caidas/{model_name}" ) if __name__ == "__main__": _a : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X2_64.pth', type=str, help='URL of the original Swin2SR checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument('--push_to_hub', action='store_true', help='Whether to push the converted model to the hub.') _a : List[Any] = parser.parse_args() convert_swinasr_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
84
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def a_ ( __magic_name__ ) -> Tuple: """simple docstring""" snake_case , snake_case : Any = image.size snake_case , snake_case : List[str] = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 snake_case : List[str] = image.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) snake_case : Dict = np.array(__magic_name__ ).astype(np.floataa ) / 255.0 snake_case : Optional[Any] = image[None].transpose(0 , 3 , 1 , 2 ) snake_case : Tuple = torch.from_numpy(__magic_name__ ) return 2.0 * image - 1.0 class a_ ( a ): def __init__( self : Optional[Any] , UpperCAmelCase__ : VQModel , UpperCAmelCase__ : UNetaDModel , UpperCAmelCase__ : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): """simple docstring""" super().__init__() self.register_modules(vqvae=UpperCAmelCase__ , unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ ) @torch.no_grad() def __call__( self : Any , UpperCAmelCase__ : Union[torch.Tensor, PIL.Image.Image] = None , UpperCAmelCase__ : Optional[int] = 1 , UpperCAmelCase__ : Optional[int] = 100 , UpperCAmelCase__ : Optional[float] = 0.0 , UpperCAmelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase__ : Optional[str] = "pil" , UpperCAmelCase__ : bool = True , ): """simple docstring""" if isinstance(UpperCAmelCase__ , PIL.Image.Image ): snake_case : Optional[int] = 1 elif isinstance(UpperCAmelCase__ , torch.Tensor ): snake_case : Any = image.shape[0] else: raise ValueError(F"`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(UpperCAmelCase__ )}" ) if isinstance(UpperCAmelCase__ , PIL.Image.Image ): snake_case : Optional[Any] = preprocess(UpperCAmelCase__ ) snake_case , snake_case : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image snake_case : List[Any] = (batch_size, self.unet.config.in_channels // 2, height, width) snake_case : str = next(self.unet.parameters() ).dtype snake_case : Dict = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=self.device , dtype=UpperCAmelCase__ ) snake_case : Any = image.to(device=self.device , dtype=UpperCAmelCase__ ) # set timesteps and move to the correct device self.scheduler.set_timesteps(UpperCAmelCase__ , device=self.device ) snake_case : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler snake_case : Union[str, Any] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] snake_case : Any = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) snake_case : Optional[Any] = {} if accepts_eta: snake_case : Dict = eta for t in self.progress_bar(UpperCAmelCase__ ): # concat latents and low resolution image in the channel dimension. snake_case : Optional[int] = torch.cat([latents, image] , dim=1 ) snake_case : str = self.scheduler.scale_model_input(UpperCAmelCase__ , UpperCAmelCase__ ) # predict the noise residual snake_case : int = self.unet(UpperCAmelCase__ , UpperCAmelCase__ ).sample # compute the previous noisy sample x_t -> x_t-1 snake_case : Any = self.scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample # decode the image latents with the VQVAE snake_case : Optional[int] = self.vqvae.decode(UpperCAmelCase__ ).sample snake_case : int = torch.clamp(UpperCAmelCase__ , -1.0 , 1.0 ) snake_case : Dict = image / 2 + 0.5 snake_case : int = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case : Any = self.numpy_to_pil(UpperCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase__ )
84
1
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class a_ ( _lowerCamelCase ): def __init__( self : Any , UpperCAmelCase__ : str , UpperCAmelCase__ : str = None , UpperCAmelCase__ : Optional[Any] = None , UpperCAmelCase__ : int = True , UpperCAmelCase__ : str = None , UpperCAmelCase__ : List[Any] = False , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : Tuple = True , UpperCAmelCase__ : Optional[Any] = "arrow" , **UpperCAmelCase__ : List[Any] , ): """simple docstring""" super().__init__( split=A__ , features=A__ , cache_dir=A__ , keep_in_memory=A__ , streaming=A__ , **A__ , ) snake_case : int = load_from_cache_file snake_case : Optional[Any] = file_format snake_case : List[Any] = Spark( df=A__ , features=A__ , cache_dir=A__ , working_dir=A__ , **A__ , ) def lowerCAmelCase( self : Dict ): """simple docstring""" if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) snake_case : Union[str, Any] = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=A__ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
700
import os import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers.models.realm.configuration_realm import RealmConfig from transformers.models.realm.retrieval_realm import _REALM_BLOCK_RECORDS_FILENAME, RealmRetriever from transformers.models.realm.tokenization_realm import VOCAB_FILES_NAMES, RealmTokenizer class a_ ( a ): def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[Any] = tempfile.mkdtemp() snake_case : Dict = 5 # Realm tok snake_case : str = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''test''', '''question''', '''this''', '''is''', '''the''', '''first''', '''second''', '''third''', '''fourth''', '''fifth''', '''record''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] snake_case : Tuple = os.path.join(self.tmpdirname , '''realm_tokenizer''' ) os.makedirs(UpperCAmelCase__ , exist_ok=UpperCAmelCase__ ) snake_case : Any = os.path.join(UpperCAmelCase__ , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) snake_case : Tuple = os.path.join(self.tmpdirname , '''realm_block_records''' ) os.makedirs(UpperCAmelCase__ , exist_ok=UpperCAmelCase__ ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" return RealmTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''realm_tokenizer''' ) ) def lowerCAmelCase( self : Any ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Any = RealmConfig(num_block_records=self.num_block_records ) return config def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Optional[int] = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''question''': ['''foo''', '''bar'''], '''answers''': [['''Foo''', '''Bar'''], ['''Bar''']], } ) return dataset def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Dict = np.array( [ b'''This is the first record''', b'''This is the second record''', b'''This is the third record''', b'''This is the fourth record''', b'''This is the fifth record''', b'''This is a longer longer longer record''', ] , dtype=UpperCAmelCase__ , ) return block_records def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Tuple = RealmRetriever( block_records=self.get_dummy_block_records() , tokenizer=self.get_tokenizer() , ) return retriever def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[str] = self.get_config() snake_case : Optional[Any] = self.get_dummy_retriever() snake_case : Optional[int] = retriever.tokenizer snake_case : Dict = np.array([0, 3] , dtype='''long''' ) snake_case : Optional[int] = tokenizer(['''Test question'''] ).input_ids snake_case : Union[str, Any] = tokenizer( ['''the fourth'''] , add_special_tokens=UpperCAmelCase__ , return_token_type_ids=UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , ).input_ids snake_case : Optional[Any] = config.reader_seq_len snake_case , snake_case , snake_case , snake_case : List[str] = retriever( UpperCAmelCase__ , UpperCAmelCase__ , answer_ids=UpperCAmelCase__ , max_length=UpperCAmelCase__ , return_tensors='''np''' ) self.assertEqual(len(UpperCAmelCase__ ) , 2 ) self.assertEqual(len(UpperCAmelCase__ ) , 2 ) self.assertEqual(len(UpperCAmelCase__ ) , 2 ) self.assertEqual(concat_inputs.input_ids.shape , (2, 10) ) self.assertEqual(concat_inputs.attention_mask.shape , (2, 10) ) self.assertEqual(concat_inputs.token_type_ids.shape , (2, 10) ) self.assertEqual(concat_inputs.special_tokens_mask.shape , (2, 10) ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[0] ) , ['''[CLS]''', '''test''', '''question''', '''[SEP]''', '''this''', '''is''', '''the''', '''first''', '''record''', '''[SEP]'''] , ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[1] ) , ['''[CLS]''', '''test''', '''question''', '''[SEP]''', '''this''', '''is''', '''the''', '''fourth''', '''record''', '''[SEP]'''] , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[Any] = self.get_config() snake_case : Optional[int] = self.get_dummy_retriever() snake_case : List[str] = retriever.tokenizer snake_case : Optional[Any] = np.array([0, 3, 5] , dtype='''long''' ) snake_case : Optional[int] = tokenizer(['''Test question'''] ).input_ids snake_case : Any = tokenizer( ['''the fourth''', '''longer longer'''] , add_special_tokens=UpperCAmelCase__ , return_token_type_ids=UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , ).input_ids snake_case : List[Any] = config.reader_seq_len snake_case , snake_case , snake_case , snake_case : Union[str, Any] = retriever( UpperCAmelCase__ , UpperCAmelCase__ , answer_ids=UpperCAmelCase__ , max_length=UpperCAmelCase__ , return_tensors='''np''' ) self.assertEqual([False, True, True] , UpperCAmelCase__ ) self.assertEqual([[-1, -1, -1], [6, -1, -1], [6, 7, 8]] , UpperCAmelCase__ ) self.assertEqual([[-1, -1, -1], [7, -1, -1], [7, 8, 9]] , UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : int = self.get_dummy_retriever() retriever.save_pretrained(os.path.join(self.tmpdirname , '''realm_block_records''' ) ) # Test local path snake_case : Optional[Any] = retriever.from_pretrained(os.path.join(self.tmpdirname , '''realm_block_records''' ) ) self.assertEqual(retriever.block_records[0] , b'''This is the first record''' ) # Test mocked remote path with patch('''transformers.models.realm.retrieval_realm.hf_hub_download''' ) as mock_hf_hub_download: snake_case : Any = os.path.join( os.path.join(self.tmpdirname , '''realm_block_records''' ) , _REALM_BLOCK_RECORDS_FILENAME ) snake_case : Any = RealmRetriever.from_pretrained('''google/realm-cc-news-pretrained-openqa''' ) self.assertEqual(retriever.block_records[0] , b'''This is the first record''' )
84
0
# # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def a_ ( *__magic_name__ ) -> int: """simple docstring""" with open(__A , '''r''' ) as fh: fcntl.flock(__A , fcntl.LOCK_EX ) try: print(*__A ) finally: fcntl.flock(__A , fcntl.LOCK_UN ) _a : Dict = int(os.environ['LOCAL_RANK']) torch.cuda.set_device(local_rank) _a : str = torch.device('cuda', local_rank) _a : Optional[int] = socket.gethostname() _a : int = f"[{hostname}-{local_rank}]" try: # test distributed dist.init_process_group('nccl') dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank _a : str = dist.get_rank() _a : Tuple = dist.get_world_size() printflock(f"{gpu} is OK (global rank: {rank}/{world_size})") dist.barrier() if rank == 0: printflock(f"pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}") except Exception: printflock(f"{gpu} is broken") raise
701
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _a : str = {'configuration_encoder_decoder': ['EncoderDecoderConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = ['EncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Dict = ['TFEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = ['FlaxEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys _a : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
84
0
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore _a : Union[str, Any] = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" _a : Optional[Any] = [file for file in filepaths if file != file.lower()] if upper_files: print(f"{len(upper_files)} files contain uppercase characters:") print('\n'.join(upper_files) + '\n') _a : List[str] = [file for file in filepaths if ' ' in file] if space_files: print(f"{len(space_files)} files contain space characters:") print('\n'.join(space_files) + '\n') _a : List[Any] = [file for file in filepaths if '-' in file] if hyphen_files: print(f"{len(hyphen_files)} files contain hyphen characters:") print('\n'.join(hyphen_files) + '\n') _a : List[str] = [file for file in filepaths if os.sep not in file] if nodir_files: print(f"{len(nodir_files)} files are not in a directory:") print('\n'.join(nodir_files) + '\n') _a : str = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
702
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer _a : str = logging.get_logger(__name__) _a : Optional[int] = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} _a : Optional[Any] = { 'vocab_file': { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json' ), }, } _a : Union[str, Any] = { 'yjernite/retribert-base-uncased': 512, } _a : Tuple = { 'yjernite/retribert-base-uncased': {'do_lower_case': True}, } class a_ ( a ): A__ : List[str] = VOCAB_FILES_NAMES A__ : Any = PRETRAINED_VOCAB_FILES_MAP A__ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Any = PRETRAINED_INIT_CONFIGURATION A__ : Optional[Any] = RetriBertTokenizer A__ : Any = ['input_ids', 'attention_mask'] def __init__( self : Optional[int] , UpperCAmelCase__ : Any=None , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Dict="[UNK]" , UpperCAmelCase__ : str="[SEP]" , UpperCAmelCase__ : Union[str, Any]="[PAD]" , UpperCAmelCase__ : Dict="[CLS]" , UpperCAmelCase__ : Optional[Any]="[MASK]" , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Optional[int]=None , **UpperCAmelCase__ : Dict , ): """simple docstring""" super().__init__( UpperCAmelCase__ , tokenizer_file=UpperCAmelCase__ , do_lower_case=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , tokenize_chinese_chars=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ , **UpperCAmelCase__ , ) snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , UpperCAmelCase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' , UpperCAmelCase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , UpperCAmelCase__ ) != tokenize_chinese_chars ): snake_case : int = getattr(UpperCAmelCase__ , normalizer_state.pop('''type''' ) ) snake_case : List[Any] = do_lower_case snake_case : Union[str, Any] = strip_accents snake_case : int = tokenize_chinese_chars snake_case : int = normalizer_class(**UpperCAmelCase__ ) snake_case : Union[str, Any] = do_lower_case def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Union[str, Any]=None ): """simple docstring""" snake_case : str = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ): """simple docstring""" snake_case : List[Any] = [self.sep_token_id] snake_case : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCAmelCase( self : str , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ): """simple docstring""" snake_case : Tuple = self._tokenizer.model.save(UpperCAmelCase__ , name=UpperCAmelCase__ ) return tuple(UpperCAmelCase__ )
84
0
import importlib.metadata import warnings from copy import deepcopy from packaging import version from ..utils import logging from .import_utils import is_accelerate_available, is_bitsandbytes_available if is_bitsandbytes_available(): import bitsandbytes as bnb import torch import torch.nn as nn from ..pytorch_utils import ConvaD if is_accelerate_available(): from accelerate import init_empty_weights from accelerate.utils import find_tied_parameters _a : int = logging.get_logger(__name__) def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__=None , __magic_name__=None ) -> Optional[Any]: """simple docstring""" if "." in tensor_name: snake_case : Dict = tensor_name.split('''.''' ) for split in splits[:-1]: snake_case : Tuple = getattr(__magic_name__ , __magic_name__ ) if new_module is None: raise ValueError(F"{module} has no attribute {split}." ) snake_case : Union[str, Any] = new_module snake_case : Optional[int] = splits[-1] if tensor_name not in module._parameters and tensor_name not in module._buffers: raise ValueError(F"{module} does not have a parameter or a buffer named {tensor_name}." ) snake_case : Union[str, Any] = tensor_name in module._buffers snake_case : Optional[int] = getattr(__magic_name__ , __magic_name__ ) if old_value.device == torch.device('''meta''' ) and device not in ["meta", torch.device('''meta''' )] and value is None: raise ValueError(F"{tensor_name} is on the meta device, we need a `value` to put in on {device}." ) snake_case : str = False snake_case : Optional[Any] = False if is_buffer or not is_bitsandbytes_available(): snake_case : Union[str, Any] = False snake_case : List[str] = False else: snake_case : Dict = hasattr(bnb.nn , '''Params4bit''' ) and isinstance(module._parameters[tensor_name] , bnb.nn.Paramsabit ) snake_case : List[str] = isinstance(module._parameters[tensor_name] , bnb.nn.IntaParams ) if is_abit or is_abit: snake_case : Dict = module._parameters[tensor_name] if param.device.type != "cuda": if value is None: snake_case : Dict = old_value.to(__magic_name__ ) elif isinstance(__magic_name__ , torch.Tensor ): snake_case : Dict = value.to('''cpu''' ) if value.dtype == torch.inta: snake_case : Dict = version.parse(importlib.metadata.version('''bitsandbytes''' ) ) > version.parse( '''0.37.2''' ) if not is_abit_serializable: raise ValueError( '''Detected int8 weights but the version of bitsandbytes is not compatible with int8 serialization. ''' '''Make sure to download the latest `bitsandbytes` version. `pip install --upgrade bitsandbytes`.''' ) else: snake_case : int = torch.tensor(__magic_name__ , device='''cpu''' ) # Support models using `Conv1D` in place of `nn.Linear` (e.g. gpt2) by transposing the weight matrix prior to quantization. # Since weights are saved in the correct "orientation", we skip transposing when loading. if issubclass(module.source_cls , __magic_name__ ) and fpaa_statistics is None: snake_case : List[str] = new_value.T snake_case : Union[str, Any] = old_value.__dict__ if is_abit: snake_case : Tuple = bnb.nn.IntaParams(__magic_name__ , requires_grad=__magic_name__ , **__magic_name__ ).to(__magic_name__ ) elif is_abit: snake_case : Tuple = bnb.nn.Paramsabit(__magic_name__ , requires_grad=__magic_name__ , **__magic_name__ ).to(__magic_name__ ) snake_case : Any = new_value if fpaa_statistics is not None: setattr(module.weight , '''SCB''' , fpaa_statistics.to(__magic_name__ ) ) else: if value is None: snake_case : List[str] = old_value.to(__magic_name__ ) elif isinstance(__magic_name__ , torch.Tensor ): snake_case : Union[str, Any] = value.to(__magic_name__ ) else: snake_case : Optional[Any] = torch.tensor(__magic_name__ , device=__magic_name__ ) if is_buffer: snake_case : Dict = new_value else: snake_case : Union[str, Any] = nn.Parameter(__magic_name__ , requires_grad=old_value.requires_grad ) snake_case : List[Any] = new_value def a_ ( __magic_name__ , __magic_name__=None , __magic_name__=None , __magic_name__=None , __magic_name__=False ) -> Dict: """simple docstring""" for name, module in model.named_children(): if current_key_name is None: snake_case : List[Any] = [] current_key_name.append(__magic_name__ ) if (isinstance(__magic_name__ , nn.Linear ) or isinstance(__magic_name__ , __magic_name__ )) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` if not any(key in '''.'''.join(__magic_name__ ) for key in modules_to_not_convert ): with init_empty_weights(): if isinstance(__magic_name__ , __magic_name__ ): snake_case , snake_case : Union[str, Any] = module.weight.shape else: snake_case : List[str] = module.in_features snake_case : Dict = module.out_features if quantization_config.quantization_method() == "llm_int8": snake_case : Union[str, Any] = bnb.nn.LinearabitLt( __magic_name__ , __magic_name__ , module.bias is not None , has_fpaa_weights=quantization_config.llm_inta_has_fpaa_weight , threshold=quantization_config.llm_inta_threshold , ) snake_case : Optional[int] = True else: if ( quantization_config.llm_inta_skip_modules is not None and name in quantization_config.llm_inta_skip_modules ): pass else: snake_case : int = bnb.nn.Linearabit( __magic_name__ , __magic_name__ , module.bias is not None , quantization_config.bnb_abit_compute_dtype , compress_statistics=quantization_config.bnb_abit_use_double_quant , quant_type=quantization_config.bnb_abit_quant_type , ) snake_case : List[str] = True # Store the module class in case we need to transpose the weight later snake_case : Any = type(__magic_name__ ) # Force requires grad to False to avoid unexpected errors model._modules[name].requires_grad_(__magic_name__ ) if len(list(module.children() ) ) > 0: snake_case , snake_case : List[Any] = _replace_with_bnb_linear( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , has_been_replaced=__magic_name__ , ) # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def a_ ( __magic_name__ , __magic_name__=None , __magic_name__=None , __magic_name__=None ) -> Optional[int]: """simple docstring""" snake_case : List[str] = ['''lm_head'''] if modules_to_not_convert is None else modules_to_not_convert snake_case , snake_case : List[str] = _replace_with_bnb_linear( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) if not has_been_replaced: logger.warning( '''You are loading your model in 8bit or 4bit but no linear modules were found in your model.''' ''' Please double check your model architecture, or submit an issue on github if you think this is''' ''' a bug.''' ) return model def a_ ( *__magic_name__ , **__magic_name__ ) -> Optional[int]: """simple docstring""" warnings.warn( '''`replace_8bit_linear` will be deprecated in a future version, please use `replace_with_bnb_linear` instead''' , __magic_name__ , ) return replace_with_bnb_linear(*__magic_name__ , **__magic_name__ ) def a_ ( *__magic_name__ , **__magic_name__ ) -> Dict: """simple docstring""" warnings.warn( '''`set_module_8bit_tensor_to_device` will be deprecated in a future version, please use `set_module_quantized_tensor_to_device` instead''' , __magic_name__ , ) return set_module_quantized_tensor_to_device(*__magic_name__ , **__magic_name__ ) def a_ ( __magic_name__ ) -> Union[str, Any]: """simple docstring""" snake_case : int = deepcopy(__magic_name__ ) # this has 0 cost since it is done inside `init_empty_weights` context manager` tied_model.tie_weights() snake_case : str = find_tied_parameters(__magic_name__ ) # For compatibility with Accelerate < 0.18 if isinstance(__magic_name__ , __magic_name__ ): snake_case : str = sum(list(tied_params.values() ) , [] ) + list(tied_params.keys() ) else: snake_case : Union[str, Any] = sum(__magic_name__ , [] ) snake_case : Dict = len(__magic_name__ ) > 0 # Check if it is a base model snake_case : str = not hasattr(__magic_name__ , model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head snake_case : Optional[int] = list(model.named_children() ) snake_case : List[Any] = [list_modules[-1][0]] # add last module together with tied weights snake_case : str = set(__magic_name__ ) - set(__magic_name__ ) snake_case : Optional[int] = list(set(__magic_name__ ) ) + list(__magic_name__ ) # remove ".weight" from the keys snake_case : List[Any] = ['''.weight''', '''.bias'''] snake_case : List[Any] = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: snake_case : str = name.replace(__magic_name__ , '''''' ) filtered_module_names.append(__magic_name__ ) return filtered_module_names
703
import string import numpy def a_ ( __magic_name__ , __magic_name__ ) -> int: """simple docstring""" return b if a == 0 else greatest_common_divisor(b % a , __magic_name__ ) class a_ : A__ : List[Any] = string.ascii_uppercase + string.digits # This cipher takes alphanumerics into account # i.e. a total of 36 characters # take x and return x % len(key_string) A__ : List[str] = numpy.vectorize(lambda a : x % 36 ) A__ : Dict = numpy.vectorize(a ) def __init__( self : List[str] , UpperCAmelCase__ : numpy.ndarray ): """simple docstring""" snake_case : int = self.modulus(UpperCAmelCase__ ) # mod36 calc's on the encrypt key self.check_determinant() # validate the determinant of the encryption key snake_case : List[str] = encrypt_key.shape[0] def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : str ): """simple docstring""" return self.key_string.index(UpperCAmelCase__ ) def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : int ): """simple docstring""" return self.key_string[round(UpperCAmelCase__ )] def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[Any] = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: snake_case : Tuple = det % len(self.key_string ) snake_case : Tuple = len(self.key_string ) if greatest_common_divisor(UpperCAmelCase__ , len(self.key_string ) ) != 1: snake_case : List[Any] = ( F"determinant modular {req_l} of encryption key({det}) " F"is not co prime w.r.t {req_l}.\nTry another key." ) raise ValueError(UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Optional[int] = [char for char in text.upper() if char in self.key_string] snake_case : Optional[int] = chars[-1] while len(UpperCAmelCase__ ) % self.break_key != 0: chars.append(UpperCAmelCase__ ) return "".join(UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Optional[int] = self.process_text(text.upper() ) snake_case : Optional[int] = '''''' for i in range(0 , len(UpperCAmelCase__ ) - self.break_key + 1 , self.break_key ): snake_case : int = text[i : i + self.break_key] snake_case : int = [self.replace_letters(UpperCAmelCase__ ) for char in batch] snake_case : Tuple = numpy.array([vec] ).T snake_case : Optional[Any] = self.modulus(self.encrypt_key.dot(UpperCAmelCase__ ) ).T.tolist()[ 0 ] snake_case : Dict = ''''''.join( self.replace_digits(UpperCAmelCase__ ) for num in batch_encrypted ) encrypted += encrypted_batch return encrypted def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Optional[int] = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: snake_case : int = det % len(self.key_string ) snake_case : Dict = None for i in range(len(self.key_string ) ): if (det * i) % len(self.key_string ) == 1: snake_case : Any = i break snake_case : Any = ( det_inv * numpy.linalg.det(self.encrypt_key ) * numpy.linalg.inv(self.encrypt_key ) ) return self.to_int(self.modulus(UpperCAmelCase__ ) ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : str ): """simple docstring""" snake_case : Any = self.make_decrypt_key() snake_case : Optional[Any] = self.process_text(text.upper() ) snake_case : int = '''''' for i in range(0 , len(UpperCAmelCase__ ) - self.break_key + 1 , self.break_key ): snake_case : Any = text[i : i + self.break_key] snake_case : int = [self.replace_letters(UpperCAmelCase__ ) for char in batch] snake_case : List[str] = numpy.array([vec] ).T snake_case : Optional[Any] = self.modulus(decrypt_key.dot(UpperCAmelCase__ ) ).T.tolist()[0] snake_case : int = ''''''.join( self.replace_digits(UpperCAmelCase__ ) for num in batch_decrypted ) decrypted += decrypted_batch return decrypted def a_ ( ) -> None: """simple docstring""" snake_case : Any = int(input('''Enter the order of the encryption key: ''' ) ) snake_case : List[Any] = [] print('''Enter each row of the encryption key with space separated integers''' ) for _ in range(__magic_name__ ): snake_case : Optional[Any] = [int(__magic_name__ ) for x in input().split()] hill_matrix.append(__magic_name__ ) snake_case : List[str] = HillCipher(numpy.array(__magic_name__ ) ) print('''Would you like to encrypt or decrypt some text? (1 or 2)''' ) snake_case : int = input('''\n1. Encrypt\n2. Decrypt\n''' ) if option == "1": snake_case : List[Any] = input('''What text would you like to encrypt?: ''' ) print('''Your encrypted text is:''' ) print(hc.encrypt(__magic_name__ ) ) elif option == "2": snake_case : int = input('''What text would you like to decrypt?: ''' ) print('''Your decrypted text is:''' ) print(hc.decrypt(__magic_name__ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
84
0
import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_cuda from accelerate.utils import ProjectConfiguration, set_seed _a : Dict = logging.getLogger(__name__) def a_ ( __magic_name__=2 , __magic_name__=3 , __magic_name__=16 , __magic_name__ = 10 , __magic_name__ = 2 ) -> Tuple: """simple docstring""" def get_dataset(__magic_name__ ): snake_case : int = torch.randn(batch_size * n_batches , 1 ) return TensorDataset(_lowercase , a * x + b + 0.1 * torch.randn(batch_size * n_batches , 1 ) ) snake_case : Dict = get_dataset(_lowercase ) snake_case : Dict = get_dataset(_lowercase ) snake_case : str = DataLoader(_lowercase , shuffle=_lowercase , batch_size=_lowercase , num_workers=4 ) snake_case : Any = DataLoader(_lowercase , shuffle=_lowercase , batch_size=_lowercase , num_workers=4 ) return (train_dataloader, valid_dataloader) def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__=None ) -> Dict: """simple docstring""" snake_case : Any = [] for epoch in range(_lowercase ): # Train quickly model.train() for batch in dataloader: snake_case : List[Any] = batch snake_case : Dict = model(_lowercase ) snake_case : Any = torch.nn.functional.mse_loss(_lowercase , _lowercase ) accelerator.backward(_lowercase ) optimizer.step() optimizer.zero_grad() rands.append(random.random() ) # Introduce some randomness if scheduler is not None: scheduler.step() return rands class a_ ( nn.Module ): def __init__( self : Optional[int] ): """simple docstring""" super().__init__() snake_case : Any = nn.Parameter(torch.randn(1 ) ) snake_case : Optional[Any] = nn.Parameter(torch.randn(1 ) ) def lowerCAmelCase( self : str , UpperCAmelCase__ : List[Any] ): """simple docstring""" return x * self.a + self.b class a_ ( unittest.TestCase ): def lowerCAmelCase( self : str ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) snake_case : Union[str, Any] = DummyModel() snake_case : List[Any] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) snake_case : Union[str, Any] = dummy_dataloaders() snake_case : List[Any] = ProjectConfiguration(total_limit=1 , project_dir=UpperCamelCase_ , automatic_checkpoint_naming=UpperCamelCase_ ) # Train baseline snake_case : List[Any] = Accelerator(project_config=UpperCamelCase_ ) snake_case : Union[str, Any] = accelerator.prepare( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # Save initial accelerator.save_state() # Save second state accelerator.save_state() self.assertEqual(len(os.listdir(accelerator.project_dir ) ) , 1 ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) snake_case : Any = DummyModel() snake_case : Union[str, Any] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) snake_case : Dict = dummy_dataloaders() # Train baseline snake_case : List[Any] = Accelerator() snake_case : Dict = accelerator.prepare( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # Save initial snake_case : Any = os.path.join(UpperCamelCase_ , '''initial''' ) accelerator.save_state(UpperCamelCase_ ) (snake_case) : Union[str, Any] = model.a.item(), model.b.item() snake_case : Optional[Any] = optimizer.state_dict() snake_case : int = train(3 , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) (snake_case) : str = model.a.item(), model.b.item() snake_case : Union[str, Any] = optimizer.state_dict() # Train partially set_seed(42 ) snake_case : int = DummyModel() snake_case : int = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) snake_case : Any = dummy_dataloaders() snake_case : List[str] = Accelerator() snake_case : List[str] = accelerator.prepare( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) accelerator.load_state(UpperCamelCase_ ) (snake_case) : List[Any] = model.a.item(), model.b.item() snake_case : int = optimizer.state_dict() self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) snake_case : str = train(2 , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # Save everything snake_case : Tuple = os.path.join(UpperCamelCase_ , '''checkpoint''' ) accelerator.save_state(UpperCamelCase_ ) # Load everything back in and make sure all states work accelerator.load_state(UpperCamelCase_ ) test_rands += train(1 , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) (snake_case) : str = model.a.item(), model.b.item() snake_case : int = optimizer.state_dict() self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) snake_case : Optional[Any] = DummyModel() snake_case : List[str] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) snake_case : Dict = dummy_dataloaders() snake_case : Union[str, Any] = ProjectConfiguration(automatic_checkpoint_naming=UpperCamelCase_ ) # Train baseline snake_case : Optional[Any] = Accelerator(project_dir=UpperCamelCase_ , project_config=UpperCamelCase_ ) snake_case : int = accelerator.prepare( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # Save initial accelerator.save_state() (snake_case) : int = model.a.item(), model.b.item() snake_case : Dict = optimizer.state_dict() snake_case : Optional[int] = train(3 , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) (snake_case) : Optional[int] = model.a.item(), model.b.item() snake_case : List[str] = optimizer.state_dict() # Train partially set_seed(42 ) snake_case : Tuple = DummyModel() snake_case : List[Any] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) snake_case : List[Any] = dummy_dataloaders() snake_case : Optional[Any] = ProjectConfiguration(iteration=1 , automatic_checkpoint_naming=UpperCamelCase_ ) snake_case : Optional[int] = Accelerator(project_dir=UpperCamelCase_ , project_config=UpperCamelCase_ ) snake_case : Dict = accelerator.prepare( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) accelerator.load_state(os.path.join(UpperCamelCase_ , '''checkpoints''' , '''checkpoint_0''' ) ) (snake_case) : str = model.a.item(), model.b.item() snake_case : Tuple = optimizer.state_dict() self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) snake_case : List[str] = train(2 , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # Save everything accelerator.save_state() # Load everything back in and make sure all states work accelerator.load_state(os.path.join(UpperCamelCase_ , '''checkpoints''' , '''checkpoint_1''' ) ) test_rands += train(1 , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) (snake_case) : Any = model.a.item(), model.b.item() snake_case : List[Any] = optimizer.state_dict() self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) def lowerCAmelCase( self : str ): """simple docstring""" snake_case : str = torch.tensor([1, 2, 3] ) snake_case : Union[str, Any] = torch.tensor([2, 3, 4] ) snake_case : Union[str, Any] = DummyModel() snake_case : Union[str, Any] = torch.optim.Adam(net.parameters() ) snake_case : int = Accelerator() with self.assertRaises(UpperCamelCase_ ) as ve: accelerator.register_for_checkpointing(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) snake_case : int = str(ve.exception ) self.assertTrue('''Item at index 0''' in message ) self.assertTrue('''Item at index 1''' in message ) self.assertFalse('''Item at index 2''' in message ) self.assertFalse('''Item at index 3''' in message ) def lowerCAmelCase( self : Tuple ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) snake_case : Dict = DummyModel() snake_case : List[str] = torch.optim.Adam(params=model.parameters() , lr=1e-3 ) snake_case : List[str] = torch.optim.lr_scheduler.StepLR(UpperCamelCase_ , step_size=1 , gamma=0.99 ) snake_case : str = dummy_dataloaders() snake_case : Any = ProjectConfiguration(automatic_checkpoint_naming=UpperCamelCase_ ) # Train baseline snake_case : List[str] = Accelerator(project_dir=UpperCamelCase_ , project_config=UpperCamelCase_ ) snake_case : List[str] = accelerator.prepare( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # Save initial accelerator.save_state() snake_case : str = scheduler.state_dict() train(3 , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) self.assertNotEqual(UpperCamelCase_ , scheduler.state_dict() ) # Load everything back in and make sure all states work accelerator.load_state(os.path.join(UpperCamelCase_ , '''checkpoints''' , '''checkpoint_0''' ) ) self.assertEqual(UpperCamelCase_ , scheduler.state_dict() ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) snake_case : Optional[Any] = DummyModel() snake_case : List[str] = ProjectConfiguration(automatic_checkpoint_naming=UpperCamelCase_ , total_limit=2 ) # Train baseline snake_case : List[str] = Accelerator(project_dir=UpperCamelCase_ , project_config=UpperCamelCase_ ) snake_case : List[Any] = accelerator.prepare(UpperCamelCase_ ) # Save 3 states: for _ in range(11 ): accelerator.save_state() self.assertTrue(not os.path.exists(os.path.join(UpperCamelCase_ , '''checkpoints''' , '''checkpoint_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase_ , '''checkpoints''' , '''checkpoint_9''' ) ) ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase_ , '''checkpoints''' , '''checkpoint_10''' ) ) ) @require_cuda def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : List[str] = ['torchrun', F"--nproc_per_node={torch.cuda.device_count()}", inspect.getfile(self.__class__ )] execute_subprocess_async(UpperCamelCase_ , env=os.environ.copy() ) if __name__ == "__main__": _a : Union[str, Any] = '''/tmp/accelerate/state_checkpointing''' _a : Any = DummyModel() _a : int = torch.optim.Adam(params=model.parameters(), lr=1e-3) _a : str = torch.optim.lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.99) _a : int = dummy_dataloaders() _a : Dict = ProjectConfiguration(automatic_checkpoint_naming=True) # Train baseline _a : Optional[Any] = Accelerator(project_dir=savedir, project_config=project_config, mixed_precision='no') if accelerator.process_index == 0: if os.path.exists(savedir): shutil.rmtree(savedir) os.makedirs(savedir) _a : Optional[Any] = accelerator.prepare( model, optimizer, train_dataloader, valid_dataloader, scheduler ) _a : Any = accelerator.prepare(model, optimizer) train(3, model, train_dataloader, optimizer, accelerator, scheduler) # Check that the intial optimizer is loaded on the GPU for group in optimizer.param_groups: _a : List[Any] = group['''params'''][0].device break assert param_device.type == accelerator.device.type _a : Optional[Any] = model.cpu() accelerator.wait_for_everyone() accelerator.save_state() accelerator.wait_for_everyone() # Check CPU state accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='cpu') for group in optimizer.param_groups: _a : Tuple = group['''params'''][0].device break assert ( param_device.type == torch.device('cpu').type ), f"Loaded optimizer states did not match, expected to be loaded on the CPU but got {param_device}" # Check device state model.to(accelerator.device) accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='on_device') for group in optimizer.param_groups: _a : Optional[int] = group['''params'''][0].device break assert ( param_device.type == accelerator.device.type ), f"Loaded optimizer states did not match, expected to be loaded on {accelerator.device} but got {param_device}" # Check error with pytest.raises(TypeError, match='Unsupported optimizer map location passed'): accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='invalid') accelerator.wait_for_everyone() if accelerator.process_index == 0: shutil.rmtree(savedir) accelerator.wait_for_everyone()
704
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class a_ ( a ): A__ : List[Any] = 'Salesforce/blip-image-captioning-base' A__ : Dict = ( 'This is a tool that generates a description of an image. It takes an input named `image` which should be the ' 'image to caption, and returns a text that contains the description in English.' ) A__ : str = 'image_captioner' A__ : Dict = AutoModelForVisionaSeq A__ : Optional[Any] = ['image'] A__ : List[str] = ['text'] def __init__( self : List[str] , *UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" requires_backends(self , ['''vision'''] ) super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : "Image" ): """simple docstring""" return self.pre_processor(images=UpperCAmelCase__ , return_tensors='''pt''' ) def lowerCAmelCase( self : Any , UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" return self.model.generate(**UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] , UpperCAmelCase__ : List[Any] ): """simple docstring""" return self.pre_processor.batch_decode(UpperCAmelCase__ , skip_special_tokens=UpperCAmelCase__ )[0].strip()
84
0
'''simple docstring''' import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(lowercase__ ) , 'Tatoeba directory does not exist.' ) class a_ ( unittest.TestCase ): @cached_property def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Any = tempfile.mkdtemp() return TatoebaConverter(save_dir=__lowerCamelCase ) @slow def lowerCAmelCase( self : str ): """simple docstring""" self.resolver.convert_models(['''heb-eng'''] ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : List[Any] = self.resolver.write_model_card('''opus-mt-he-en''' , dry_run=__lowerCamelCase ) assert mmeta["long_pair"] == "heb-eng"
705
def a_ ( __magic_name__ ) -> bool: """simple docstring""" if p < 2: raise ValueError('''p should not be less than 2!''' ) elif p == 2: return True snake_case : int = 4 snake_case : Optional[Any] = (1 << p) - 1 for _ in range(p - 2 ): snake_case : Optional[Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
84
0
from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax _a : Union[str, Any] = logging.get_logger(__name__) @add_end_docstrings(__lowerCamelCase ) class a_ ( __lowerCamelCase ): def __init__( self : List[str] , **UpperCAmelCase__ : Any ): """simple docstring""" super().__init__(**UpperCAmelCase_ ) requires_backends(self , '''vision''' ) self.check_model_type( TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if self.framework == '''tf''' else MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING ) def __call__( self : Optional[int] , UpperCAmelCase__ : Union[str, List[str], "Image", List["Image"]] , **UpperCAmelCase__ : Dict ): """simple docstring""" return super().__call__(UpperCAmelCase_ , **UpperCAmelCase_ ) def lowerCAmelCase( self : List[Any] , **UpperCAmelCase__ : Dict ): """simple docstring""" snake_case : List[str] = {} if "candidate_labels" in kwargs: snake_case : Dict = kwargs['candidate_labels'] if "hypothesis_template" in kwargs: snake_case : Tuple = kwargs['hypothesis_template'] return preprocess_params, {}, {} def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Optional[int]="This is a photo of {}." ): """simple docstring""" snake_case : Union[str, Any] = load_image(UpperCAmelCase_ ) snake_case : List[str] = self.image_processor(images=[image] , return_tensors=self.framework ) snake_case : Dict = candidate_labels snake_case : Any = [hypothesis_template.format(UpperCAmelCase_ ) for x in candidate_labels] snake_case : Union[str, Any] = self.tokenizer(UpperCAmelCase_ , return_tensors=self.framework , padding=UpperCAmelCase_ ) snake_case : Union[str, Any] = [text_inputs] return inputs def lowerCAmelCase( self : str , UpperCAmelCase__ : List[str] ): """simple docstring""" snake_case : Union[str, Any] = model_inputs.pop('''candidate_labels''' ) snake_case : Optional[Any] = model_inputs.pop('''text_inputs''' ) if isinstance(text_inputs[0] , UpperCAmelCase_ ): snake_case : List[str] = text_inputs[0] else: # Batching case. snake_case : List[Any] = text_inputs[0][0] snake_case : Dict = self.model(**UpperCAmelCase_ , **UpperCAmelCase_ ) snake_case : str = { 'candidate_labels': candidate_labels, 'logits': outputs.logits_per_image, } return model_outputs def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : Tuple ): """simple docstring""" snake_case : List[Any] = model_outputs.pop('''candidate_labels''' ) snake_case : Any = model_outputs['logits'][0] if self.framework == "pt": snake_case : Union[str, Any] = logits.softmax(dim=-1 ).squeeze(-1 ) snake_case : str = probs.tolist() if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): snake_case : Dict = [scores] elif self.framework == "tf": snake_case : str = stable_softmax(UpperCAmelCase_ , axis=-1 ) snake_case : Optional[int] = probs.numpy().tolist() else: raise ValueError(F"Unsupported framework: {self.framework}" ) snake_case : Dict = [ {'score': score, 'label': candidate_label} for score, candidate_label in sorted(zip(UpperCAmelCase_ , UpperCAmelCase_ ) , key=lambda UpperCAmelCase__ : -x[0] ) ] return result
706
from sklearn.metrics import fa_score import datasets _a : List[str] = '\nThe F1 score is the harmonic mean of the precision and recall. It can be computed with the equation:\nF1 = 2 * (precision * recall) / (precision + recall)\n' _a : Dict = '\nArgs:\n predictions (`list` of `int`): Predicted labels.\n references (`list` of `int`): Ground truth labels.\n labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None.\n pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1.\n average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`.\n\n - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary.\n - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives.\n - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.\n - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall.\n - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification).\n sample_weight (`list` of `float`): Sample weights Defaults to None.\n\nReturns:\n f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better.\n\nExamples:\n\n Example 1-A simple binary example\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0])\n >>> print(results)\n {\'f1\': 0.5}\n\n Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0)\n >>> print(round(results[\'f1\'], 2))\n 0.67\n\n Example 3-The same simple binary example as in Example 1, but with `sample_weight` included.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3])\n >>> print(round(results[\'f1\'], 2))\n 0.35\n\n Example 4-A multiclass example, with different values for the `average` input.\n >>> predictions = [0, 2, 1, 0, 0, 1]\n >>> references = [0, 1, 2, 0, 1, 2]\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro")\n >>> print(round(results[\'f1\'], 2))\n 0.33\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=None)\n >>> print(results)\n {\'f1\': array([0.8, 0. , 0. ])}\n' _a : List[Any] = '\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P.\n and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and\n Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.},\n journal={Journal of Machine Learning Research},\n volume={12},\n pages={2825--2830},\n year={2011}\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): def lowerCAmelCase( self : Any ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''int32''' ) ), '''references''': datasets.Sequence(datasets.Value('''int32''' ) ), } if self.config_name == '''multilabel''' else { '''predictions''': datasets.Value('''int32''' ), '''references''': datasets.Value('''int32''' ), } ) , reference_urls=['''https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html'''] , ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Optional[Any]=1 , UpperCAmelCase__ : List[str]="binary" , UpperCAmelCase__ : str=None ): """simple docstring""" snake_case : List[Any] = fa_score( UpperCAmelCase__ , UpperCAmelCase__ , labels=UpperCAmelCase__ , pos_label=UpperCAmelCase__ , average=UpperCAmelCase__ , sample_weight=UpperCAmelCase__ ) return {"f1": float(UpperCAmelCase__ ) if score.size == 1 else score}
84
0
import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class a_ ( UpperCAmelCase_ ): A__ : List[str] = (KDPMaDiscreteScheduler,) A__ : Dict = 10 def lowerCAmelCase( self : List[Any] , **UpperCAmelCase__ : Optional[Any] ): """simple docstring""" snake_case : str = { 'num_train_timesteps': 1_100, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', } config.update(**_lowercase ) return config def lowerCAmelCase( self : Dict ): """simple docstring""" for timesteps in [10, 50, 100, 1_000]: self.check_over_configs(num_train_timesteps=_lowercase ) def lowerCAmelCase( self : Tuple ): """simple docstring""" for beta_start, beta_end in zip([0.0_0001, 0.0001, 0.001] , [0.0002, 0.002, 0.02] ): self.check_over_configs(beta_start=_lowercase , beta_end=_lowercase ) def lowerCAmelCase( self : Dict ): """simple docstring""" for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=_lowercase ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_lowercase ) def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : Dict = self.scheduler_classes[0] snake_case : int = self.get_scheduler_config(prediction_type='''v_prediction''' ) snake_case : Union[str, Any] = scheduler_class(**_lowercase ) scheduler.set_timesteps(self.num_inference_steps ) snake_case : Optional[int] = self.dummy_model() snake_case : Union[str, Any] = self.dummy_sample_deter * scheduler.init_noise_sigma snake_case : Union[str, Any] = sample.to(_lowercase ) for i, t in enumerate(scheduler.timesteps ): snake_case : Optional[int] = scheduler.scale_model_input(_lowercase , _lowercase ) snake_case : Tuple = model(_lowercase , _lowercase ) snake_case : Any = scheduler.step(_lowercase , _lowercase , _lowercase ) snake_case : str = output.prev_sample snake_case : Optional[Any] = torch.sum(torch.abs(_lowercase ) ) snake_case : List[Any] = torch.mean(torch.abs(_lowercase ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6_9_3_4e-0_7 ) < 1e-2 assert abs(result_mean.item() - 6.1_1_1_2e-1_0 ) < 1e-3 else: # CUDA assert abs(result_sum.item() - 4.6_9_3_4_2_8_6_5_0_1_7_0_9_7_2e-0_7 ) < 1e-2 assert abs(result_mean.item() - 0.0002 ) < 1e-3 def lowerCAmelCase( self : Dict ): """simple docstring""" if torch_device == "mps": return snake_case : str = self.scheduler_classes[0] snake_case : Tuple = self.get_scheduler_config() snake_case : Tuple = scheduler_class(**_lowercase ) scheduler.set_timesteps(self.num_inference_steps ) snake_case : Optional[int] = self.dummy_model() snake_case : List[str] = self.dummy_sample_deter * scheduler.init_noise_sigma snake_case : str = sample.to(_lowercase ) for i, t in enumerate(scheduler.timesteps ): snake_case : Dict = scheduler.scale_model_input(_lowercase , _lowercase ) snake_case : Union[str, Any] = model(_lowercase , _lowercase ) snake_case : List[Any] = scheduler.step(_lowercase , _lowercase , _lowercase ) snake_case : List[Any] = output.prev_sample snake_case : Union[str, Any] = torch.sum(torch.abs(_lowercase ) ) snake_case : str = torch.mean(torch.abs(_lowercase ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 20.4125 ) < 1e-2 assert abs(result_mean.item() - 0.0266 ) < 1e-3 else: # CUDA assert abs(result_sum.item() - 20.4125 ) < 1e-2 assert abs(result_mean.item() - 0.0266 ) < 1e-3 def lowerCAmelCase( self : str ): """simple docstring""" if torch_device == "mps": return snake_case : str = self.scheduler_classes[0] snake_case : Dict = self.get_scheduler_config() snake_case : Optional[int] = scheduler_class(**_lowercase ) scheduler.set_timesteps(self.num_inference_steps , device=_lowercase ) snake_case : Dict = self.dummy_model() snake_case : List[Any] = self.dummy_sample_deter.to(_lowercase ) * scheduler.init_noise_sigma for t in scheduler.timesteps: snake_case : Tuple = scheduler.scale_model_input(_lowercase , _lowercase ) snake_case : Union[str, Any] = model(_lowercase , _lowercase ) snake_case : Dict = scheduler.step(_lowercase , _lowercase , _lowercase ) snake_case : Union[str, Any] = output.prev_sample snake_case : Any = torch.sum(torch.abs(_lowercase ) ) snake_case : Optional[Any] = torch.mean(torch.abs(_lowercase ) ) if str(_lowercase ).startswith('''cpu''' ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 20.4125 ) < 1e-2 assert abs(result_mean.item() - 0.0266 ) < 1e-3 else: # CUDA assert abs(result_sum.item() - 20.4125 ) < 1e-2 assert abs(result_mean.item() - 0.0266 ) < 1e-3
707
def a_ ( __magic_name__ ) -> int: """simple docstring""" if not isinstance(__magic_name__ , __magic_name__ ): raise TypeError('''only integers accepted as input''' ) else: snake_case : str = str(abs(__magic_name__ ) ) snake_case : Optional[Any] = [list(__magic_name__ ) for char in range(len(__magic_name__ ) )] for index in range(len(__magic_name__ ) ): num_transpositions[index].pop(__magic_name__ ) return max( int(''''''.join(list(__magic_name__ ) ) ) for transposition in num_transpositions ) if __name__ == "__main__": __import__('doctest').testmod()
84
0
import pytest from datasets.parallel import ParallelBackendConfig, parallel_backend from datasets.utils.py_utils import map_nested from .utils import require_dill_gt_0_3_2, require_joblibspark, require_not_windows def a_ ( __magic_name__ ) -> Optional[int]: # picklable for multiprocessing """simple docstring""" return i + 1 @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows def a_ ( ) -> Any: """simple docstring""" with parallel_backend('''spark''' ): assert ParallelBackendConfig.backend_name == "spark" snake_case : List[str] = [1, 2, 3] with pytest.raises(__A ): with parallel_backend('''unsupported backend''' ): map_nested(__A , __A , num_proc=2 ) with pytest.raises(__A ): with parallel_backend('''unsupported backend''' ): map_nested(__A , __A , num_proc=-1 ) @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows @pytest.mark.parametrize('''num_proc''' , [2, -1] ) def a_ ( __magic_name__ ) -> Optional[Any]: """simple docstring""" snake_case : Any = [1, 2] snake_case : Dict = {'''a''': 1, '''b''': 2} snake_case : str = {'''a''': [1, 2], '''b''': [3, 4]} snake_case : List[str] = {'''a''': {'''1''': 1}, '''b''': 2} snake_case : Union[str, Any] = {'''a''': 1, '''b''': 2, '''c''': 3, '''d''': 4} snake_case : str = [2, 3] snake_case : Dict = {'''a''': 2, '''b''': 3} snake_case : int = {'''a''': [2, 3], '''b''': [4, 5]} snake_case : Tuple = {'''a''': {'''1''': 2}, '''b''': 3} snake_case : Tuple = {'''a''': 2, '''b''': 3, '''c''': 4, '''d''': 5} with parallel_backend('''spark''' ): assert map_nested(__A , __A , num_proc=__A ) == expected_map_nested_sa assert map_nested(__A , __A , num_proc=__A ) == expected_map_nested_sa assert map_nested(__A , __A , num_proc=__A ) == expected_map_nested_sa assert map_nested(__A , __A , num_proc=__A ) == expected_map_nested_sa assert map_nested(__A , __A , num_proc=__A ) == expected_map_nested_sa
708
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class a_ : def __init__( self : Union[str, Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any]=99 , UpperCAmelCase__ : Dict=13 , UpperCAmelCase__ : int=7 , UpperCAmelCase__ : Any=9 , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Tuple=False , UpperCAmelCase__ : Tuple=32 , UpperCAmelCase__ : Dict=5 , UpperCAmelCase__ : Optional[int]=4 , UpperCAmelCase__ : List[Any]=37 , UpperCAmelCase__ : Union[str, Any]=8 , UpperCAmelCase__ : Optional[Any]=0.1 , UpperCAmelCase__ : str=0.002 , UpperCAmelCase__ : str=1 , UpperCAmelCase__ : Any=0 , UpperCAmelCase__ : Union[str, Any]=0 , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Optional[Any]=None , ): """simple docstring""" snake_case : Union[str, Any] = parent snake_case : Union[str, Any] = batch_size snake_case : Any = encoder_seq_length snake_case : str = decoder_seq_length # For common tests snake_case : Optional[int] = self.decoder_seq_length snake_case : Optional[Any] = is_training snake_case : List[Any] = use_attention_mask snake_case : Union[str, Any] = use_labels snake_case : Any = vocab_size snake_case : Optional[int] = hidden_size snake_case : List[str] = num_hidden_layers snake_case : Union[str, Any] = num_attention_heads snake_case : Any = d_ff snake_case : Any = relative_attention_num_buckets snake_case : Optional[Any] = dropout_rate snake_case : int = initializer_factor snake_case : Optional[Any] = eos_token_id snake_case : Dict = pad_token_id snake_case : Optional[Any] = decoder_start_token_id snake_case : Union[str, Any] = None snake_case : List[str] = decoder_layers def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" return TaConfig.from_pretrained('''google/umt5-base''' ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Any , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : List[Any]=None , ): """simple docstring""" if attention_mask is None: snake_case : Union[str, Any] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: snake_case : Any = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: snake_case : List[Any] = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=UpperCAmelCase__ ) if decoder_head_mask is None: snake_case : Tuple = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase__ ) if cross_attn_head_mask is None: snake_case : Union[str, Any] = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=UpperCAmelCase__ ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input snake_case : List[str] = input_ids.clamp(self.pad_token_id + 1 ) snake_case : List[str] = decoder_input_ids.clamp(self.pad_token_id + 1 ) snake_case : str = self.get_config() snake_case : Tuple = config.num_attention_heads snake_case : List[Any] = self.prepare_inputs_dict(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) return config, input_dict def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : List[str] = self.prepare_config_and_inputs() return config, inputs_dict def lowerCAmelCase( self : Dict ): """simple docstring""" return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def lowerCAmelCase( self : Tuple ): """simple docstring""" return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def lowerCAmelCase( self : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Tuple , ): """simple docstring""" snake_case : str = UMTaModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : str = model( input_ids=UpperCAmelCase__ , decoder_input_ids=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , decoder_attention_mask=UpperCAmelCase__ , ) snake_case : int = model(input_ids=UpperCAmelCase__ , decoder_input_ids=UpperCAmelCase__ ) snake_case : int = result.last_hidden_state snake_case : Dict = result.past_key_values snake_case : Dict = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(UpperCAmelCase__ ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def lowerCAmelCase( self : int , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[str] , ): """simple docstring""" snake_case : int = UMTaModel(config=UpperCAmelCase__ ).get_decoder().to(UpperCAmelCase__ ).eval() # first forward pass snake_case : List[Any] = model(UpperCAmelCase__ , use_cache=UpperCAmelCase__ ) snake_case : List[Any] = model(UpperCAmelCase__ ) snake_case : Any = model(UpperCAmelCase__ , use_cache=UpperCAmelCase__ ) self.parent.assertTrue(len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) ) self.parent.assertTrue(len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) + 1 ) snake_case , snake_case : List[str] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case : Any = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and snake_case : Union[str, Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case : Any = model(UpperCAmelCase__ )['''last_hidden_state'''] snake_case : Optional[Any] = model(UpperCAmelCase__ , past_key_values=UpperCAmelCase__ )['''last_hidden_state'''] # select random slice snake_case : Tuple = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case : Union[str, Any] = output_from_no_past[:, -1, random_slice_idx].detach() snake_case : Tuple = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1e-3 ) ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : int , ): """simple docstring""" snake_case : int = UMTaModel(config=UpperCAmelCase__ ).to(UpperCAmelCase__ ).half().eval() snake_case : str = model(**UpperCAmelCase__ )['''last_hidden_state'''] self.parent.assertFalse(torch.isnan(UpperCAmelCase__ ).any().item() ) @require_torch class a_ ( a , a , a , unittest.TestCase ): A__ : str = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) A__ : str = (UMTaForConditionalGeneration,) if is_torch_available() else () A__ : Any = ( { 'conversational': UMTaForConditionalGeneration, 'feature-extraction': UMTaModel, 'summarization': UMTaForConditionalGeneration, 'text2text-generation': UMTaForConditionalGeneration, 'translation': UMTaForConditionalGeneration, 'question-answering': UMTaForQuestionAnswering, } if is_torch_available() else {} ) A__ : Dict = True A__ : List[str] = False A__ : Optional[int] = False A__ : Optional[int] = True A__ : List[str] = True # The small UMT5 model needs higher percentages for CPU/MP tests A__ : int = [0.8, 0.9] def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Union[str, Any] = UMTaModelTester(self ) @unittest.skip('''Test has a segmentation fault on torch 1.8.0''' ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Tuple = self.model_tester.prepare_config_and_inputs() snake_case : Optional[Any] = UMTaModel(config_and_inputs[0] ).to(UpperCAmelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( UpperCAmelCase__ , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"{tmpdirname}/t5_test.onnx" , export_params=UpperCAmelCase__ , opset_version=9 , input_names=['''input_ids''', '''decoder_input_ids'''] , ) @unittest.skipIf(torch_device == '''cpu''' , '''Cant do half precision''' ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*UpperCAmelCase__ ) def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : Optional[int] = ['''encoder_attentions''', '''decoder_attentions''', '''cross_attentions'''] snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() snake_case : int = config_and_inputs[0] snake_case : Union[str, Any] = UMTaForConditionalGeneration(UpperCAmelCase__ ).eval() model.to(UpperCAmelCase__ ) snake_case : str = { '''head_mask''': torch.zeros(config.num_layers , config.num_heads , device=UpperCAmelCase__ ), '''decoder_head_mask''': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ), '''cross_attn_head_mask''': torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ), } for attn_name, (name, mask) in zip(UpperCAmelCase__ , head_masking.items() ): snake_case : int = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": snake_case : List[str] = torch.ones( config.num_decoder_layers , config.num_heads , device=UpperCAmelCase__ ) snake_case : Union[str, Any] = model.generate( config_and_inputs[1]['''input_ids'''] , num_beams=1 , max_length=3 , output_attentions=UpperCAmelCase__ , return_dict_in_generate=UpperCAmelCase__ , **UpperCAmelCase__ , ) # We check the state of decoder_attentions and cross_attentions just from the last step snake_case : List[str] = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('''Does not work on the tiny model as we keep hitting edge cases.''' ) def lowerCAmelCase( self : Any ): """simple docstring""" pass @require_torch @require_sentencepiece @require_tokenizers class a_ ( unittest.TestCase ): @slow @unittest.skip( '''Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged''' ) def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Optional[Any] = UMTaForConditionalGeneration.from_pretrained('''google/umt5-small''' , return_dict=UpperCAmelCase__ ).to(UpperCAmelCase__ ) snake_case : int = AutoTokenizer.from_pretrained('''google/umt5-small''' , use_fast=UpperCAmelCase__ , legacy=UpperCAmelCase__ ) snake_case : List[str] = [ '''Bonjour monsieur <extra_id_0> bien <extra_id_1>.''', '''No se como puedo <extra_id_0>.''', '''This is the reason why we <extra_id_0> them.''', '''The <extra_id_0> walks in <extra_id_1>, seats''', '''A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.''', ] snake_case : Dict = tokenizer(UpperCAmelCase__ , return_tensors='''pt''' , padding=UpperCAmelCase__ ).input_ids # fmt: off snake_case : Optional[Any] = torch.tensor( [ [ 38_530, 210_703, 256_299, 1_410, 256_298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25_922, 256_299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1_460, 339, 312, 19_014, 10_620, 758, 256_299, 2_355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256_299, 14_869, 281, 301, 256_298, 275, 119_983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256_299, 14_869, 281, 2_234, 289, 2_275, 333,61_391, 289, 256_298, 543, 256_297, 168_714, 329, 256_296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : List[Any] = model.generate(input_ids.to(UpperCAmelCase__ ) ) snake_case : int = [ '''<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>''', '''<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', '''<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>''', ] snake_case : Tuple = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ )
84
0
import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class a_ ( a ): A__ : str = ['image_processor', 'tokenizer'] A__ : int = 'BlipImageProcessor' A__ : List[str] = 'AutoTokenizer' def __init__( self : List[str] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Dict ): """simple docstring""" super().__init__(_lowerCAmelCase , _lowerCAmelCase ) # add QFormer tokenizer snake_case : Optional[int] = qformer_tokenizer def __call__( self : Tuple , UpperCAmelCase__ : List[str] = None , UpperCAmelCase__ : Optional[Any] = None , UpperCAmelCase__ : int = True , UpperCAmelCase__ : List[Any] = False , UpperCAmelCase__ : int = None , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : Optional[int] = 0 , UpperCAmelCase__ : int = None , UpperCAmelCase__ : str = None , UpperCAmelCase__ : List[str] = False , UpperCAmelCase__ : Union[str, Any] = False , UpperCAmelCase__ : str = False , UpperCAmelCase__ : Optional[Any] = False , UpperCAmelCase__ : Dict = False , UpperCAmelCase__ : str = True , UpperCAmelCase__ : Tuple = None , **UpperCAmelCase__ : int , ): """simple docstring""" if images is None and text is None: raise ValueError('''You have to specify at least images or text.''' ) snake_case : int = BatchFeature() if text is not None: snake_case : List[str] = self.tokenizer( text=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , max_length=_lowerCAmelCase , stride=_lowerCAmelCase , pad_to_multiple_of=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase , return_overflowing_tokens=_lowerCAmelCase , return_special_tokens_mask=_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase , return_length=_lowerCAmelCase , verbose=_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase , ) encoding.update(_lowerCAmelCase ) snake_case : Optional[int] = self.qformer_tokenizer( text=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , max_length=_lowerCAmelCase , stride=_lowerCAmelCase , pad_to_multiple_of=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase , return_overflowing_tokens=_lowerCAmelCase , return_special_tokens_mask=_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase , return_length=_lowerCAmelCase , verbose=_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase , ) snake_case : List[Any] = qformer_text_encoding.pop('''input_ids''' ) snake_case : int = qformer_text_encoding.pop('''attention_mask''' ) if images is not None: snake_case : Dict = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase ) encoding.update(_lowerCAmelCase ) return encoding def lowerCAmelCase( self : Optional[int] , *UpperCAmelCase__ : Tuple , **UpperCAmelCase__ : str ): """simple docstring""" return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def lowerCAmelCase( self : Union[str, Any] , *UpperCAmelCase__ : List[Any] , **UpperCAmelCase__ : str ): """simple docstring""" return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : List[Any] = self.tokenizer.model_input_names snake_case : Dict = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def lowerCAmelCase( self : Union[str, Any] , UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : int ): """simple docstring""" if os.path.isfile(_lowerCAmelCase ): raise ValueError(F"Provided path ({save_directory}) should be a directory, not a file" ) os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) snake_case : Tuple = os.path.join(_lowerCAmelCase , '''qformer_tokenizer''' ) self.qformer_tokenizer.save_pretrained(_lowerCAmelCase ) return super().save_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) @classmethod def lowerCAmelCase( cls : int , UpperCAmelCase__ : Tuple , **UpperCAmelCase__ : List[str] ): """simple docstring""" snake_case : Union[str, Any] = AutoTokenizer.from_pretrained(_lowerCAmelCase , subfolder='''qformer_tokenizer''' ) snake_case : str = cls._get_arguments_from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) args.append(_lowerCAmelCase ) return cls(*_lowerCAmelCase )
709
import torch from diffusers import DiffusionPipeline class a_ ( a ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[Any] ): """simple docstring""" super().__init__() self.register_modules(unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ ) def __call__( self : Optional[int] ): """simple docstring""" snake_case : Any = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) snake_case : Dict = 1 snake_case : Optional[Any] = self.unet(UpperCAmelCase__ , UpperCAmelCase__ ).sample snake_case : List[Any] = self.scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample snake_case : List[Any] = scheduler_output - scheduler_output + torch.ones_like(UpperCAmelCase__ ) return result
84
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, TensorType _a : List[str] = logging.get_logger(__name__) _a : Any = { "openai/imagegpt-small": "", "openai/imagegpt-medium": "", "openai/imagegpt-large": "", } class a_ ( a ): A__ : List[str] = "imagegpt" A__ : str = ["past_key_values"] A__ : Tuple = { "hidden_size": "n_embd", "max_position_embeddings": "n_positions", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Optional[int] , UpperCAmelCase__ : List[Any]=512 + 1 , UpperCAmelCase__ : str=32 * 32 , UpperCAmelCase__ : Union[str, Any]=512 , UpperCAmelCase__ : Optional[int]=24 , UpperCAmelCase__ : Tuple=8 , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : int="quick_gelu" , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : Tuple=0.1 , UpperCAmelCase__ : Optional[Any]=0.1 , UpperCAmelCase__ : Optional[Any]=1e-5 , UpperCAmelCase__ : int=0.02 , UpperCAmelCase__ : List[str]=True , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Dict=False , UpperCAmelCase__ : Union[str, Any]=False , UpperCAmelCase__ : List[str]=False , **UpperCAmelCase__ : Any , ): """simple docstring""" snake_case : Dict = vocab_size snake_case : Dict = n_positions snake_case : Tuple = n_embd snake_case : List[str] = n_layer snake_case : Any = n_head snake_case : Optional[Any] = n_inner snake_case : Dict = activation_function snake_case : Optional[Any] = resid_pdrop snake_case : List[str] = embd_pdrop snake_case : str = attn_pdrop snake_case : int = layer_norm_epsilon snake_case : Tuple = initializer_range snake_case : List[str] = scale_attn_weights snake_case : int = use_cache snake_case : Tuple = scale_attn_by_inverse_layer_idx snake_case : Optional[int] = reorder_and_upcast_attn snake_case : List[str] = tie_word_embeddings super().__init__(tie_word_embeddings=UpperCamelCase_ , **UpperCamelCase_ ) class a_ ( a ): @property def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" return OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''sequence'''}), ] ) def lowerCAmelCase( self : int , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Tuple = 1 , UpperCAmelCase__ : Dict = -1 , UpperCAmelCase__ : str = False , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : Optional[Any] = 3 , UpperCAmelCase__ : int = 32 , UpperCAmelCase__ : Union[str, Any] = 32 , ): """simple docstring""" snake_case : Tuple = self._generate_dummy_images(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) snake_case : Optional[int] = dict(preprocessor(images=UpperCamelCase_ , return_tensors=UpperCamelCase_ ) ) return inputs
710
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class a_ ( a ): A__ : List[str] = ['image_processor', 'tokenizer'] A__ : Any = 'CLIPImageProcessor' A__ : Optional[int] = ('CLIPTokenizer', 'CLIPTokenizerFast') def __init__( self : Union[str, Any] , UpperCAmelCase__ : str=None , UpperCAmelCase__ : Union[str, Any]=None , **UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : Optional[int] = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , UpperCAmelCase__ , ) snake_case : List[Any] = kwargs.pop('''feature_extractor''' ) snake_case : Optional[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(UpperCAmelCase__ , UpperCAmelCase__ ) def __call__( self : Any , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : int=None , **UpperCAmelCase__ : Union[str, Any] ): """simple docstring""" if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: snake_case : int = self.tokenizer(UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , **UpperCAmelCase__ ) if images is not None: snake_case : Dict = self.image_processor(UpperCAmelCase__ , return_tensors=UpperCAmelCase__ , **UpperCAmelCase__ ) if text is not None and images is not None: snake_case : Tuple = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**UpperCAmelCase__ ) , tensor_type=UpperCAmelCase__ ) def lowerCAmelCase( self : List[str] , *UpperCAmelCase__ : Dict , **UpperCAmelCase__ : int ): """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : str ): """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : int = self.tokenizer.model_input_names snake_case : Tuple = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCAmelCase( self : Tuple ): """simple docstring""" warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , UpperCAmelCase__ , ) return self.image_processor_class @property def lowerCAmelCase( self : Any ): """simple docstring""" warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , UpperCAmelCase__ , ) return self.image_processor
84
0
'''simple docstring''' import heapq as hq import math from collections.abc import Iterator class a_ : def __init__( self : Tuple , UpperCAmelCase__ : Any ): """simple docstring""" snake_case : Any = str(id_ ) snake_case : Optional[int] = None snake_case : Optional[int] = None snake_case : Union[str, Any] = [] snake_case : Optional[Any] = {} # {vertex:distance} def __lt__( self : int , UpperCAmelCase__ : List[Any] ): """simple docstring""" return self.key < other.key def __repr__( self : List[str] ): """simple docstring""" return self.id def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : Optional[int] ): """simple docstring""" self.neighbors.append(__A ) def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : int ): """simple docstring""" snake_case : Optional[Any] = weight def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , _lowerCAmelCase ) graph[b - 1].add_edge(graph[a - 1] , _lowerCAmelCase ) def a_ ( __magic_name__ , __magic_name__ ) -> Union[str, Any]: """simple docstring""" snake_case : str = [] for u in graph: snake_case : Any = math.inf snake_case : List[str] = None snake_case : Tuple = 0 snake_case : Tuple = graph[:] while q: snake_case : List[str] = min(_lowerCAmelCase ) q.remove(_lowerCAmelCase ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): snake_case : int = u snake_case : Optional[Any] = u.edges[v.id] for i in range(1 , len(_lowerCAmelCase ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def a_ ( __magic_name__ , __magic_name__ ) -> Union[str, Any]: """simple docstring""" for u in graph: snake_case : Any = math.inf snake_case : Tuple = None snake_case : int = 0 snake_case : Dict = list(_lowerCAmelCase ) hq.heapify(_lowerCAmelCase ) while h: snake_case : Dict = hq.heappop(_lowerCAmelCase ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): snake_case : List[Any] = u snake_case : List[str] = u.edges[v.id] hq.heapify(_lowerCAmelCase ) for i in range(1 , len(_lowerCAmelCase ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def a_ ( ) -> Optional[Any]: """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
711
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets _a : Optional[Any] = '\\n@inproceedings{pillutla-etal:mauve:neurips2021,\n title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers},\n author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid},\n booktitle = {NeurIPS},\n year = {2021}\n}\n\n' _a : str = '\\nMAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure.\n\nMAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences.\n\nFor details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021).\n\nThis metrics is a wrapper around the official implementation of MAUVE:\nhttps://github.com/krishnap25/mauve\n' _a : List[Any] = '\nCalculates MAUVE scores between two lists of generated text and reference text.\nArgs:\n predictions: list of generated text to score. Each predictions\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\nOptional Args:\n num_buckets: the size of the histogram to quantize P and Q. Options: \'auto\' (default) or an integer\n pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1\n kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9\n kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5\n kmeans_max_iter: maximum number of k-means iterations. Default 500\n featurize_model_name: name of the model from which features are obtained. Default \'gpt2-large\' Use one of [\'gpt2\', \'gpt2-medium\', \'gpt2-large\', \'gpt2-xl\'].\n device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU\n max_text_length: maximum number of tokens to consider. Default 1024\n divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25\n mauve_scaling_factor: "c" from the paper. Default 5.\n verbose: If True (default), print running time updates\n seed: random seed to initialize k-means cluster assignments.\nReturns:\n mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer,\n frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer,\n divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve,\n p_hist: a discrete distribution, which is a quantized version of the text distribution p_text,\n q_hist: same as above, but with q_text.\nExamples:\n\n >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest\n >>> import datasets\n >>> mauve = datasets.load_metric(\'mauve\')\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP\n >>> print(out.mauve) # doctest: +SKIP\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): def lowerCAmelCase( self : Any ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://github.com/krishnap25/mauve''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/krishnap25/mauve'''] , reference_urls=[ '''https://arxiv.org/abs/2102.01454''', '''https://github.com/krishnap25/mauve''', ] , ) def lowerCAmelCase( self : int , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : List[str]="auto" , UpperCAmelCase__ : Tuple=-1 , UpperCAmelCase__ : Optional[int]=0.9 , UpperCAmelCase__ : List[Any]=5 , UpperCAmelCase__ : List[Any]=500 , UpperCAmelCase__ : Union[str, Any]="gpt2-large" , UpperCAmelCase__ : Optional[Any]=-1 , UpperCAmelCase__ : int=1_024 , UpperCAmelCase__ : List[Any]=25 , UpperCAmelCase__ : Union[str, Any]=5 , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : List[Any]=25 , ): """simple docstring""" snake_case : List[str] = compute_mauve( p_text=UpperCAmelCase__ , q_text=UpperCAmelCase__ , p_features=UpperCAmelCase__ , q_features=UpperCAmelCase__ , p_tokens=UpperCAmelCase__ , q_tokens=UpperCAmelCase__ , num_buckets=UpperCAmelCase__ , pca_max_data=UpperCAmelCase__ , kmeans_explained_var=UpperCAmelCase__ , kmeans_num_redo=UpperCAmelCase__ , kmeans_max_iter=UpperCAmelCase__ , featurize_model_name=UpperCAmelCase__ , device_id=UpperCAmelCase__ , max_text_length=UpperCAmelCase__ , divergence_curve_discretization_size=UpperCAmelCase__ , mauve_scaling_factor=UpperCAmelCase__ , verbose=UpperCAmelCase__ , seed=UpperCAmelCase__ , ) return out
84
0
import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class a_ ( UpperCamelCase_ , unittest.TestCase ): A__ : List[Any] = RoCBertTokenizer A__ : Tuple = None A__ : Dict = False A__ : int = True A__ : Any = filter_non_english def lowerCAmelCase( self : List[Any] ): """simple docstring""" super().setUp() snake_case : Any = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """你""", """好""", """是""", """谁""", """a""", """b""", """c""", """d"""] snake_case : List[Any] = {} snake_case : List[Any] = {} for i, value in enumerate(_a ): snake_case : List[Any] = i snake_case : str = i snake_case : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) snake_case : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''word_shape_file'''] ) snake_case : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''word_pronunciation_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) with open(self.word_shape_file , '''w''' , encoding='''utf-8''' ) as word_shape_writer: json.dump(_a , _a , ensure_ascii=_a ) with open(self.word_pronunciation_file , '''w''' , encoding='''utf-8''' ) as word_pronunciation_writer: json.dump(_a , _a , ensure_ascii=_a ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : str = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) snake_case : Any = tokenizer.tokenize('''你好[SEP]你是谁''' ) self.assertListEqual(_a , ['''你''', '''好''', '''[SEP]''', '''你''', '''是''', '''谁'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(_a ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(_a ) , [5, 6, 2, 5, 7, 8] ) def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Dict = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize('''ah\u535A\u63A8zz''' ) , ['''ah''', '''\u535A''', '''\u63A8''', '''zz'''] ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : List[str] = RoCBertBasicTokenizer(do_lower_case=_a ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) , ['''hello''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def lowerCAmelCase( self : int ): """simple docstring""" snake_case : str = RoCBertBasicTokenizer(do_lower_case=_a , strip_accents=_a ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hällo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''h\u00E9llo'''] ) def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : List[Any] = RoCBertBasicTokenizer(do_lower_case=_a , strip_accents=_a ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : Optional[Any] = RoCBertBasicTokenizer(do_lower_case=_a ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : int = RoCBertBasicTokenizer(do_lower_case=_a ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def lowerCAmelCase( self : str ): """simple docstring""" snake_case : int = RoCBertBasicTokenizer(do_lower_case=_a , strip_accents=_a ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''HäLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : Optional[Any] = RoCBertBasicTokenizer(do_lower_case=_a , strip_accents=_a ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''HaLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Union[str, Any] = RoCBertBasicTokenizer(do_lower_case=_a , never_split=['''[UNK]'''] ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? [UNK]''' ) , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''', '''[UNK]'''] ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[Any] = ["""[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing"""] snake_case : Optional[Any] = {} for i, token in enumerate(_a ): snake_case : Optional[Any] = i snake_case : List[str] = RoCBertWordpieceTokenizer(vocab=_a , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''unwanted running''' ) , ['''un''', '''##want''', '''##ed''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.tokenize('''unwantedX running''' ) , ['''[UNK]''', '''runn''', '''##ing'''] ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" self.assertTrue(_is_whitespace(''' ''' ) ) self.assertTrue(_is_whitespace('''\t''' ) ) self.assertTrue(_is_whitespace('''\r''' ) ) self.assertTrue(_is_whitespace('''\n''' ) ) self.assertTrue(_is_whitespace('''\u00A0''' ) ) self.assertFalse(_is_whitespace('''A''' ) ) self.assertFalse(_is_whitespace('''-''' ) ) def lowerCAmelCase( self : str ): """simple docstring""" self.assertTrue(_is_control('''\u0005''' ) ) self.assertFalse(_is_control('''A''' ) ) self.assertFalse(_is_control(''' ''' ) ) self.assertFalse(_is_control('''\t''' ) ) self.assertFalse(_is_control('''\r''' ) ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" self.assertTrue(_is_punctuation('''-''' ) ) self.assertTrue(_is_punctuation('''$''' ) ) self.assertTrue(_is_punctuation('''`''' ) ) self.assertTrue(_is_punctuation('''.''' ) ) self.assertFalse(_is_punctuation('''A''' ) ) self.assertFalse(_is_punctuation(''' ''' ) ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[int] = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(_a ) for t in ['''Test''', '''\xad''', '''test''']] , [['''[UNK]'''], [], ['''[UNK]''']] ) if self.test_rust_tokenizer: snake_case : Optional[Any] = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(_a ) for t in ['''Test''', '''\xad''', '''test''']] , [['''[UNK]'''], [], ['''[UNK]''']] ) def lowerCAmelCase( self : Any ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): snake_case : List[Any] = self.rust_tokenizer_class.from_pretrained(_a , **_a ) snake_case : Dict = F"A, naïve {tokenizer_r.mask_token} AllenNLP sentence." snake_case : List[Any] = tokenizer_r.encode_plus( _a , return_attention_mask=_a , return_token_type_ids=_a , return_offsets_mapping=_a , add_special_tokens=_a , ) snake_case : Optional[int] = tokenizer_r.do_lower_case if hasattr(_a , '''do_lower_case''' ) else False snake_case : Optional[int] = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """A"""), ((1, 2), ""","""), ((3, 5), """na"""), ((5, 6), """##ï"""), ((6, 8), """##ve"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """Allen"""), ((21, 23), """##NL"""), ((23, 24), """##P"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """a"""), ((1, 2), ""","""), ((3, 8), """naive"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """allen"""), ((21, 23), """##nl"""), ((23, 24), """##p"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens['''input_ids'''] ) ) self.assertEqual([e[0] for e in expected_results] , tokens['''offset_mapping'''] ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Optional[int] = ["""的""", """人""", """有"""] snake_case : Dict = """""".join(_a ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): snake_case : Union[str, Any] = True snake_case : Tuple = self.tokenizer_class.from_pretrained(_a , **_a ) snake_case : str = self.rust_tokenizer_class.from_pretrained(_a , **_a ) snake_case : Optional[int] = tokenizer_p.encode(_a , add_special_tokens=_a ) snake_case : List[Any] = tokenizer_r.encode(_a , add_special_tokens=_a ) snake_case : Tuple = tokenizer_r.convert_ids_to_tokens(_a ) snake_case : Any = tokenizer_p.convert_ids_to_tokens(_a ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(_a , _a ) self.assertListEqual(_a , _a ) snake_case : Any = False snake_case : Tuple = self.rust_tokenizer_class.from_pretrained(_a , **_a ) snake_case : Union[str, Any] = self.tokenizer_class.from_pretrained(_a , **_a ) snake_case : str = tokenizer_r.encode(_a , add_special_tokens=_a ) snake_case : Dict = tokenizer_p.encode(_a , add_special_tokens=_a ) snake_case : Optional[int] = tokenizer_r.convert_ids_to_tokens(_a ) snake_case : Optional[int] = tokenizer_p.convert_ids_to_tokens(_a ) # it is expected that only the first Chinese character is not preceded by "##". snake_case : Tuple = [ F"##{token}" if idx != 0 else token for idx, token in enumerate(_a ) ] self.assertListEqual(_a , _a ) self.assertListEqual(_a , _a ) @slow def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : Optional[Any] = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) snake_case : Union[str, Any] = tokenizer.encode('''你好''' , add_special_tokens=_a ) snake_case : Any = tokenizer.encode('''你是谁''' , add_special_tokens=_a ) snake_case : Optional[Any] = tokenizer.build_inputs_with_special_tokens(_a ) snake_case : Any = tokenizer.build_inputs_with_special_tokens(_a , _a ) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : int = self.get_tokenizers(do_lower_case=_a ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): snake_case : Any = """你好,你是谁""" snake_case : Tuple = tokenizer.tokenize(_a ) snake_case : List[Any] = tokenizer.convert_tokens_to_ids(_a ) snake_case : List[Any] = tokenizer.convert_tokens_to_shape_ids(_a ) snake_case : List[str] = tokenizer.convert_tokens_to_pronunciation_ids(_a ) snake_case : int = tokenizer.prepare_for_model( _a , _a , _a , add_special_tokens=_a ) snake_case : List[str] = tokenizer.encode_plus(_a , add_special_tokens=_a ) self.assertEqual(_a , _a )
712
import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" if "cls_token" in name: snake_case : Tuple = name.replace('''cls_token''' , '''vit.embeddings.cls_token''' ) if "mask_token" in name: snake_case : Optional[int] = name.replace('''mask_token''' , '''decoder.mask_token''' ) if "decoder_pos_embed" in name: snake_case : List[str] = name.replace('''decoder_pos_embed''' , '''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: snake_case : List[str] = name.replace('''pos_embed''' , '''vit.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: snake_case : List[Any] = name.replace('''patch_embed.proj''' , '''vit.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: snake_case : int = name.replace('''patch_embed.norm''' , '''vit.embeddings.norm''' ) if "decoder_blocks" in name: snake_case : int = name.replace('''decoder_blocks''' , '''decoder.decoder_layers''' ) if "blocks" in name: snake_case : Optional[Any] = name.replace('''blocks''' , '''vit.encoder.layer''' ) if "attn.proj" in name: snake_case : str = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: snake_case : Any = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: snake_case : List[str] = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: snake_case : Tuple = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: snake_case : Tuple = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: snake_case : Tuple = name.replace('''mlp.fc2''' , '''output.dense''' ) if "decoder_embed" in name: snake_case : Dict = name.replace('''decoder_embed''' , '''decoder.decoder_embed''' ) if "decoder_norm" in name: snake_case : Dict = name.replace('''decoder_norm''' , '''decoder.decoder_norm''' ) if "decoder_pred" in name: snake_case : Dict = name.replace('''decoder_pred''' , '''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name: snake_case : Optional[int] = name.replace('''norm.weight''' , '''vit.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name: snake_case : List[str] = name.replace('''norm.bias''' , '''vit.layernorm.bias''' ) return name def a_ ( __magic_name__ , __magic_name__ ) -> str: """simple docstring""" for key in orig_state_dict.copy().keys(): snake_case : Union[str, Any] = orig_state_dict.pop(__magic_name__ ) if "qkv" in key: snake_case : Optional[int] = key.split('''.''' ) snake_case : int = int(key_split[1] ) if "decoder_blocks" in key: snake_case : List[str] = config.decoder_hidden_size snake_case : List[Any] = '''decoder.decoder_layers.''' if "weight" in key: snake_case : str = val[:dim, :] snake_case : Optional[Any] = val[dim : dim * 2, :] snake_case : Any = val[-dim:, :] elif "bias" in key: snake_case : Optional[Any] = val[:dim] snake_case : List[Any] = val[dim : dim * 2] snake_case : List[Any] = val[-dim:] else: snake_case : Optional[int] = config.hidden_size snake_case : Tuple = '''vit.encoder.layer.''' if "weight" in key: snake_case : Optional[Any] = val[:dim, :] snake_case : str = val[dim : dim * 2, :] snake_case : Union[str, Any] = val[-dim:, :] elif "bias" in key: snake_case : Tuple = val[:dim] snake_case : int = val[dim : dim * 2] snake_case : Optional[Any] = val[-dim:] else: snake_case : Optional[Any] = val return orig_state_dict def a_ ( __magic_name__ , __magic_name__ ) -> Any: """simple docstring""" snake_case : List[str] = ViTMAEConfig() if "large" in checkpoint_url: snake_case : str = 1_024 snake_case : Tuple = 4_096 snake_case : Optional[Any] = 24 snake_case : List[Any] = 16 elif "huge" in checkpoint_url: snake_case : Tuple = 14 snake_case : int = 1_280 snake_case : Dict = 5_120 snake_case : Tuple = 32 snake_case : Optional[Any] = 16 snake_case : Optional[Any] = ViTMAEForPreTraining(__magic_name__ ) snake_case : Optional[Any] = torch.hub.load_state_dict_from_url(__magic_name__ , map_location='''cpu''' )['''model'''] snake_case : int = ViTMAEImageProcessor(size=config.image_size ) snake_case : Dict = convert_state_dict(__magic_name__ , __magic_name__ ) model.load_state_dict(__magic_name__ ) model.eval() snake_case : Tuple = '''https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg''' snake_case : List[Any] = Image.open(requests.get(__magic_name__ , stream=__magic_name__ ).raw ) snake_case : Dict = ViTMAEImageProcessor(size=config.image_size ) snake_case : str = image_processor(images=__magic_name__ , return_tensors='''pt''' ) # forward pass torch.manual_seed(2 ) snake_case : Union[str, Any] = model(**__magic_name__ ) snake_case : Optional[Any] = outputs.logits if "large" in checkpoint_url: snake_case : Any = torch.tensor( [[-0.7309, -0.7128, -1.0169], [-1.0161, -0.9058, -1.1878], [-1.0478, -0.9411, -1.1911]] ) elif "huge" in checkpoint_url: snake_case : List[Any] = torch.tensor( [[-1.1599, -0.9199, -1.2221], [-1.1952, -0.9269, -1.2307], [-1.2143, -0.9337, -1.2262]] ) else: snake_case : Dict = torch.tensor( [[-0.9192, -0.8481, -1.1259], [-1.1349, -1.0034, -1.2599], [-1.1757, -1.0429, -1.2726]] ) # verify logits assert torch.allclose(logits[0, :3, :3] , __magic_name__ , atol=1e-4 ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(__magic_name__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__magic_name__ ) if __name__ == "__main__": _a : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth', type=str, help='URL of the checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _a : str = parser.parse_args() convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
84
0
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetaImageProcessor class a_ ( unittest.TestCase ): def __init__( self : str , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : str=7 , UpperCAmelCase__ : Optional[Any]=3 , UpperCAmelCase__ : Tuple=30 , UpperCAmelCase__ : int=400 , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Optional[Any]=None , UpperCAmelCase__ : str=True , UpperCAmelCase__ : Union[str, Any]=[0.5, 0.5, 0.5] , UpperCAmelCase__ : Tuple=[0.5, 0.5, 0.5] , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Any=1 / 255 , UpperCAmelCase__ : Any=True , ): """simple docstring""" snake_case : int = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 1_333} snake_case : List[str] = parent snake_case : List[Any] = batch_size snake_case : Optional[int] = num_channels snake_case : Any = min_resolution snake_case : Optional[Any] = max_resolution snake_case : Any = do_resize snake_case : List[Any] = size snake_case : Union[str, Any] = do_normalize snake_case : Dict = image_mean snake_case : Tuple = image_std snake_case : int = do_rescale snake_case : Any = rescale_factor snake_case : List[Any] = do_pad def lowerCAmelCase( self : Dict ): """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def lowerCAmelCase( self : int , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Tuple=False ): """simple docstring""" if not batched: snake_case : List[Any] = image_inputs[0] if isinstance(UpperCAmelCase__ , Image.Image ): snake_case , snake_case : List[Any] = image.size else: snake_case , snake_case : str = image.shape[1], image.shape[2] if w < h: snake_case : List[Any] = int(self.size['''shortest_edge'''] * h / w ) snake_case : Any = self.size['''shortest_edge'''] elif w > h: snake_case : Tuple = self.size['''shortest_edge'''] snake_case : Tuple = int(self.size['''shortest_edge'''] * w / h ) else: snake_case : Union[str, Any] = self.size['''shortest_edge'''] snake_case : Optional[Any] = self.size['''shortest_edge'''] else: snake_case : Optional[int] = [] for image in image_inputs: snake_case , snake_case : Optional[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) snake_case : Dict = max(UpperCAmelCase__ , key=lambda UpperCAmelCase__ : item[0] )[0] snake_case : Union[str, Any] = max(UpperCAmelCase__ , key=lambda UpperCAmelCase__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a_ ( _snake_case , unittest.TestCase ): A__ : Union[str, Any] = DetaImageProcessor if is_vision_available() else None def lowerCAmelCase( self : int ): """simple docstring""" snake_case : Any = DetaImageProcessingTester(self ) @property def lowerCAmelCase( self : int ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase__ , '''image_mean''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''image_std''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''do_normalize''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''do_resize''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''do_rescale''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''do_pad''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''size''' ) ) def lowerCAmelCase( self : str ): """simple docstring""" snake_case : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 1_333} ) self.assertEqual(image_processor.do_pad , UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" pass def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , Image.Image ) # Test not batched input snake_case : str = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values snake_case , snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(UpperCAmelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case , snake_case : Dict = self.image_processor_tester.get_expected_values(UpperCAmelCase__ , batched=UpperCAmelCase__ ) snake_case : Dict = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , numpify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , np.ndarray ) # Test not batched input snake_case : List[str] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values snake_case , snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(UpperCAmelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case : Optional[Any] = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values snake_case , snake_case : Tuple = self.image_processor_tester.get_expected_values(UpperCAmelCase__ , batched=UpperCAmelCase__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , torchify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , torch.Tensor ) # Test not batched input snake_case : int = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values snake_case , snake_case : int = self.image_processor_tester.get_expected_values(UpperCAmelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case : Optional[int] = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values snake_case , snake_case : Dict = self.image_processor_tester.get_expected_values(UpperCAmelCase__ , batched=UpperCAmelCase__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : Optional[int] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: snake_case : Union[str, Any] = json.loads(f.read() ) snake_case : Optional[int] = {'''image_id''': 39_769, '''annotations''': target} # encode them snake_case : Optional[int] = DetaImageProcessor() snake_case : Optional[int] = image_processing(images=UpperCAmelCase__ , annotations=UpperCAmelCase__ , return_tensors='''pt''' ) # verify pixel values snake_case : Union[str, Any] = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding['''pixel_values'''].shape , UpperCAmelCase__ ) snake_case : Optional[int] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , UpperCAmelCase__ , atol=1e-4 ) ) # verify area snake_case : Dict = torch.tensor([5_887.9_600, 11_250.2_061, 489_353.8_438, 837_122.7_500, 147_967.5_156, 165_732.3_438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , UpperCAmelCase__ ) ) # verify boxes snake_case : Any = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , UpperCAmelCase__ ) snake_case : Optional[Any] = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , UpperCAmelCase__ , atol=1e-3 ) ) # verify image_id snake_case : Any = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , UpperCAmelCase__ ) ) # verify is_crowd snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , UpperCAmelCase__ ) ) # verify class_labels snake_case : List[str] = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , UpperCAmelCase__ ) ) # verify orig_size snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , UpperCAmelCase__ ) ) # verify size snake_case : List[Any] = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , UpperCAmelCase__ ) ) @slow def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : Dict = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: snake_case : List[str] = json.loads(f.read() ) snake_case : int = {'''file_name''': '''000000039769.png''', '''image_id''': 39_769, '''segments_info''': target} snake_case : Tuple = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them snake_case : List[Any] = DetaImageProcessor(format='''coco_panoptic''' ) snake_case : str = image_processing(images=UpperCAmelCase__ , annotations=UpperCAmelCase__ , masks_path=UpperCAmelCase__ , return_tensors='''pt''' ) # verify pixel values snake_case : Optional[int] = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding['''pixel_values'''].shape , UpperCAmelCase__ ) snake_case : int = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , UpperCAmelCase__ , atol=1e-4 ) ) # verify area snake_case : Any = torch.tensor([147_979.6_875, 165_527.0_469, 484_638.5_938, 11_292.9_375, 5_879.6_562, 7_634.1_147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , UpperCAmelCase__ ) ) # verify boxes snake_case : Any = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , UpperCAmelCase__ ) snake_case : Union[str, Any] = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , UpperCAmelCase__ , atol=1e-3 ) ) # verify image_id snake_case : int = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , UpperCAmelCase__ ) ) # verify is_crowd snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , UpperCAmelCase__ ) ) # verify class_labels snake_case : str = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , UpperCAmelCase__ ) ) # verify masks snake_case : List[str] = 822_873 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , UpperCAmelCase__ ) # verify orig_size snake_case : List[Any] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , UpperCAmelCase__ ) ) # verify size snake_case : int = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , UpperCAmelCase__ ) )
713
import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _a : Optional[Any] = 16 _a : Union[str, Any] = 32 def a_ ( __magic_name__ , __magic_name__ = 16 ) -> Dict: """simple docstring""" snake_case : Tuple = AutoTokenizer.from_pretrained('''bert-base-cased''' ) snake_case : Any = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(__magic_name__ ): # max_length=None => use the model max length (it's actually the default) snake_case : Union[str, Any] = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=__magic_name__ , max_length=__magic_name__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): snake_case : Union[str, Any] = datasets.map( __magic_name__ , batched=__magic_name__ , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library snake_case : Optional[Any] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(__magic_name__ ): # On TPU it's best to pad everything to the same length or training will be very slow. snake_case : str = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": snake_case : Tuple = 16 elif accelerator.mixed_precision != "no": snake_case : Dict = 8 else: snake_case : Union[str, Any] = None return tokenizer.pad( __magic_name__ , padding='''longest''' , max_length=__magic_name__ , pad_to_multiple_of=__magic_name__ , return_tensors='''pt''' , ) # Instantiate dataloaders. snake_case : str = DataLoader( tokenized_datasets['''train'''] , shuffle=__magic_name__ , collate_fn=__magic_name__ , batch_size=__magic_name__ ) snake_case : List[str] = DataLoader( tokenized_datasets['''validation'''] , shuffle=__magic_name__ , collate_fn=__magic_name__ , batch_size=__magic_name__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders _a : Optional[int] = mocked_dataloaders # noqa: F811 def a_ ( __magic_name__ , __magic_name__ ) -> Optional[Any]: """simple docstring""" if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , __magic_name__ ) == "1": snake_case : Optional[int] = 2 # Initialize accelerator snake_case : int = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs snake_case : Dict = config['''lr'''] snake_case : Any = int(config['''num_epochs'''] ) snake_case : List[str] = int(config['''seed'''] ) snake_case : List[Any] = int(config['''batch_size'''] ) snake_case : Tuple = evaluate.load('''glue''' , '''mrpc''' ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=__magic_name__ ) def inner_training_loop(__magic_name__ ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(__magic_name__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) snake_case : str = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=__magic_name__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). snake_case : Optional[int] = model.to(accelerator.device ) # Instantiate optimizer snake_case : Optional[int] = AdamW(params=model.parameters() , lr=__magic_name__ ) snake_case , snake_case : List[Any] = get_dataloaders(__magic_name__ , __magic_name__ ) # Instantiate scheduler snake_case : int = get_linear_schedule_with_warmup( optimizer=__magic_name__ , num_warmup_steps=100 , num_training_steps=(len(__magic_name__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. snake_case , snake_case , snake_case , snake_case , snake_case : Tuple = accelerator.prepare( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) # Now we train the model for epoch in range(__magic_name__ ): model.train() for step, batch in enumerate(__magic_name__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) snake_case : int = model(**__magic_name__ ) snake_case : Optional[int] = outputs.loss accelerator.backward(__magic_name__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__magic_name__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): snake_case : List[str] = model(**__magic_name__ ) snake_case : List[Any] = outputs.logits.argmax(dim=-1 ) snake_case , snake_case : Dict = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=__magic_name__ , references=__magic_name__ , ) snake_case : Tuple = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"epoch {epoch}:" , __magic_name__ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def a_ ( ) -> Union[str, Any]: """simple docstring""" snake_case : int = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=__magic_name__ , default=__magic_name__ , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) snake_case : Optional[Any] = parser.parse_args() snake_case : Optional[int] = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(__magic_name__ , __magic_name__ ) if __name__ == "__main__": main()
84
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _a : Dict = logging.get_logger(__name__) _a : str = { 'uclanlp/visualbert-vqa': 'https://huggingface.co/uclanlp/visualbert-vqa/resolve/main/config.json', 'uclanlp/visualbert-vqa-pre': 'https://huggingface.co/uclanlp/visualbert-vqa-pre/resolve/main/config.json', 'uclanlp/visualbert-vqa-coco-pre': ( 'https://huggingface.co/uclanlp/visualbert-vqa-coco-pre/resolve/main/config.json' ), 'uclanlp/visualbert-vcr': 'https://huggingface.co/uclanlp/visualbert-vcr/resolve/main/config.json', 'uclanlp/visualbert-vcr-pre': 'https://huggingface.co/uclanlp/visualbert-vcr-pre/resolve/main/config.json', 'uclanlp/visualbert-vcr-coco-pre': ( 'https://huggingface.co/uclanlp/visualbert-vcr-coco-pre/resolve/main/config.json' ), 'uclanlp/visualbert-nlvr2': 'https://huggingface.co/uclanlp/visualbert-nlvr2/resolve/main/config.json', 'uclanlp/visualbert-nlvr2-pre': 'https://huggingface.co/uclanlp/visualbert-nlvr2-pre/resolve/main/config.json', 'uclanlp/visualbert-nlvr2-coco-pre': ( 'https://huggingface.co/uclanlp/visualbert-nlvr2-coco-pre/resolve/main/config.json' ) # See all VisualBERT models at https://huggingface.co/models?filter=visual_bert } class a_ ( a ): A__ : List[str] = 'visual_bert' def __init__( self : Union[str, Any] , UpperCAmelCase__ : Tuple=30_522 , UpperCAmelCase__ : List[str]=768 , UpperCAmelCase__ : List[Any]=512 , UpperCAmelCase__ : List[Any]=12 , UpperCAmelCase__ : Optional[Any]=12 , UpperCAmelCase__ : Any=3_072 , UpperCAmelCase__ : List[Any]="gelu" , UpperCAmelCase__ : Any=0.1 , UpperCAmelCase__ : int=0.1 , UpperCAmelCase__ : int=512 , UpperCAmelCase__ : Dict=2 , UpperCAmelCase__ : Optional[int]=0.02 , UpperCAmelCase__ : Dict=1e-1_2 , UpperCAmelCase__ : Any=False , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : List[str]=1 , UpperCAmelCase__ : Optional[int]=0 , UpperCAmelCase__ : Optional[int]=2 , **UpperCAmelCase__ : Optional[int] , ): """simple docstring""" super().__init__(pad_token_id=_UpperCAmelCase , bos_token_id=_UpperCAmelCase , eos_token_id=_UpperCAmelCase , **_UpperCAmelCase ) snake_case : Union[str, Any] = vocab_size snake_case : List[Any] = max_position_embeddings snake_case : Any = hidden_size snake_case : Optional[int] = visual_embedding_dim snake_case : Any = num_hidden_layers snake_case : str = num_attention_heads snake_case : List[Any] = intermediate_size snake_case : Union[str, Any] = hidden_act snake_case : Union[str, Any] = hidden_dropout_prob snake_case : Optional[Any] = attention_probs_dropout_prob snake_case : List[Any] = initializer_range snake_case : str = type_vocab_size snake_case : Dict = layer_norm_eps snake_case : int = bypass_transformer snake_case : List[Any] = special_visual_initialize
714
from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract _a : Dict = logging.get_logger(__name__) def a_ ( __magic_name__ , __magic_name__ , __magic_name__ ) -> List[Any]: """simple docstring""" return [ int(1_000 * (box[0] / width) ), int(1_000 * (box[1] / height) ), int(1_000 * (box[2] / width) ), int(1_000 * (box[3] / height) ), ] def a_ ( __magic_name__ , __magic_name__ , __magic_name__ = None ) -> str: """simple docstring""" snake_case : Any = tesseract_config if tesseract_config is not None else '''''' # apply OCR snake_case : str = to_pil_image(__magic_name__ ) snake_case , snake_case : Union[str, Any] = pil_image.size snake_case : List[Any] = pytesseract.image_to_data(__magic_name__ , lang=__magic_name__ , output_type='''dict''' , config=__magic_name__ ) snake_case , snake_case , snake_case , snake_case , snake_case : Optional[Any] = data['''text'''], data['''left'''], data['''top'''], data['''width'''], data['''height'''] # filter empty words and corresponding coordinates snake_case : Union[str, Any] = [idx for idx, word in enumerate(__magic_name__ ) if not word.strip()] snake_case : Union[str, Any] = [word for idx, word in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : Optional[Any] = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] snake_case : int = [coord for idx, coord in enumerate(__magic_name__ ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format snake_case : List[Any] = [] for x, y, w, h in zip(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): snake_case : Optional[int] = [x, y, x + w, y + h] actual_boxes.append(__magic_name__ ) # finally, normalize the bounding boxes snake_case : List[Any] = [] for box in actual_boxes: normalized_boxes.append(normalize_box(__magic_name__ , __magic_name__ , __magic_name__ ) ) assert len(__magic_name__ ) == len(__magic_name__ ), "Not as many words as there are bounding boxes" return words, normalized_boxes class a_ ( a ): A__ : int = ['pixel_values'] def __init__( self : Optional[int] , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : PILImageResampling = PILImageResampling.BILINEAR , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[str] = "" , **UpperCAmelCase__ : int , ): """simple docstring""" super().__init__(**UpperCAmelCase__ ) snake_case : Any = size if size is not None else {'''height''': 224, '''width''': 224} snake_case : Tuple = get_size_dict(UpperCAmelCase__ ) snake_case : Dict = do_resize snake_case : str = size snake_case : Optional[int] = resample snake_case : Union[str, Any] = apply_ocr snake_case : int = ocr_lang snake_case : Union[str, Any] = tesseract_config def lowerCAmelCase( self : List[str] , UpperCAmelCase__ : np.ndarray , UpperCAmelCase__ : Dict[str, int] , UpperCAmelCase__ : PILImageResampling = PILImageResampling.BILINEAR , UpperCAmelCase__ : Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase__ : Any , ): """simple docstring""" snake_case : Dict = get_size_dict(UpperCAmelCase__ ) if "height" not in size or "width" not in size: raise ValueError(F"The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}" ) snake_case : Tuple = (size['''height'''], size['''width''']) return resize(UpperCAmelCase__ , size=UpperCAmelCase__ , resample=UpperCAmelCase__ , data_format=UpperCAmelCase__ , **UpperCAmelCase__ ) def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : ImageInput , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Dict[str, int] = None , UpperCAmelCase__ : PILImageResampling = None , UpperCAmelCase__ : bool = None , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[Union[str, TensorType]] = None , UpperCAmelCase__ : ChannelDimension = ChannelDimension.FIRST , **UpperCAmelCase__ : List[Any] , ): """simple docstring""" snake_case : Tuple = do_resize if do_resize is not None else self.do_resize snake_case : List[Any] = size if size is not None else self.size snake_case : Tuple = get_size_dict(UpperCAmelCase__ ) snake_case : str = resample if resample is not None else self.resample snake_case : Optional[int] = apply_ocr if apply_ocr is not None else self.apply_ocr snake_case : Any = ocr_lang if ocr_lang is not None else self.ocr_lang snake_case : Optional[int] = tesseract_config if tesseract_config is not None else self.tesseract_config snake_case : List[str] = make_list_of_images(UpperCAmelCase__ ) if not valid_images(UpperCAmelCase__ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) # All transformations expect numpy arrays. snake_case : Any = [to_numpy_array(UpperCAmelCase__ ) for image in images] if apply_ocr: requires_backends(self , '''pytesseract''' ) snake_case : Optional[int] = [] snake_case : Union[str, Any] = [] for image in images: snake_case , snake_case : List[Any] = apply_tesseract(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) words_batch.append(UpperCAmelCase__ ) boxes_batch.append(UpperCAmelCase__ ) if do_resize: snake_case : Any = [self.resize(image=UpperCAmelCase__ , size=UpperCAmelCase__ , resample=UpperCAmelCase__ ) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) snake_case : int = [flip_channel_order(UpperCAmelCase__ ) for image in images] snake_case : Optional[Any] = [to_channel_dimension_format(UpperCAmelCase__ , UpperCAmelCase__ ) for image in images] snake_case : List[Any] = BatchFeature(data={'''pixel_values''': images} , tensor_type=UpperCAmelCase__ ) if apply_ocr: snake_case : Dict = words_batch snake_case : Dict = boxes_batch return data
84
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a : str = logging.get_logger(__name__) _a : Any = { "xlm-roberta-base": "https://huggingface.co/xlm-roberta-base/resolve/main/config.json", "xlm-roberta-large": "https://huggingface.co/xlm-roberta-large/resolve/main/config.json", "xlm-roberta-large-finetuned-conll02-dutch": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll02-spanish": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll03-english": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll03-german": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/config.json" ), } class a_ ( __lowerCAmelCase ): A__ : List[Any] = 'xlm-roberta' def __init__( self : List[str] , UpperCAmelCase__ : Any=30_522 , UpperCAmelCase__ : Optional[Any]=768 , UpperCAmelCase__ : Any=12 , UpperCAmelCase__ : Optional[int]=12 , UpperCAmelCase__ : Optional[Any]=3_072 , UpperCAmelCase__ : Any="gelu" , UpperCAmelCase__ : Tuple=0.1 , UpperCAmelCase__ : Tuple=0.1 , UpperCAmelCase__ : Tuple=512 , UpperCAmelCase__ : str=2 , UpperCAmelCase__ : Dict=0.02 , UpperCAmelCase__ : List[str]=1e-1_2 , UpperCAmelCase__ : Optional[int]=1 , UpperCAmelCase__ : Union[str, Any]=0 , UpperCAmelCase__ : str=2 , UpperCAmelCase__ : Tuple="absolute" , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Tuple=None , **UpperCAmelCase__ : Dict , ): """simple docstring""" super().__init__(pad_token_id=_UpperCamelCase , bos_token_id=_UpperCamelCase , eos_token_id=_UpperCamelCase , **_UpperCamelCase ) snake_case : Union[str, Any] = vocab_size snake_case : Tuple = hidden_size snake_case : Union[str, Any] = num_hidden_layers snake_case : int = num_attention_heads snake_case : List[str] = hidden_act snake_case : List[Any] = intermediate_size snake_case : Optional[Any] = hidden_dropout_prob snake_case : str = attention_probs_dropout_prob snake_case : Tuple = max_position_embeddings snake_case : Union[str, Any] = type_vocab_size snake_case : Dict = initializer_range snake_case : Any = layer_norm_eps snake_case : Union[str, Any] = position_embedding_type snake_case : Tuple = use_cache snake_case : int = classifier_dropout class a_ ( __lowerCAmelCase ): @property def lowerCAmelCase( self : List[str] ): """simple docstring""" if self.task == "multiple-choice": snake_case : Union[str, Any] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: snake_case : Union[str, Any] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
715
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class a_ : def __init__( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Union[str, Any]=13 , UpperCAmelCase__ : Union[str, Any]=2 , UpperCAmelCase__ : List[Any]=24 , UpperCAmelCase__ : Union[str, Any]=16 , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : str=True , UpperCAmelCase__ : int=32 , UpperCAmelCase__ : Tuple=5 , UpperCAmelCase__ : Dict=4 , UpperCAmelCase__ : Optional[int]=37 , UpperCAmelCase__ : Optional[int]="gelu" , UpperCAmelCase__ : Dict=0.1 , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : Optional[int]=10 , UpperCAmelCase__ : Any=0.02 , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : List[Any]=2 , UpperCAmelCase__ : Optional[Any]=2 , ): """simple docstring""" snake_case : Tuple = parent snake_case : Dict = batch_size snake_case : str = patch_size snake_case : Union[str, Any] = max_length snake_case : str = num_mel_bins snake_case : Any = is_training snake_case : Union[str, Any] = use_labels snake_case : Tuple = hidden_size snake_case : Dict = num_hidden_layers snake_case : Any = num_attention_heads snake_case : Any = intermediate_size snake_case : List[Any] = hidden_act snake_case : str = hidden_dropout_prob snake_case : str = attention_probs_dropout_prob snake_case : str = type_sequence_label_size snake_case : Optional[int] = initializer_range snake_case : str = scope snake_case : int = frequency_stride snake_case : Union[str, Any] = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) snake_case : Optional[int] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 snake_case : Any = (self.max_length - self.patch_size) // self.time_stride + 1 snake_case : Union[str, Any] = frequency_out_dimension * time_out_dimension snake_case : Union[str, Any] = num_patches + 2 def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Optional[int] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) snake_case : str = None if self.use_labels: snake_case : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case : List[str] = self.get_config() return config, input_values, labels def lowerCAmelCase( self : Any ): """simple docstring""" return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCAmelCase__ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def lowerCAmelCase( self : Tuple , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[int] ): """simple docstring""" snake_case : str = ASTModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() snake_case : Any = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : int = self.prepare_config_and_inputs() ( ( snake_case ) , ( snake_case ) , ( snake_case ) , ) : int = config_and_inputs snake_case : Tuple = {'''input_values''': input_values} return config, inputs_dict @require_torch class a_ ( a , a , unittest.TestCase ): A__ : List[Any] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) A__ : int = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) A__ : Optional[Any] = False A__ : Dict = False A__ : int = False A__ : Optional[int] = False def lowerCAmelCase( self : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : int ): """simple docstring""" if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[int] = ASTModelTester(self ) snake_case : Optional[int] = ConfigTester(self , config_class=UpperCAmelCase__ , has_text_modality=UpperCAmelCase__ , hidden_size=37 ) def lowerCAmelCase( self : List[str] ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def lowerCAmelCase( self : Tuple ): """simple docstring""" pass def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Optional[Any] = model_class(UpperCAmelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase__ , nn.Linear ) ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case , snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case : Any = model_class(UpperCAmelCase__ ) snake_case : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case : str = [*signature.parameters.keys()] snake_case : List[str] = ['''input_values'''] self.assertListEqual(arg_names[:1] , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case : List[str] = ASTModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) def a_ ( ) -> Dict: """simple docstring""" snake_case : Dict = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) snake_case , snake_case : int = torchaudio.load(__magic_name__ ) return audio, sampling_rate @require_torch @require_torchaudio class a_ ( unittest.TestCase ): @cached_property def lowerCAmelCase( self : Any ): """simple docstring""" return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def lowerCAmelCase( self : Tuple ): """simple docstring""" snake_case : List[str] = self.default_feature_extractor snake_case : str = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(UpperCAmelCase__ ) snake_case : str = self.default_feature_extractor snake_case , snake_case : int = prepare_audio() snake_case : Optional[int] = audio.squeeze().numpy() snake_case : Optional[Any] = feature_extractor(UpperCAmelCase__ , sampling_rate=UpperCAmelCase__ , return_tensors='''pt''' ).to(UpperCAmelCase__ ) # forward pass with torch.no_grad(): snake_case : Union[str, Any] = model(**UpperCAmelCase__ ) # verify the logits snake_case : Any = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase__ ) snake_case : str = torch.tensor([-0.8760, -7.0042, -8.6602] ).to(UpperCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase__ , atol=1e-4 ) )
84
0
import argparse import random import joblib import numpy as np import torch from igf.igf import ( SecondaryLearner, collect_objective_set, compute_perplexity, generate_datasets, load_gpta, recopy_gpta, set_seed, train_secondary_learner, ) from torch.utils.data import DataLoader, RandomSampler from transformers import GPTaLMHeadModel def a_ ( __magic_name__=32 , __magic_name__=10 , __magic_name__=100 , __magic_name__=1_026 , __magic_name__=True , __magic_name__="data/tokenized_stories_train_wikitext103.jbl" , __magic_name__="igf_context_pairs.jbl" , ) -> str: """simple docstring""" set_seed(3 ) # generate train_data and objective_set snake_case : int = generate_datasets( _lowerCamelCase , _lowerCamelCase , number=_lowerCamelCase , min_len=1_026 , trim=_lowerCamelCase ) # keeps model same across runs set_seed(4 ) # model, lm_optimizer, lm_scheduler = recopy_gpt2(model, device, max_steps) # store original model weights # can we train on GPU? snake_case : int = torch.device('''cuda:0''' if torch.cuda.is_available() else '''cpu''' ) # load pretrained model snake_case : Dict = load_gpta('''gpt2''' ).to(_lowerCamelCase ) print('''computing perplexity on objective set''' ) snake_case : str = compute_perplexity(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ).item() print('''perplexity on objective set:''' , _lowerCamelCase ) # collect igf pairs and save to file demo.jbl collect_objective_set(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # clean up, delete model and data we don't need anymore del model, train_data, objective_set torch.cuda.empty_cache() def a_ ( __magic_name__ , __magic_name__=15 , __magic_name__=128 , __magic_name__=100 , __magic_name__="igf_model.pt" , ) -> Optional[int]: """simple docstring""" set_seed(42 ) # Load pre-trained model snake_case : int = GPTaLMHeadModel.from_pretrained('''gpt2''' ) # Initialize secondary learner to use embedding weights of model snake_case : Optional[int] = SecondaryLearner(_lowerCamelCase ) # Train secondary learner snake_case : int = train_secondary_learner( _lowerCamelCase , _lowerCamelCase , max_epochs=_lowerCamelCase , batch_size=_lowerCamelCase , eval_freq=100 , igf_model_path=_lowerCamelCase , ) del model, secondary_learner_train_data torch.cuda.empty_cache() return secondary_learner def a_ ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__=32 , __magic_name__=1_000 , __magic_name__=16 , __magic_name__=1.0 , __magic_name__=recopy_gpta , __magic_name__=None , __magic_name__=10 , __magic_name__="gpt2_finetuned.pt" , ) -> Any: """simple docstring""" snake_case : str = torch.device('''cuda:0''' if torch.cuda.is_available() else '''cpu''' ) snake_case : Tuple = RandomSampler(_lowerCamelCase ) snake_case : Tuple = DataLoader(_lowerCamelCase , sampler=_lowerCamelCase ) snake_case : List[str] = max_steps // (len(_lowerCamelCase )) + 1 snake_case : Any = 0 snake_case : Any = torch.zeros((1, context_len) , dtype=torch.long , device=_lowerCamelCase ) snake_case : Tuple = recopy_model(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) model.train() if secondary_learner is not None: secondary_learner.to(_lowerCamelCase ) secondary_learner.eval() snake_case : Tuple = [] snake_case : int = 0 snake_case : Optional[int] = [] snake_case : str = [] # Compute the performance of the transformer model at the beginning snake_case : Any = compute_perplexity(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) test_perps.append(_lowerCamelCase ) print('''Test perplexity, step''' , _lowerCamelCase , ''':''' , _lowerCamelCase ) for epoch in range(int(_lowerCamelCase ) ): for step, example in enumerate(_lowerCamelCase ): torch.cuda.empty_cache() snake_case : List[str] = random.randint(0 , example.size(2 ) - context_len - 1 ) snake_case : List[Any] = example[0, 0, start : start + context_len] lm_optimizer.zero_grad() snake_case : Optional[int] = model(_lowerCamelCase , labels=_lowerCamelCase ) snake_case : List[str] = True if secondary_learner is not None: snake_case : Dict = secondary_learner.forward( torch.tensor(_lowerCamelCase , dtype=torch.long , device=_lowerCamelCase ).unsqueeze(0 ) )[0].item() observed_qs.append(float(_lowerCamelCase ) ) # Here we implement the simple non-constant threshold for the predicted IG(X) value # We will decay the selectivity of our secondary learner filter from # 1 standard deviation above average to 1 below average after 10 batches. if global_step == 10: snake_case : Union[str, Any] = -1 if predicted_q < threshold: snake_case : List[Any] = False # If we passed the filter, add the context to the batch! if do_backprop: contexts.append(np.array(context.cpu() ) ) snake_case : Optional[Any] = outputs[0] lm_loss.backward() examples += 1 del outputs # Once the batch is filled with enough contexts, backprop on the batch. if examples == batch_size: torch.cuda.empty_cache() snake_case : str = 0 # Do LM backprop torch.nn.utils.clip_grad_norm_(model.parameters() , 3.0 ) lm_optimizer.step() lm_scheduler.step() # Update learning rate schedule global_step += 1 # Compute the performance of the transformer model at this batch if global_step % eval_interval == 0: snake_case : Tuple = compute_perplexity(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) test_perps.append(_lowerCamelCase ) print('''Test perplexity, step''' , _lowerCamelCase , ''':''' , _lowerCamelCase ) # Break out of the loop after 60 batches if max_steps > 0 and global_step > 60: break if max_steps > 0 and global_step > 60: break # save finetuned transformer model torch.save(model.state_dict() , _lowerCamelCase ) torch.cuda.empty_cache() # Do some cleaning up so we can reinitialize for the next run of this function del lm_optimizer del lm_scheduler return model def a_ ( ) -> List[str]: """simple docstring""" snake_case : Tuple = argparse.ArgumentParser(description='''Fine-tune a transformer model with IGF on a language modeling task''' ) # Required parameters parser.add_argument( '''--data_dir''' , default=_lowerCamelCase , type=_lowerCamelCase , required=_lowerCamelCase , help='''The input data dir. Should contain data files for WikiText.''' , ) parser.add_argument( '''--model_name_or_path''' , default=_lowerCamelCase , type=_lowerCamelCase , required=_lowerCamelCase , help='''Path to pretrained model or model identifier from huggingface.co/models''' , ) parser.add_argument( '''--data_file''' , type=_lowerCamelCase , default=_lowerCamelCase , help=( '''A jbl file containing tokenized data which can be split as objective dataset, ''' '''train_dataset and test_dataset.''' ) , ) parser.add_argument( '''--igf_data_file''' , type=_lowerCamelCase , default=_lowerCamelCase , help='''A jbl file containing the context and information gain pairs to train secondary learner.''' , ) parser.add_argument( '''--output_dir''' , default=_lowerCamelCase , type=_lowerCamelCase , required=_lowerCamelCase , help='''The output directory where the final fine-tuned model is stored.''' , ) parser.add_argument( '''--tokenizer_name''' , default=_lowerCamelCase , type=_lowerCamelCase , help='''Pretrained tokenizer name or path if not the same as model_name''' , ) parser.add_argument('''--seed''' , type=_lowerCamelCase , default=_lowerCamelCase , help='''A seed for reproducible training.''' ) parser.add_argument( '''--context_len''' , default=32 , type=_lowerCamelCase , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--size_objective_set''' , default=100 , type=_lowerCamelCase , help='''number of articles that are long enough to be used as our objective set''' , ) parser.add_argument( '''--eval_freq''' , default=100 , type=_lowerCamelCase , help='''secondary model evaluation is triggered at eval_freq''' ) parser.add_argument('''--max_steps''' , default=1_000 , type=_lowerCamelCase , help='''To calculate training epochs''' ) parser.add_argument( '''--secondary_learner_batch_size''' , default=128 , type=_lowerCamelCase , help='''batch size of training data for secondary learner''' , ) parser.add_argument( '''--batch_size''' , default=16 , type=_lowerCamelCase , help='''batch size of training data of language model(gpt2) ''' ) parser.add_argument( '''--eval_interval''' , default=10 , type=_lowerCamelCase , help=( '''decay the selectivity of our secondary learner filter from''' '''1 standard deviation above average to 1 below average after 10 batches''' ) , ) parser.add_argument( '''--number''' , default=100 , type=_lowerCamelCase , help='''The number of examples split to be used as objective_set/test_data''' ) parser.add_argument( '''--min_len''' , default=1_026 , type=_lowerCamelCase , help='''The minimum length of the article to be used as objective set''' ) parser.add_argument( '''--secondary_learner_max_epochs''' , default=15 , type=_lowerCamelCase , help='''number of epochs to train secondary learner''' ) parser.add_argument('''--trim''' , default=_lowerCamelCase , type=_lowerCamelCase , help='''truncate the example if it exceeds context length''' ) parser.add_argument( '''--threshold''' , default=1.0 , type=_lowerCamelCase , help=( '''The threshold value used by secondary learner to filter the train_data and allow only''' ''' informative data as input to the model''' ) , ) parser.add_argument('''--finetuned_model_name''' , default='''gpt2_finetuned.pt''' , type=_lowerCamelCase , help='''finetuned_model_name''' ) parser.add_argument( '''--recopy_model''' , default=_lowerCamelCase , type=_lowerCamelCase , help='''Reset the model to the original pretrained GPT-2 weights after each iteration''' , ) # function calls # Collecting *n* pairs of context and information gain(X, IG(X)) for training the secondary learner generate_n_pairs( context_len=32 , max_steps=10 , size_objective_set=100 , min_len=1_026 , trim=_lowerCamelCase , data_file='''data/tokenized_stories_train_wikitext103.jbl''' , igf_data_file='''igf_context_pairs.jbl''' , ) # Load train data for secondary learner snake_case : Optional[int] = joblib.load('''data/IGF_values.jbl''' ) # Train secondary learner snake_case : str = training_secondary_learner( _lowerCamelCase , secondary_learner_max_epochs=15 , secondary_learner_batch_size=128 , eval_freq=100 , igf_model_path='''igf_model.pt''' , ) # load pretrained gpt2 model snake_case : Any = GPTaLMHeadModel.from_pretrained('''gpt2''' ) set_seed(42 ) # Generate train and test data to train and evaluate gpt2 model snake_case : Any = generate_datasets( context_len=32 , file='''data/tokenized_stories_train_wikitext103.jbl''' , number=100 , min_len=1_026 , trim=_lowerCamelCase ) # fine-tuning of the gpt2 model using igf (Information Gain Filtration) finetune( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , context_len=32 , max_steps=1_000 , batch_size=16 , threshold=1.0 , recopy_model=_lowerCamelCase , secondary_learner=_lowerCamelCase , eval_interval=10 , finetuned_model_name='''gpt2_finetuned.pt''' , ) if __name__ == "__main__": main()
716
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _a : Union[str, Any] = logging.getLogger(__name__) def a_ ( __magic_name__ , __magic_name__ ) -> Optional[int]: """simple docstring""" return (preds == labels).mean() @dataclass class a_ : A__ : str = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) A__ : Optional[str] = field( default=a , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) @dataclass class a_ : A__ : str = field(metadata={'help': 'The name of the task to train on: ' + ', '.join(processors.keys() )} ) A__ : str = field(metadata={'help': 'Should contain the data files for the task.'} ) A__ : int = field( default=128 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) A__ : bool = field( default=a , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) def a_ ( ) -> Dict: """simple docstring""" snake_case : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) snake_case , snake_case , snake_case : Tuple = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"Output directory ({training_args.output_dir}) already exists and is not empty. Use" ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , __magic_name__ ) # Set seed set_seed(training_args.seed ) try: snake_case : int = processors[data_args.task_name]() snake_case : List[str] = processor.get_labels() snake_case : str = len(__magic_name__ ) except KeyError: raise ValueError('''Task not found: %s''' % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=__magic_name__ , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) snake_case : str = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) snake_case : Any = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=__magic_name__ , cache_dir=model_args.cache_dir , ) # Get datasets snake_case : Optional[int] = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=__magic_name__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) snake_case : Any = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=__magic_name__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(__magic_name__ ) -> Dict: snake_case : str = np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(__magic_name__ , p.label_ids )} # Data collator snake_case : Dict = DataCollatorWithPadding(__magic_name__ , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer snake_case : List[Any] = Trainer( model=__magic_name__ , args=__magic_name__ , train_dataset=__magic_name__ , eval_dataset=__magic_name__ , compute_metrics=__magic_name__ , data_collator=__magic_name__ , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation snake_case : Optional[int] = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) snake_case : Optional[Any] = trainer.evaluate() snake_case : Union[str, Any] = os.path.join(training_args.output_dir , '''eval_results.txt''' ) if trainer.is_world_master(): with open(__magic_name__ , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(''' %s = %s''' , __magic_name__ , __magic_name__ ) writer.write('''%s = %s\n''' % (key, value) ) results.update(__magic_name__ ) return results def a_ ( __magic_name__ ) -> List[Any]: """simple docstring""" main() if __name__ == "__main__": main()
84
0
_a : Any = { 0: '0', 1: '1', 2: '2', 3: '3', 4: '4', 5: '5', 6: '6', 7: '7', 8: '8', 9: '9', 10: 'a', 11: 'b', 12: 'c', 13: 'd', 14: 'e', 15: 'f', } def a_ ( __magic_name__ ) -> List[str]: """simple docstring""" assert type(_lowerCamelCase ) in (int, float) and decimal == int(_lowerCamelCase ) snake_case : Optional[int] = int(_lowerCamelCase ) snake_case : Dict = "" snake_case : Dict = False if decimal < 0: snake_case : List[str] = True decimal *= -1 while decimal > 0: snake_case : Optional[Any] = divmod(_lowerCamelCase , 16 ) snake_case : Any = values[remainder] + hexadecimal snake_case : Tuple = "0x" + hexadecimal if negative: snake_case : Optional[Any] = "-" + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
717
import re def a_ ( __magic_name__ ) -> bool: """simple docstring""" snake_case : List[str] = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(__magic_name__ , __magic_name__ ) ) if __name__ == "__main__": _a : Any = '0094702343221' print(is_sri_lankan_phone_number(phone))
84
0
import os from collections import namedtuple import pytest from datasets import ClassLabel, Features, Sequence, Value from datasets.commands.test import TestCommand from datasets.info import DatasetInfo, DatasetInfosDict _a : Union[str, Any] = namedtuple( '_TestCommandArgs', [ 'dataset', 'name', 'cache_dir', 'data_dir', 'all_configs', 'save_infos', 'ignore_verifications', 'force_redownload', 'clear_cache', ], defaults=[None, None, None, False, False, False, False, False], ) def a_ ( __magic_name__ , __magic_name__ ) -> Tuple: """simple docstring""" return (abs(source - target ) / target) < 0.01 @pytest.mark.integration def a_ ( __magic_name__ ) -> int: """simple docstring""" snake_case : Optional[Any] = _TestCommandArgs(dataset=a_ , all_configs=a_ , save_infos=a_ ) snake_case : int = TestCommand(*a_ ) test_command.run() snake_case : Optional[int] = os.path.join(a_ , '''README.md''' ) assert os.path.exists(a_ ) snake_case : Dict = DatasetInfosDict.from_directory(a_ ) snake_case : Any = DatasetInfosDict( { '''default''': DatasetInfo( features=Features( { '''tokens''': Sequence(Value('''string''' ) ), '''ner_tags''': Sequence( ClassLabel(names=['''O''', '''B-PER''', '''I-PER''', '''B-ORG''', '''I-ORG''', '''B-LOC''', '''I-LOC'''] ) ), '''langs''': Sequence(Value('''string''' ) ), '''spans''': Sequence(Value('''string''' ) ), } ) , splits=[ { '''name''': '''train''', '''num_bytes''': 2_351_563, '''num_examples''': 10_000, }, { '''name''': '''validation''', '''num_bytes''': 238_418, '''num_examples''': 1_000, }, ] , download_size=3_940_680 , dataset_size=2_589_981 , ) } ) assert dataset_infos.keys() == expected_dataset_infos.keys() for key in DatasetInfo._INCLUDED_INFO_IN_YAML: snake_case : Any = getattr(dataset_infos['''default'''] , a_ ), getattr(expected_dataset_infos['''default'''] , a_ ) if key == "num_bytes": assert is_apercent_close(a_ , a_ ) elif key == "splits": assert list(a_ ) == list(a_ ) for split in result: assert result[split].name == expected[split].name assert result[split].num_examples == expected[split].num_examples assert is_apercent_close(result[split].num_bytes , expected[split].num_bytes ) else: result == expected
718
import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class a_ ( unittest.TestCase ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[int]=7 , UpperCAmelCase__ : Union[str, Any]=3 , UpperCAmelCase__ : int=18 , UpperCAmelCase__ : Optional[int]=30 , UpperCAmelCase__ : Optional[int]=400 , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : int=True , ): """simple docstring""" snake_case : int = size if size is not None else {'''height''': 18, '''width''': 18} snake_case : Optional[Any] = parent snake_case : Any = batch_size snake_case : Any = num_channels snake_case : Union[str, Any] = image_size snake_case : Dict = min_resolution snake_case : Dict = max_resolution snake_case : int = do_resize snake_case : List[str] = size snake_case : List[Any] = apply_ocr def lowerCAmelCase( self : int ): """simple docstring""" return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class a_ ( a , unittest.TestCase ): A__ : List[Any] = LayoutLMvaImageProcessor if is_pytesseract_available() else None def lowerCAmelCase( self : Dict ): """simple docstring""" snake_case : Optional[Any] = LayoutLMvaImageProcessingTester(self ) @property def lowerCAmelCase( self : Dict ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase( self : List[Any] ): """simple docstring""" snake_case : List[str] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase__ , '''do_resize''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''size''' ) ) self.assertTrue(hasattr(UpperCAmelCase__ , '''apply_ocr''' ) ) def lowerCAmelCase( self : Optional[int] ): """simple docstring""" snake_case : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 18, '''width''': 18} ) snake_case : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) def lowerCAmelCase( self : str ): """simple docstring""" pass def lowerCAmelCase( self : Dict ): """simple docstring""" # Initialize image_processing snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , Image.Image ) # Test not batched input snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) self.assertIsInstance(encoding.words , UpperCAmelCase__ ) self.assertIsInstance(encoding.boxes , UpperCAmelCase__ ) # Test batched snake_case : Dict = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" # Initialize image_processing snake_case : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , numpify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , np.ndarray ) # Test not batched input snake_case : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched snake_case : List[Any] = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" # Initialize image_processing snake_case : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , torchify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , torch.Tensor ) # Test not batched input snake_case : List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched snake_case : Tuple = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" # with apply_OCR = True snake_case : int = LayoutLMvaImageProcessor() from datasets import load_dataset snake_case : List[Any] = load_dataset('''hf-internal-testing/fixtures_docvqa''' , split='''test''' ) snake_case : List[Any] = Image.open(ds[0]['''file'''] ).convert('''RGB''' ) snake_case : Any = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 snake_case : Optional[Any] = [['''11:14''', '''to''', '''11:39''', '''a.m''', '''11:39''', '''to''', '''11:44''', '''a.m.''', '''11:44''', '''a.m.''', '''to''', '''12:25''', '''p.m.''', '''12:25''', '''to''', '''12:58''', '''p.m.''', '''12:58''', '''to''', '''4:00''', '''p.m.''', '''2:00''', '''to''', '''5:00''', '''p.m.''', '''Coffee''', '''Break''', '''Coffee''', '''will''', '''be''', '''served''', '''for''', '''men''', '''and''', '''women''', '''in''', '''the''', '''lobby''', '''adjacent''', '''to''', '''exhibit''', '''area.''', '''Please''', '''move''', '''into''', '''exhibit''', '''area.''', '''(Exhibits''', '''Open)''', '''TRRF''', '''GENERAL''', '''SESSION''', '''(PART''', '''|)''', '''Presiding:''', '''Lee''', '''A.''', '''Waller''', '''TRRF''', '''Vice''', '''President''', '''“Introductory''', '''Remarks”''', '''Lee''', '''A.''', '''Waller,''', '''TRRF''', '''Vice''', '''Presi-''', '''dent''', '''Individual''', '''Interviews''', '''with''', '''TRRF''', '''Public''', '''Board''', '''Members''', '''and''', '''Sci-''', '''entific''', '''Advisory''', '''Council''', '''Mem-''', '''bers''', '''Conducted''', '''by''', '''TRRF''', '''Treasurer''', '''Philip''', '''G.''', '''Kuehn''', '''to''', '''get''', '''answers''', '''which''', '''the''', '''public''', '''refrigerated''', '''warehousing''', '''industry''', '''is''', '''looking''', '''for.''', '''Plus''', '''questions''', '''from''', '''the''', '''floor.''', '''Dr.''', '''Emil''', '''M.''', '''Mrak,''', '''University''', '''of''', '''Cal-''', '''ifornia,''', '''Chairman,''', '''TRRF''', '''Board;''', '''Sam''', '''R.''', '''Cecil,''', '''University''', '''of''', '''Georgia''', '''College''', '''of''', '''Agriculture;''', '''Dr.''', '''Stanley''', '''Charm,''', '''Tufts''', '''University''', '''School''', '''of''', '''Medicine;''', '''Dr.''', '''Robert''', '''H.''', '''Cotton,''', '''ITT''', '''Continental''', '''Baking''', '''Company;''', '''Dr.''', '''Owen''', '''Fennema,''', '''University''', '''of''', '''Wis-''', '''consin;''', '''Dr.''', '''Robert''', '''E.''', '''Hardenburg,''', '''USDA.''', '''Questions''', '''and''', '''Answers''', '''Exhibits''', '''Open''', '''Capt.''', '''Jack''', '''Stoney''', '''Room''', '''TRRF''', '''Scientific''', '''Advisory''', '''Council''', '''Meeting''', '''Ballroom''', '''Foyer''']] # noqa: E231 snake_case : Union[str, Any] = [[[141, 57, 214, 69], [228, 58, 252, 69], [141, 75, 216, 88], [230, 79, 280, 88], [142, 260, 218, 273], [230, 261, 255, 273], [143, 279, 218, 290], [231, 282, 290, 291], [143, 342, 218, 354], [231, 345, 289, 355], [202, 362, 227, 373], [143, 379, 220, 392], [231, 382, 291, 394], [144, 714, 220, 726], [231, 715, 256, 726], [144, 732, 220, 745], [232, 736, 291, 747], [144, 769, 218, 782], [231, 770, 256, 782], [141, 788, 202, 801], [215, 791, 274, 804], [143, 826, 204, 838], [215, 826, 240, 838], [142, 844, 202, 857], [215, 847, 274, 859], [334, 57, 427, 69], [440, 57, 522, 69], [369, 75, 461, 88], [469, 75, 516, 88], [528, 76, 562, 88], [570, 76, 667, 88], [675, 75, 711, 87], [721, 79, 778, 88], [789, 75, 840, 88], [369, 97, 470, 107], [484, 94, 507, 106], [518, 94, 562, 107], [576, 94, 655, 110], [668, 94, 792, 109], [804, 95, 829, 107], [369, 113, 465, 125], [477, 116, 547, 125], [562, 113, 658, 125], [671, 116, 748, 125], [761, 113, 811, 125], [369, 131, 465, 143], [477, 133, 548, 143], [563, 130, 698, 145], [710, 130, 802, 146], [336, 171, 412, 183], [423, 171, 572, 183], [582, 170, 716, 184], [728, 171, 817, 187], [829, 171, 844, 186], [338, 197, 482, 212], [507, 196, 557, 209], [569, 196, 595, 208], [610, 196, 702, 209], [505, 214, 583, 226], [595, 214, 656, 227], [670, 215, 807, 227], [335, 259, 543, 274], [556, 259, 708, 272], [372, 279, 422, 291], [435, 279, 460, 291], [474, 279, 574, 292], [587, 278, 664, 291], [676, 278, 738, 291], [751, 279, 834, 291], [372, 298, 434, 310], [335, 341, 483, 354], [497, 341, 655, 354], [667, 341, 728, 354], [740, 341, 825, 354], [335, 360, 430, 372], [442, 360, 534, 372], [545, 359, 687, 372], [697, 360, 754, 372], [765, 360, 823, 373], [334, 378, 428, 391], [440, 378, 577, 394], [590, 378, 705, 391], [720, 378, 801, 391], [334, 397, 400, 409], [370, 416, 529, 429], [544, 416, 576, 432], [587, 416, 665, 428], [677, 416, 814, 429], [372, 435, 452, 450], [465, 434, 495, 447], [511, 434, 600, 447], [611, 436, 637, 447], [649, 436, 694, 451], [705, 438, 824, 447], [369, 453, 452, 466], [464, 454, 509, 466], [522, 453, 611, 469], [625, 453, 792, 469], [370, 472, 556, 488], [570, 472, 684, 487], [697, 472, 718, 485], [732, 472, 835, 488], [369, 490, 411, 503], [425, 490, 484, 503], [496, 490, 635, 506], [645, 490, 707, 503], [718, 491, 761, 503], [771, 490, 840, 503], [336, 510, 374, 521], [388, 510, 447, 522], [460, 510, 489, 521], [503, 510, 580, 522], [592, 509, 736, 525], [745, 509, 770, 522], [781, 509, 840, 522], [338, 528, 434, 541], [448, 528, 596, 541], [609, 527, 687, 540], [700, 528, 792, 541], [336, 546, 397, 559], [407, 546, 431, 559], [443, 546, 525, 560], [537, 546, 680, 562], [688, 546, 714, 559], [722, 546, 837, 562], [336, 565, 449, 581], [461, 565, 485, 577], [497, 565, 665, 581], [681, 565, 718, 577], [732, 565, 837, 580], [337, 584, 438, 597], [452, 583, 521, 596], [535, 584, 677, 599], [690, 583, 787, 596], [801, 583, 825, 596], [338, 602, 478, 615], [492, 602, 530, 614], [543, 602, 638, 615], [650, 602, 676, 614], [688, 602, 788, 615], [802, 602, 843, 614], [337, 621, 502, 633], [516, 621, 615, 637], [629, 621, 774, 636], [789, 621, 827, 633], [337, 639, 418, 652], [432, 640, 571, 653], [587, 639, 731, 655], [743, 639, 769, 652], [780, 639, 841, 652], [338, 658, 440, 673], [455, 658, 491, 670], [508, 658, 602, 671], [616, 658, 638, 670], [654, 658, 835, 674], [337, 677, 429, 689], [337, 714, 482, 726], [495, 714, 548, 726], [561, 714, 683, 726], [338, 770, 461, 782], [474, 769, 554, 785], [489, 788, 562, 803], [576, 788, 643, 801], [656, 787, 751, 804], [764, 788, 844, 801], [334, 825, 421, 838], [430, 824, 574, 838], [584, 824, 723, 841], [335, 844, 450, 857], [464, 843, 583, 860], [628, 862, 755, 875], [769, 861, 848, 878]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , UpperCAmelCase__ ) self.assertListEqual(encoding.boxes , UpperCAmelCase__ ) # with apply_OCR = False snake_case : str = LayoutLMvaImageProcessor(apply_ocr=UpperCAmelCase__ ) snake_case : Optional[Any] = image_processing(UpperCAmelCase__ , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) )
84
0
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging _a : Tuple = logging.get_logger(__name__) class a_ ( _UpperCAmelCase ): A__ : Tuple = ['audio_values', 'audio_mask'] def __init__( self : Optional[Any] , UpperCAmelCase__ : Optional[Any]=2_048 , UpperCAmelCase__ : Any=1 , UpperCAmelCase__ : Any=[16, 16] , UpperCAmelCase__ : Optional[int]=128 , UpperCAmelCase__ : Union[str, Any]=44_100 , UpperCAmelCase__ : List[str]=86 , UpperCAmelCase__ : int=2_048 , UpperCAmelCase__ : Optional[Any]=0.0 , **UpperCAmelCase__ : Dict , ): """simple docstring""" super().__init__( feature_size=__UpperCamelCase , sampling_rate=__UpperCamelCase , padding_value=__UpperCamelCase , **__UpperCamelCase , ) snake_case : Tuple = spectrogram_length snake_case : Any = num_channels snake_case : Union[str, Any] = patch_size snake_case : Tuple = feature_size // self.patch_size[1] snake_case : Optional[Any] = n_fft snake_case : List[str] = sampling_rate // hop_length_to_sampling_rate snake_case : str = sampling_rate snake_case : List[str] = padding_value snake_case : Dict = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=__UpperCamelCase , min_frequency=0.0 , max_frequency=22_050.0 , sampling_rate=__UpperCamelCase , norm='''slaney''' , mel_scale='''slaney''' , ).T def lowerCAmelCase( self : Dict , UpperCAmelCase__ : np.array ): """simple docstring""" snake_case : Tuple = spectrogram( __UpperCamelCase , window_function(self.n_fft , '''hann''' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='''dB''' , db_range=80.0 , ) snake_case : int = log_spec[:, :-1] snake_case : Union[str, Any] = log_spec - 20.0 snake_case : Tuple = np.clip(log_spec / 40.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : Optional[Any] , UpperCAmelCase__ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , UpperCAmelCase__ : Optional[Union[str, TensorType]] = None , UpperCAmelCase__ : Optional[bool] = True , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : bool = False , **UpperCAmelCase__ : str , ): """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( '''This feature extractor is set to support sampling rate''' F" of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled" F" with {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''' ) snake_case : Dict = isinstance(__UpperCamelCase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}" ) snake_case : Dict = is_batched_numpy or ( isinstance(__UpperCamelCase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: snake_case : Optional[int] = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(__UpperCamelCase , np.ndarray ): snake_case : List[str] = np.asarray(__UpperCamelCase , dtype=np.floataa ) elif isinstance(__UpperCamelCase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): snake_case : Union[str, Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: snake_case : Dict = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis snake_case : Any = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , __UpperCamelCase ): snake_case : Optional[int] = [np.asarray(__UpperCamelCase , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask snake_case : List[str] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: snake_case : Dict = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] snake_case : Tuple = np.array(__UpperCamelCase ).astype(np.floataa ) # convert into correct format for padding snake_case : Optional[Any] = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch snake_case : Union[str, Any] = np.ones([len(__UpperCamelCase ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) snake_case : Dict = padded_audio_features * self.padding_value for i in range(len(__UpperCamelCase ) ): snake_case : Optional[int] = audio_features[i] snake_case : Any = feature # return as BatchFeature if return_attention_mask: snake_case : Any = {'''audio_values''': padded_audio_features, '''audio_mask''': audio_mask} else: snake_case : str = {'''audio_values''': padded_audio_features} snake_case : int = BatchFeature(data=__UpperCamelCase , tensor_type=__UpperCamelCase ) return encoded_inputs
719
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _a : Tuple = logging.get_logger(__name__) # pylint: disable=invalid-name _a : Dict = '\n Examples:\n ```py\n >>> import torch\n >>> import numpy as np\n\n >>> from diffusers import KandinskyV22PriorPipeline, KandinskyV22ControlnetPipeline\n >>> from transformers import pipeline\n >>> from diffusers.utils import load_image\n\n\n >>> def make_hint(image, depth_estimator):\n ... image = depth_estimator(image)["depth"]\n ... image = np.array(image)\n ... image = image[:, :, None]\n ... image = np.concatenate([image, image, image], axis=2)\n ... detected_map = torch.from_numpy(image).float() / 255.0\n ... hint = detected_map.permute(2, 0, 1)\n ... return hint\n\n\n >>> depth_estimator = pipeline("depth-estimation")\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16\n ... )\n >>> pipe_prior = pipe_prior.to("cuda")\n\n >>> pipe = KandinskyV22ControlnetPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-controlnet-depth", torch_dtype=torch.float16\n ... )\n >>> pipe = pipe.to("cuda")\n\n\n >>> img = load_image(\n ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main"\n ... "/kandinsky/cat.png"\n ... ).resize((768, 768))\n\n >>> hint = make_hint(img, depth_estimator).unsqueeze(0).half().to("cuda")\n\n >>> prompt = "A robot, 4k photo"\n >>> negative_prior_prompt = "lowres, text, error, cropped, worst quality, low quality, jpeg artifacts, ugly, duplicate, morbid, mutilated, out of frame, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, mutation, deformed, blurry, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, username, watermark, signature"\n\n >>> generator = torch.Generator(device="cuda").manual_seed(43)\n\n >>> image_emb, zero_image_emb = pipe_prior(\n ... prompt=prompt, negative_prompt=negative_prior_prompt, generator=generator\n ... ).to_tuple()\n\n >>> images = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... hint=hint,\n ... num_inference_steps=50,\n ... generator=generator,\n ... height=768,\n ... width=768,\n ... ).images\n\n >>> images[0].save("robot_cat.png")\n ```\n' def a_ ( __magic_name__ , __magic_name__ , __magic_name__=8 ) -> str: """simple docstring""" snake_case : List[Any] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 snake_case : Tuple = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class a_ ( a ): def __init__( self : Optional[int] , UpperCAmelCase__ : UNetaDConditionModel , UpperCAmelCase__ : DDPMScheduler , UpperCAmelCase__ : VQModel , ): """simple docstring""" super().__init__() self.register_modules( unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ , movq=UpperCAmelCase__ , ) snake_case : List[Any] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def lowerCAmelCase( self : int , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Any ): """simple docstring""" if latents is None: snake_case : int = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=UpperCAmelCase__ , dtype=UpperCAmelCase__ ) else: if latents.shape != shape: raise ValueError(F"Unexpected latents shape, got {latents.shape}, expected {shape}" ) snake_case : Optional[Any] = latents.to(UpperCAmelCase__ ) snake_case : List[Any] = latents * scheduler.init_noise_sigma return latents def lowerCAmelCase( self : Dict , UpperCAmelCase__ : Optional[int]=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) snake_case : Union[str, Any] = torch.device(F"cuda:{gpu_id}" ) snake_case : Dict = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : List[Any] , UpperCAmelCase__ : Any=0 ): """simple docstring""" if is_accelerate_available() and is_accelerate_version('''>=''' , '''0.17.0.dev0''' ): from accelerate import cpu_offload_with_hook else: raise ImportError('''`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.''' ) snake_case : Optional[int] = torch.device(F"cuda:{gpu_id}" ) if self.device.type != "cpu": self.to('''cpu''' , silence_dtype_warnings=UpperCAmelCase__ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) snake_case : List[str] = None for cpu_offloaded_model in [self.unet, self.movq]: snake_case , snake_case : Optional[int] = cpu_offload_with_hook(UpperCAmelCase__ , UpperCAmelCase__ , prev_module_hook=UpperCAmelCase__ ) # We'll offload the last model manually. snake_case : Tuple = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" if not hasattr(self.unet , '''_hf_hook''' ): return self.device for module in self.unet.modules(): if ( hasattr(UpperCAmelCase__ , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(UpperCAmelCase__ ) def __call__( self : List[str] , UpperCAmelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , UpperCAmelCase__ : torch.FloatTensor , UpperCAmelCase__ : int = 512 , UpperCAmelCase__ : int = 512 , UpperCAmelCase__ : int = 100 , UpperCAmelCase__ : float = 4.0 , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase__ : Optional[torch.FloatTensor] = None , UpperCAmelCase__ : Optional[str] = "pil" , UpperCAmelCase__ : bool = True , ): """simple docstring""" snake_case : Optional[int] = self._execution_device snake_case : Union[str, Any] = guidance_scale > 1.0 if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Any = torch.cat(UpperCAmelCase__ , dim=0 ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : Union[str, Any] = torch.cat(UpperCAmelCase__ , dim=0 ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case : int = torch.cat(UpperCAmelCase__ , dim=0 ) snake_case : List[Any] = image_embeds.shape[0] * num_images_per_prompt if do_classifier_free_guidance: snake_case : Dict = image_embeds.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Optional[Any] = negative_image_embeds.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Tuple = hint.repeat_interleave(UpperCAmelCase__ , dim=0 ) snake_case : Optional[Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=UpperCAmelCase__ ) snake_case : List[str] = torch.cat([hint, hint] , dim=0 ).to(dtype=self.unet.dtype , device=UpperCAmelCase__ ) self.scheduler.set_timesteps(UpperCAmelCase__ , device=UpperCAmelCase__ ) snake_case : str = self.scheduler.timesteps snake_case : Optional[Any] = self.movq.config.latent_channels snake_case , snake_case : Optional[Any] = downscale_height_and_width(UpperCAmelCase__ , UpperCAmelCase__ , self.movq_scale_factor ) # create initial latent snake_case : Dict = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , self.scheduler , ) for i, t in enumerate(self.progress_bar(UpperCAmelCase__ ) ): # expand the latents if we are doing classifier free guidance snake_case : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents snake_case : Optional[int] = {'''image_embeds''': image_embeds, '''hint''': hint} snake_case : Any = self.unet( sample=UpperCAmelCase__ , timestep=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , added_cond_kwargs=UpperCAmelCase__ , return_dict=UpperCAmelCase__ , )[0] if do_classifier_free_guidance: snake_case , snake_case : Dict = noise_pred.split(latents.shape[1] , dim=1 ) snake_case , snake_case : Any = noise_pred.chunk(2 ) snake_case , snake_case : Dict = variance_pred.chunk(2 ) snake_case : str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) snake_case : List[str] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , '''variance_type''' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): snake_case , snake_case : Tuple = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 snake_case : List[Any] = self.scheduler.step( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , generator=UpperCAmelCase__ , )[0] # post-processing snake_case : List[Any] = self.movq.decode(UpperCAmelCase__ , force_not_quantize=UpperCAmelCase__ )['''sample'''] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}" ) if output_type in ["np", "pil"]: snake_case : Optional[Any] = image * 0.5 + 0.5 snake_case : int = image.clamp(0 , 1 ) snake_case : List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": snake_case : str = self.numpy_to_pil(UpperCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase__ )
84
0
def a_ ( __magic_name__ = 50_000_000 ) -> int: """simple docstring""" snake_case : Optional[Any] = set() snake_case : int = int((limit - 24) ** (1 / 2) ) snake_case : Optional[int] = set(range(3 , prime_square_limit + 1 , 2 ) ) primes.add(2 ) for p in range(3 , prime_square_limit + 1 , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , prime_square_limit + 1 , __magic_name__ ) ) ) for primea in primes: snake_case : Optional[Any] = primea * primea for primea in primes: snake_case : int = primea * primea * primea if square + cube >= limit - 16: break for primea in primes: snake_case : int = primea * primea * primea * primea snake_case : str = square + cube + tetr if total >= limit: break ret.add(__magic_name__ ) return len(__magic_name__ ) if __name__ == "__main__": print(f"{solution() = }")
720
import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _a : List[Any] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class a_ ( a , unittest.TestCase ): A__ : Dict = ReformerTokenizer A__ : Optional[int] = ReformerTokenizerFast A__ : str = True A__ : Tuple = False A__ : str = True def lowerCAmelCase( self : List[Any] ): """simple docstring""" super().setUp() snake_case : str = ReformerTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase( self : Any ): """simple docstring""" snake_case : int = '''<s>''' snake_case : List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase__ ) , UpperCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase__ ) , UpperCAmelCase__ ) def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Any = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(UpperCAmelCase__ ) , 1_000 ) def lowerCAmelCase( self : List[Any] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def lowerCAmelCase( self : Dict ): """simple docstring""" if not self.test_rust_tokenizer: return snake_case : Any = self.get_tokenizer() snake_case : str = self.get_rust_tokenizer() snake_case : Tuple = '''I was born in 92000, and this is falsé.''' snake_case : str = tokenizer.tokenize(UpperCAmelCase__ ) snake_case : int = rust_tokenizer.tokenize(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : Union[str, Any] = tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) snake_case : List[str] = rust_tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) snake_case : List[str] = self.get_rust_tokenizer() snake_case : Optional[int] = tokenizer.encode(UpperCAmelCase__ ) snake_case : Optional[Any] = rust_tokenizer.encode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def lowerCAmelCase( self : Dict , UpperCAmelCase__ : List[Any]=15 ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): snake_case : str = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) # Simple input snake_case : Union[str, Any] = '''This is a simple input''' snake_case : List[str] = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case : int = ('''This is a simple input''', '''This is a pair''') snake_case : int = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Simple input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Simple input self.assertRaises( UpperCAmelCase__ , tokenizer_r.batch_encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' , ) # Pair input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Pair input self.assertRaises(UpperCAmelCase__ , tokenizer_r.encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' ) # Pair input self.assertRaises( UpperCAmelCase__ , tokenizer_r.batch_encode_plus , UpperCAmelCase__ , max_length=UpperCAmelCase__ , padding='''max_length''' , ) def lowerCAmelCase( self : str ): """simple docstring""" pass def lowerCAmelCase( self : Union[str, Any] ): """simple docstring""" snake_case : Union[str, Any] = ReformerTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) snake_case : List[str] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(UpperCAmelCase__ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , [285, 46, 10, 170, 382] , ) snake_case : int = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) snake_case : int = tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) snake_case : List[Any] = tokenizer.convert_ids_to_tokens(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def lowerCAmelCase( self : Tuple ): """simple docstring""" return ReformerTokenizer.from_pretrained('''google/reformer-crime-and-punishment''' ) @slow def lowerCAmelCase( self : List[str] ): """simple docstring""" snake_case : Any = '''Hello World!''' snake_case : Optional[Any] = [126, 32, 262, 152, 38, 72, 287] self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @slow def lowerCAmelCase( self : Optional[Any] ): """simple docstring""" snake_case : Optional[Any] = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) snake_case : Dict = [ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(UpperCAmelCase__ , self.big_tokenizer.encode(UpperCAmelCase__ ) ) @require_torch @slow def lowerCAmelCase( self : List[Any] ): """simple docstring""" import torch from transformers import ReformerConfig, ReformerModel # Build sequence snake_case : Any = list(self.big_tokenizer.get_vocab().keys() )[:10] snake_case : Union[str, Any] = ''' '''.join(UpperCAmelCase__ ) snake_case : Optional[int] = self.big_tokenizer.encode_plus(UpperCAmelCase__ , return_tensors='''pt''' ) snake_case : List[str] = self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors='''pt''' ) snake_case : Optional[Any] = ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) snake_case : Tuple = encoded_sequence['''input_ids'''].shape snake_case : List[Any] = ReformerModel(UpperCAmelCase__ ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**UpperCAmelCase__ ) model(**UpperCAmelCase__ ) @slow def lowerCAmelCase( self : Optional[int] ): """simple docstring""" # fmt: off snake_case : Tuple = {'''input_ids''': [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 snake_case : Tuple = [ '''This is a very simple sentence.''', '''The quick brown fox jumps over the lazy dog.''', ] self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase__ , model_name='''google/reformer-crime-and-punishment''' , revision='''0e6c3decb8211d49bf881013425dc8b0448b3f5a''' , padding=UpperCAmelCase__ , sequences=UpperCAmelCase__ , )
84
0