code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" from math import asin, atan, cos, radians, sin, sqrt, tan A: str = 637_8137.0 A: List[Any] = 635_6752.31_4245 A: Any = 6_3_7_8_1_3_7 def _snake_case ( UpperCamelCase : float , UpperCamelCase : float , UpperCamelCase : float , UpperCamelCase : float ): UpperCAmelCase : Optional[int] = (AXIS_A - AXIS_B) / AXIS_A UpperCAmelCase : Optional[int] = atan((1 - flattening) * tan(radians(UpperCamelCase ) ) ) UpperCAmelCase : Optional[Any] = atan((1 - flattening) * tan(radians(UpperCamelCase ) ) ) UpperCAmelCase : List[Any] = radians(UpperCamelCase ) UpperCAmelCase : Dict = radians(UpperCamelCase ) # Equation UpperCAmelCase : Dict = sin((phi_a - phi_a) / 2 ) UpperCAmelCase : List[Any] = sin((lambda_a - lambda_a) / 2 ) # Square both values sin_sq_phi *= sin_sq_phi sin_sq_lambda *= sin_sq_lambda UpperCAmelCase : List[str] = sqrt(sin_sq_phi + (cos(UpperCamelCase ) * cos(UpperCamelCase ) * sin_sq_lambda) ) return 2 * RADIUS * asin(UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
109
"""simple docstring""" from __future__ import annotations from statistics import mean def lowercase (_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): __lowerCAmelCase = [0] * no_of_processes __lowerCAmelCase = [0] * no_of_processes # Initialize remaining_time to waiting_time. for i in range(_lowerCAmelCase ): __lowerCAmelCase = burst_time[i] __lowerCAmelCase = [] __lowerCAmelCase = 0 __lowerCAmelCase = 0 # When processes are not completed, # A process whose arrival time has passed \ # and has remaining execution time is put into the ready_process. # The shortest process in the ready_process, target_process is executed. while completed != no_of_processes: __lowerCAmelCase = [] __lowerCAmelCase = -1 for i in range(_lowerCAmelCase ): if (arrival_time[i] <= total_time) and (remaining_time[i] > 0): ready_process.append(_lowerCAmelCase ) if len(_lowerCAmelCase ) > 0: __lowerCAmelCase = ready_process[0] for i in ready_process: if remaining_time[i] < remaining_time[target_process]: __lowerCAmelCase = i total_time += burst_time[target_process] completed += 1 __lowerCAmelCase = 0 __lowerCAmelCase = ( total_time - arrival_time[target_process] - burst_time[target_process] ) else: total_time += 1 return waiting_time def lowercase (_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): __lowerCAmelCase = [0] * no_of_processes for i in range(_lowerCAmelCase ): __lowerCAmelCase = burst_time[i] + waiting_time[i] return turn_around_time if __name__ == "__main__": print('''[TEST CASE 01]''') SCREAMING_SNAKE_CASE_ = 4 SCREAMING_SNAKE_CASE_ = [2, 5, 3, 7] SCREAMING_SNAKE_CASE_ = [0, 0, 0, 0] SCREAMING_SNAKE_CASE_ = calculate_waitingtime(arrival_time, burst_time, no_of_processes) SCREAMING_SNAKE_CASE_ = calculate_turnaroundtime( burst_time, no_of_processes, waiting_time ) # Printing the Result print('''PID\tBurst Time\tArrival Time\tWaiting Time\tTurnaround Time''') for i, process_id in enumerate(list(range(1, 5))): print( F"{process_id}\t{burst_time[i]}\t\t\t{arrival_time[i]}\t\t\t\t" F"{waiting_time[i]}\t\t\t\t{turn_around_time[i]}" ) print(F"\nAverage waiting time = {mean(waiting_time):.5f}") print(F"Average turnaround time = {mean(turn_around_time):.5f}")
301
0
from __future__ import annotations def SCREAMING_SNAKE_CASE__ ( snake_case_ ) -> Tuple: """simple docstring""" return len(set(a__ ) ) == len(a__ ) if __name__ == "__main__": import doctest doctest.testmod()
371
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ : Union[str, Any] = logging.get_logger(__name__) UpperCamelCase__ : Union[str, Any] = { """hustvl/yolos-small""": """https://huggingface.co/hustvl/yolos-small/resolve/main/config.json""", # See all YOLOS models at https://huggingface.co/models?filter=yolos } class lowerCamelCase_ ( a_ ): SCREAMING_SNAKE_CASE_ = 'yolos' def __init__( self : Union[str, Any] ,__lowerCamelCase : int=7_68 ,__lowerCamelCase : Dict=12 ,__lowerCamelCase : Union[str, Any]=12 ,__lowerCamelCase : List[Any]=30_72 ,__lowerCamelCase : int="gelu" ,__lowerCamelCase : int=0.0 ,__lowerCamelCase : str=0.0 ,__lowerCamelCase : Optional[Any]=0.02 ,__lowerCamelCase : int=1e-12 ,__lowerCamelCase : Any=[5_12, 8_64] ,__lowerCamelCase : Tuple=16 ,__lowerCamelCase : int=3 ,__lowerCamelCase : Tuple=True ,__lowerCamelCase : Optional[int]=1_00 ,__lowerCamelCase : List[Any]=True ,__lowerCamelCase : List[str]=False ,__lowerCamelCase : int=1 ,__lowerCamelCase : List[Any]=5 ,__lowerCamelCase : Optional[int]=2 ,__lowerCamelCase : int=5 ,__lowerCamelCase : str=2 ,__lowerCamelCase : Tuple=0.1 ,**__lowerCamelCase : List[Any] ,): '''simple docstring''' super().__init__(**__lowerCamelCase ) a = hidden_size a = num_hidden_layers a = num_attention_heads a = intermediate_size a = hidden_act a = hidden_dropout_prob a = attention_probs_dropout_prob a = initializer_range a = layer_norm_eps a = image_size a = patch_size a = num_channels a = qkv_bias a = num_detection_tokens a = use_mid_position_embeddings a = auxiliary_loss # Hungarian matcher a = class_cost a = bbox_cost a = giou_cost # Loss coefficients a = bbox_loss_coefficient a = giou_loss_coefficient a = eos_coefficient class lowerCamelCase_ ( a_ ): SCREAMING_SNAKE_CASE_ = version.parse('1.11' ) @property def SCREAMING_SNAKE_CASE_ ( self : str ): '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ): '''simple docstring''' return 1e-4 @property def SCREAMING_SNAKE_CASE_ ( self : str ): '''simple docstring''' return 12
330
0
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() a_ : str = logging.get_logger(__name__) a_ : List[Any] = { """post_extract_proj""": """feature_projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.upsample.0""": """encoder.upsample.projection""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """layer_norm""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", } def a_ ( __snake_case : Dict , __snake_case : List[str] , __snake_case : Union[str, Any] , __snake_case : Optional[Any] , __snake_case : Union[str, Any] ) -> List[Any]: """simple docstring""" for attribute in key.split('''.''' ): lowerCamelCase_ =getattr(__snake_case , __snake_case ) if weight_type is not None: lowerCamelCase_ =getattr(__snake_case , __snake_case ).shape else: lowerCamelCase_ =hf_pointer.shape assert hf_shape == value.shape, ( F'''Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": lowerCamelCase_ =value elif weight_type == "weight_g": lowerCamelCase_ =value elif weight_type == "weight_v": lowerCamelCase_ =value elif weight_type == "bias": lowerCamelCase_ =value else: lowerCamelCase_ =value logger.info(F'''{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.''' ) def a_ ( __snake_case : int , __snake_case : Optional[int] , __snake_case : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ =[] lowerCamelCase_ =fairseq_model.state_dict() lowerCamelCase_ =hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): lowerCamelCase_ =False if "conv_layers" in name: load_conv_layer( __snake_case , __snake_case , __snake_case , __snake_case , hf_model.config.feat_extract_norm == '''group''' , ) lowerCamelCase_ =True else: for key, mapped_key in MAPPING.items(): lowerCamelCase_ ='''sew.''' + mapped_key if (is_finetuned and mapped_key != '''lm_head''') else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: lowerCamelCase_ =True if "*" in mapped_key: lowerCamelCase_ =name.split(__snake_case )[0].split('''.''' )[-2] lowerCamelCase_ =mapped_key.replace('''*''' , __snake_case ) if "weight_g" in name: lowerCamelCase_ ='''weight_g''' elif "weight_v" in name: lowerCamelCase_ ='''weight_v''' elif "weight" in name: lowerCamelCase_ ='''weight''' elif "bias" in name: lowerCamelCase_ ='''bias''' else: lowerCamelCase_ =None set_recursively(__snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) continue if not is_used: unused_weights.append(__snake_case ) logger.warning(F'''Unused weights: {unused_weights}''' ) def a_ ( __snake_case : Tuple , __snake_case : Any , __snake_case : Optional[Any] , __snake_case : List[Any] , __snake_case : Dict ) -> str: """simple docstring""" lowerCamelCase_ =full_name.split('''conv_layers.''' )[-1] lowerCamelCase_ =name.split('''.''' ) lowerCamelCase_ =int(items[0] ) lowerCamelCase_ =int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) lowerCamelCase_ =value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) lowerCamelCase_ =value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) lowerCamelCase_ =value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) lowerCamelCase_ =value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(__snake_case ) def a_ ( __snake_case : List[str] , __snake_case : List[Any] ) -> Tuple: """simple docstring""" lowerCamelCase_ =SEWConfig() if is_finetuned: lowerCamelCase_ =model.wav_encoder.wav_model.cfg else: lowerCamelCase_ =model.cfg lowerCamelCase_ =fs_config.conv_bias lowerCamelCase_ =eval(fs_config.conv_feature_layers ) lowerCamelCase_ =[x[0] for x in conv_layers] lowerCamelCase_ =[x[1] for x in conv_layers] lowerCamelCase_ =[x[2] for x in conv_layers] lowerCamelCase_ ='''gelu''' lowerCamelCase_ ='''layer''' if fs_config.extractor_mode == '''layer_norm''' else '''group''' lowerCamelCase_ =0.0 lowerCamelCase_ =fs_config.activation_fn.name lowerCamelCase_ =fs_config.encoder_embed_dim lowerCamelCase_ =0.0_2 lowerCamelCase_ =fs_config.encoder_ffn_embed_dim lowerCamelCase_ =1e-5 lowerCamelCase_ =fs_config.encoder_layerdrop lowerCamelCase_ =fs_config.encoder_attention_heads lowerCamelCase_ =fs_config.conv_pos_groups lowerCamelCase_ =fs_config.conv_pos lowerCamelCase_ =len(__snake_case ) lowerCamelCase_ =fs_config.encoder_layers lowerCamelCase_ =fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: lowerCamelCase_ =model.cfg lowerCamelCase_ =fs_config.final_dropout lowerCamelCase_ =fs_config.layerdrop lowerCamelCase_ =fs_config.activation_dropout lowerCamelCase_ =fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 lowerCamelCase_ =fs_config.attention_dropout lowerCamelCase_ =fs_config.dropout_input lowerCamelCase_ =fs_config.dropout lowerCamelCase_ =fs_config.mask_channel_length lowerCamelCase_ =fs_config.mask_channel_prob lowerCamelCase_ =fs_config.mask_length lowerCamelCase_ =fs_config.mask_prob lowerCamelCase_ ='''Wav2Vec2FeatureExtractor''' lowerCamelCase_ ='''Wav2Vec2CTCTokenizer''' return config @torch.no_grad() def a_ ( __snake_case : Union[str, Any] , __snake_case : Tuple , __snake_case : str=None , __snake_case : str=None , __snake_case : Dict=True ) -> List[str]: """simple docstring""" if is_finetuned: lowerCamelCase_, lowerCamelCase_, lowerCamelCase_ =fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: lowerCamelCase_, lowerCamelCase_, lowerCamelCase_ =fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: lowerCamelCase_ =SEWConfig.from_pretrained(__snake_case ) else: lowerCamelCase_ =convert_config(model[0] , __snake_case ) lowerCamelCase_ =model[0].eval() lowerCamelCase_ =True if config.feat_extract_norm == '''layer''' else False lowerCamelCase_ =WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=__snake_case , return_attention_mask=__snake_case , ) if is_finetuned: if dict_path: lowerCamelCase_ =Dictionary.load(__snake_case ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq lowerCamelCase_ =target_dict.pad_index lowerCamelCase_ =target_dict.bos_index lowerCamelCase_ =target_dict.pad_index lowerCamelCase_ =target_dict.bos_index lowerCamelCase_ =target_dict.eos_index lowerCamelCase_ =len(target_dict.symbols ) lowerCamelCase_ =os.path.join(__snake_case , '''vocab.json''' ) if not os.path.isdir(__snake_case ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(__snake_case ) ) return os.makedirs(__snake_case , exist_ok=__snake_case ) with open(__snake_case , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(target_dict.indices , __snake_case ) lowerCamelCase_ =WavaVecaCTCTokenizer( __snake_case , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=__snake_case , ) lowerCamelCase_ =WavaVecaProcessor(feature_extractor=__snake_case , tokenizer=__snake_case ) processor.save_pretrained(__snake_case ) lowerCamelCase_ =SEWForCTC(__snake_case ) else: lowerCamelCase_ =SEWModel(__snake_case ) feature_extractor.save_pretrained(__snake_case ) recursively_load_weights(__snake_case , __snake_case , __snake_case ) hf_model.save_pretrained(__snake_case ) if __name__ == "__main__": a_ : List[Any] = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--is_finetuned""", action="""store_true""", help="""Whether the model to convert is a fine-tuned model or not""" ) a_ : List[Any] = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
75
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class lowercase__ ( snake_case__, unittest.TestCase ): _UpperCAmelCase :Union[str, Any] = KandinskyVaaControlnetImgaImgPipeline _UpperCAmelCase :List[Any] = ["image_embeds", "negative_image_embeds", "image", "hint"] _UpperCAmelCase :List[str] = ["image_embeds", "negative_image_embeds", "image", "hint"] _UpperCAmelCase :Dict = [ "generator", "height", "width", "strength", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] _UpperCAmelCase :str = False @property def UpperCAmelCase__ ( self : Tuple ): return 32 @property def UpperCAmelCase__ ( self : List[Any] ): return 32 @property def UpperCAmelCase__ ( self : Dict ): return self.time_input_dim @property def UpperCAmelCase__ ( self : int ): return self.time_input_dim * 4 @property def UpperCAmelCase__ ( self : Optional[int] ): return 100 @property def UpperCAmelCase__ ( self : int ): torch.manual_seed(0 ) lowerCamelCase_ : Union[str, Any] ={ "in_channels": 8, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image_hint", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } lowerCamelCase_ : Union[str, Any] =UNetaDConditionModel(**snake_case__ ) return model @property def UpperCAmelCase__ ( self : Any ): return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def UpperCAmelCase__ ( self : int ): torch.manual_seed(0 ) lowerCamelCase_ : int =VQModel(**self.dummy_movq_kwargs ) return model def UpperCAmelCase__ ( self : Dict ): lowerCamelCase_ : Optional[int] =self.dummy_unet lowerCamelCase_ : Optional[Any] =self.dummy_movq lowerCamelCase_ : Optional[Any] ={ "num_train_timesteps": 1000, "beta_schedule": "linear", "beta_start": 0.00_085, "beta_end": 0.012, "clip_sample": False, "set_alpha_to_one": False, "steps_offset": 0, "prediction_type": "epsilon", "thresholding": False, } lowerCamelCase_ : Optional[Any] =DDIMScheduler(**snake_case__ ) lowerCamelCase_ : Optional[Any] ={ "unet": unet, "scheduler": scheduler, "movq": movq, } return components def UpperCAmelCase__ ( self : Union[str, Any] , snake_case__ : str , snake_case__ : str=0 ): lowerCamelCase_ : Optional[int] =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) lowerCamelCase_ : Optional[Any] =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( snake_case__ ) # create init_image lowerCamelCase_ : List[Any] =floats_tensor((1, 3, 64, 64) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) lowerCamelCase_ : Union[str, Any] =image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCamelCase_ : Tuple =Image.fromarray(np.uinta(snake_case__ ) ).convert("RGB" ).resize((256, 256) ) # create hint lowerCamelCase_ : Dict =floats_tensor((1, 3, 64, 64) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) if str(snake_case__ ).startswith("mps" ): lowerCamelCase_ : List[Any] =torch.manual_seed(snake_case__ ) else: lowerCamelCase_ : List[str] =torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) lowerCamelCase_ : Dict ={ "image": init_image, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "hint": hint, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 10, "guidance_scale": 7.0, "strength": 0.2, "output_type": "np", } return inputs def UpperCAmelCase__ ( self : Optional[Any] ): lowerCamelCase_ : Any ="cpu" lowerCamelCase_ : Dict =self.get_dummy_components() lowerCamelCase_ : Dict =self.pipeline_class(**snake_case__ ) lowerCamelCase_ : str =pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) lowerCamelCase_ : Optional[Any] =pipe(**self.get_dummy_inputs(snake_case__ ) ) lowerCamelCase_ : Dict =output.images lowerCamelCase_ : Dict =pipe( **self.get_dummy_inputs(snake_case__ ) , return_dict=snake_case__ , )[0] lowerCamelCase_ : List[str] =image[0, -3:, -3:, -1] lowerCamelCase_ : Optional[int] =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCamelCase_ : Union[str, Any] =np.array( [0.54_985_034, 0.55_509_365, 0.52_561_504, 0.5_570_494, 0.5_593_818, 0.5_263_979, 0.50_285_643, 0.5_069_846, 0.51_196_736] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class lowercase__ ( unittest.TestCase ): def UpperCAmelCase__ ( self : str ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self : int ): lowerCamelCase_ : List[Any] =load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy" ) lowerCamelCase_ : Optional[int] =load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png" ) lowerCamelCase_ : Optional[int] =init_image.resize((512, 512) ) lowerCamelCase_ : int =load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/hint_image_cat.png" ) lowerCamelCase_ : Any =torch.from_numpy(np.array(snake_case__ ) ).float() / 255.0 lowerCamelCase_ : Union[str, Any] =hint.permute(2 , 0 , 1 ).unsqueeze(0 ) lowerCamelCase_ : str ="A robot, 4k photo" lowerCamelCase_ : List[Any] =KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(snake_case__ ) lowerCamelCase_ : Any =KandinskyVaaControlnetImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-controlnet-depth" , torch_dtype=torch.floataa ) lowerCamelCase_ : List[str] =pipeline.to(snake_case__ ) pipeline.set_progress_bar_config(disable=snake_case__ ) lowerCamelCase_ : Tuple =torch.Generator(device="cpu" ).manual_seed(0 ) lowerCamelCase_ , lowerCamelCase_ : Tuple =pipe_prior( snake_case__ , image=snake_case__ , strength=0.85 , generator=snake_case__ , negative_prompt="" , ).to_tuple() lowerCamelCase_ : str =pipeline( image=snake_case__ , image_embeds=snake_case__ , negative_image_embeds=snake_case__ , hint=snake_case__ , generator=snake_case__ , num_inference_steps=100 , height=512 , width=512 , strength=0.5 , output_type="np" , ) lowerCamelCase_ : Optional[Any] =output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(snake_case__ , snake_case__ )
144
0
'''simple docstring''' import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def __magic_name__ ( A ) -> Union[str, Any]: snake_case , snake_case = image.size snake_case , snake_case = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 snake_case = image.resize((w, h) , resample=PIL_INTERPOLATION['lanczos'] ) snake_case = np.array(A ).astype(np.floataa ) / 255.0 snake_case = image[None].transpose(0 , 3 , 1 , 2 ) snake_case = torch.from_numpy(A ) return 2.0 * image - 1.0 class lowerCamelCase ( __lowerCAmelCase ): def __init__( self, lowercase_, lowercase_, lowercase_, ) -> Dict: super().__init__() self.register_modules(vqvae=lowercase_, unet=lowercase_, scheduler=lowercase_ ) @torch.no_grad() def __call__( self, lowercase_ = None, lowercase_ = 1, lowercase_ = 100, lowercase_ = 0.0, lowercase_ = None, lowercase_ = "pil", lowercase_ = True, ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowercase_, PIL.Image.Image ): snake_case = 1 elif isinstance(lowercase_, torch.Tensor ): snake_case = image.shape[0] else: raise ValueError(F'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowercase_ )}''' ) if isinstance(lowercase_, PIL.Image.Image ): snake_case = preprocess(lowercase_ ) snake_case , snake_case = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image snake_case = (batch_size, self.unet.config.in_channels // 2, height, width) snake_case = next(self.unet.parameters() ).dtype snake_case = randn_tensor(lowercase_, generator=lowercase_, device=self.device, dtype=lowercase_ ) snake_case = image.to(device=self.device, dtype=lowercase_ ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowercase_, device=self.device ) snake_case = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler snake_case = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] snake_case = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) snake_case = {} if accepts_eta: snake_case = eta for t in self.progress_bar(lowercase_ ): # concat latents and low resolution image in the channel dimension. snake_case = torch.cat([latents, image], dim=1 ) snake_case = self.scheduler.scale_model_input(lowercase_, lowercase_ ) # predict the noise residual snake_case = self.unet(lowercase_, lowercase_ ).sample # compute the previous noisy sample x_t -> x_t-1 snake_case = self.scheduler.step(lowercase_, lowercase_, lowercase_, **lowercase_ ).prev_sample # decode the image latents with the VQVAE snake_case = self.vqvae.decode(lowercase_ ).sample snake_case = torch.clamp(lowercase_, -1.0, 1.0 ) snake_case = image / 2 + 0.5 snake_case = image.cpu().permute(0, 2, 3, 1 ).numpy() if output_type == "pil": snake_case = self.numpy_to_pil(lowercase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowercase_ )
332
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import TransformeraDModel, VQDiffusionPipeline, VQDiffusionScheduler, VQModel from diffusers.pipelines.vq_diffusion.pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings from diffusers.utils import load_numpy, slow, torch_device from diffusers.utils.testing_utils import require_torch_gpu lowerCAmelCase_ = False class lowerCamelCase ( unittest.TestCase ): def _lowerCamelCase ( self ) -> Union[str, Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowerCamelCase ( self ) -> List[Any]: return 12 @property def _lowerCamelCase ( self ) -> Dict: return 12 @property def _lowerCamelCase ( self ) -> List[Any]: return 32 @property def _lowerCamelCase ( self ) -> List[Any]: torch.manual_seed(0 ) snake_case = VQModel( block_out_channels=[32, 64], in_channels=3, out_channels=3, down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'], up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'], latent_channels=3, num_vq_embeddings=self.num_embed, vq_embed_dim=3, ) return model @property def _lowerCamelCase ( self ) -> List[Any]: snake_case = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def _lowerCamelCase ( self ) -> Tuple: torch.manual_seed(0 ) snake_case = CLIPTextConfig( bos_token_id=0, eos_token_id=2, hidden_size=self.text_embedder_hidden_size, intermediate_size=37, layer_norm_eps=1E-05, num_attention_heads=4, num_hidden_layers=5, pad_token_id=1, vocab_size=1000, ) return CLIPTextModel(lowercase_ ) @property def _lowerCamelCase ( self ) -> str: torch.manual_seed(0 ) snake_case = 12 snake_case = 12 snake_case = { 'attention_bias': True, 'cross_attention_dim': 32, 'attention_head_dim': height * width, 'num_attention_heads': 1, 'num_vector_embeds': self.num_embed, 'num_embeds_ada_norm': self.num_embeds_ada_norm, 'norm_num_groups': 32, 'sample_size': width, 'activation_fn': 'geglu-approximate', } snake_case = TransformeraDModel(**lowercase_ ) return model def _lowerCamelCase ( self ) -> Tuple: snake_case = 'cpu' snake_case = self.dummy_vqvae snake_case = self.dummy_text_encoder snake_case = self.dummy_tokenizer snake_case = self.dummy_transformer snake_case = VQDiffusionScheduler(self.num_embed ) snake_case = LearnedClassifierFreeSamplingEmbeddings(learnable=lowercase_ ) snake_case = VQDiffusionPipeline( vqvae=lowercase_, text_encoder=lowercase_, tokenizer=lowercase_, transformer=lowercase_, scheduler=lowercase_, learned_classifier_free_sampling_embeddings=lowercase_, ) snake_case = pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) snake_case = 'teddy bear playing in the pool' snake_case = torch.Generator(device=lowercase_ ).manual_seed(0 ) snake_case = pipe([prompt], generator=lowercase_, num_inference_steps=2, output_type='np' ) snake_case = output.images snake_case = torch.Generator(device=lowercase_ ).manual_seed(0 ) snake_case = pipe( [prompt], generator=lowercase_, output_type='np', return_dict=lowercase_, num_inference_steps=2 )[0] snake_case = image[0, -3:, -3:, -1] snake_case = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 24, 24, 3) snake_case = np.array([0.6_551, 0.6_168, 0.5_008, 0.5_676, 0.5_659, 0.4_295, 0.6_073, 0.5_599, 0.4_992] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def _lowerCamelCase ( self ) -> Optional[Any]: snake_case = 'cpu' snake_case = self.dummy_vqvae snake_case = self.dummy_text_encoder snake_case = self.dummy_tokenizer snake_case = self.dummy_transformer snake_case = VQDiffusionScheduler(self.num_embed ) snake_case = LearnedClassifierFreeSamplingEmbeddings( learnable=lowercase_, hidden_size=self.text_embedder_hidden_size, length=tokenizer.model_max_length ) snake_case = VQDiffusionPipeline( vqvae=lowercase_, text_encoder=lowercase_, tokenizer=lowercase_, transformer=lowercase_, scheduler=lowercase_, learned_classifier_free_sampling_embeddings=lowercase_, ) snake_case = pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) snake_case = 'teddy bear playing in the pool' snake_case = torch.Generator(device=lowercase_ ).manual_seed(0 ) snake_case = pipe([prompt], generator=lowercase_, num_inference_steps=2, output_type='np' ) snake_case = output.images snake_case = torch.Generator(device=lowercase_ ).manual_seed(0 ) snake_case = pipe( [prompt], generator=lowercase_, output_type='np', return_dict=lowercase_, num_inference_steps=2 )[0] snake_case = image[0, -3:, -3:, -1] snake_case = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 24, 24, 3) snake_case = np.array([0.6_693, 0.6_075, 0.4_959, 0.5_701, 0.5_583, 0.4_333, 0.6_171, 0.5_684, 0.4_988] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2.0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class lowerCamelCase ( unittest.TestCase ): def _lowerCamelCase ( self ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self ) -> str: snake_case = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/vq_diffusion/teddy_bear_pool_classifier_free_sampling.npy' ) snake_case = VQDiffusionPipeline.from_pretrained('microsoft/vq-diffusion-ithq' ) snake_case = pipeline.to(lowercase_ ) pipeline.set_progress_bar_config(disable=lowercase_ ) # requires GPU generator for gumbel softmax # don't use GPU generator in tests though snake_case = torch.Generator(device=lowercase_ ).manual_seed(0 ) snake_case = pipeline( 'teddy bear playing in the pool', num_images_per_prompt=1, generator=lowercase_, output_type='np', ) snake_case = output.images[0] assert image.shape == (256, 256, 3) assert np.abs(expected_image - image ).max() < 2.0
332
1
import logging from transformers.configuration_utils import PretrainedConfig UpperCamelCase = logging.getLogger(__name__) class snake_case_ ( __A ): __A : List[Any] = "masked_bert" def __init__( self : Optional[int] , lowercase_ : List[str]=3_05_22 , lowercase_ : Optional[Any]=7_68 , lowercase_ : Tuple=12 , lowercase_ : str=12 , lowercase_ : List[Any]=30_72 , lowercase_ : Union[str, Any]="gelu" , lowercase_ : Dict=0.1 , lowercase_ : Union[str, Any]=0.1 , lowercase_ : List[Any]=5_12 , lowercase_ : List[str]=2 , lowercase_ : List[str]=0.02 , lowercase_ : int=1E-12 , lowercase_ : Any=0 , lowercase_ : List[Any]="topK" , lowercase_ : List[str]="constant" , lowercase_ : Dict=0.0 , **lowercase_ : Union[str, Any] , ) -> Optional[Any]: super().__init__(pad_token_id=lowercase_ , **lowercase_ ) lowercase__ : int = vocab_size lowercase__ : Optional[int] = hidden_size lowercase__ : Any = num_hidden_layers lowercase__ : Dict = num_attention_heads lowercase__ : Optional[int] = hidden_act lowercase__ : Any = intermediate_size lowercase__ : Optional[Any] = hidden_dropout_prob lowercase__ : int = attention_probs_dropout_prob lowercase__ : str = max_position_embeddings lowercase__ : str = type_vocab_size lowercase__ : Optional[int] = initializer_range lowercase__ : Optional[int] = layer_norm_eps lowercase__ : Tuple = pruning_method lowercase__ : Union[str, Any] = mask_init lowercase__ : List[str] = mask_scale
87
"""simple docstring""" from math import pi, sqrt, tan def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> float: '''simple docstring''' if side_length < 0: raise ValueError("""surface_area_cube() only accepts non-negative values""" ) return 6 * side_length**2 def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> float: '''simple docstring''' if length < 0 or breadth < 0 or height < 0: raise ValueError("""surface_area_cuboid() only accepts non-negative values""" ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> float: '''simple docstring''' if radius < 0: raise ValueError("""surface_area_sphere() only accepts non-negative values""" ) return 4 * pi * radius**2 def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> float: '''simple docstring''' if radius < 0: raise ValueError("""surface_area_hemisphere() only accepts non-negative values""" ) return 3 * pi * radius**2 def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError("""surface_area_cone() only accepts non-negative values""" ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> float: '''simple docstring''' if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( """surface_area_conical_frustum() only accepts non-negative values""" ) lowercase_ = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError("""surface_area_cylinder() only accepts non-negative values""" ) return 2 * pi * radius * (height + radius) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> float: '''simple docstring''' if torus_radius < 0 or tube_radius < 0: raise ValueError("""surface_area_torus() only accepts non-negative values""" ) if torus_radius < tube_radius: raise ValueError( """surface_area_torus() does not support spindle or self intersecting tori""" ) return 4 * pow(__lowerCAmelCase , 2 ) * torus_radius * tube_radius def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> float: '''simple docstring''' if length < 0 or width < 0: raise ValueError("""area_rectangle() only accepts non-negative values""" ) return length * width def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> float: '''simple docstring''' if side_length < 0: raise ValueError("""area_square() only accepts non-negative values""" ) return side_length**2 def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError("""area_triangle() only accepts non-negative values""" ) return (base * height) / 2 def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> float: '''simple docstring''' if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError("""area_triangle_three_sides() only accepts non-negative values""" ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError("""Given three sides do not form a triangle""" ) lowercase_ = (sidea + sidea + sidea) / 2 lowercase_ = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError("""area_parallelogram() only accepts non-negative values""" ) return base * height def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> float: '''simple docstring''' if basea < 0 or basea < 0 or height < 0: raise ValueError("""area_trapezium() only accepts non-negative values""" ) return 1 / 2 * (basea + basea) * height def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> float: '''simple docstring''' if radius < 0: raise ValueError("""area_circle() only accepts non-negative values""" ) return pi * radius**2 def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> float: '''simple docstring''' if radius_x < 0 or radius_y < 0: raise ValueError("""area_ellipse() only accepts non-negative values""" ) return pi * radius_x * radius_y def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> float: '''simple docstring''' if diagonal_a < 0 or diagonal_a < 0: raise ValueError("""area_rhombus() only accepts non-negative values""" ) return 1 / 2 * diagonal_a * diagonal_a def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> float: '''simple docstring''' if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or sides < 3: raise ValueError( """area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides""" ) elif length < 0: raise ValueError( """area_reg_polygon() only accepts non-negative values as \ length of a side""" ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print("[DEMO] Areas of various geometric shapes: \n") print(F"Rectangle: {area_rectangle(10, 20) = }") print(F"Square: {area_square(10) = }") print(F"Triangle: {area_triangle(10, 10) = }") print(F"Triangle: {area_triangle_three_sides(5, 12, 13) = }") print(F"Parallelogram: {area_parallelogram(10, 20) = }") print(F"Rhombus: {area_rhombus(10, 20) = }") print(F"Trapezium: {area_trapezium(10, 20, 30) = }") print(F"Circle: {area_circle(20) = }") print(F"Ellipse: {area_ellipse(10, 20) = }") print("\nSurface Areas of various geometric shapes: \n") print(F"Cube: {surface_area_cube(20) = }") print(F"Cuboid: {surface_area_cuboid(10, 20, 30) = }") print(F"Sphere: {surface_area_sphere(20) = }") print(F"Hemisphere: {surface_area_hemisphere(20) = }") print(F"Cone: {surface_area_cone(10, 20) = }") print(F"Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }") print(F"Cylinder: {surface_area_cylinder(10, 20) = }") print(F"Torus: {surface_area_torus(20, 10) = }") print(F"Equilateral Triangle: {area_reg_polygon(3, 10) = }") print(F"Square: {area_reg_polygon(4, 10) = }") print(F"Reqular Pentagon: {area_reg_polygon(5, 10) = }")
136
0
"""simple docstring""" def _a ( _snake_case ): """simple docstring""" if bit_count < 0: raise ValueError("""The given input must be positive""" ) # get the generated string sequence UpperCAmelCase = gray_code_sequence_string(_snake_case ) # # convert them to integers for i in range(len(_snake_case ) ): UpperCAmelCase = int(sequence[i] , 2 ) return sequence def _a ( _snake_case ): """simple docstring""" if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] UpperCAmelCase = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits UpperCAmelCase = gray_code_sequence_string(bit_count - 1 ) UpperCAmelCase = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): UpperCAmelCase = """0""" + smaller_sequence[i] sequence.append(_snake_case ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): UpperCAmelCase = """1""" + smaller_sequence[i] sequence.append(_snake_case ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
234
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL _UpperCamelCase = logging.get_logger(__name__) def _a ( _snake_case ): """simple docstring""" if isinstance(_snake_case , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(_snake_case , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(_snake_case ): return [[videos]] raise ValueError(F'''Could not make batched video from {videos}''' ) class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = ['''pixel_values'''] def __init__( self ,A = True ,A = None ,A = PILImageResampling.BILINEAR ,A = True ,A = None ,A = True ,A = 1 / 255 ,A = True ,A = True ,A = None ,A = None ,**A ,): super().__init__(**A ) UpperCAmelCase = size if size is not None else {"""shortest_edge""": 256} UpperCAmelCase = get_size_dict(A ,default_to_square=A ) UpperCAmelCase = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} UpperCAmelCase = get_size_dict(A ,param_name="""crop_size""" ) UpperCAmelCase = do_resize UpperCAmelCase = size UpperCAmelCase = do_center_crop UpperCAmelCase = crop_size UpperCAmelCase = resample UpperCAmelCase = do_rescale UpperCAmelCase = rescale_factor UpperCAmelCase = offset UpperCAmelCase = do_normalize UpperCAmelCase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase = image_std if image_std is not None else IMAGENET_STANDARD_STD def _UpperCamelCase ( self ,A ,A ,A = PILImageResampling.BILINEAR ,A = None ,**A ,): UpperCAmelCase = get_size_dict(A ,default_to_square=A ) if "shortest_edge" in size: UpperCAmelCase = get_resize_output_image_size(A ,size["""shortest_edge"""] ,default_to_square=A ) elif "height" in size and "width" in size: UpperCAmelCase = (size["""height"""], size["""width"""]) else: raise ValueError(F'''Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}''' ) return resize(A ,size=A ,resample=A ,data_format=A ,**A ) def _UpperCamelCase ( self ,A ,A ,A = None ,**A ,): UpperCAmelCase = get_size_dict(A ) if "height" not in size or "width" not in size: raise ValueError(F'''Size must have \'height\' and \'width\' as keys. Got {size.keys()}''' ) return center_crop(A ,size=(size["""height"""], size["""width"""]) ,data_format=A ,**A ) def _UpperCamelCase ( self ,A ,A ,A = True ,A = None ,**A ,): UpperCAmelCase = image.astype(np.floataa ) if offset: UpperCAmelCase = image - (scale / 2) return rescale(A ,scale=A ,data_format=A ,**A ) def _UpperCamelCase ( self ,A ,A ,A ,A = None ,**A ,): return normalize(A ,mean=A ,std=A ,data_format=A ,**A ) def _UpperCamelCase ( self ,A ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = ChannelDimension.FIRST ,): if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) if offset and not do_rescale: raise ValueError("""For offset, do_rescale must also be set to True.""" ) # All transformations expect numpy arrays. UpperCAmelCase = to_numpy_array(A ) if do_resize: UpperCAmelCase = self.resize(image=A ,size=A ,resample=A ) if do_center_crop: UpperCAmelCase = self.center_crop(A ,size=A ) if do_rescale: UpperCAmelCase = self.rescale(image=A ,scale=A ,offset=A ) if do_normalize: UpperCAmelCase = self.normalize(image=A ,mean=A ,std=A ) UpperCAmelCase = to_channel_dimension_format(A ,A ) return image def _UpperCamelCase ( self ,A ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = ChannelDimension.FIRST ,**A ,): UpperCAmelCase = do_resize if do_resize is not None else self.do_resize UpperCAmelCase = resample if resample is not None else self.resample UpperCAmelCase = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase = offset if offset is not None else self.offset UpperCAmelCase = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase = image_mean if image_mean is not None else self.image_mean UpperCAmelCase = image_std if image_std is not None else self.image_std UpperCAmelCase = size if size is not None else self.size UpperCAmelCase = get_size_dict(A ,default_to_square=A ) UpperCAmelCase = crop_size if crop_size is not None else self.crop_size UpperCAmelCase = get_size_dict(A ,param_name="""crop_size""" ) if not valid_images(A ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) UpperCAmelCase = make_batched(A ) UpperCAmelCase = [ [ self._preprocess_image( image=A ,do_resize=A ,size=A ,resample=A ,do_center_crop=A ,crop_size=A ,do_rescale=A ,rescale_factor=A ,offset=A ,do_normalize=A ,image_mean=A ,image_std=A ,data_format=A ,) for img in video ] for video in videos ] UpperCAmelCase = {"""pixel_values""": videos} return BatchFeature(data=A ,tensor_type=A )
234
1
import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel _snake_case = False _snake_case = True _snake_case = False if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') _snake_case = parser.parse_args() _snake_case = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } _snake_case = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } _snake_case = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: _snake_case = reader.read() _snake_case = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): _snake_case = UNetaDModel(**config) else: _snake_case = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel _snake_case = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) _snake_case = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: _snake_case = config[key] del config[key] _snake_case = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] _snake_case = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: _snake_case = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) _snake_case = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue _snake_case = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: _snake_case = param_value _snake_case = True if not has_changed: _snake_case = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
157
import random import timeit from functools import wraps from typing import Callable, Optional from ..configuration_utils import PretrainedConfig from ..models.auto.modeling_tf_auto import TF_MODEL_MAPPING, TF_MODEL_WITH_LM_HEAD_MAPPING from ..utils import is_pyanvml_available, is_tf_available, logging from .benchmark_utils import ( Benchmark, Memory, MemorySummary, measure_peak_memory_cpu, start_memory_tracing, stop_memory_tracing, ) if is_tf_available(): import tensorflow as tf from tensorflow.python.framework.errors_impl import ResourceExhaustedError from .benchmark_args_tf import TensorFlowBenchmarkArguments if is_pyanvml_available(): import pyanvml.pyanvml as nvml _snake_case = logging.get_logger(__name__) def _UpperCamelCase ( snake_case__, snake_case__ ) -> List[str]: def run_func(snake_case__ ): @wraps(snake_case__ ) def run_in_eager_mode(*snake_case__, **snake_case__ ): return func(*snake_case__, **snake_case__ ) @wraps(snake_case__ ) @tf.function(experimental_compile=snake_case__ ) def run_in_graph_mode(*snake_case__, **snake_case__ ): return func(*snake_case__, **snake_case__ ) if do_eager_mode is True: if use_xla is not False: raise ValueError( "Cannot run model in XLA, if `args.eager_mode` is set to `True`. Please set `args.eager_mode=False`." ) return run_in_eager_mode else: return run_in_graph_mode return run_func def _UpperCamelCase ( snake_case__, snake_case__, snake_case__ ) -> ["tf.Tensor"]: __UpperCAmelCase : str = random.Random() __UpperCAmelCase : str = [rng.randint(0, vocab_size - 1 ) for i in range(batch_size * sequence_length )] return tf.constant(snake_case__, shape=(batch_size, sequence_length), dtype=tf.intaa ) class _snake_case ( _lowercase ): lowerCamelCase__: TensorFlowBenchmarkArguments lowerCamelCase__: PretrainedConfig lowerCamelCase__: str = "TensorFlow" @property def _lowerCamelCase ( self: int ) -> Any: return tf.__version__ def _lowerCamelCase ( self: Dict , __lowerCamelCase: str , __lowerCamelCase: int , __lowerCamelCase: int ) -> float: # initialize GPU on separate process __UpperCAmelCase : List[Any] = self.args.strategy if strategy is None: raise ValueError("A device strategy has to be initialized before using TensorFlow." ) __UpperCAmelCase : int = self._prepare_inference_func(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) return self._measure_speed(_inference ) def _lowerCamelCase ( self: Tuple , __lowerCamelCase: str , __lowerCamelCase: int , __lowerCamelCase: int ) -> float: __UpperCAmelCase : Union[str, Any] = self.args.strategy if strategy is None: raise ValueError("A device strategy has to be initialized before using TensorFlow." ) __UpperCAmelCase : Dict = self._prepare_train_func(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) return self._measure_speed(_train ) def _lowerCamelCase ( self: Union[str, Any] , __lowerCamelCase: str , __lowerCamelCase: int , __lowerCamelCase: int ) -> [Memory, Optional[MemorySummary]]: # initialize GPU on separate process if self.args.is_gpu: tf.config.experimental.set_memory_growth(self.args.gpu_list[self.args.device_idx] , __lowerCamelCase ) __UpperCAmelCase : List[str] = self.args.strategy if strategy is None: raise ValueError("A device strategy has to be initialized before using TensorFlow." ) __UpperCAmelCase : int = self._prepare_inference_func(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) return self._measure_memory(_inference ) def _lowerCamelCase ( self: str , __lowerCamelCase: str , __lowerCamelCase: int , __lowerCamelCase: int ) -> [Memory, Optional[MemorySummary]]: if self.args.is_gpu: tf.config.experimental.set_memory_growth(self.args.gpu_list[self.args.device_idx] , __lowerCamelCase ) __UpperCAmelCase : int = self.args.strategy if strategy is None: raise ValueError("A device strategy has to be initialized before using TensorFlow." ) __UpperCAmelCase : int = self._prepare_train_func(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) return self._measure_memory(_train ) def _lowerCamelCase ( self: int , __lowerCamelCase: str , __lowerCamelCase: int , __lowerCamelCase: int ) -> Callable[[], None]: __UpperCAmelCase : Union[str, Any] = self.config_dict[model_name] if self.args.fpaa: raise NotImplementedError("Mixed precision is currently not supported." ) __UpperCAmelCase : int = ( hasattr(__lowerCamelCase , "architectures" ) and isinstance(config.architectures , __lowerCamelCase ) and len(config.architectures ) > 0 ) if not self.args.only_pretrain_model and has_model_class_in_config: try: __UpperCAmelCase : int = "TF" + config.architectures[0] # prepend 'TF' for tensorflow model __UpperCAmelCase : Dict = __import__("transformers" , fromlist=[model_class] ) __UpperCAmelCase : str = getattr(__lowerCamelCase , __lowerCamelCase ) __UpperCAmelCase : Optional[Any] = model_cls(__lowerCamelCase ) except ImportError: raise ImportError( f'''{model_class} does not exist. If you just want to test the pretrained model, you might want to''' " set `--only_pretrain_model` or `args.only_pretrain_model=True`." ) else: __UpperCAmelCase : int = TF_MODEL_MAPPING[config.__class__](__lowerCamelCase ) # encoder-decoder has vocab size saved differently __UpperCAmelCase : List[str] = config.vocab_size if hasattr(__lowerCamelCase , "vocab_size" ) else config.encoder.vocab_size __UpperCAmelCase : Dict = random_input_ids(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_decoder_forward(): return model(__lowerCamelCase , decoder_input_ids=__lowerCamelCase , training=__lowerCamelCase ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_forward(): return model(__lowerCamelCase , training=__lowerCamelCase ) __UpperCAmelCase : int = encoder_decoder_forward if config.is_encoder_decoder else encoder_forward return _inference def _lowerCamelCase ( self: List[str] , __lowerCamelCase: str , __lowerCamelCase: int , __lowerCamelCase: int ) -> Callable[[], None]: __UpperCAmelCase : Any = self.config_dict[model_name] if self.args.eager_mode is not False: raise ValueError("Training cannot be done in eager mode. Please make sure that `args.eager_mode = False`." ) if self.args.fpaa: raise NotImplementedError("Mixed precision is currently not supported." ) __UpperCAmelCase : Tuple = ( hasattr(__lowerCamelCase , "architectures" ) and isinstance(config.architectures , __lowerCamelCase ) and len(config.architectures ) > 0 ) if not self.args.only_pretrain_model and has_model_class_in_config: try: __UpperCAmelCase : Dict = "TF" + config.architectures[0] # prepend 'TF' for tensorflow model __UpperCAmelCase : Optional[Any] = __import__("transformers" , fromlist=[model_class] ) __UpperCAmelCase : int = getattr(__lowerCamelCase , __lowerCamelCase ) __UpperCAmelCase : Any = model_cls(__lowerCamelCase ) except ImportError: raise ImportError( f'''{model_class} does not exist. If you just want to test the pretrained model, you might want to''' " set `--only_pretrain_model` or `args.only_pretrain_model=True`." ) else: __UpperCAmelCase : Union[str, Any] = TF_MODEL_WITH_LM_HEAD_MAPPING[config.__class__](__lowerCamelCase ) # encoder-decoder has vocab size saved differently __UpperCAmelCase : List[Any] = config.vocab_size if hasattr(__lowerCamelCase , "vocab_size" ) else config.encoder.vocab_size __UpperCAmelCase : Dict = random_input_ids(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_decoder_train(): __UpperCAmelCase : List[Any] = model(__lowerCamelCase , decoder_input_ids=__lowerCamelCase , labels=__lowerCamelCase , training=__lowerCamelCase )[0] __UpperCAmelCase : Optional[Any] = tf.gradients(__lowerCamelCase , model.trainable_variables ) return gradients @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_train(): __UpperCAmelCase : Optional[Any] = model(__lowerCamelCase , labels=__lowerCamelCase , training=__lowerCamelCase )[0] __UpperCAmelCase : List[Any] = tf.gradients(__lowerCamelCase , model.trainable_variables ) return gradients __UpperCAmelCase : Optional[int] = encoder_decoder_train if config.is_encoder_decoder else encoder_train return _train def _lowerCamelCase ( self: List[Any] , __lowerCamelCase: Any ) -> float: with self.args.strategy.scope(): try: if self.args.is_tpu or self.args.use_xla: # run additional 10 times to stabilize compilation for tpu logger.info("Do inference on TPU. Running model 5 times to stabilize compilation" ) timeit.repeat(__lowerCamelCase , repeat=1 , number=5 ) # as written in https://docs.python.org/2/library/timeit.html#timeit.Timer.repeat, min should be taken rather than the average __UpperCAmelCase : List[str] = timeit.repeat( __lowerCamelCase , repeat=self.args.repeat , number=10 , ) return min(__lowerCamelCase ) / 10.0 except ResourceExhaustedError as e: self.print_fn(f'''Doesn\'t fit on GPU. {e}''' ) def _lowerCamelCase ( self: Optional[int] , __lowerCamelCase: Callable[[], None] ) -> [Memory, MemorySummary]: logger.info( "Note that TensorFlow allocates more memory than " "it might need to speed up computation. " "The memory reported here corresponds to the memory " "reported by `nvidia-smi`, which can vary depending " "on total available memory on the GPU that is used." ) with self.args.strategy.scope(): try: if self.args.trace_memory_line_by_line: if not self.args.eager_mode: raise ValueError( "`args.eager_mode` is set to `False`. Make sure to run model in eager mode to measure memory" " consumption line by line." ) __UpperCAmelCase : Union[str, Any] = start_memory_tracing("transformers" ) if self.args.is_tpu: # tpu raise NotImplementedError( "Memory Benchmarking is currently not implemented for TPU. Please disable memory benchmarking" " with `args.memory=False`" ) elif self.args.is_gpu: # gpu if not is_pyanvml_available(): logger.warning( "py3nvml not installed, we won't log GPU memory usage. " "Install py3nvml (pip install py3nvml) to log information about GPU." ) __UpperCAmelCase : Union[str, Any] = "N/A" else: logger.info( "Measuring total GPU usage on GPU device. Make sure to not have additional processes" " running on the same GPU." ) # init nvml nvml.nvmlInit() func() __UpperCAmelCase : str = nvml.nvmlDeviceGetHandleByIndex(self.args.device_idx ) __UpperCAmelCase : List[Any] = nvml.nvmlDeviceGetMemoryInfo(__lowerCamelCase ) __UpperCAmelCase : List[Any] = meminfo.used __UpperCAmelCase : List[Any] = Memory(__lowerCamelCase ) # shutdown nvml nvml.nvmlShutdown() else: # cpu if self.args.trace_memory_line_by_line: logger.info( "When enabling line by line tracing, the max peak memory for CPU is inaccurate in" " TensorFlow." ) __UpperCAmelCase : Tuple = None else: __UpperCAmelCase : str = measure_peak_memory_cpu(__lowerCamelCase ) __UpperCAmelCase : Optional[int] = Memory(__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else memory_bytes if self.args.trace_memory_line_by_line: __UpperCAmelCase : str = stop_memory_tracing(__lowerCamelCase ) if memory is None: __UpperCAmelCase : Tuple = summary.total else: __UpperCAmelCase : Union[str, Any] = None return memory, summary except ResourceExhaustedError as e: self.print_fn(f'''Doesn\'t fit on GPU. {e}''' ) return "N/A", None
157
1
import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print('''Googling.....''') __A = '''https://www.google.com/search?q=''' + ''' '''.join(sys.argv[1:]) __A = requests.get(url, headers={'''UserAgent''': UserAgent().random}) # res.raise_for_status() with open('''project1a.html''', '''wb''') as out_file: # only for knowing the class for data in res.iter_content(1_00_00): out_file.write(data) __A = BeautifulSoup(res.text, '''html.parser''') __A = list(soup.select('''.eZt8xd'''))[:5] print(len(links)) for link in links: if link.text == "Maps": webbrowser.open(link.get('''href''')) else: webbrowser.open(f'''https://google.com{link.get("href")}''')
278
__A = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def snake_case_(_UpperCamelCase ) -> bytes: """simple docstring""" if not isinstance(_UpperCamelCase , _UpperCamelCase ): _snake_case = F"""a bytes-like object is required, not '{data.__class__.__name__}'""" raise TypeError(_UpperCamelCase ) _snake_case = ''''''.join(bin(_UpperCamelCase )[2:].zfill(8 ) for byte in data ) _snake_case = len(_UpperCamelCase ) % 6 != 0 if padding_needed: # The padding that will be added later _snake_case = b'''=''' * ((6 - len(_UpperCamelCase ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(_UpperCamelCase ) % 6) else: _snake_case = b'''''' # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(_UpperCamelCase ) , 6 ) ).encode() + padding ) def snake_case_(_UpperCamelCase ) -> bytes: """simple docstring""" if not isinstance(_UpperCamelCase , _UpperCamelCase ) and not isinstance(_UpperCamelCase , _UpperCamelCase ): _snake_case = ( '''argument should be a bytes-like object or ASCII string, ''' F"""not '{encoded_data.__class__.__name__}'""" ) raise TypeError(_UpperCamelCase ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(_UpperCamelCase , _UpperCamelCase ): try: _snake_case = encoded_data.decode('''utf-8''' ) except UnicodeDecodeError: raise ValueError('''base64 encoded data should only contain ASCII characters''' ) _snake_case = encoded_data.count('''=''' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(_UpperCamelCase ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one _snake_case = encoded_data[:-padding] _snake_case = ''''''.join( bin(B64_CHARSET.index(_UpperCamelCase ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: _snake_case = ''''''.join( bin(B64_CHARSET.index(_UpperCamelCase ) )[2:].zfill(6 ) for char in encoded_data ) _snake_case = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(_UpperCamelCase ) , 8 ) ] return bytes(_UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
278
1
'''simple docstring''' import os import sys import unittest a_ : Optional[Any] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import get_test_info # noqa: E402 from get_test_info import ( # noqa: E402 get_model_to_test_mapping, get_model_to_tester_mapping, get_test_to_tester_mapping, ) a_ : List[str] = os.path.join("""tests""", """models""", """bert""", """test_modeling_bert.py""") a_ : Tuple = os.path.join("""tests""", """models""", """blip""", """test_modeling_blip.py""") class snake_case ( unittest.TestCase ): """simple docstring""" def snake_case ( self ): """simple docstring""" lowerCamelCase_ = get_test_to_tester_mapping(UpperCamelCase ) lowerCamelCase_ = get_test_to_tester_mapping(UpperCamelCase ) lowerCamelCase_ = {"BertModelTest": "BertModelTester"} lowerCamelCase_ = { "BlipModelTest": "BlipModelTester", "BlipTextImageModelTest": "BlipTextImageModelsModelTester", "BlipTextModelTest": "BlipTextModelTester", "BlipTextRetrievalModelTest": "BlipTextRetrievalModelTester", "BlipVQAModelTest": "BlipVQAModelTester", "BlipVisionModelTest": "BlipVisionModelTester", } self.assertEqual(get_test_info.to_json(UpperCamelCase ) , UpperCamelCase ) self.assertEqual(get_test_info.to_json(UpperCamelCase ) , UpperCamelCase ) def snake_case ( self ): """simple docstring""" lowerCamelCase_ = get_model_to_test_mapping(UpperCamelCase ) lowerCamelCase_ = get_model_to_test_mapping(UpperCamelCase ) lowerCamelCase_ = { "BertForMaskedLM": ["BertModelTest"], "BertForMultipleChoice": ["BertModelTest"], "BertForNextSentencePrediction": ["BertModelTest"], "BertForPreTraining": ["BertModelTest"], "BertForQuestionAnswering": ["BertModelTest"], "BertForSequenceClassification": ["BertModelTest"], "BertForTokenClassification": ["BertModelTest"], "BertLMHeadModel": ["BertModelTest"], "BertModel": ["BertModelTest"], } lowerCamelCase_ = { "BlipForConditionalGeneration": ["BlipTextImageModelTest"], "BlipForImageTextRetrieval": ["BlipTextRetrievalModelTest"], "BlipForQuestionAnswering": ["BlipVQAModelTest"], "BlipModel": ["BlipModelTest"], "BlipTextModel": ["BlipTextModelTest"], "BlipVisionModel": ["BlipVisionModelTest"], } self.assertEqual(get_test_info.to_json(UpperCamelCase ) , UpperCamelCase ) self.assertEqual(get_test_info.to_json(UpperCamelCase ) , UpperCamelCase ) def snake_case ( self ): """simple docstring""" lowerCamelCase_ = get_model_to_tester_mapping(UpperCamelCase ) lowerCamelCase_ = get_model_to_tester_mapping(UpperCamelCase ) lowerCamelCase_ = { "BertForMaskedLM": ["BertModelTester"], "BertForMultipleChoice": ["BertModelTester"], "BertForNextSentencePrediction": ["BertModelTester"], "BertForPreTraining": ["BertModelTester"], "BertForQuestionAnswering": ["BertModelTester"], "BertForSequenceClassification": ["BertModelTester"], "BertForTokenClassification": ["BertModelTester"], "BertLMHeadModel": ["BertModelTester"], "BertModel": ["BertModelTester"], } lowerCamelCase_ = { "BlipForConditionalGeneration": ["BlipTextImageModelsModelTester"], "BlipForImageTextRetrieval": ["BlipTextRetrievalModelTester"], "BlipForQuestionAnswering": ["BlipVQAModelTester"], "BlipModel": ["BlipModelTester"], "BlipTextModel": ["BlipTextModelTester"], "BlipVisionModel": ["BlipVisionModelTester"], } self.assertEqual(get_test_info.to_json(UpperCamelCase ) , UpperCamelCase ) self.assertEqual(get_test_info.to_json(UpperCamelCase ) , UpperCamelCase )
55
'''simple docstring''' import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging a_ : Dict = logging.get_logger(__name__) class snake_case ( lowercase ): """simple docstring""" def __init__( self , UpperCamelCase ): """simple docstring""" super().__init__() lowerCamelCase_ = nn.ModuleList(UpperCamelCase ) def snake_case ( self , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = None , UpperCamelCase = None , UpperCamelCase = None , UpperCamelCase = None , UpperCamelCase = False , UpperCamelCase = True , ): """simple docstring""" for i, (image, scale, controlnet) in enumerate(zip(UpperCamelCase , UpperCamelCase , self.nets ) ): lowerCamelCase_ ,lowerCamelCase_ = controlnet( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , ) # merge samples if i == 0: lowerCamelCase_ ,lowerCamelCase_ = down_samples, mid_sample else: lowerCamelCase_ = [ samples_prev + samples_curr for samples_prev, samples_curr in zip(UpperCamelCase , UpperCamelCase ) ] mid_block_res_sample += mid_sample return down_block_res_samples, mid_block_res_sample def snake_case ( self , UpperCamelCase , UpperCamelCase = True , UpperCamelCase = None , UpperCamelCase = False , UpperCamelCase = None , ): """simple docstring""" lowerCamelCase_ = 0 lowerCamelCase_ = save_directory for controlnet in self.nets: controlnet.save_pretrained( UpperCamelCase , is_main_process=UpperCamelCase , save_function=UpperCamelCase , safe_serialization=UpperCamelCase , variant=UpperCamelCase , ) idx += 1 lowerCamelCase_ = model_path_to_save + f'''_{idx}''' @classmethod def snake_case ( cls , UpperCamelCase , **UpperCamelCase ): """simple docstring""" lowerCamelCase_ = 0 lowerCamelCase_ = [] # load controlnet and append to list until no controlnet directory exists anymore # first controlnet has to be saved under `./mydirectory/controlnet` to be compliant with `DiffusionPipeline.from_prertained` # second, third, ... controlnets have to be saved under `./mydirectory/controlnet_1`, `./mydirectory/controlnet_2`, ... lowerCamelCase_ = pretrained_model_path while os.path.isdir(UpperCamelCase ): lowerCamelCase_ = ControlNetModel.from_pretrained(UpperCamelCase , **UpperCamelCase ) controlnets.append(UpperCamelCase ) idx += 1 lowerCamelCase_ = pretrained_model_path + f'''_{idx}''' logger.info(f'''{len(UpperCamelCase )} controlnets loaded from {pretrained_model_path}.''' ) if len(UpperCamelCase ) == 0: raise ValueError( f'''No ControlNets found under {os.path.dirname(UpperCamelCase )}. Expected at least {pretrained_model_path + "_0"}.''' ) return cls(UpperCamelCase )
55
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import MvpTokenizer _lowercase : str = logging.get_logger(__name__) _lowercase : Optional[int] = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} # See all MVP models at https://huggingface.co/models?filter=mvp _lowercase : Optional[int] = { "vocab_file": { "RUCAIBox/mvp": "https://huggingface.co/RUCAIBox/mvp/resolve/main/vocab.json", }, "added_tokens.json": { "RUCAIBox/mvp": "https://huggingface.co/RUCAIBox/mvp/resolve/main/added_tokens.json", }, "merges_file": { "RUCAIBox/mvp": "https://huggingface.co/RUCAIBox/mvp/resolve/main/merges.txt", }, "tokenizer_file": { "RUCAIBox/mvp": "https://huggingface.co/RUCAIBox/mvp/resolve/main/tokenizer.json", }, } _lowercase : str = { "RUCAIBox/mvp": 1_0_2_4, } class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' _a = VOCAB_FILES_NAMES _a = PRETRAINED_VOCAB_FILES_MAP _a = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _a = ['input_ids', 'attention_mask'] _a = MvpTokenizer def __init__( self : Optional[Any], lowerCamelCase : Any=None, lowerCamelCase : List[str]=None, lowerCamelCase : int=None, lowerCamelCase : int="replace", lowerCamelCase : List[Any]="<s>", lowerCamelCase : Optional[Any]="</s>", lowerCamelCase : Dict="</s>", lowerCamelCase : Union[str, Any]="<s>", lowerCamelCase : int="<unk>", lowerCamelCase : List[str]="<pad>", lowerCamelCase : List[str]="<mask>", lowerCamelCase : Union[str, Any]=False, lowerCamelCase : Dict=True, **lowerCamelCase : Optional[Any], )-> Optional[Any]: super().__init__( lowerCamelCase, lowerCamelCase, tokenizer_file=lowerCamelCase, errors=lowerCamelCase, bos_token=lowerCamelCase, eos_token=lowerCamelCase, sep_token=lowerCamelCase, cls_token=lowerCamelCase, unk_token=lowerCamelCase, pad_token=lowerCamelCase, mask_token=lowerCamelCase, add_prefix_space=lowerCamelCase, trim_offsets=lowerCamelCase, **lowerCamelCase, ) lowerCamelCase__ : List[Any] =json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''', lowerCamelCase ) != add_prefix_space: lowerCamelCase__ : Optional[Any] =getattr(lowerCamelCase, pre_tok_state.pop('''type''' ) ) lowerCamelCase__ : str =add_prefix_space lowerCamelCase__ : List[Any] =pre_tok_class(**lowerCamelCase ) lowerCamelCase__ : str =add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowerCamelCase__ : int ='''post_processor''' lowerCamelCase__ : Optional[Any] =getattr(self.backend_tokenizer, lowerCamelCase, lowerCamelCase ) if tokenizer_component_instance: lowerCamelCase__ : int =json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowerCamelCase__ : Optional[int] =tuple(state['''sep'''] ) if "cls" in state: lowerCamelCase__ : str =tuple(state['''cls'''] ) lowerCamelCase__ : List[Any] =False if state.get('''add_prefix_space''', lowerCamelCase ) != add_prefix_space: lowerCamelCase__ : Optional[Any] =add_prefix_space lowerCamelCase__ : Dict =True if state.get('''trim_offsets''', lowerCamelCase ) != trim_offsets: lowerCamelCase__ : List[str] =trim_offsets lowerCamelCase__ : List[Any] =True if changes_to_apply: lowerCamelCase__ : Tuple =getattr(lowerCamelCase, state.pop('''type''' ) ) lowerCamelCase__ : Optional[Any] =component_class(**lowerCamelCase ) setattr(self.backend_tokenizer, lowerCamelCase, lowerCamelCase ) @property def snake_case ( self : Tuple )-> str: if self._mask_token is None: if self.verbose: logger.error('''Using mask_token, but it is not set yet.''' ) return None return str(self._mask_token ) @mask_token.setter def snake_case ( self : Tuple, lowerCamelCase : List[Any] )-> List[str]: lowerCamelCase__ : Dict =AddedToken(lowerCamelCase, lstrip=lowerCamelCase, rstrip=lowerCamelCase ) if isinstance(lowerCamelCase, lowerCamelCase ) else value lowerCamelCase__ : Union[str, Any] =value def snake_case ( self : Tuple, *lowerCamelCase : Optional[Any], **lowerCamelCase : int )-> BatchEncoding: lowerCamelCase__ : Any =kwargs.get('''is_split_into_words''', lowerCamelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' '''to use it with pretokenized inputs.''' ) return super()._batch_encode_plus(*lowerCamelCase, **lowerCamelCase ) def snake_case ( self : Any, *lowerCamelCase : str, **lowerCamelCase : Dict )-> BatchEncoding: lowerCamelCase__ : Dict =kwargs.get('''is_split_into_words''', lowerCamelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' '''to use it with pretokenized inputs.''' ) return super()._encode_plus(*lowerCamelCase, **lowerCamelCase ) def snake_case ( self : List[str], lowerCamelCase : str, lowerCamelCase : Optional[str] = None )-> Tuple[str]: lowerCamelCase__ : str =self._tokenizer.model.save(lowerCamelCase, name=lowerCamelCase ) return tuple(lowerCamelCase ) def snake_case ( self : Optional[int], lowerCamelCase : List[Any], lowerCamelCase : List[str]=None )-> int: lowerCamelCase__ : Optional[Any] =[self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def snake_case ( self : List[Any], lowerCamelCase : List[int], lowerCamelCase : Optional[List[int]] = None )-> List[int]: lowerCamelCase__ : Optional[Any] =[self.sep_token_id] lowerCamelCase__ : int =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
272
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings _lowercase : Tuple = r"\n [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and\n can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information.\n\n Args:\n title_sep (`str`, *optional*, defaults to `\" / \"`):\n Separator inserted between the title and the text of the retrieved document when calling [`RagRetriever`].\n doc_sep (`str`, *optional*, defaults to `\" // \"`):\n Separator inserted between the text of the retrieved document and the original input when calling\n [`RagRetriever`].\n n_docs (`int`, *optional*, defaults to 5):\n Number of documents to retrieve.\n max_combined_length (`int`, *optional*, defaults to 300):\n Max length of contextualized input returned by [`~RagRetriever.__call__`].\n retrieval_vector_size (`int`, *optional*, defaults to 768):\n Dimensionality of the document embeddings indexed by [`RagRetriever`].\n retrieval_batch_size (`int`, *optional*, defaults to 8):\n Retrieval batch size, defined as the number of queries issues concurrently to the faiss index encapsulated\n [`RagRetriever`].\n dataset (`str`, *optional*, defaults to `\"wiki_dpr\"`):\n A dataset identifier of the indexed dataset in HuggingFace Datasets (list all available datasets and ids\n using `datasets.list_datasets()`).\n dataset_split (`str`, *optional*, defaults to `\"train\"`)\n Which split of the `dataset` to load.\n index_name (`str`, *optional*, defaults to `\"compressed\"`)\n The index name of the index associated with the `dataset`. One can choose between `\"legacy\"`, `\"exact\"` and\n `\"compressed\"`.\n index_path (`str`, *optional*)\n The path to the serialized faiss index on disk.\n passages_path (`str`, *optional*):\n A path to text passages compatible with the faiss index. Required if using\n [`~models.rag.retrieval_rag.LegacyIndex`]\n use_dummy_dataset (`bool`, *optional*, defaults to `False`)\n Whether to load a \"dummy\" variant of the dataset specified by `dataset`.\n label_smoothing (`float`, *optional*, defaults to 0.0):\n Only relevant if `return_loss` is set to `True`. Controls the `epsilon` parameter value for label smoothing\n in the loss calculation. If set to 0, no label smoothing is performed.\n do_marginalize (`bool`, *optional*, defaults to `False`):\n If `True`, the logits are marginalized over all documents by making use of\n `torch.nn.functional.log_softmax`.\n reduce_loss (`bool`, *optional*, defaults to `False`):\n Whether or not to reduce the NLL loss using the `torch.Tensor.sum` operation.\n do_deduplication (`bool`, *optional*, defaults to `True`):\n Whether or not to deduplicate the generations from different context documents for a given input. Has to be\n set to `False` if used while training with distributed backend.\n exclude_bos_score (`bool`, *optional*, defaults to `False`):\n Whether or not to disregard the BOS token when computing the loss.\n output_retrieved(`bool`, *optional*, defaults to `False`):\n If set to `True`, `retrieved_doc_embeds`, `retrieved_doc_ids`, `context_input_ids` and\n `context_attention_mask` are returned. See returned tensors for more detail.\n use_cache (`bool`, *optional*, defaults to `True`):\n Whether or not the model should return the last key/values attentions (not used by all models).\n forced_eos_token_id (`int`, *optional*):\n The id of the token to force as the last generated token when `max_length` is reached. Usually set to\n `eos_token_id`.\n" @add_start_docstrings(lowerCAmelCase_ ) class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' _a = 'rag' _a = True def __init__( self : Optional[int], lowerCamelCase : Union[str, Any]=None, lowerCamelCase : Union[str, Any]=True, lowerCamelCase : Dict=None, lowerCamelCase : Optional[Any]=None, lowerCamelCase : Any=None, lowerCamelCase : List[Any]=None, lowerCamelCase : Optional[Any]=None, lowerCamelCase : Any=" / ", lowerCamelCase : Union[str, Any]=" // ", lowerCamelCase : List[Any]=5, lowerCamelCase : int=300, lowerCamelCase : Optional[Any]=768, lowerCamelCase : Optional[Any]=8, lowerCamelCase : Tuple="wiki_dpr", lowerCamelCase : Tuple="train", lowerCamelCase : Union[str, Any]="compressed", lowerCamelCase : List[str]=None, lowerCamelCase : Any=None, lowerCamelCase : Union[str, Any]=False, lowerCamelCase : Dict=False, lowerCamelCase : Optional[int]=0.0, lowerCamelCase : Any=True, lowerCamelCase : Dict=False, lowerCamelCase : Tuple=False, lowerCamelCase : List[str]=False, lowerCamelCase : List[Any]=True, lowerCamelCase : Optional[Any]=None, **lowerCamelCase : str, )-> List[Any]: super().__init__( bos_token_id=lowerCamelCase, pad_token_id=lowerCamelCase, eos_token_id=lowerCamelCase, decoder_start_token_id=lowerCamelCase, forced_eos_token_id=lowerCamelCase, is_encoder_decoder=lowerCamelCase, prefix=lowerCamelCase, vocab_size=lowerCamelCase, **lowerCamelCase, ) assert ( "question_encoder" in kwargs and "generator" in kwargs ), "Config has to be initialized with question_encoder and generator config" lowerCamelCase__ : Tuple =kwargs.pop('''question_encoder''' ) lowerCamelCase__ : int =question_encoder_config.pop('''model_type''' ) lowerCamelCase__ : Dict =kwargs.pop('''generator''' ) lowerCamelCase__ : Tuple =decoder_config.pop('''model_type''' ) from ..auto.configuration_auto import AutoConfig lowerCamelCase__ : Tuple =AutoConfig.for_model(lowerCamelCase, **lowerCamelCase ) lowerCamelCase__ : Dict =AutoConfig.for_model(lowerCamelCase, **lowerCamelCase ) lowerCamelCase__ : Any =reduce_loss lowerCamelCase__ : Tuple =label_smoothing lowerCamelCase__ : List[str] =exclude_bos_score lowerCamelCase__ : Dict =do_marginalize lowerCamelCase__ : Union[str, Any] =title_sep lowerCamelCase__ : Dict =doc_sep lowerCamelCase__ : List[Any] =n_docs lowerCamelCase__ : List[str] =max_combined_length lowerCamelCase__ : List[Any] =dataset lowerCamelCase__ : int =dataset_split lowerCamelCase__ : List[Any] =index_name lowerCamelCase__ : int =retrieval_vector_size lowerCamelCase__ : Dict =retrieval_batch_size lowerCamelCase__ : str =passages_path lowerCamelCase__ : Any =index_path lowerCamelCase__ : List[Any] =use_dummy_dataset lowerCamelCase__ : Optional[int] =output_retrieved lowerCamelCase__ : List[str] =do_deduplication lowerCamelCase__ : Tuple =use_cache if self.forced_eos_token_id is None: lowerCamelCase__ : int =getattr(self.generator, '''forced_eos_token_id''', lowerCamelCase ) @classmethod def snake_case ( cls : List[str], lowerCamelCase : PretrainedConfig, lowerCamelCase : PretrainedConfig, **lowerCamelCase : int )-> PretrainedConfig: return cls(question_encoder=question_encoder_config.to_dict(), generator=generator_config.to_dict(), **lowerCamelCase ) def snake_case ( self : Optional[int] )-> int: lowerCamelCase__ : Union[str, Any] =copy.deepcopy(self.__dict__ ) lowerCamelCase__ : Optional[int] =self.question_encoder.to_dict() lowerCamelCase__ : Tuple =self.generator.to_dict() lowerCamelCase__ : Optional[Any] =self.__class__.model_type return output
272
1
import logging import os from .state import PartialState class lowercase__ ( logging.LoggerAdapter ): @staticmethod def UpperCAmelCase ( __UpperCAmelCase )-> Optional[int]: '''simple docstring''' lowerCAmelCase__ = PartialState() return not main_process_only or (main_process_only and state.is_main_process) def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , *__UpperCAmelCase , **__UpperCAmelCase )-> List[str]: '''simple docstring''' if PartialState._shared_state == {}: raise RuntimeError( "You must initialize the accelerate state by calling either `PartialState()` or `Accelerator()` before using the logging utility." ) lowerCAmelCase__ = kwargs.pop("main_process_only" , UpperCAmelCase__ ) lowerCAmelCase__ = kwargs.pop("in_order" , UpperCAmelCase__ ) if self.isEnabledFor(UpperCAmelCase__ ): if self._should_log(UpperCAmelCase__ ): lowerCAmelCase__ , lowerCAmelCase__ = self.process(UpperCAmelCase__ , UpperCAmelCase__ ) self.logger.log(UpperCAmelCase__ , UpperCAmelCase__ , *UpperCAmelCase__ , **UpperCAmelCase__ ) elif in_order: lowerCAmelCase__ = PartialState() for i in range(state.num_processes ): if i == state.process_index: lowerCAmelCase__ , lowerCAmelCase__ = self.process(UpperCAmelCase__ , UpperCAmelCase__ ) self.logger.log(UpperCAmelCase__ , UpperCAmelCase__ , *UpperCAmelCase__ , **UpperCAmelCase__ ) state.wait_for_everyone() def _a ( UpperCamelCase_ : str , UpperCamelCase_ : str = None ) -> int: """simple docstring""" if log_level is None: lowerCAmelCase__ = os.environ.get("ACCELERATE_LOG_LEVEL" , UpperCamelCase_ ) lowerCAmelCase__ = logging.getLogger(UpperCamelCase_ ) if log_level is not None: logger.setLevel(log_level.upper() ) logger.root.setLevel(log_level.upper() ) return MultiProcessAdapter(UpperCamelCase_ , {} )
340
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case =logging.get_logger(__name__) __snake_case ={ """facebook/s2t-wav2vec2-large-en-de""": ( """https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json""" ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class UpperCAmelCase_ ( __lowercase ): lowerCamelCase : Union[str, Any] = '''speech_to_text_2''' lowerCamelCase : Any = ['''past_key_values'''] lowerCamelCase : Optional[Any] = {'''num_attention_heads''': '''decoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Optional[int] , UpperCAmelCase__ : Optional[Any]=1_0_0_0_0 , UpperCAmelCase__ : int=6 , UpperCAmelCase__ : Optional[Any]=2_0_4_8 , UpperCAmelCase__ : Union[str, Any]=4 , UpperCAmelCase__ : List[Any]=0.0 , UpperCAmelCase__ : str=True , UpperCAmelCase__ : str="relu" , UpperCAmelCase__ : Any=2_5_6 , UpperCAmelCase__ : Optional[Any]=0.1 , UpperCAmelCase__ : Any=0.0 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : List[Any]=0.02 , UpperCAmelCase__ : List[Any]=2 , UpperCAmelCase__ : str=True , UpperCAmelCase__ : List[str]=1 , UpperCAmelCase__ : Any=0 , UpperCAmelCase__ : Dict=2 , UpperCAmelCase__ : int=1_0_2_4 , **UpperCAmelCase__ : Optional[Any] , ) -> Dict: lowerCAmelCase = vocab_size lowerCAmelCase = d_model lowerCAmelCase = decoder_ffn_dim lowerCAmelCase = decoder_layers lowerCAmelCase = decoder_attention_heads lowerCAmelCase = dropout lowerCAmelCase = attention_dropout lowerCAmelCase = activation_dropout lowerCAmelCase = activation_function lowerCAmelCase = init_std lowerCAmelCase = decoder_layerdrop lowerCAmelCase = use_cache lowerCAmelCase = decoder_layers lowerCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True lowerCAmelCase = max_target_positions super().__init__( pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , decoder_start_token_id=UpperCAmelCase__ , **UpperCAmelCase__ , )
4
0
'''simple docstring''' import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class snake_case__ ( unittest.TestCase ): def A_ ( self : Optional[Any] ) -> Any: '''simple docstring''' # A mock response for an HTTP head request to emulate server down __snake_case : List[Any] = mock.Mock() __snake_case : Any = 500 __snake_case : Optional[int] = {} __snake_case : Optional[int] = HTTPError __snake_case : List[str] = {} # Download this model to make sure it's in the cache. __snake_case : Union[str, Any] = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=__a ) as mock_head: __snake_case : Tuple = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def A_ ( self : List[Any] ) -> List[str]: '''simple docstring''' # A mock response for an HTTP head request to emulate server down __snake_case : Union[str, Any] = mock.Mock() __snake_case : Tuple = 500 __snake_case : int = {} __snake_case : List[str] = HTTPError __snake_case : List[Any] = {} # Download this model to make sure it's in the cache. __snake_case : str = GPTaTokenizerFast.from_pretrained('gpt2' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=__a ) as mock_head: __snake_case : Dict = GPTaTokenizerFast.from_pretrained('gpt2' ) # This check we did call the fake head request mock_head.assert_called() def A_ ( self : List[Any] ) -> str: '''simple docstring''' # This test is for deprecated behavior and can be removed in v5 try: __snake_case : Optional[int] = tempfile.mktemp() with open(__a , 'wb' ) as f: http_get('https://huggingface.co/albert-base-v1/resolve/main/spiece.model' , __a ) __snake_case : Any = AlbertTokenizer.from_pretrained(__a ) finally: os.remove(__a ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile('tokenizer.json' ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open('tokenizer.json' , 'wb' ) as f: http_get('https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json' , __a ) __snake_case : Optional[int] = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 1000 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove('tokenizer.json' ) def A_ ( self : List[str] ) -> Any: '''simple docstring''' # This test is for deprecated behavior and can be removed in v5 __snake_case : int = AlbertTokenizer.from_pretrained('https://huggingface.co/albert-base-v1/resolve/main/spiece.model' ) @is_staging_test class snake_case__ ( unittest.TestCase ): A__ = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''bla''', '''blou'''] @classmethod def A_ ( cls : int ) -> Dict: '''simple docstring''' __snake_case : Union[str, Any] = TOKEN HfFolder.save_token(__a ) @classmethod def A_ ( cls : List[Any] ) -> Dict: '''simple docstring''' try: delete_repo(token=cls._token , repo_id='test-tokenizer' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-tokenizer-org' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-tokenizer' ) except HTTPError: pass def A_ ( self : List[Any] ) -> Tuple: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: __snake_case : Optional[Any] = os.path.join(__a , 'vocab.txt' ) with open(__a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) __snake_case : int = BertTokenizer(__a ) tokenizer.push_to_hub('test-tokenizer' , use_auth_token=self._token ) __snake_case : Union[str, Any] = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='test-tokenizer' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__a , repo_id='test-tokenizer' , push_to_hub=__a , use_auth_token=self._token ) __snake_case : Union[str, Any] = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def A_ ( self : Optional[int] ) -> List[str]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: __snake_case : str = os.path.join(__a , 'vocab.txt' ) with open(__a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) __snake_case : Union[str, Any] = BertTokenizer(__a ) tokenizer.push_to_hub('valid_org/test-tokenizer-org' , use_auth_token=self._token ) __snake_case : List[str] = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='valid_org/test-tokenizer-org' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( __a , repo_id='valid_org/test-tokenizer-org' , push_to_hub=__a , use_auth_token=self._token ) __snake_case : Dict = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def A_ ( self : Optional[Any] ) -> Any: '''simple docstring''' CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case : Union[str, Any] = os.path.join(__a , 'vocab.txt' ) with open(__a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) __snake_case : Optional[Any] = CustomTokenizer(__a ) # No fast custom tokenizer tokenizer.push_to_hub('test-dynamic-tokenizer' , use_auth_token=self._token ) __snake_case : Union[str, Any] = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=__a ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizer' ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case : Dict = os.path.join(__a , 'vocab.txt' ) with open(__a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) __snake_case : Dict = BertTokenizerFast.from_pretrained(__a ) bert_tokenizer.save_pretrained(__a ) __snake_case : Optional[int] = CustomTokenizerFast.from_pretrained(__a ) tokenizer.push_to_hub('test-dynamic-tokenizer' , use_auth_token=self._token ) __snake_case : Union[str, Any] = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=__a ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizerFast' ) __snake_case : int = AutoTokenizer.from_pretrained( f'''{USER}/test-dynamic-tokenizer''' , use_fast=__a , trust_remote_code=__a ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizer' ) class snake_case__ ( unittest.TestCase ): def A_ ( self : str ) -> Optional[Any]: '''simple docstring''' __snake_case : Any = Trie() trie.add('Hello 友達' ) self.assertEqual(trie.data , {'H': {'e': {'l': {'l': {'o': {' ': {'友': {'達': {'': 1}}}}}}}}} ) trie.add('Hello' ) trie.data self.assertEqual(trie.data , {'H': {'e': {'l': {'l': {'o': {'': 1, ' ': {'友': {'達': {'': 1}}}}}}}}} ) def A_ ( self : Optional[Any] ) -> List[str]: '''simple docstring''' __snake_case : Optional[Any] = Trie() self.assertEqual(trie.split('[CLS] This is a extra_id_100' ) , ['[CLS] This is a extra_id_100'] ) trie.add('[CLS]' ) trie.add('extra_id_1' ) trie.add('extra_id_100' ) self.assertEqual(trie.split('[CLS] This is a extra_id_100' ) , ['[CLS]', ' This is a ', 'extra_id_100'] ) def A_ ( self : int ) -> Union[str, Any]: '''simple docstring''' __snake_case : Any = Trie() trie.add('A' ) self.assertEqual(trie.split('ABC' ) , ['A', 'BC'] ) self.assertEqual(trie.split('BCA' ) , ['BC', 'A'] ) def A_ ( self : List[Any] ) -> Optional[int]: '''simple docstring''' __snake_case : Dict = Trie() trie.add('TOKEN]' ) trie.add('[SPECIAL_TOKEN]' ) self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]' ) , ['This is something ', '[SPECIAL_TOKEN]'] ) def A_ ( self : str ) -> List[Any]: '''simple docstring''' __snake_case : Dict = Trie() trie.add('A' ) trie.add('P' ) trie.add('[SPECIAL_TOKEN]' ) self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]' ) , ['This is something ', '[SPECIAL_TOKEN]'] ) def A_ ( self : Dict ) -> Union[str, Any]: '''simple docstring''' __snake_case : int = Trie() trie.add('AB' ) trie.add('B' ) trie.add('C' ) self.assertEqual(trie.split('ABC' ) , ['AB', 'C'] ) def A_ ( self : Optional[Any] ) -> Optional[Any]: '''simple docstring''' __snake_case : Any = Trie() trie.add('ABC' ) trie.add('B' ) trie.add('CD' ) self.assertEqual(trie.split('ABCD' ) , ['ABC', 'D'] ) def A_ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' # Even if the offsets are wrong, we necessarily output correct string # parts. __snake_case : Optional[Any] = Trie() __snake_case : Dict = trie.cut_text('ABC' , [0, 0, 2, 1, 2, 3] ) self.assertEqual(__a , ['AB', 'C'] )
0
'''simple docstring''' def a_ ( _UpperCAmelCase : int ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError('The given input must be positive' ) # get the generated string sequence __snake_case : Optional[Any] = gray_code_sequence_string(_UpperCAmelCase ) # # convert them to integers for i in range(len(_UpperCAmelCase ) ): __snake_case : Optional[Any] = int(sequence[i] ,2 ) return sequence def a_ ( _UpperCAmelCase : int ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] __snake_case : Dict = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits __snake_case : Dict = gray_code_sequence_string(bit_count - 1 ) __snake_case : Any = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): __snake_case : str = '0' + smaller_sequence[i] sequence.append(_UpperCAmelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): __snake_case : Any = '1' + smaller_sequence[i] sequence.append(_UpperCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
0
1
"""simple docstring""" import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class UpperCAmelCase_ : @staticmethod def _UpperCAmelCase ( *a , **a ) -> Optional[int]: pass def a_ ( _lowerCAmelCase : Image ): '''simple docstring''' lowercase__ : Union[str, Any] = hashlib.mda(image.tobytes() ) return m.hexdigest()[:10] def a_ ( _lowerCAmelCase : Image ): '''simple docstring''' lowercase__ : Union[str, Any] = np.array(_lowerCAmelCase ) lowercase__ : Any = npimg.shape return {"hash": hashimage(_lowerCAmelCase ), "shape": shape} @is_pipeline_test @require_vision @require_torch class UpperCAmelCase_ ( unittest.TestCase): lowerCamelCase__ : Optional[Any] = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items()) if MODEL_FOR_MASK_GENERATION_MAPPING else [])) lowerCamelCase__ : Dict = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items()) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else [])) def _UpperCAmelCase ( self , a , a , a ) -> List[Any]: lowercase__ : List[str] = MaskGenerationPipeline(model=a , image_processor=a ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def _UpperCAmelCase ( self , a , a ) -> Optional[Any]: pass @require_tf @unittest.skip('Image segmentation not implemented in TF' ) def _UpperCAmelCase ( self ) -> Optional[Any]: pass @slow @require_torch def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : int = pipeline('mask-generation' , model='facebook/sam-vit-huge' ) lowercase__ : List[str] = image_segmenter('http://images.cocodataset.org/val2017/000000039769.jpg' , points_per_batch=2_5_6 ) # Shortening by hashing lowercase__ : int = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(a ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(a , decimals=4 ) , [ {'mask': {'hash': '115ad19f5f', 'shape': (4_8_0, 6_4_0)}, 'scores': 1.0_444}, {'mask': {'hash': '6affa964c6', 'shape': (4_8_0, 6_4_0)}, 'scores': 1.021}, {'mask': {'hash': 'dfe28a0388', 'shape': (4_8_0, 6_4_0)}, 'scores': 1.0_167}, {'mask': {'hash': 'c0a5f4a318', 'shape': (4_8_0, 6_4_0)}, 'scores': 1.0_132}, {'mask': {'hash': 'fe8065c197', 'shape': (4_8_0, 6_4_0)}, 'scores': 1.0_053}, {'mask': {'hash': 'e2d0b7a0b7', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_967}, {'mask': {'hash': '453c7844bd', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.993}, {'mask': {'hash': '3d44f2926d', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_909}, {'mask': {'hash': '64033ddc3f', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_879}, {'mask': {'hash': '801064ff79', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_834}, {'mask': {'hash': '6172f276ef', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_716}, {'mask': {'hash': 'b49e60e084', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_612}, {'mask': {'hash': 'a811e775fd', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_599}, {'mask': {'hash': 'a6a8ebcf4b', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_552}, {'mask': {'hash': '9d8257e080', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_532}, {'mask': {'hash': '32de6454a8', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_516}, {'mask': {'hash': 'af3d4af2c8', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_499}, {'mask': {'hash': '3c6db475fb', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_483}, {'mask': {'hash': 'c290813fb9', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_464}, {'mask': {'hash': 'b6f0b8f606', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.943}, {'mask': {'hash': '92ce16bfdf', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.943}, {'mask': {'hash': 'c749b25868', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_408}, {'mask': {'hash': 'efb6cab859', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_335}, {'mask': {'hash': '1ff2eafb30', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_326}, {'mask': {'hash': '788b798e24', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.9_262}, {'mask': {'hash': 'abea804f0e', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.8_999}, {'mask': {'hash': '7b9e8ddb73', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.8_986}, {'mask': {'hash': 'cd24047c8a', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.8_984}, {'mask': {'hash': '6943e6bcbd', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.8_873}, {'mask': {'hash': 'b5f47c9191', 'shape': (4_8_0, 6_4_0)}, 'scores': 0.8_871} ] , ) # fmt: on @require_torch @slow def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : Tuple = 'facebook/sam-vit-huge' lowercase__ : Union[str, Any] = pipeline('mask-generation' , model=a ) lowercase__ : Tuple = image_segmenter( 'http://images.cocodataset.org/val2017/000000039769.jpg' , pred_iou_thresh=1 , points_per_batch=2_5_6 ) # Shortening by hashing lowercase__ : List[Any] = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(a ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(a , decimals=4 ) , [ {'mask': {'hash': '115ad19f5f', 'shape': (4_8_0, 6_4_0)}, 'scores': 1.0_444}, {'mask': {'hash': '6affa964c6', 'shape': (4_8_0, 6_4_0)}, 'scores': 1.0_210}, {'mask': {'hash': 'dfe28a0388', 'shape': (4_8_0, 6_4_0)}, 'scores': 1.0_167}, {'mask': {'hash': 'c0a5f4a318', 'shape': (4_8_0, 6_4_0)}, 'scores': 1.0_132}, {'mask': {'hash': 'fe8065c197', 'shape': (4_8_0, 6_4_0)}, 'scores': 1.0_053}, ] , )
77
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[List[PIL.Image.Image], np.ndarray] lowerCamelCase__ : Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.26.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version(">=", "0.0.12") ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : np.ndarray lowerCamelCase__ : List[bool] from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
77
1
"""simple docstring""" from collections import UserDict from typing import Union import numpy as np import requests from ..utils import ( add_end_docstrings, logging, ) from .audio_classification import ffmpeg_read from .base import PIPELINE_INIT_ARGS, Pipeline _lowerCAmelCase : int = logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase_ ) class __magic_name__ ( lowerCAmelCase_ ): def __init__( self , **__snake_case ) -> Dict: '''simple docstring''' super().__init__(**__snake_case ) if self.framework != "pt": raise ValueError(f'The {self.__class__} is only available in PyTorch.' ) # No specific FOR_XXX available yet def __call__( self , __snake_case , **__snake_case ) -> Union[str, Any]: '''simple docstring''' return super().__call__(__snake_case , **__snake_case ) def __magic_name__ ( self , **__snake_case ) -> str: '''simple docstring''' __a ={} if "candidate_labels" in kwargs: __a =kwargs['candidate_labels'] if "hypothesis_template" in kwargs: __a =kwargs['hypothesis_template'] return preprocess_params, {}, {} def __magic_name__ ( self , __snake_case , __snake_case=None , __snake_case="This is a sound of {}." ) -> Optional[int]: '''simple docstring''' if isinstance(__snake_case , __snake_case ): if audio.startswith('http://' ) or audio.startswith('https://' ): # We need to actually check for a real protocol, otherwise it's impossible to use a local file # like http_huggingface_co.png __a =requests.get(__snake_case ).content else: with open(__snake_case , 'rb' ) as f: __a =f.read() if isinstance(__snake_case , __snake_case ): __a =ffmpeg_read(__snake_case , self.feature_extractor.sampling_rate ) if not isinstance(__snake_case , np.ndarray ): raise ValueError('We expect a numpy ndarray as input' ) if len(audio.shape ) != 1: raise ValueError('We expect a single channel audio input for ZeroShotAudioClassificationPipeline' ) __a =self.feature_extractor( [audio] , sampling_rate=self.feature_extractor.sampling_rate , return_tensors='pt' ) __a =candidate_labels __a =[hypothesis_template.format(__snake_case ) for x in candidate_labels] __a =self.tokenizer(__snake_case , return_tensors=self.framework , padding=__snake_case ) __a =[text_inputs] return inputs def __magic_name__ ( self , __snake_case ) -> List[str]: '''simple docstring''' __a =model_inputs.pop('candidate_labels' ) __a =model_inputs.pop('text_inputs' ) if isinstance(text_inputs[0] , __snake_case ): __a =text_inputs[0] else: # Batching case. __a =text_inputs[0][0] __a =self.model(**__snake_case , **__snake_case ) __a ={ 'candidate_labels': candidate_labels, 'logits': outputs.logits_per_audio, } return model_outputs def __magic_name__ ( self , __snake_case ) -> Optional[int]: '''simple docstring''' __a =model_outputs.pop('candidate_labels' ) __a =model_outputs['logits'][0] if self.framework == "pt": __a =logits.softmax(dim=0 ) __a =probs.tolist() else: raise ValueError('`tf` framework not supported.' ) __a =[ {'score': score, 'label': candidate_label} for score, candidate_label in sorted(zip(__snake_case , __snake_case ) , key=lambda __snake_case : -x[0] ) ] return result
371
from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake _lowerCAmelCase : Optional[Any] = numpy.array([0, 0]) _lowerCAmelCase : Dict = numpy.array([0.5, 0.8660254]) _lowerCAmelCase : Any = numpy.array([1, 0]) _lowerCAmelCase : int = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def UpperCamelCase_( _snake_case : list[numpy.ndarray] , _snake_case : int ): """simple docstring""" __a =initial_vectors for _ in range(_snake_case ): __a =iteration_step(_snake_case ) return vectors def UpperCamelCase_( _snake_case : list[numpy.ndarray] ): """simple docstring""" __a =[] for i, start_vector in enumerate(vectors[:-1] ): __a =vectors[i + 1] new_vectors.append(_snake_case ) __a =end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def UpperCamelCase_( _snake_case : numpy.ndarray , _snake_case : float ): """simple docstring""" __a =numpy.radians(_snake_case ) __a , __a =numpy.cos(_snake_case ), numpy.sin(_snake_case ) __a =numpy.array(((c, -s), (s, c)) ) return numpy.dot(_snake_case , _snake_case ) def UpperCamelCase_( _snake_case : list[numpy.ndarray] ): """simple docstring""" __a =plt.gca() axes.set_aspect('equal' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() __a , __a =zip(*_snake_case ) plt.plot(_snake_case , _snake_case ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() _lowerCAmelCase : List[Any] = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
308
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging UpperCAmelCase_ : List[str] = logging.get_logger(__name__) class lowerCAmelCase__ ( UpperCAmelCase__ ): '''simple docstring''' __UpperCamelCase = ["pixel_values"] def __init__( self : Tuple , lowercase_ : bool = True , lowercase_ : Optional[Dict[str, int]] = None , lowercase_ : PILImageResampling = PILImageResampling.BILINEAR , lowercase_ : bool = True , lowercase_ : Dict[str, int] = None , lowercase_ : bool = True , lowercase_ : Union[int, float] = 1 / 255 , lowercase_ : bool = True , lowercase_ : Optional[Union[float, List[float]]] = None , lowercase_ : Optional[Union[float, List[float]]] = None , **lowercase_ : str , ): '''simple docstring''' super().__init__(**lowercase_) SCREAMING_SNAKE_CASE_ : Dict = size if size is not None else {'''shortest_edge''': 256} SCREAMING_SNAKE_CASE_ : str = get_size_dict(lowercase_ , default_to_square=lowercase_) SCREAMING_SNAKE_CASE_ : Dict = crop_size if crop_size is not None else {'''height''': 224, '''width''': 224} SCREAMING_SNAKE_CASE_ : int = get_size_dict(lowercase_) SCREAMING_SNAKE_CASE_ : List[Any] = do_resize SCREAMING_SNAKE_CASE_ : int = size SCREAMING_SNAKE_CASE_ : List[Any] = resample SCREAMING_SNAKE_CASE_ : int = do_center_crop SCREAMING_SNAKE_CASE_ : List[Any] = crop_size SCREAMING_SNAKE_CASE_ : List[Any] = do_rescale SCREAMING_SNAKE_CASE_ : Optional[int] = rescale_factor SCREAMING_SNAKE_CASE_ : str = do_normalize SCREAMING_SNAKE_CASE_ : int = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN SCREAMING_SNAKE_CASE_ : List[Any] = image_std if image_std is not None else IMAGENET_STANDARD_STD def _SCREAMING_SNAKE_CASE ( self : str , lowercase_ : np.ndarray , lowercase_ : Dict[str, int] , lowercase_ : PILImageResampling = PILImageResampling.BICUBIC , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : Optional[int] , ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[Any] = get_size_dict(lowercase_ , default_to_square=lowercase_) if "shortest_edge" not in size: raise ValueError(F'The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}') SCREAMING_SNAKE_CASE_ : int = get_resize_output_image_size(lowercase_ , size=size['''shortest_edge'''] , default_to_square=lowercase_) return resize(lowercase_ , size=lowercase_ , resample=lowercase_ , data_format=lowercase_ , **lowercase_) def _SCREAMING_SNAKE_CASE ( self : Optional[int] , lowercase_ : np.ndarray , lowercase_ : Dict[str, int] , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : int , ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[str] = get_size_dict(lowercase_) return center_crop(lowercase_ , size=(size['''height'''], size['''width''']) , data_format=lowercase_ , **lowercase_) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowercase_ : np.ndarray , lowercase_ : float , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : List[str]): '''simple docstring''' return rescale(lowercase_ , scale=lowercase_ , data_format=lowercase_ , **lowercase_) def _SCREAMING_SNAKE_CASE ( self : int , lowercase_ : np.ndarray , lowercase_ : Union[float, List[float]] , lowercase_ : Union[float, List[float]] , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : str , ): '''simple docstring''' return normalize(lowercase_ , mean=lowercase_ , std=lowercase_ , data_format=lowercase_ , **lowercase_) def _SCREAMING_SNAKE_CASE ( self : int , lowercase_ : ImageInput , lowercase_ : Optional[bool] = None , lowercase_ : Dict[str, int] = None , lowercase_ : PILImageResampling = None , lowercase_ : bool = None , lowercase_ : Dict[str, int] = None , lowercase_ : Optional[bool] = None , lowercase_ : Optional[float] = None , lowercase_ : Optional[bool] = None , lowercase_ : Optional[Union[float, List[float]]] = None , lowercase_ : Optional[Union[float, List[float]]] = None , lowercase_ : Optional[Union[str, TensorType]] = None , lowercase_ : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowercase_ : Dict , ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE_ : Union[str, Any] = size if size is not None else self.size SCREAMING_SNAKE_CASE_ : int = get_size_dict(lowercase_ , default_to_square=lowercase_) SCREAMING_SNAKE_CASE_ : Dict = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE_ : Dict = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE_ : List[str] = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE_ : Any = get_size_dict(lowercase_) SCREAMING_SNAKE_CASE_ : Dict = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE_ : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE_ : str = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE_ : List[Any] = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE_ : Optional[int] = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE_ : List[Any] = make_list_of_images(lowercase_) if not valid_images(lowercase_): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''') if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''') if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''') if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''') if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''') # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE_ : Union[str, Any] = [to_numpy_array(lowercase_) for image in images] if do_resize: SCREAMING_SNAKE_CASE_ : Union[str, Any] = [self.resize(image=lowercase_ , size=lowercase_ , resample=lowercase_) for image in images] if do_center_crop: SCREAMING_SNAKE_CASE_ : Tuple = [self.center_crop(image=lowercase_ , size=lowercase_) for image in images] if do_rescale: SCREAMING_SNAKE_CASE_ : int = [self.rescale(image=lowercase_ , scale=lowercase_) for image in images] if do_normalize: SCREAMING_SNAKE_CASE_ : Optional[int] = [self.normalize(image=lowercase_ , mean=lowercase_ , std=lowercase_) for image in images] SCREAMING_SNAKE_CASE_ : Any = [to_channel_dimension_format(lowercase_ , lowercase_) for image in images] SCREAMING_SNAKE_CASE_ : Dict = {'''pixel_values''': images} return BatchFeature(data=lowercase_ , tensor_type=lowercase_)
91
def _snake_case( SCREAMING_SNAKE_CASE__ : float , SCREAMING_SNAKE_CASE__ : float , SCREAMING_SNAKE_CASE__ : float , SCREAMING_SNAKE_CASE__ : float , SCREAMING_SNAKE_CASE__ : float , ) -> float: '''simple docstring''' A__ = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError('All input parameters must be positive' ) if any(p > 1 for p in parameters[1:4] ): raise ValueError('Relative densities cannot be greater than one' ) else: A__ = 1 - (matter_density + radiation_density + dark_energy) A__ = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) A__ = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation lowercase_ = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1e-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
7
0
def _UpperCamelCase (a__ :int , a__ :int ): """simple docstring""" while second != 0: UpperCamelCase__ = first & second first ^= second UpperCamelCase__ = c << 1 return first if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = int(input("Enter the first number: ").strip()) UpperCamelCase__ = int(input("Enter the second number: ").strip()) print(f"""{add(first, second) = }""")
362
import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch.nn import CrossEntropyLoss from torch.utils.data import DataLoader, TensorDataset from utils_ner import TokenClassificationTask UpperCamelCase__ = logging.getLogger(__name__) class __SCREAMING_SNAKE_CASE ( _a ): snake_case : str = """token-classification""" def __init__( self , __lowerCAmelCase ): if type(__lowerCAmelCase ) == dict: UpperCamelCase__ = Namespace(**__lowerCAmelCase ) UpperCamelCase__ = import_module("""tasks""" ) try: UpperCamelCase__ = getattr(__lowerCAmelCase , hparams.task_type ) UpperCamelCase__ = token_classification_task_clazz() except AttributeError: raise ValueError( f"""Task {hparams.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. """ f"""Available tasks classes are: {TokenClassificationTask.__subclasses__()}""" ) UpperCamelCase__ = self.token_classification_task.get_labels(hparams.labels ) UpperCamelCase__ = CrossEntropyLoss().ignore_index super().__init__(__lowerCAmelCase , len(self.labels ) , self.mode ) def _lowerCamelCase ( self , **__lowerCAmelCase ): return self.model(**__lowerCAmelCase ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase ): UpperCamelCase__ = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type != "distilbert": UpperCamelCase__ = ( batch[2] if self.config.model_type in ["""bert""", """xlnet"""] else None ) # XLM and RoBERTa don"t use token_type_ids UpperCamelCase__ = self(**__lowerCAmelCase ) UpperCamelCase__ = outputs[0] # tensorboard_logs = {"loss": loss, "rate": self.lr_scheduler.get_last_lr()[-1]} return {"loss": loss} def _lowerCamelCase ( self ): UpperCamelCase__ = self.hparams for mode in ["train", "dev", "test"]: UpperCamelCase__ = self._feature_file(__lowerCAmelCase ) if os.path.exists(__lowerCAmelCase ) and not args.overwrite_cache: logger.info("""Loading features from cached file %s""" , __lowerCAmelCase ) UpperCamelCase__ = torch.load(__lowerCAmelCase ) else: logger.info("""Creating features from dataset file at %s""" , args.data_dir ) UpperCamelCase__ = self.token_classification_task.read_examples_from_file(args.data_dir , __lowerCAmelCase ) UpperCamelCase__ = self.token_classification_task.convert_examples_to_features( __lowerCAmelCase , self.labels , args.max_seq_length , self.tokenizer , cls_token_at_end=bool(self.config.model_type in ["""xlnet"""] ) , cls_token=self.tokenizer.cls_token , cls_token_segment_id=2 if self.config.model_type in ["""xlnet"""] else 0 , sep_token=self.tokenizer.sep_token , sep_token_extra=__lowerCAmelCase , pad_on_left=bool(self.config.model_type in ["""xlnet"""] ) , pad_token=self.tokenizer.pad_token_id , pad_token_segment_id=self.tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info("""Saving features into cached file %s""" , __lowerCAmelCase ) torch.save(__lowerCAmelCase , __lowerCAmelCase ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = False ): UpperCamelCase__ = self._feature_file(__lowerCAmelCase ) logger.info("""Loading features from cached file %s""" , __lowerCAmelCase ) UpperCamelCase__ = torch.load(__lowerCAmelCase ) UpperCamelCase__ = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) UpperCamelCase__ = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) if features[0].token_type_ids is not None: UpperCamelCase__ = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) else: UpperCamelCase__ = torch.tensor([0 for f in features] , dtype=torch.long ) # HACK(we will not use this anymore soon) UpperCamelCase__ = torch.tensor([f.label_ids for f in features] , dtype=torch.long ) return DataLoader( TensorDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) , batch_size=__lowerCAmelCase ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase ): """Compute validation""" "" UpperCamelCase__ = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type != "distilbert": UpperCamelCase__ = ( batch[2] if self.config.model_type in ["""bert""", """xlnet"""] else None ) # XLM and RoBERTa don"t use token_type_ids UpperCamelCase__ = self(**__lowerCAmelCase ) UpperCamelCase__ , UpperCamelCase__ = outputs[:2] UpperCamelCase__ = logits.detach().cpu().numpy() UpperCamelCase__ = inputs["""labels"""].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def _lowerCamelCase ( self , __lowerCAmelCase ): UpperCamelCase__ = torch.stack([x["""val_loss"""] for x in outputs] ).mean() UpperCamelCase__ = np.concatenate([x["""pred"""] for x in outputs] , axis=0 ) UpperCamelCase__ = np.argmax(__lowerCAmelCase , axis=2 ) UpperCamelCase__ = np.concatenate([x["""target"""] for x in outputs] , axis=0 ) UpperCamelCase__ = dict(enumerate(self.labels ) ) UpperCamelCase__ = [[] for _ in range(out_label_ids.shape[0] )] UpperCamelCase__ = [[] for _ in range(out_label_ids.shape[0] )] for i in range(out_label_ids.shape[0] ): for j in range(out_label_ids.shape[1] ): if out_label_ids[i, j] != self.pad_token_label_id: out_label_list[i].append(label_map[out_label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) UpperCamelCase__ = { """val_loss""": val_loss_mean, """accuracy_score""": accuracy_score(__lowerCAmelCase , __lowerCAmelCase ), """precision""": precision_score(__lowerCAmelCase , __lowerCAmelCase ), """recall""": recall_score(__lowerCAmelCase , __lowerCAmelCase ), """f1""": fa_score(__lowerCAmelCase , __lowerCAmelCase ), } UpperCamelCase__ = dict(results.items() ) UpperCamelCase__ = results return ret, preds_list, out_label_list def _lowerCamelCase ( self , __lowerCAmelCase ): # when stable UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = self._eval_end(__lowerCAmelCase ) UpperCamelCase__ = ret["""log"""] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def _lowerCamelCase ( self , __lowerCAmelCase ): # updating to test_epoch_end instead of deprecated test_end UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = self._eval_end(__lowerCAmelCase ) # Converting to the dict required by pl # https://github.com/PyTorchLightning/pytorch-lightning/blob/master/\ # pytorch_lightning/trainer/logging.py#L139 UpperCamelCase__ = ret["""log"""] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def _lowerCamelCase ( __lowerCAmelCase , __lowerCAmelCase ): # Add NER specific options BaseTransformer.add_model_specific_args(__lowerCAmelCase , __lowerCAmelCase ) parser.add_argument( """--task_type""" , default="""NER""" , type=__lowerCAmelCase , help="""Task type to fine tune in training (e.g. NER, POS, etc)""" ) parser.add_argument( """--max_seq_length""" , default=128 , type=__lowerCAmelCase , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--labels""" , default="""""" , type=__lowerCAmelCase , help="""Path to a file containing all labels. If not specified, CoNLL-2003 labels are used.""" , ) parser.add_argument( """--gpus""" , default=0 , type=__lowerCAmelCase , help="""The number of GPUs allocated for this, it is by default 0 meaning none""" , ) parser.add_argument( """--overwrite_cache""" , action="""store_true""" , help="""Overwrite the cached training and evaluation sets""" ) return parser if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() add_generic_args(parser, os.getcwd()) UpperCamelCase__ = NERTransformer.add_model_specific_args(parser, os.getcwd()) UpperCamelCase__ = parser.parse_args() UpperCamelCase__ = NERTransformer(args) UpperCamelCase__ = generic_train(model, args) if args.do_predict: # See https://github.com/huggingface/transformers/issues/3159 # pl use this default format to create a checkpoint: # https://github.com/PyTorchLightning/pytorch-lightning/blob/master\ # /pytorch_lightning/callbacks/model_checkpoint.py#L322 UpperCamelCase__ = sorted(glob.glob(os.path.join(args.output_dir, "checkpoint-epoch=*.ckpt"), recursive=True)) UpperCamelCase__ = model.load_from_checkpoint(checkpoints[-1]) trainer.test(model)
87
0
import os from typing import Dict, List, Tuple, TypeVar, Union UpperCamelCase__ = TypeVar("""T""") UpperCamelCase__ = Union[List[T], Tuple[T, ...]] UpperCamelCase__ = Union[T, List[T], Dict[str, T]] UpperCamelCase__ = Union[str, bytes, os.PathLike]
92
from __future__ import annotations import collections import tempfile import unittest import numpy as np from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import is_tf_available, is_vision_available from ...test_modeling_tf_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_tf_bert import TFBertModelTester from ..clip.test_modeling_tf_clip import TFCLIPVisionModelTester from ..deit.test_modeling_tf_deit import TFDeiTModelTester from ..roberta.test_modeling_tf_roberta import TFRobertaModelTester from ..vit.test_modeling_tf_vit import TFViTModelTester if is_tf_available(): from transformers import ( TFBertModel, TFCLIPVisionModel, TFDeiTModel, TFRobertaModel, TFVisionTextDualEncoderModel, TFViTModel, VisionTextDualEncoderConfig, ) if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor def _a ( SCREAMING_SNAKE_CASE_ : Union[str, Any] ): if isinstance(SCREAMING_SNAKE_CASE_ , collections.abc.Iterable ): return x return (x, x) @require_tf class a__ : def __SCREAMING_SNAKE_CASE( self , _A , _A ): """simple docstring""" pass def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" pass def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" pass def __SCREAMING_SNAKE_CASE( self , _A , _A , _A , _A , _A=None , **_A ): """simple docstring""" __lowerCAmelCase = VisionTextDualEncoderConfig.from_vision_text_configs(_A , _A ) __lowerCAmelCase = TFVisionTextDualEncoderModel(_A ) __lowerCAmelCase = model(input_ids=_A , pixel_values=_A , attention_mask=_A ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], config.projection_dim) ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A , _A , _A=None , **_A ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.get_vision_text_model(_A , _A ) __lowerCAmelCase = TFVisionTextDualEncoderModel(vision_model=_A , text_model=_A ) __lowerCAmelCase = model(input_ids=_A , pixel_values=_A , attention_mask=_A ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim) ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A , _A , _A=None , **_A ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.get_vision_text_model(_A , _A ) __lowerCAmelCase = {"vision_model": vision_model, "text_model": text_model} __lowerCAmelCase = TFVisionTextDualEncoderModel.from_vision_text_pretrained(**_A ) __lowerCAmelCase = model(input_ids=_A , pixel_values=_A , attention_mask=_A ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim) ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A , _A , _A=None , **_A ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.get_vision_text_model(_A , _A ) __lowerCAmelCase = TFVisionTextDualEncoderModel(vision_model=_A , text_model=_A ) __lowerCAmelCase = model(input_ids=_A , pixel_values=_A , attention_mask=_A ) __lowerCAmelCase = output[0].numpy() with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_A ) __lowerCAmelCase = TFVisionTextDualEncoderModel.from_pretrained(_A ) __lowerCAmelCase = model(input_ids=_A , pixel_values=_A , attention_mask=_A ) __lowerCAmelCase = after_output[0].numpy() __lowerCAmelCase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(_A , 1E-5 ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A , _A , _A=None , **_A ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.get_vision_text_model(_A , _A ) __lowerCAmelCase = TFVisionTextDualEncoderModel(vision_model=_A , text_model=_A ) __lowerCAmelCase = model( input_ids=_A , pixel_values=_A , attention_mask=_A , output_attentions=_A ) __lowerCAmelCase = output.vision_model_output.attentions self.assertEqual(len(_A ) , vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) __lowerCAmelCase = to_atuple(vision_model.config.image_size ) __lowerCAmelCase = to_atuple(vision_model.config.patch_size ) __lowerCAmelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) __lowerCAmelCase = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) __lowerCAmelCase = output.text_model_output.attentions self.assertEqual(len(_A ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A ): """simple docstring""" __lowerCAmelCase = np.abs((a - b) ).max() self.assertLessEqual(_A , _A , f"""Difference between torch and flax is {diff} (>= {tol}).""" ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_model(**_A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**_A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**_A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.prepare_config_and_inputs() self.check_save_load(**_A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**_A ) @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.get_pretrained_model_and_inputs() __lowerCAmelCase = model_a(**_A ) __lowerCAmelCase = outputs[0].numpy() with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(_A ) __lowerCAmelCase = TFVisionTextDualEncoderModel.from_pretrained(_A ) __lowerCAmelCase = model_a(**_A ) __lowerCAmelCase = after_outputs[0].numpy() __lowerCAmelCase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(_A , 1E-5 ) @require_tf class a__ ( snake_case__ , unittest.TestCase ): def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = TFVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-vit" , "hf-internal-testing/tiny-random-bert" ) __lowerCAmelCase = 1_3 __lowerCAmelCase = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) __lowerCAmelCase = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) __lowerCAmelCase = random_attention_mask([batch_size, 4] ) __lowerCAmelCase = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def __SCREAMING_SNAKE_CASE( self , _A , _A ): """simple docstring""" __lowerCAmelCase = TFViTModel(_A , name="vision_model" ) __lowerCAmelCase = TFBertModel(_A , name="text_model" ) return vision_model, text_model def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = TFViTModelTester(self ) __lowerCAmelCase = TFBertModelTester(self ) __lowerCAmelCase = vit_model_tester.prepare_config_and_inputs() __lowerCAmelCase = bert_model_tester.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = vision_config_and_inputs ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class a__ ( snake_case__ , unittest.TestCase ): def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = TFVisionTextDualEncoderModel.from_vision_text_pretrained( "Rocketknight1/tiny-random-deit-tf" , "hf-internal-testing/tiny-random-roberta" ) __lowerCAmelCase = 1_3 __lowerCAmelCase = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) __lowerCAmelCase = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) __lowerCAmelCase = random_attention_mask([batch_size, 4] ) __lowerCAmelCase = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def __SCREAMING_SNAKE_CASE( self , _A , _A , _A , _A , _A=None , **_A ): """simple docstring""" __lowerCAmelCase , __lowerCAmelCase = self.get_vision_text_model(_A , _A ) __lowerCAmelCase = TFVisionTextDualEncoderModel(vision_model=_A , text_model=_A ) __lowerCAmelCase = model( input_ids=_A , pixel_values=_A , attention_mask=_A , output_attentions=_A ) __lowerCAmelCase = output.vision_model_output.attentions self.assertEqual(len(_A ) , vision_config.num_hidden_layers ) # in DEiT, the seq_len equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) __lowerCAmelCase = to_atuple(vision_model.config.image_size ) __lowerCAmelCase = to_atuple(vision_model.config.patch_size ) __lowerCAmelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) __lowerCAmelCase = num_patches + 2 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) __lowerCAmelCase = output.text_model_output.attentions self.assertEqual(len(_A ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def __SCREAMING_SNAKE_CASE( self , _A , _A ): """simple docstring""" __lowerCAmelCase = TFDeiTModel(_A , name="vision_model" ) __lowerCAmelCase = TFRobertaModel(_A , name="text_model" ) return vision_model, text_model def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = TFDeiTModelTester(self ) __lowerCAmelCase = TFRobertaModelTester(self ) __lowerCAmelCase = vit_model_tester.prepare_config_and_inputs() __lowerCAmelCase = bert_model_tester.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = vision_config_and_inputs ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class a__ ( snake_case__ , unittest.TestCase ): def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = TFVisionTextDualEncoderModel.from_vision_text_pretrained( "Rocketknight1/tiny-random-clip-tf" , "hf-internal-testing/tiny-random-bert" ) __lowerCAmelCase = 1_3 __lowerCAmelCase = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) __lowerCAmelCase = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) __lowerCAmelCase = random_attention_mask([batch_size, 4] ) __lowerCAmelCase = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def __SCREAMING_SNAKE_CASE( self , _A , _A ): """simple docstring""" __lowerCAmelCase = TFCLIPVisionModel(_A , name="vision_model" ) __lowerCAmelCase = TFBertModel(_A , name="text_model" ) return vision_model, text_model def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = TFCLIPVisionModelTester(self ) __lowerCAmelCase = TFBertModelTester(self ) __lowerCAmelCase = clip_model_tester.prepare_config_and_inputs() __lowerCAmelCase = bert_model_tester.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase = vision_config_and_inputs ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_vision @require_tf class a__ ( unittest.TestCase ): @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = TFVisionTextDualEncoderModel.from_pretrained( "clip-italian/clip-italian" , logit_scale_init_value=1.0 , from_pt=_A ) __lowerCAmelCase = VisionTextDualEncoderProcessor.from_pretrained("clip-italian/clip-italian" ) __lowerCAmelCase = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) __lowerCAmelCase = processor( text=["una foto di un gatto", "una foto di un cane"] , images=_A , padding=_A , return_tensors="np" ) __lowerCAmelCase = model(**_A ) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) __lowerCAmelCase = np.array([[1.2_28_47_27, 0.3_10_41_22]] ) self.assertTrue(np.allclose(outputs.logits_per_image.numpy() , _A , atol=1E-3 ) )
92
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCAmelCase = { """configuration_time_series_transformer""": [ """TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TimeSeriesTransformerConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = [ """TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TimeSeriesTransformerForPrediction""", """TimeSeriesTransformerModel""", """TimeSeriesTransformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys _lowerCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
353
'''simple docstring''' import math import random from typing import Any from .hill_climbing import SearchProblem def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase = True , UpperCamelCase = math.inf , UpperCamelCase = -math.inf , UpperCamelCase = math.inf , UpperCamelCase = -math.inf , UpperCamelCase = False , UpperCamelCase = 100 , UpperCamelCase = 0.01 , UpperCamelCase = 1 , ): """simple docstring""" lowerCAmelCase__ : int = False lowerCAmelCase__ : Optional[Any] = search_prob lowerCAmelCase__ : Tuple = start_temperate lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : str = 0 lowerCAmelCase__ : Tuple = None while not search_end: lowerCAmelCase__ : List[str] = current_state.score() if best_state is None or current_score > best_state.score(): lowerCAmelCase__ : List[str] = current_state scores.append(UpperCamelCase ) iterations += 1 lowerCAmelCase__ : Dict = None lowerCAmelCase__ : List[Any] = current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to lowerCAmelCase__ : Tuple = random.randint(0 , len(UpperCamelCase ) - 1 ) # picking a random neighbor lowerCAmelCase__ : str = neighbors.pop(UpperCamelCase ) lowerCAmelCase__ : Optional[Any] = picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: lowerCAmelCase__ : Union[str, Any] = change * -1 # in case we are finding minimum if change > 0: # improves the solution lowerCAmelCase__ : Tuple = picked_neighbor else: lowerCAmelCase__ : List[Any] = (math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability lowerCAmelCase__ : List[Any] = picked_neighbor lowerCAmelCase__ : Optional[int] = current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor lowerCAmelCase__ : str = True else: lowerCAmelCase__ : Optional[int] = next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(UpperCamelCase ) , UpperCamelCase ) plt.xlabel("""Iterations""" ) plt.ylabel("""Function values""" ) plt.show() return best_state if __name__ == "__main__": def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase ): """simple docstring""" return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) _lowerCAmelCase = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) _lowerCAmelCase = simulated_annealing( prob, find_max=False, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( '''The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) # starting the problem with initial coordinates (12, 47) _lowerCAmelCase = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) _lowerCAmelCase = simulated_annealing( prob, find_max=True, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( '''The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase ): """simple docstring""" return (3 * x**2) - (6 * y) _lowerCAmelCase = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) _lowerCAmelCase = simulated_annealing(prob, find_max=False, visualization=True) print( '''The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F"""{local_min.score()}""" ) _lowerCAmelCase = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) _lowerCAmelCase = simulated_annealing(prob, find_max=True, visualization=True) print( '''The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F"""{local_min.score()}""" )
184
0
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def lowercase__ ( __snake_case : List[str] , __snake_case : Tuple , __snake_case : List[str] , __snake_case : Dict ): '''simple docstring''' if isinstance(__snake_case , __snake_case ): UpperCAmelCase_ : Dict = np.full((len(__snake_case ), sequence_length, 2) , __snake_case ) else: UpperCAmelCase_ : Dict = np.full((len(__snake_case ), sequence_length) , __snake_case ) for i, tensor in enumerate(__snake_case ): if padding_side == "right": if isinstance(__snake_case , __snake_case ): UpperCAmelCase_ : Optional[Any] = tensor[:sequence_length] else: UpperCAmelCase_ : str = tensor[:sequence_length] else: if isinstance(__snake_case , __snake_case ): UpperCAmelCase_ : str = tensor[:sequence_length] else: UpperCAmelCase_ : Union[str, Any] = tensor[:sequence_length] return out_tensor.tolist() def lowercase__ ( __snake_case : Optional[int] ): '''simple docstring''' UpperCAmelCase_ : Optional[Any] = ord(__snake_case ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 123 and cp <= 126): return True UpperCAmelCase_ : str = unicodedata.category(__snake_case ) if cat.startswith('P' ): return True return False @dataclass class lowerCamelCase (_snake_case ): '''simple docstring''' _snake_case : Optional[Any] = 4_2 _snake_case : List[str] = True _snake_case : List[Any] = None _snake_case : int = None _snake_case : Union[str, Any] = -1_0_0 _snake_case : Optional[int] = '''pt''' def __UpperCAmelCase ( self , _UpperCamelCase ) -> Tuple: import torch UpperCAmelCase_ : Dict = 'label' if 'label' in features[0].keys() else 'labels' UpperCAmelCase_ : Any = [feature[label_name] for feature in features] if label_name in features[0].keys() else None UpperCAmelCase_ : str = self.tokenizer.pad( __UpperCAmelCase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='pt' if labels is None else None , ) if labels is None: return batch UpperCAmelCase_ : Dict = torch.tensor(batch['entity_ids'] ).shape[1] UpperCAmelCase_ : int = self.tokenizer.padding_side if padding_side == "right": UpperCAmelCase_ : Optional[int] = [ list(__UpperCAmelCase ) + [self.label_pad_token_id] * (sequence_length - len(__UpperCAmelCase )) for label in labels ] else: UpperCAmelCase_ : Optional[Any] = [ [self.label_pad_token_id] * (sequence_length - len(__UpperCAmelCase )) + list(__UpperCAmelCase ) for label in labels ] UpperCAmelCase_ : Optional[int] = [feature['ner_tags'] for feature in features] UpperCAmelCase_ : Union[str, Any] = padding_tensor(__UpperCAmelCase , -1 , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase_ : List[str] = [feature['original_entity_spans'] for feature in features] UpperCAmelCase_ : List[Any] = padding_tensor(__UpperCAmelCase , (-1, -1) , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase_ : Union[str, Any] = {k: torch.tensor(__UpperCAmelCase , dtype=torch.intaa ) for k, v in batch.items()} return batch
29
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_electra import ElectraTokenizer UpperCAmelCase__ = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} UpperCAmelCase__ = { "vocab_file": { "google/electra-small-generator": ( "https://huggingface.co/google/electra-small-generator/resolve/main/vocab.txt" ), "google/electra-base-generator": "https://huggingface.co/google/electra-base-generator/resolve/main/vocab.txt", "google/electra-large-generator": ( "https://huggingface.co/google/electra-large-generator/resolve/main/vocab.txt" ), "google/electra-small-discriminator": ( "https://huggingface.co/google/electra-small-discriminator/resolve/main/vocab.txt" ), "google/electra-base-discriminator": ( "https://huggingface.co/google/electra-base-discriminator/resolve/main/vocab.txt" ), "google/electra-large-discriminator": ( "https://huggingface.co/google/electra-large-discriminator/resolve/main/vocab.txt" ), }, "tokenizer_file": { "google/electra-small-generator": ( "https://huggingface.co/google/electra-small-generator/resolve/main/tokenizer.json" ), "google/electra-base-generator": ( "https://huggingface.co/google/electra-base-generator/resolve/main/tokenizer.json" ), "google/electra-large-generator": ( "https://huggingface.co/google/electra-large-generator/resolve/main/tokenizer.json" ), "google/electra-small-discriminator": ( "https://huggingface.co/google/electra-small-discriminator/resolve/main/tokenizer.json" ), "google/electra-base-discriminator": ( "https://huggingface.co/google/electra-base-discriminator/resolve/main/tokenizer.json" ), "google/electra-large-discriminator": ( "https://huggingface.co/google/electra-large-discriminator/resolve/main/tokenizer.json" ), }, } UpperCAmelCase__ = { "google/electra-small-generator": 512, "google/electra-base-generator": 512, "google/electra-large-generator": 512, "google/electra-small-discriminator": 512, "google/electra-base-discriminator": 512, "google/electra-large-discriminator": 512, } UpperCAmelCase__ = { "google/electra-small-generator": {"do_lower_case": True}, "google/electra-base-generator": {"do_lower_case": True}, "google/electra-large-generator": {"do_lower_case": True}, "google/electra-small-discriminator": {"do_lower_case": True}, "google/electra-base-discriminator": {"do_lower_case": True}, "google/electra-large-discriminator": {"do_lower_case": True}, } class lowercase_ ( lowercase ): '''simple docstring''' __snake_case = VOCAB_FILES_NAMES __snake_case = PRETRAINED_VOCAB_FILES_MAP __snake_case = PRETRAINED_INIT_CONFIGURATION __snake_case = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case = ElectraTokenizer def __init__( self : Dict , __UpperCAmelCase : int=None , __UpperCAmelCase : str=None , __UpperCAmelCase : Optional[int]=True , __UpperCAmelCase : str="[UNK]" , __UpperCAmelCase : Any="[SEP]" , __UpperCAmelCase : str="[PAD]" , __UpperCAmelCase : Optional[Any]="[CLS]" , __UpperCAmelCase : Union[str, Any]="[MASK]" , __UpperCAmelCase : List[str]=True , __UpperCAmelCase : Tuple=None , **__UpperCAmelCase : Optional[int] , ) ->str: """simple docstring""" super().__init__( __UpperCAmelCase , tokenizer_file=__UpperCAmelCase , do_lower_case=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , tokenize_chinese_chars=__UpperCAmelCase , strip_accents=__UpperCAmelCase , **__UpperCAmelCase , ) a = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , __UpperCAmelCase ) != do_lower_case or normalizer_state.get('''strip_accents''' , __UpperCAmelCase ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , __UpperCAmelCase ) != tokenize_chinese_chars ): a = getattr(__UpperCAmelCase , normalizer_state.pop('''type''' ) ) a = do_lower_case a = strip_accents a = tokenize_chinese_chars a = normalizer_class(**__UpperCAmelCase ) a = do_lower_case def __lowerCAmelCase ( self : List[Any] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Tuple=None ) ->str: """simple docstring""" a = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __lowerCAmelCase ( self : Optional[Any] , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) ->List[int]: """simple docstring""" a = [self.sep_token_id] a = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __lowerCAmelCase ( self : List[str] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) ->Tuple[str]: """simple docstring""" a = self._tokenizer.model.save(__UpperCAmelCase , name=__UpperCAmelCase ) return tuple(__UpperCAmelCase )
0
0
"""simple docstring""" import unittest from transformers import PegasusTokenizer, PegasusTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin a_ = get_tests_dir("""fixtures/test_sentencepiece_no_bos.model""") @require_sentencepiece @require_tokenizers class __snake_case ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = PegasusTokenizer _lowerCamelCase = PegasusTokenizerFast _lowerCamelCase = True _lowerCamelCase = True def UpperCamelCase__( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __A : Union[str, Any] = PegasusTokenizer(__lowerCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def UpperCamelCase__( self ): '''simple docstring''' return PegasusTokenizer.from_pretrained('''google/pegasus-large''' ) def UpperCamelCase__( self , **__lowerCamelCase ): '''simple docstring''' return PegasusTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' return ("This is a test", "This is a test") def UpperCamelCase__( self ): '''simple docstring''' __A : List[Any] = '''</s>''' __A : Union[str, Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowerCamelCase ) , __lowerCamelCase ) def UpperCamelCase__( self ): '''simple docstring''' __A : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<pad>''' ) self.assertEqual(vocab_keys[1] , '''</s>''' ) self.assertEqual(vocab_keys[-1] , '''v''' ) self.assertEqual(len(__lowerCamelCase ) , 1103 ) def UpperCamelCase__( self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1103 ) def UpperCamelCase__( self ): '''simple docstring''' __A : Dict = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) __A : Tuple = self.tokenizer_class.from_pretrained(self.tmpdirname ) __A : Optional[int] = ( '''Let\'s see which <unk> is the better <unk_token_11> one <mask_1> It seems like this <mask_2> was important''' ''' </s> <pad> <pad> <pad>''' ) __A : List[Any] = rust_tokenizer([raw_input_str] , return_tensors=__lowerCamelCase , add_special_tokens=__lowerCamelCase ).input_ids[0] __A : List[str] = py_tokenizer([raw_input_str] , return_tensors=__lowerCamelCase , add_special_tokens=__lowerCamelCase ).input_ids[0] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def UpperCamelCase__( self ): '''simple docstring''' __A : Optional[int] = self._large_tokenizer # <mask_1> masks whole sentence while <mask_2> masks single word __A : Any = '''<mask_1> To ensure a <mask_2> flow of bank resolutions.''' __A : Optional[int] = [2, 413, 615, 114, 3, 1971, 113, 1679, 1_0710, 107, 1] __A : str = tokenizer([raw_input_str] , return_tensors=__lowerCamelCase ).input_ids[0] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def UpperCamelCase__( self ): '''simple docstring''' __A : Union[str, Any] = self._large_tokenizer # The tracebacks for the following asserts are **better** without messages or self.assertEqual assert tokenizer.vocab_size == 9_6103 assert tokenizer.pad_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.offset == 103 assert tokenizer.unk_token_id == tokenizer.offset + 2 == 105 assert tokenizer.unk_token == "<unk>" assert tokenizer.model_max_length == 1024 __A : Optional[int] = '''To ensure a smooth flow of bank resolutions.''' __A : str = [413, 615, 114, 2291, 1971, 113, 1679, 1_0710, 107, 1] __A : Dict = tokenizer([raw_input_str] , return_tensors=__lowerCamelCase ).input_ids[0] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) assert tokenizer.convert_ids_to_tokens([0, 1, 2, 3] ) == ["<pad>", "</s>", "<mask_1>", "<mask_2>"] @require_torch def UpperCamelCase__( self ): '''simple docstring''' __A : List[str] = ['''This is going to be way too long.''' * 150, '''short example'''] __A : Dict = ['''not super long but more than 5 tokens''', '''tiny'''] __A : List[str] = self._large_tokenizer(__lowerCamelCase , padding=__lowerCamelCase , truncation=__lowerCamelCase , return_tensors='''pt''' ) __A : int = self._large_tokenizer( text_target=__lowerCamelCase , max_length=5 , padding=__lowerCamelCase , truncation=__lowerCamelCase , return_tensors='''pt''' ) assert batch.input_ids.shape == (2, 1024) assert batch.attention_mask.shape == (2, 1024) assert targets["input_ids"].shape == (2, 5) assert len(__lowerCamelCase ) == 2 # input_ids, attention_mask. @slow def UpperCamelCase__( self ): '''simple docstring''' __A : int = {'''input_ids''': [[3_8979, 143, 1_8485, 606, 130, 2_6669, 8_7686, 121, 5_4189, 1129, 111, 2_6669, 8_7686, 121, 9114, 1_4787, 121, 1_3249, 158, 592, 956, 121, 1_4621, 3_1576, 143, 6_2613, 108, 9688, 930, 4_3430, 1_1562, 6_2613, 304, 108, 1_1443, 897, 108, 9314, 1_7415, 6_3399, 108, 1_1443, 7614, 1_8316, 118, 4284, 7148, 1_2430, 143, 1400, 2_5703, 158, 111, 4284, 7148, 1_1772, 143, 2_1297, 1064, 158, 122, 204, 3506, 1754, 1133, 1_4787, 1581, 115, 3_3224, 4482, 111, 1355, 110, 2_9173, 317, 5_0833, 108, 2_0147, 9_4665, 111, 7_7198, 107, 1], [110, 6_2613, 117, 638, 112, 1133, 121, 2_0098, 1355, 7_9050, 1_3872, 135, 1596, 5_3541, 1352, 141, 1_3039, 5542, 124, 302, 518, 111, 268, 2956, 115, 149, 4427, 107, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [139, 1235, 2799, 1_8289, 1_7780, 204, 109, 9474, 1296, 107, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowerCamelCase , model_name='''google/bigbird-pegasus-large-arxiv''' , revision='''ba85d0851d708441f91440d509690f1ab6353415''' , ) @require_sentencepiece @require_tokenizers class __snake_case ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = PegasusTokenizer _lowerCamelCase = PegasusTokenizerFast _lowerCamelCase = True _lowerCamelCase = True def UpperCamelCase__( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __A : List[str] = PegasusTokenizer(__lowerCamelCase , offset=0 , mask_token_sent=__lowerCamelCase , mask_token='''[MASK]''' ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def UpperCamelCase__( self ): '''simple docstring''' return PegasusTokenizer.from_pretrained('''google/bigbird-pegasus-large-arxiv''' ) def UpperCamelCase__( self , **__lowerCamelCase ): '''simple docstring''' return PegasusTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' return ("This is a test", "This is a test") def UpperCamelCase__( self ): '''simple docstring''' __A : List[Any] = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) __A : int = self.tokenizer_class.from_pretrained(self.tmpdirname ) __A : Optional[Any] = ( '''Let\'s see which <unk> is the better <unk_token> one [MASK] It seems like this [MASK] was important </s>''' ''' <pad> <pad> <pad>''' ) __A : List[Any] = rust_tokenizer([raw_input_str] , return_tensors=__lowerCamelCase , add_special_tokens=__lowerCamelCase ).input_ids[0] __A : int = py_tokenizer([raw_input_str] , return_tensors=__lowerCamelCase , add_special_tokens=__lowerCamelCase ).input_ids[0] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) @require_torch def UpperCamelCase__( self ): '''simple docstring''' __A : Optional[Any] = ['''This is going to be way too long.''' * 1000, '''short example'''] __A : List[Any] = ['''not super long but more than 5 tokens''', '''tiny'''] __A : List[str] = self._large_tokenizer(__lowerCamelCase , padding=__lowerCamelCase , truncation=__lowerCamelCase , return_tensors='''pt''' ) __A : Dict = self._large_tokenizer( text_target=__lowerCamelCase , max_length=5 , padding=__lowerCamelCase , truncation=__lowerCamelCase , return_tensors='''pt''' ) assert batch.input_ids.shape == (2, 4096) assert batch.attention_mask.shape == (2, 4096) assert targets["input_ids"].shape == (2, 5) assert len(__lowerCamelCase ) == 2 # input_ids, attention_mask. def UpperCamelCase__( self ): '''simple docstring''' __A : str = ( '''This is an example string that is used to test the original TF implementation against the HF''' ''' implementation''' ) __A : Optional[Any] = self._large_tokenizer(__lowerCamelCase ).input_ids self.assertListEqual( __lowerCamelCase , [182, 117, 142, 587, 4211, 120, 117, 263, 112, 804, 109, 856, 2_5016, 3137, 464, 109, 2_6955, 3137, 1] , )
291
"""simple docstring""" import numpy as np import qiskit def __lowercase ( snake_case_ : int = 8 ,snake_case_ : int | None = None ) ->str: '''simple docstring''' __A : str = np.random.default_rng(seed=snake_case_ ) # Roughly 25% of the qubits will contribute to the key. # So we take more than we need. __A : str = 6 * key_len # Measurement basis for Alice's qubits. __A : Any = rng.integers(2 ,size=snake_case_ ) # The set of states Alice will prepare. __A : Any = rng.integers(2 ,size=snake_case_ ) # Measurement basis for Bob's qubits. __A : str = rng.integers(2 ,size=snake_case_ ) # Quantum Circuit to simulate BB84 __A : Dict = qiskit.QuantumCircuit(snake_case_ ,name='''BB84''' ) # Alice prepares her qubits according to rules above. for index, _ in enumerate(snake_case_ ): if alice_state[index] == 1: bbaa_circ.x(snake_case_ ) if alice_basis[index] == 1: bbaa_circ.h(snake_case_ ) bbaa_circ.barrier() # Bob measures the received qubits according to rules above. for index, _ in enumerate(snake_case_ ): if bob_basis[index] == 1: bbaa_circ.h(snake_case_ ) bbaa_circ.barrier() bbaa_circ.measure_all() # Simulate the quantum circuit. __A : List[str] = qiskit.Aer.get_backend('''aer_simulator''' ) # We only need to run one shot because the key is unique. # Multiple shots will produce the same key. __A : List[str] = qiskit.execute(snake_case_ ,snake_case_ ,shots=1 ,seed_simulator=snake_case_ ) # Returns the result of measurement. __A : Union[str, Any] = job.result().get_counts(snake_case_ ).most_frequent() # Extracting the generated key from the simulation results. # Only keep measurement results where Alice and Bob chose the same basis. __A : int = ''''''.join( [ result_bit for alice_basis_bit, bob_basis_bit, result_bit in zip( snake_case_ ,snake_case_ ,snake_case_ ) if alice_basis_bit == bob_basis_bit ] ) # Get final key. Pad with 0 if too short, otherwise truncate. __A : Union[str, Any] = gen_key[:key_len] if len(snake_case_ ) >= key_len else gen_key.ljust(snake_case_ ,'''0''' ) return key if __name__ == "__main__": print(f'''The generated key is : {bbaa(8, seed=0)}''') from doctest import testmod testmod()
291
1
from __future__ import annotations import numpy as np def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> str: return np.maximum(0 , __lowerCAmelCase ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
212
import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device lowerCamelCase : Tuple =False class __a ( unittest.TestCase ): pass @slow @require_torch_gpu class __a ( unittest.TestCase ): def __lowercase ( self : int ): '''simple docstring''' UpperCamelCase__ : Optional[Any] = VersatileDiffusionImageVariationPipeline.from_pretrained("shi-labs/versatile-diffusion" ) pipe.to(SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Union[str, Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) UpperCamelCase__ : List[Any] = torch.manual_seed(0 ) UpperCamelCase__ : List[Any] = pipe( image=SCREAMING_SNAKE_CASE , generator=SCREAMING_SNAKE_CASE , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" , ).images UpperCamelCase__ : Union[str, Any] = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) UpperCamelCase__ : Optional[Any] = np.array([0.0_4_4_1, 0.0_4_6_9, 0.0_5_0_7, 0.0_5_7_5, 0.0_6_3_2, 0.0_6_5_0, 0.0_8_6_5, 0.0_9_0_9, 0.0_9_4_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
189
0
'''simple docstring''' from pathlib import Path import numpy as np from PIL import Image def _A ( snake_case ) -> np.ndarray: _lowercase , _lowercase , _lowercase : List[str] = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.2989 * r + 0.5870 * g + 0.1140 * b def _A ( snake_case ) -> np.ndarray: return (gray > 1_27) & (gray <= 2_55) def _A ( snake_case , snake_case ) -> np.ndarray: _lowercase : Dict = np.zeros_like(snake_case ) _lowercase : Union[str, Any] = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image _lowercase : str = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): _lowercase : Dict = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() _lowercase : List[Any] = int(summation > 0 ) return output if __name__ == "__main__": # read original image _snake_case = Path(__file__).resolve().parent / 'image_data' / 'lena.jpg' _snake_case = np.array(Image.open(lena_path)) # kernel to be applied _snake_case = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) _snake_case = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image _snake_case = Image.fromarray(output).convert('RGB') pil_img.save('result_dilation.png')
199
'''simple docstring''' from __future__ import annotations def _A ( snake_case , snake_case = None , snake_case = None ) -> None: if start is None: _lowercase : Dict = 0 if end is None: _lowercase : List[Any] = len(snake_case ) - 1 if start >= end: return _lowercase : int = (start + end) // 2 slowsort(snake_case , snake_case , snake_case ) slowsort(snake_case , mid + 1 , snake_case ) if sequence[end] < sequence[mid]: _lowercase , _lowercase : Any = sequence[mid], sequence[end] slowsort(snake_case , snake_case , end - 1 ) if __name__ == "__main__": from doctest import testmod testmod()
199
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class _A ( unittest.TestCase ): @slow def A__ ( self ): """simple docstring""" lowercase = TFCamembertModel.from_pretrained("""jplu/tf-camembert-base""" ) lowercase = tf.convert_to_tensor( [[5, 121, 11, 660, 16, 730, 2_5543, 110, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" lowercase = model(__lowerCAmelCase )["""last_hidden_state"""] lowercase = tf.TensorShape((1, 10, 768) ) self.assertEqual(output.shape , __lowerCAmelCase ) # compare the actual values for a slice. lowercase = tf.convert_to_tensor( [[[-0.0_2_5_4, 0.0_2_3_5, 0.1_0_2_7], [0.0_6_0_6, -0.1_8_1_1, -0.0_4_1_8], [-0.1_5_6_1, -0.1_1_2_7, 0.2_6_8_7]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
197
"""simple docstring""" from __future__ import annotations import numpy as np def UpperCAmelCase__ ( lowerCAmelCase__ :list[float] ) -> Optional[Any]: '''simple docstring''' return np.maximum(0 , lowerCAmelCase__ ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
197
1
import json import os import shutil import tempfile from unittest import TestCase from transformers import BartTokenizer, BartTokenizerFast, DPRQuestionEncoderTokenizer, DPRQuestionEncoderTokenizerFast from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_tokenizers, require_torch, slow from transformers.utils import is_datasets_available, is_faiss_available, is_torch_available if is_torch_available() and is_datasets_available() and is_faiss_available(): from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.tokenization_rag import RagTokenizer @require_faiss @require_torch class __magic_name__ ( SCREAMING_SNAKE_CASE__ ): def UpperCAmelCase_ ( self )-> List[str]: UpperCamelCase_ = tempfile.mkdtemp() UpperCamelCase_ = 8 # DPR tok UpperCamelCase_ = [ "[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] UpperCamelCase_ = os.path.join(self.tmpdirname , "dpr_tokenizer" ) os.makedirs(_SCREAMING_SNAKE_CASE , exist_ok=_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = os.path.join(_SCREAMING_SNAKE_CASE , DPR_VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) # BART tok UpperCamelCase_ = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] UpperCamelCase_ = dict(zip(_SCREAMING_SNAKE_CASE , range(len(_SCREAMING_SNAKE_CASE ) ) ) ) UpperCamelCase_ = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] UpperCamelCase_ = {"unk_token": "<unk>"} UpperCamelCase_ = os.path.join(self.tmpdirname , "bart_tokenizer" ) os.makedirs(_SCREAMING_SNAKE_CASE , exist_ok=_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = os.path.join(_SCREAMING_SNAKE_CASE , BART_VOCAB_FILES_NAMES["vocab_file"] ) UpperCamelCase_ = os.path.join(_SCREAMING_SNAKE_CASE , BART_VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(_SCREAMING_SNAKE_CASE ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(_SCREAMING_SNAKE_CASE ) ) def UpperCAmelCase_ ( self )-> DPRQuestionEncoderTokenizer: return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , "dpr_tokenizer" ) ) def UpperCAmelCase_ ( self )-> BartTokenizer: return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , "bart_tokenizer" ) ) def UpperCAmelCase_ ( self )-> Tuple: shutil.rmtree(self.tmpdirname ) @require_tokenizers def UpperCAmelCase_ ( self )-> Optional[Any]: UpperCamelCase_ = os.path.join(self.tmpdirname , "rag_tokenizer" ) UpperCamelCase_ = RagConfig(question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() ) UpperCamelCase_ = RagTokenizer(question_encoder=self.get_dpr_tokenizer() , generator=self.get_bart_tokenizer() ) rag_config.save_pretrained(_SCREAMING_SNAKE_CASE ) rag_tokenizer.save_pretrained(_SCREAMING_SNAKE_CASE ) UpperCamelCase_ = RagTokenizer.from_pretrained(_SCREAMING_SNAKE_CASE , config=_SCREAMING_SNAKE_CASE ) self.assertIsInstance(new_rag_tokenizer.question_encoder , _SCREAMING_SNAKE_CASE ) self.assertEqual(new_rag_tokenizer.question_encoder.get_vocab() , rag_tokenizer.question_encoder.get_vocab() ) self.assertIsInstance(new_rag_tokenizer.generator , _SCREAMING_SNAKE_CASE ) self.assertEqual(new_rag_tokenizer.generator.get_vocab() , rag_tokenizer.generator.get_vocab() ) @slow def UpperCAmelCase_ ( self )-> List[Any]: UpperCamelCase_ = RagTokenizer.from_pretrained("facebook/rag-token-nq" ) UpperCamelCase_ = [ "who got the first nobel prize in physics", "when is the next deadpool movie being released", "which mode is used for short wave broadcast service", "who is the owner of reading football club", "when is the next scandal episode coming out", "when is the last time the philadelphia won the superbowl", "what is the most current adobe flash player version", "how many episodes are there in dragon ball z", "what is the first step in the evolution of the eye", "where is gall bladder situated in human body", "what is the main mineral in lithium batteries", "who is the president of usa right now", "where do the greasers live in the outsiders", "panda is a national animal of which country", "what is the name of manchester united stadium", ] UpperCamelCase_ = tokenizer(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) @slow def UpperCAmelCase_ ( self )-> Any: UpperCamelCase_ = RagTokenizer.from_pretrained("facebook/rag-sequence-nq" ) UpperCamelCase_ = [ "who got the first nobel prize in physics", "when is the next deadpool movie being released", "which mode is used for short wave broadcast service", "who is the owner of reading football club", "when is the next scandal episode coming out", "when is the last time the philadelphia won the superbowl", "what is the most current adobe flash player version", "how many episodes are there in dragon ball z", "what is the first step in the evolution of the eye", "where is gall bladder situated in human body", "what is the main mineral in lithium batteries", "who is the president of usa right now", "where do the greasers live in the outsiders", "panda is a national animal of which country", "what is the name of manchester united stadium", ] UpperCamelCase_ = tokenizer(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE )
364
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING import torch from ..models.auto import AutoModelForVisualQuestionAnswering, AutoProcessor from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class __magic_name__ ( snake_case ): UpperCamelCase_ :List[Any] = """dandelin/vilt-b32-finetuned-vqa""" UpperCamelCase_ :Dict = ( """This is a tool that answers a question about an image. It takes an input named `image` which should be the """ """image containing the information, as well as a `question` which should be the question in English. It """ """returns a text that is the answer to the question.""" ) UpperCamelCase_ :Optional[int] = """image_qa""" UpperCamelCase_ :int = AutoProcessor UpperCamelCase_ :Tuple = AutoModelForVisualQuestionAnswering UpperCamelCase_ :Optional[int] = ["""image""", """text"""] UpperCamelCase_ :Tuple = ["""text"""] def __init__( self , *_lowercase , **_lowercase )-> Union[str, Any]: requires_backends(self , ["vision"] ) super().__init__(*_lowercase , **_lowercase ) def UpperCAmelCase_ ( self , _lowercase , _lowercase )-> str: return self.pre_processor(_lowercase , _lowercase , return_tensors="pt" ) def UpperCAmelCase_ ( self , _lowercase )-> str: with torch.no_grad(): return self.model(**_lowercase ).logits def UpperCAmelCase_ ( self , _lowercase )-> List[Any]: UpperCamelCase_ = outputs.argmax(-1 ).item() return self.model.config.idalabel[idx]
60
0
import unittest from knapsack import greedy_knapsack as kp class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def UpperCamelCase ( self ): A__ = [10, 20, 30, 40, 50, 60] A__ = [2, 4, 6, 8, 10, 12] A__ = 100 self.assertEqual(kp.calc_profit(A_,A_,A_ ),210 ) def UpperCamelCase ( self ): self.assertRaisesRegex(A_,'''max_weight must greater than zero.''' ) def UpperCamelCase ( self ): self.assertRaisesRegex(A_,'''Weight can not be negative.''' ) def UpperCamelCase ( self ): self.assertRaisesRegex(A_,'''Profit can not be negative.''' ) def UpperCamelCase ( self ): self.assertRaisesRegex(A_,'''max_weight must greater than zero.''' ) def UpperCamelCase ( self ): self.assertRaisesRegex( A_,'''The length of profit and weight must be same.''' ) if __name__ == "__main__": unittest.main()
193
from __future__ import annotations from functools import lru_cache from math import ceil __lowerCamelCase : str = 100 __lowerCamelCase : Any = set(range(3, NUM_PRIMES, 2)) primes.add(2) __lowerCamelCase : int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=100 ) def A_ ( _lowerCAmelCase ) -> set[int]: if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} UpperCamelCase : set[int] = set() UpperCamelCase : int UpperCamelCase : int for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def A_ ( _lowerCAmelCase = 5000 ) -> int | None: for number_to_partition in range(1 , _lowerCAmelCase ): if len(partition(_lowerCAmelCase ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(f"""{solution() = }""")
52
0
"""simple docstring""" import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to properly calculate the metrics on the # validation dataset when in a distributed system, and builds off the # `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _snake_case = 16 _snake_case = 32 def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ = 1_6 ): '''simple docstring''' _a : str = AutoTokenizer.from_pretrained("""bert-base-cased""" ) _a : Dict = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(UpperCamelCase__ ): # max_length=None => use the model max length (it's actually the default) _a : Optional[int] = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=UpperCamelCase__ , max_length=UpperCamelCase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _a : Tuple = datasets.map( UpperCamelCase__ , batched=UpperCamelCase__ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _a : List[Any] = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(UpperCamelCase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. _a : Union[str, Any] = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _a : int = 1_6 elif accelerator.mixed_precision != "no": _a : int = 8 else: _a : str = None return tokenizer.pad( UpperCamelCase__ , padding="""longest""" , max_length=UpperCamelCase__ , pad_to_multiple_of=UpperCamelCase__ , return_tensors="""pt""" , ) # Instantiate dataloaders. _a : int = DataLoader( tokenized_datasets["""train"""] , shuffle=UpperCamelCase__ , collate_fn=UpperCamelCase__ , batch_size=UpperCamelCase__ ) _a : List[str] = DataLoader( tokenized_datasets["""validation"""] , shuffle=UpperCamelCase__ , collate_fn=UpperCamelCase__ , batch_size=UpperCamelCase__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders _snake_case = mocked_dataloaders # noqa: F811 def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , UpperCamelCase__ ) == "1": _a : str = 2 # Initialize accelerator _a : int = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _a : Any = config["""lr"""] _a : Union[str, Any] = int(config["""num_epochs"""] ) _a : str = int(config["""seed"""] ) _a : List[Any] = int(config["""batch_size"""] ) _a : Tuple = evaluate.load("""glue""" , """mrpc""" ) # If the batch size is too big we use gradient accumulation _a : Optional[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: _a : Optional[Any] = batch_size // MAX_GPU_BATCH_SIZE _a : str = MAX_GPU_BATCH_SIZE set_seed(UpperCamelCase__ ) _a : Optional[int] = get_dataloaders(UpperCamelCase__ , UpperCamelCase__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _a : int = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=UpperCamelCase__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _a : List[str] = model.to(accelerator.device ) # Instantiate optimizer _a : List[str] = AdamW(params=model.parameters() , lr=UpperCamelCase__ ) # Instantiate scheduler _a : List[str] = get_linear_schedule_with_warmup( optimizer=UpperCamelCase__ , num_warmup_steps=1_0_0 , num_training_steps=(len(UpperCamelCase__ ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _a : Optional[Any] = accelerator.prepare( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Now we train the model for epoch in range(UpperCamelCase__ ): model.train() for step, batch in enumerate(UpperCamelCase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _a : Optional[Any] = model(**UpperCamelCase__ ) _a : str = outputs.loss _a : Optional[int] = loss / gradient_accumulation_steps accelerator.backward(UpperCamelCase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() _a : Union[str, Any] = 0 for step, batch in enumerate(UpperCamelCase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Dict = model(**UpperCamelCase__ ) _a : Optional[Any] = outputs.logits.argmax(dim=-1 ) _a : int = accelerator.gather((predictions, batch["""labels"""]) ) # New Code # # First we check if it's a distributed system if accelerator.use_distributed: # Then see if we're on the last batch of our eval dataloader if step == len(UpperCamelCase__ ) - 1: # Last batch needs to be truncated on distributed systems as it contains additional samples _a : str = predictions[: len(eval_dataloader.dataset ) - samples_seen] _a : int = references[: len(eval_dataloader.dataset ) - samples_seen] else: # Otherwise we add the number of samples seen samples_seen += references.shape[0] # All of this can be avoided if you use `Accelerator.gather_for_metrics` instead of `Accelerator.gather`: # accelerator.gather_for_metrics((predictions, batch["labels"])) metric.add_batch( predictions=UpperCamelCase__ , references=UpperCamelCase__ , ) _a : int = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"""epoch {epoch}:""" , UpperCamelCase__ ) def lowerCAmelCase__ ( ): '''simple docstring''' _a : Tuple = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=UpperCamelCase__ , default=UpperCamelCase__ , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) _a : Optional[Any] = parser.parse_args() _a : Tuple = {"""lr""": 2e-5, """num_epochs""": 3, """seed""": 4_2, """batch_size""": 1_6} training_function(UpperCamelCase__ , UpperCamelCase__ ) if __name__ == "__main__": main()
350
"""simple docstring""" _snake_case = 8.31_44_62 # Unit - J mol-1 K-1 def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if moles < 0 or kelvin < 0 or volume < 0: raise ValueError("""Invalid inputs. Enter positive value.""" ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / volume def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if moles < 0 or kelvin < 0 or pressure < 0: raise ValueError("""Invalid inputs. Enter positive value.""" ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / pressure if __name__ == "__main__": from doctest import testmod testmod()
324
0
'''simple docstring''' from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def UpperCAmelCase ( a_ ) -> Dict[str, torch.Tensor]: """simple docstring""" A_ : List[str] = [] A_ : Dict = [] A_ : List[Any] = [] for rt in rc.restypes: A_ : Tuple = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) A_ : Union[str, Any] = {name: i for i, name in enumerate(a_ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 1_4 ) restype_atomaa_to_atomaa_list.append([0] * 3_7 ) restype_atomaa_mask_list.append([0.0] * 1_4 ) A_ : Tuple = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) A_ : Optional[int] = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) A_ : List[Any] = torch.tensor( a_ , dtype=torch.floataa , device=protein["""aatype"""].device , ) A_ : Optional[int] = protein["""aatype"""].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein A_ : Dict = restype_atomaa_to_atomaa[protein_aatype] A_ : Optional[Any] = restype_atomaa_mask[protein_aatype] A_ : Any = residx_atomaa_mask A_ : List[str] = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back A_ : Tuple = restype_atomaa_to_atomaa[protein_aatype] A_ : Tuple = residx_atomaa_to_atomaa.long() # create the corresponding mask A_ : Optional[Any] = torch.zeros([2_1, 3_7] , dtype=torch.floataa , device=protein["""aatype"""].device ) for restype, restype_letter in enumerate(rc.restypes ): A_ : Optional[Any] = rc.restype_atoa[restype_letter] A_ : Any = rc.residue_atoms[restype_name] for atom_name in atom_names: A_ : Any = rc.atom_order[atom_name] A_ : Optional[int] = 1 A_ : Optional[int] = restype_atomaa_mask[protein_aatype] A_ : Dict = residx_atomaa_mask return protein def UpperCAmelCase ( a_ ) -> Dict[str, np.ndarray]: """simple docstring""" A_ : Union[str, Any] = tree_map(lambda a_ : torch.tensor(a_ , device=batch["""aatype"""].device ) , a_ , np.ndarray ) A_ : Optional[int] = tensor_tree_map(lambda a_ : np.array(a_ ) , make_atomaa_masks(a_ ) ) return out
344
'''simple docstring''' import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=30 , _lowerCamelCase=2 , _lowerCamelCase=3 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=32 , _lowerCamelCase=5 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=10 , _lowerCamelCase=0.02 , _lowerCamelCase=3 , _lowerCamelCase=None , _lowerCamelCase=2 , ) -> str: A_ : Optional[int] = parent A_ : Dict = batch_size A_ : List[Any] = image_size A_ : Optional[int] = patch_size A_ : List[str] = num_channels A_ : List[Any] = is_training A_ : Union[str, Any] = use_labels A_ : Union[str, Any] = hidden_size A_ : str = num_hidden_layers A_ : List[str] = num_attention_heads A_ : Union[str, Any] = intermediate_size A_ : Any = hidden_act A_ : Optional[Any] = hidden_dropout_prob A_ : List[Any] = attention_probs_dropout_prob A_ : Dict = type_sequence_label_size A_ : Optional[int] = initializer_range A_ : str = scope A_ : Optional[Any] = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) A_ : Tuple = (image_size // patch_size) ** 2 A_ : Union[str, Any] = num_patches + 2 def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Dict = None if self.use_labels: A_ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Optional[Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self ) -> int: return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : List[str] = DeiTModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Dict = model(_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : int = DeiTForMaskedImageModeling(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images A_ : Dict = 1 A_ : Optional[int] = DeiTForMaskedImageModeling(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : int = model(_lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: A_ : Tuple = self.type_sequence_label_size A_ : Tuple = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images A_ : Dict = 1 A_ : Any = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : str = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : List[str] = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : List[Any] = self.prepare_config_and_inputs() ( ( A_ ) , ( A_ ) , ( A_ ) , ) : Union[str, Any] = config_and_inputs A_ : Tuple = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) lowerCamelCase = ( { '''feature-extraction''': DeiTModel, '''image-classification''': (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : int = DeiTModelTester(self ) A_ : str = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase , hidden_size=37 ) def UpperCAmelCase_ ( self ) -> List[str]: self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def UpperCAmelCase_ ( self ) -> Optional[int]: pass def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ , A_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[Any] = model_class(_lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) A_ : Union[str, Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCamelCase , nn.Linear ) ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[str] = model_class(_lowerCamelCase ) A_ : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : Union[str, Any] = [*signature.parameters.keys()] A_ : List[str] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=False ) -> Union[str, Any]: A_ : int = super()._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def UpperCAmelCase_ ( self ) -> Optional[Any]: if not self.model_tester.is_training: return A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A_ : Optional[Any] = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(_lowerCamelCase ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue A_ : List[str] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() A_ : List[str] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : List[str] = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> int: A_ , A_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return A_ : Any = False A_ : Union[str, Any] = True for model_class in self.all_model_classes: if model_class in get_values(_lowerCamelCase ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue A_ : List[Any] = model_class(_lowerCamelCase ) model.gradient_checkpointing_enable() model.to(_lowerCamelCase ) model.train() A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Union[str, Any] = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> Tuple: A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A_ : Optional[Any] = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(_lowerCamelCase ), *get_values(_lowerCamelCase ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F"Testing {model_class} with {problem_type['title']}" ): A_ : Dict = problem_type["""title"""] A_ : List[Any] = problem_type["""num_labels"""] A_ : List[str] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() A_ : List[Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if problem_type["num_labels"] > 1: A_ : Tuple = inputs["""labels"""].unsqueeze(1 ).repeat(1 , problem_type["""num_labels"""] ) A_ : Union[str, Any] = inputs["""labels"""].to(problem_type["""dtype"""] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=_lowerCamelCase ) as warning_list: A_ : List[str] = model(**_lowerCamelCase ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F"Something is going wrong in the regression problem: intercepted {w.message}" ) loss.backward() @slow def UpperCAmelCase_ ( self ) -> Tuple: for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : int = DeiTModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" A_ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase_ ( self ) -> Optional[Any]: return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def UpperCAmelCase_ ( self ) -> Tuple: A_ : Any = DeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ).to( _lowerCamelCase ) A_ : Optional[int] = self.default_image_processor A_ : str = prepare_img() A_ : Any = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ).to(_lowerCamelCase ) # forward pass with torch.no_grad(): A_ : Any = model(**_lowerCamelCase ) # verify the logits A_ : Tuple = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : List[Any] = torch.tensor([-1.0266, 0.1912, -1.2861] ).to(_lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = DeiTModel.from_pretrained( """facebook/deit-base-distilled-patch16-224""" , torch_dtype=torch.floataa , device_map="""auto""" ) A_ : int = self.default_image_processor A_ : List[str] = prepare_img() A_ : List[Any] = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ) A_ : Union[str, Any] = inputs.pixel_values.to(_lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): A_ : List[Any] = model(_lowerCamelCase )
344
1
import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase ( _SCREAMING_SNAKE_CASE , unittest.TestCase ): __lowercase : Dict = CodeGenTokenizer __lowercase : Tuple = CodeGenTokenizerFast __lowercase : List[Any] = True __lowercase : str = {"add_prefix_space": True} __lowercase : Optional[int] = False def __UpperCamelCase ( self ) -> List[Any]: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCamelCase = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', '<|endoftext|>', ] UpperCamelCase = dict(zip(A_ , range(len(A_ ) ) ) ) UpperCamelCase = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] UpperCamelCase = {'unk_token': '<unk>'} UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(A_ ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(A_ ) ) def __UpperCamelCase ( self , **A_ ) -> Dict: """simple docstring""" kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase ( self , **A_ ) -> List[Any]: """simple docstring""" kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase ( self , A_ ) -> str: """simple docstring""" UpperCamelCase = 'lower newer' UpperCamelCase = 'lower newer' return input_text, output_text def __UpperCamelCase ( self ) -> List[str]: """simple docstring""" UpperCamelCase = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) UpperCamelCase = 'lower newer' UpperCamelCase = ['\u0120low', 'er', '\u0120', 'n', 'e', 'w', 'er'] UpperCamelCase = tokenizer.tokenize(A_ , add_prefix_space=A_ ) self.assertListEqual(A_ , A_ ) UpperCamelCase = tokens + [tokenizer.unk_token] UpperCamelCase = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , A_ ) def __UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" if not self.test_rust_tokenizer: return UpperCamelCase = self.get_tokenizer() UpperCamelCase = self.get_rust_tokenizer(add_prefix_space=A_ ) UpperCamelCase = 'lower newer' # Testing tokenization UpperCamelCase = tokenizer.tokenize(A_ , add_prefix_space=A_ ) UpperCamelCase = rust_tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) # Testing conversion to ids without special tokens UpperCamelCase = tokenizer.encode(A_ , add_special_tokens=A_ , add_prefix_space=A_ ) UpperCamelCase = rust_tokenizer.encode(A_ , add_special_tokens=A_ ) self.assertListEqual(A_ , A_ ) # Testing conversion to ids with special tokens UpperCamelCase = self.get_rust_tokenizer(add_prefix_space=A_ ) UpperCamelCase = tokenizer.encode(A_ , add_prefix_space=A_ ) UpperCamelCase = rust_tokenizer.encode(A_ ) self.assertListEqual(A_ , A_ ) # Testing the unknown token UpperCamelCase = tokens + [rust_tokenizer.unk_token] UpperCamelCase = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(A_ ) , A_ ) def __UpperCamelCase ( self , *A_ , **A_ ) -> Optional[int]: """simple docstring""" # It's very difficult to mix/test pretokenization with byte-level # And get both CodeGen and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def __UpperCamelCase ( self , A_=15 ) -> List[str]: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): UpperCamelCase = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) # Simple input UpperCamelCase = 'This is a simple input' UpperCamelCase = ['This is a simple input 1', 'This is a simple input 2'] UpperCamelCase = ('This is a simple input', 'This is a pair') UpperCamelCase = [ ('This is a simple input 1', 'This is a simple input 2'), ('This is a simple pair 1', 'This is a simple pair 2'), ] # Simple input tests self.assertRaises(A_ , tokenizer_r.encode , A_ , max_length=A_ , padding='max_length' ) # Simple input self.assertRaises(A_ , tokenizer_r.encode_plus , A_ , max_length=A_ , padding='max_length' ) # Simple input self.assertRaises( A_ , tokenizer_r.batch_encode_plus , A_ , max_length=A_ , padding='max_length' , ) # Pair input self.assertRaises(A_ , tokenizer_r.encode , A_ , max_length=A_ , padding='max_length' ) # Pair input self.assertRaises(A_ , tokenizer_r.encode_plus , A_ , max_length=A_ , padding='max_length' ) # Pair input self.assertRaises( A_ , tokenizer_r.batch_encode_plus , A_ , max_length=A_ , padding='max_length' , ) def __UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token='<pad>' ) # Simple input UpperCamelCase = 'This is a simple input' UpperCamelCase = ['This is a simple input looooooooong', 'This is a simple input'] UpperCamelCase = ('This is a simple input', 'This is a pair') UpperCamelCase = [ ('This is a simple input loooooong', 'This is a simple input'), ('This is a simple pair loooooong', 'This is a simple pair'), ] UpperCamelCase = tokenizer.pad_token_id UpperCamelCase = tokenizer(A_ , padding='max_length' , max_length=30 , return_tensors='np' ) UpperCamelCase = tokenizer(A_ , padding=A_ , truncate=A_ , return_tensors='np' ) UpperCamelCase = tokenizer(*A_ , padding='max_length' , max_length=60 , return_tensors='np' ) UpperCamelCase = tokenizer(A_ , padding=A_ , truncate=A_ , return_tensors='np' ) # s # test single string max_length padding self.assertEqual(out_s['input_ids'].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s['input_ids'] ) self.assertTrue(0 in out_s['attention_mask'] ) # s2 # test automatic padding self.assertEqual(out_sa['input_ids'].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['input_ids'][0] ) self.assertFalse(0 in out_sa['attention_mask'][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['input_ids'][1] ) self.assertTrue(0 in out_sa['attention_mask'][1] ) # p # test single pair max_length padding self.assertEqual(out_p['input_ids'].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p['input_ids'] ) self.assertTrue(0 in out_p['attention_mask'] ) # p2 # test automatic padding pair self.assertEqual(out_pa['input_ids'].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['input_ids'][0] ) self.assertFalse(0 in out_pa['attention_mask'][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['input_ids'][1] ) self.assertTrue(0 in out_pa['attention_mask'][1] ) def __UpperCamelCase ( self ) -> str: """simple docstring""" UpperCamelCase = '$$$' UpperCamelCase = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=A_ , add_bos_token=A_ ) UpperCamelCase = 'This is a simple input' UpperCamelCase = ['This is a simple input 1', 'This is a simple input 2'] UpperCamelCase = tokenizer.bos_token_id UpperCamelCase = tokenizer(A_ ) UpperCamelCase = tokenizer(A_ ) self.assertEqual(out_s.input_ids[0] , A_ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) UpperCamelCase = tokenizer.decode(out_s.input_ids ) UpperCamelCase = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , A_ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def __UpperCamelCase ( self ) -> Dict: """simple docstring""" UpperCamelCase = CodeGenTokenizer.from_pretrained('Salesforce/codegen-350M-mono' ) UpperCamelCase = '\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#' UpperCamelCase = '\nif len_a > len_b: result = a\nelse: result = b' UpperCamelCase = tokenizer.encode(A_ ) UpperCamelCase = ['^#', re.escape('<|endoftext|>' ), '^\'\'\'', '^"""', '\n\n\n'] UpperCamelCase = tokenizer.decode(A_ , truncate_before_pattern=A_ ) self.assertEqual(A_ , A_ ) def __UpperCamelCase ( self ) -> List[str]: """simple docstring""" pass
110
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class lowercase ( unittest.TestCase ): def __UpperCamelCase ( self ) -> List[str]: """simple docstring""" UpperCamelCase = tempfile.mkdtemp() # fmt: off UpperCamelCase = ['', 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'lo', 'l</w>', 'w</w>', 'r</w>', 't</w>', 'low</w>', 'er</w>', 'lowest</w>', 'newer</w>', 'wider', '<unk>', '<|startoftext|>', '<|endoftext|>'] # fmt: on UpperCamelCase = dict(zip(A_ , range(len(A_ ) ) ) ) UpperCamelCase = ['#version: 0.2', 'l o', 'lo w</w>', 'e r</w>', ''] UpperCamelCase = {'unk_token': '<unk>'} UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(A_ ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(A_ ) ) UpperCamelCase = { 'do_resize': True, 'size': 20, 'do_center_crop': True, 'crop_size': 18, 'do_normalize': True, 'image_mean': [0.4814_5466, 0.457_8275, 0.4082_1073], 'image_std': [0.2686_2954, 0.2613_0258, 0.2757_7711], } UpperCamelCase = os.path.join(self.tmpdirname , A_ ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(A_ , A_ ) def __UpperCamelCase ( self , **A_ ) -> Union[str, Any]: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token='!' , **A_ ) def __UpperCamelCase ( self , **A_ ) -> Tuple: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token='!' , **A_ ) def __UpperCamelCase ( self , **A_ ) -> Union[str, Any]: """simple docstring""" return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase ( self ) -> List[str]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def __UpperCamelCase ( self ) -> List[Any]: """simple docstring""" UpperCamelCase = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] UpperCamelCase = [Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def __UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase = self.get_tokenizer() UpperCamelCase = self.get_rust_tokenizer() UpperCamelCase = self.get_image_processor() UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) processor_slow.save_pretrained(self.tmpdirname ) UpperCamelCase = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=A_ ) UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) processor_fast.save_pretrained(self.tmpdirname ) UpperCamelCase = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , A_ ) self.assertIsInstance(processor_fast.tokenizer , A_ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , A_ ) self.assertIsInstance(processor_fast.image_processor , A_ ) def __UpperCamelCase ( self ) -> List[Any]: """simple docstring""" UpperCamelCase = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCamelCase = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) UpperCamelCase = self.get_image_processor(do_normalize=A_ ) UpperCamelCase = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=A_ ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) def __UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase = self.get_image_processor() UpperCamelCase = self.get_tokenizer() UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase = self.prepare_image_inputs() UpperCamelCase = image_processor(A_ , return_tensors='np' ) UpperCamelCase = processor(images=A_ , return_tensors='np' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def __UpperCamelCase ( self ) -> str: """simple docstring""" UpperCamelCase = self.get_image_processor() UpperCamelCase = self.get_tokenizer() UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase = 'lower newer' UpperCamelCase = processor(text=A_ , return_tensors='np' ) UpperCamelCase = tokenizer(A_ , return_tensors='np' ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def __UpperCamelCase ( self ) -> int: """simple docstring""" UpperCamelCase = self.get_image_processor() UpperCamelCase = self.get_tokenizer() UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase = 'lower newer' UpperCamelCase = self.prepare_image_inputs() UpperCamelCase = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def __UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = 'google/owlvit-base-patch32' UpperCamelCase = OwlViTProcessor.from_pretrained(A_ ) UpperCamelCase = ['cat', 'nasa badge'] UpperCamelCase = processor(text=A_ ) UpperCamelCase = 16 self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def __UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = 'google/owlvit-base-patch32' UpperCamelCase = OwlViTProcessor.from_pretrained(A_ ) UpperCamelCase = [['cat', 'nasa badge'], ['person']] UpperCamelCase = processor(text=A_ ) UpperCamelCase = 16 UpperCamelCase = len(A_ ) UpperCamelCase = max([len(A_ ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def __UpperCamelCase ( self ) -> Dict: """simple docstring""" UpperCamelCase = 'google/owlvit-base-patch32' UpperCamelCase = OwlViTProcessor.from_pretrained(A_ ) UpperCamelCase = ['cat', 'nasa badge'] UpperCamelCase = processor(text=A_ ) UpperCamelCase = 16 UpperCamelCase = inputs['input_ids'] UpperCamelCase = [ [49_406, 2_368, 49_407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [49_406, 6_841, 11_301, 49_407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def __UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase = self.get_image_processor() UpperCamelCase = self.get_tokenizer() UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase = self.prepare_image_inputs() UpperCamelCase = self.prepare_image_inputs() UpperCamelCase = processor(images=A_ , query_images=A_ ) self.assertListEqual(list(inputs.keys() ) , ['query_pixel_values', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def __UpperCamelCase ( self ) -> int: """simple docstring""" UpperCamelCase = self.get_image_processor() UpperCamelCase = self.get_tokenizer() UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCamelCase = processor.batch_decode(A_ ) UpperCamelCase = tokenizer.batch_decode(A_ ) self.assertListEqual(A_ , A_ )
110
1
'''simple docstring''' import os from pickle import UnpicklingError from typing import Dict, Tuple import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict, unflatten_dict import transformers from .utils import logging lowerCamelCase__ = logging.get_logger(__name__) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False ): try: import torch # noqa: F401 except ImportError: logger.error( "Loading a PyTorch model in Flax, requires both PyTorch and Flax to be installed. Please see" " https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation" " instructions." ) raise if not is_sharded: _UpperCAmelCase : Optional[Any] = os.path.abspath(__lowerCAmelCase ) logger.info(F"""Loading PyTorch weights from {pt_path}""" ) _UpperCAmelCase : List[Any] = torch.load(__lowerCAmelCase , map_location="cpu" ) logger.info(F"""PyTorch checkpoint contains {sum(t.numel() for t in pt_state_dict.values() ):,} parameters.""" ) _UpperCAmelCase : List[Any] = convert_pytorch_state_dict_to_flax(__lowerCAmelCase , __lowerCAmelCase ) else: # model is sharded and pytorch_checkpoint_path already contains the list of .pt shard files _UpperCAmelCase : Tuple = convert_pytorch_sharded_state_dict_to_flax(__lowerCAmelCase , __lowerCAmelCase ) return flax_state_dict def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ): def is_key_or_prefix_key_in_dict(__lowerCAmelCase ) -> bool: return len(set(__lowerCAmelCase ) & {key, (model_prefix,) + key} ) > 0 # layer norm _UpperCAmelCase : Any = pt_tuple_key[:-1] + ("scale",) if pt_tuple_key[-1] in ["weight", "gamma"] and is_key_or_prefix_key_in_dict(__lowerCAmelCase ): return renamed_pt_tuple_key, pt_tensor # batch norm layer mean _UpperCAmelCase : Optional[Any] = pt_tuple_key[:-1] + ("mean",) if pt_tuple_key[-1] == "running_mean" and not is_key_or_prefix_key_in_dict(__lowerCAmelCase ): return renamed_pt_tuple_key, pt_tensor # batch norm layer var _UpperCAmelCase : Optional[int] = pt_tuple_key[:-1] + ("var",) if pt_tuple_key[-1] == "running_var" and not is_key_or_prefix_key_in_dict(__lowerCAmelCase ): return renamed_pt_tuple_key, pt_tensor # embedding _UpperCAmelCase : Union[str, Any] = pt_tuple_key[:-1] + ("embedding",) if pt_tuple_key[-1] == "weight" and is_key_or_prefix_key_in_dict(__lowerCAmelCase ): return renamed_pt_tuple_key, pt_tensor # conv layer _UpperCAmelCase : Tuple = pt_tuple_key[:-1] + ("kernel",) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4 and not is_key_or_prefix_key_in_dict(__lowerCAmelCase ): _UpperCAmelCase : List[str] = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer _UpperCAmelCase : Union[str, Any] = pt_tuple_key[:-1] + ("kernel",) if pt_tuple_key[-1] == "weight" and not is_key_or_prefix_key_in_dict(__lowerCAmelCase ): _UpperCAmelCase : List[Any] = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight _UpperCAmelCase : List[Any] = pt_tuple_key[:-1] + ("weight",) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias _UpperCAmelCase : List[Any] = pt_tuple_key[:-1] + ("bias",) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 _UpperCAmelCase : List[str] = None if pt_tuple_key[-3::2] == ("parametrizations", "original0"): _UpperCAmelCase : Dict = pt_tuple_key[-2] + "_g" elif pt_tuple_key[-3::2] == ("parametrizations", "original1"): _UpperCAmelCase : Tuple = pt_tuple_key[-2] + "_v" if name is not None: _UpperCAmelCase : Optional[int] = pt_tuple_key[:-3] + (name,) return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): # convert pytorch tensor to numpy _UpperCAmelCase : Tuple = {k: v.numpy() for k, v in pt_state_dict.items()} _UpperCAmelCase : str = flax_model.base_model_prefix # use params dict if the model contains batch norm layers if "params" in flax_model.params: _UpperCAmelCase : List[str] = flax_model.params["params"] else: _UpperCAmelCase : List[Any] = flax_model.params _UpperCAmelCase : List[Any] = flatten_dict(__lowerCAmelCase ) # add batch_stats keys,values to dict if "batch_stats" in flax_model.params: _UpperCAmelCase : Optional[Any] = flatten_dict(flax_model.params["batch_stats"] ) random_flax_state_dict.update(__lowerCAmelCase ) _UpperCAmelCase : List[Any] = {} _UpperCAmelCase : Dict = (model_prefix not in flax_model_params) and ( model_prefix in {k.split("." )[0] for k in pt_state_dict.keys()} ) _UpperCAmelCase : Optional[int] = (model_prefix in flax_model_params) and ( model_prefix not in {k.split("." )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): _UpperCAmelCase : List[str] = tuple(pt_key.split("." ) ) # remove base model prefix if necessary _UpperCAmelCase : Optional[int] = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: _UpperCAmelCase : Tuple = pt_tuple_key[1:] # Correctly rename weight parameters _UpperCAmelCase , _UpperCAmelCase : Tuple = rename_key_and_reshape_tensor( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # add model prefix if necessary _UpperCAmelCase : Optional[Any] = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: _UpperCAmelCase : Optional[Any] = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ F"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1] or "var" in flax_key[-1]: _UpperCAmelCase : Dict = jnp.asarray(__lowerCAmelCase ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) continue # also add unexpected weight so that warning is thrown _UpperCAmelCase : Dict = jnp.asarray(__lowerCAmelCase ) else: # also add unexpected weight so that warning is thrown _UpperCAmelCase : List[Any] = jnp.asarray(__lowerCAmelCase ) return unflatten_dict(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): import torch # Load the index _UpperCAmelCase : List[Any] = {} for shard_file in shard_filenames: # load using msgpack utils _UpperCAmelCase : Tuple = torch.load(__lowerCAmelCase ) _UpperCAmelCase : List[Any] = {k: v.numpy() for k, v in pt_state_dict.items()} _UpperCAmelCase : str = flax_model.base_model_prefix # use params dict if the model contains batch norm layers and then add batch_stats keys,values to dict if "batch_stats" in flax_model.params: _UpperCAmelCase : Optional[Any] = flax_model.params["params"] _UpperCAmelCase : List[Any] = flatten_dict(__lowerCAmelCase ) random_flax_state_dict.update(flatten_dict(flax_model.params["batch_stats"] ) ) else: _UpperCAmelCase : Optional[int] = flax_model.params _UpperCAmelCase : int = flatten_dict(__lowerCAmelCase ) _UpperCAmelCase : Any = (model_prefix not in flax_model_params) and ( model_prefix in {k.split("." )[0] for k in pt_state_dict.keys()} ) _UpperCAmelCase : Optional[Any] = (model_prefix in flax_model_params) and ( model_prefix not in {k.split("." )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): _UpperCAmelCase : Dict = tuple(pt_key.split("." ) ) # remove base model prefix if necessary _UpperCAmelCase : List[Any] = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: _UpperCAmelCase : Union[str, Any] = pt_tuple_key[1:] # Correctly rename weight parameters _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = rename_key_and_reshape_tensor( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # add model prefix if necessary _UpperCAmelCase : int = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: _UpperCAmelCase : List[Any] = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ F"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1]: _UpperCAmelCase : List[str] = jnp.asarray(__lowerCAmelCase ) continue if "var" in flax_key[-1]: _UpperCAmelCase : Dict = jnp.asarray(__lowerCAmelCase ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) continue # also add unexpected weight so that warning is thrown _UpperCAmelCase : List[Any] = jnp.asarray(__lowerCAmelCase ) else: # also add unexpected weight so that warning is thrown _UpperCAmelCase : Any = jnp.asarray(__lowerCAmelCase ) return unflatten_dict(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = os.path.abspath(__lowerCAmelCase ) logger.info(F"""Loading Flax weights from {flax_checkpoint_path}""" ) # import correct flax class _UpperCAmelCase : Optional[Any] = getattr(__lowerCAmelCase , "Flax" + model.__class__.__name__ ) # load flax weight dict with open(__lowerCAmelCase , "rb" ) as state_f: try: _UpperCAmelCase : Optional[int] = from_bytes(__lowerCAmelCase , state_f.read() ) except UnpicklingError: raise EnvironmentError(F"""Unable to convert {flax_checkpoint_path} to Flax deserializable object. """ ) return load_flax_weights_in_pytorch_model(__lowerCAmelCase , __lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): try: import torch # noqa: F401 except ImportError: logger.error( "Loading a Flax weights in PyTorch, requires both PyTorch and Flax to be installed. Please see" " https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation" " instructions." ) raise # check if we have bf16 weights _UpperCAmelCase : Any = flatten_dict(jax.tree_util.tree_map(lambda __lowerCAmelCase : x.dtype == jnp.bfloataa , __lowerCAmelCase ) ).values() if any(__lowerCAmelCase ): # convert all weights to fp32 if the are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( "Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` " "before loading those in PyTorch model." ) _UpperCAmelCase : Tuple = jax.tree_util.tree_map( lambda __lowerCAmelCase : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , __lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = flatten_dict(__lowerCAmelCase ) _UpperCAmelCase : str = pt_model.state_dict() _UpperCAmelCase : Dict = (pt_model.base_model_prefix in flax_state) and ( pt_model.base_model_prefix not in {k.split("." )[0] for k in pt_model_dict.keys()} ) _UpperCAmelCase : int = (pt_model.base_model_prefix not in flax_state) and ( pt_model.base_model_prefix in {k.split("." )[0] for k in pt_model_dict.keys()} ) # keep track of unexpected & missing keys _UpperCAmelCase : Any = [] _UpperCAmelCase : Union[str, Any] = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): _UpperCAmelCase : Optional[Any] = flax_key_tuple[0] == pt_model.base_model_prefix _UpperCAmelCase : List[Any] = ".".join((pt_model.base_model_prefix,) + flax_key_tuple ) in pt_model_dict # adapt flax_key to prepare for loading from/to base model only if load_model_with_head_into_base_model and has_base_model_prefix: _UpperCAmelCase : List[str] = flax_key_tuple[1:] elif load_base_model_into_model_with_head and require_base_model_prefix: _UpperCAmelCase : Optional[Any] = (pt_model.base_model_prefix,) + flax_key_tuple # rename flax weights to PyTorch format if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 4 and ".".join(__lowerCAmelCase ) not in pt_model_dict: # conv layer _UpperCAmelCase : Dict = flax_key_tuple[:-1] + ("weight",) _UpperCAmelCase : Tuple = jnp.transpose(__lowerCAmelCase , (3, 2, 0, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(__lowerCAmelCase ) not in pt_model_dict: # linear layer _UpperCAmelCase : str = flax_key_tuple[:-1] + ("weight",) _UpperCAmelCase : Optional[Any] = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: _UpperCAmelCase : Union[str, Any] = flax_key_tuple[:-1] + ("weight",) # adding batch stats from flax batch norm to pt elif "mean" in flax_key_tuple[-1]: _UpperCAmelCase : List[Any] = flax_key_tuple[:-1] + ("running_mean",) elif "var" in flax_key_tuple[-1]: _UpperCAmelCase : List[str] = flax_key_tuple[:-1] + ("running_var",) if "batch_stats" in flax_state: _UpperCAmelCase : str = ".".join(flax_key_tuple[1:] ) # Remove the params/batch_stats header else: _UpperCAmelCase : Dict = ".".join(__lowerCAmelCase ) # We also need to look at `pt_model_dict` and see if there are keys requiring further transformation. _UpperCAmelCase : Optional[Any] = {} # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 for key in pt_model_dict: _UpperCAmelCase : Any = key.split("." ) _UpperCAmelCase : str = None if key_components[-3::2] == ["parametrizations", "original0"]: _UpperCAmelCase : Tuple = key_components[-2] + "_g" elif key_components[-3::2] == ["parametrizations", "original1"]: _UpperCAmelCase : int = key_components[-2] + "_v" if name is not None: _UpperCAmelCase : Dict = key_components[:-3] + [name] _UpperCAmelCase : Union[str, Any] = ".".join(__lowerCAmelCase ) _UpperCAmelCase : Dict = key if flax_key in special_pt_names: _UpperCAmelCase : int = special_pt_names[flax_key] if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"""Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected """ F"""to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) else: # add weight to pytorch dict _UpperCAmelCase : List[str] = np.asarray(__lowerCAmelCase ) if not isinstance(__lowerCAmelCase , np.ndarray ) else flax_tensor _UpperCAmelCase : Optional[int] = torch.from_numpy(__lowerCAmelCase ) # remove from missing keys missing_keys.remove(__lowerCAmelCase ) else: # weight is not expected by PyTorch model unexpected_keys.append(__lowerCAmelCase ) pt_model.load_state_dict(__lowerCAmelCase ) # re-transform missing_keys to list _UpperCAmelCase : Tuple = list(__lowerCAmelCase ) if len(__lowerCAmelCase ) > 0: logger.warning( "Some weights of the Flax model were not used when initializing the PyTorch model" F""" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing""" F""" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture""" " (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This" F""" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect""" " to be exactly identical (e.g. initializing a BertForSequenceClassification model from a" " FlaxBertForSequenceClassification model)." ) else: logger.warning(F"""All Flax model weights were used when initializing {pt_model.__class__.__name__}.\n""" ) if len(__lowerCAmelCase ) > 0: logger.warning( F"""Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly""" F""" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to""" " use it for predictions and inference." ) else: logger.warning( F"""All the weights of {pt_model.__class__.__name__} were initialized from the Flax model.\n""" "If your task is similar to the task the model of the checkpoint was trained on, " F"""you can already use {pt_model.__class__.__name__} for predictions without further training.""" ) return pt_model
234
'''simple docstring''' import warnings from diffusers import StableDiffusionImgaImgPipeline # noqa F401 warnings.warn( 'The `image_to_image.py` script is outdated. Please use directly `from diffusers import' ' StableDiffusionImg2ImgPipeline` instead.' )
234
1
'''simple docstring''' def lowercase__ ( __UpperCamelCase = 1000 )-> int: UpperCamelCase = 2**power UpperCamelCase = str(__UpperCamelCase ) UpperCamelCase = list(__UpperCamelCase ) UpperCamelCase = 0 for i in list_num: sum_of_num += int(__UpperCamelCase ) return sum_of_num if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = int(input('Enter the power of 2: ').strip()) print('2 ^ ', power, ' = ', 2**power) SCREAMING_SNAKE_CASE__ = solution(power) print('Sum of the digits is: ', result)
183
'''simple docstring''' import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class a_ ( unittest.TestCase ): def A__ ( self ) -> Tuple: """simple docstring""" UpperCamelCase = ["""a""", """b""", """c"""] # Defaults to last layer if both are None UpperCamelCase ,UpperCamelCase = get_aligned_output_features_output_indices(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , ["""c"""] ) self.assertEqual(_SCREAMING_SNAKE_CASE , [2] ) # Out indices set to match out features UpperCamelCase ,UpperCamelCase = get_aligned_output_features_output_indices(["""a""", """c"""] , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , ["""a""", """c"""] ) self.assertEqual(_SCREAMING_SNAKE_CASE , [0, 2] ) # Out features set to match out indices UpperCamelCase ,UpperCamelCase = get_aligned_output_features_output_indices(_SCREAMING_SNAKE_CASE , [0, 2] , _SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , ["""a""", """c"""] ) self.assertEqual(_SCREAMING_SNAKE_CASE , [0, 2] ) # Out features selected from negative indices UpperCamelCase ,UpperCamelCase = get_aligned_output_features_output_indices(_SCREAMING_SNAKE_CASE , [-3, -1] , _SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , ["""a""", """c"""] ) self.assertEqual(_SCREAMING_SNAKE_CASE , [-3, -1] ) def A__ ( self ) -> str: """simple docstring""" with self.assertRaises(_SCREAMING_SNAKE_CASE ): verify_out_features_out_indices(["""a""", """b"""] , (0, 1) , _SCREAMING_SNAKE_CASE ) # Out features must be a list with self.assertRaises(_SCREAMING_SNAKE_CASE ): verify_out_features_out_indices(("""a""", """b""") , (0, 1) , ["""a""", """b"""] ) # Out features must be a subset of stage names with self.assertRaises(_SCREAMING_SNAKE_CASE ): verify_out_features_out_indices(["""a""", """b"""] , (0, 1) , ["""a"""] ) # Out indices must be a list or tuple with self.assertRaises(_SCREAMING_SNAKE_CASE ): verify_out_features_out_indices(_SCREAMING_SNAKE_CASE , 0 , ["""a""", """b"""] ) # Out indices must be a subset of stage names with self.assertRaises(_SCREAMING_SNAKE_CASE ): verify_out_features_out_indices(_SCREAMING_SNAKE_CASE , (0, 1) , ["""a"""] ) # Out features and out indices must be the same length with self.assertRaises(_SCREAMING_SNAKE_CASE ): verify_out_features_out_indices(["""a""", """b"""] , (0,) , ["""a""", """b""", """c"""] ) # Out features should match out indices with self.assertRaises(_SCREAMING_SNAKE_CASE ): verify_out_features_out_indices(["""a""", """b"""] , (0, 2) , ["""a""", """b""", """c"""] ) # Out features and out indices should be in order with self.assertRaises(_SCREAMING_SNAKE_CASE ): verify_out_features_out_indices(["""b""", """a"""] , (0, 1) , ["""a""", """b"""] ) # Check passes with valid inputs verify_out_features_out_indices(["""a""", """b""", """d"""] , (0, 1, -1) , ["""a""", """b""", """c""", """d"""] ) def A__ ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = BackboneMixin() UpperCamelCase = ["""a""", """b""", """c"""] UpperCamelCase = ["""a""", """c"""] UpperCamelCase = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features , ["""a""", """c"""] ) self.assertEqual(backbone.out_indices , [0, 2] ) # Check out features and indices are updated correctly UpperCamelCase = ["""a""", """b"""] self.assertEqual(backbone.out_features , ["""a""", """b"""] ) self.assertEqual(backbone.out_indices , [0, 1] ) UpperCamelCase = [-3, -1] self.assertEqual(backbone.out_features , ["""a""", """c"""] ) self.assertEqual(backbone.out_indices , [-3, -1] )
183
1
import random def lowerCamelCase__ ( a__ : int , a__ : float , a__ : bool = False ) -> dict: UpperCamelCase_ = {i: [] for i in range(a__ )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(a__ ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(a__ ): for j in range(i + 1 , a__ ): if random.random() < probability: graph[i].append(a__ ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(a__ ) return graph def lowerCamelCase__ ( a__ : int ) -> dict: return { i: [j for j in range(a__ ) if i != j] for i in range(a__ ) } if __name__ == "__main__": import doctest doctest.testmod()
122
from __future__ import annotations def lowerCamelCase__ ( a__ : int | float | str , a__ : int | float | str ) -> list[str]: if nth_term == "": return [""] UpperCamelCase_ = int(a__ ) UpperCamelCase_ = int(a__ ) UpperCamelCase_ = [] for temp in range(int(a__ ) ): series.append(f'''1 / {pow(temp + 1 , int(a__ ) )}''' if series else """1""" ) return series if __name__ == "__main__": import doctest doctest.testmod() _A = int(input('''Enter the last number (nth term) of the P-Series''')) _A = int(input('''Enter the power for P-Series''')) print('''Formula of P-Series => 1+1/2^p+1/3^p ..... 1/n^p''') print(p_series(nth_term, power))
122
1
import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _a (__magic_name__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__: Tuple = KandinskyVaaControlnetPipeline UpperCAmelCase__: str = ['''image_embeds''', '''negative_image_embeds''', '''hint'''] UpperCAmelCase__: List[Any] = ['''image_embeds''', '''negative_image_embeds''', '''hint'''] UpperCAmelCase__: str = [ '''generator''', '''height''', '''width''', '''latents''', '''guidance_scale''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] UpperCAmelCase__: Union[str, Any] = False @property def __A ( self ): return 32 @property def __A ( self ): return 32 @property def __A ( self ): return self.time_input_dim @property def __A ( self ): return self.time_input_dim * 4 @property def __A ( self ): return 100 @property def __A ( self ): torch.manual_seed(0 ) A__ : str = { """in_channels""": 8, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image_hint""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } A__ : List[Any] = UNetaDConditionModel(**A__ ) return model @property def __A ( self ): return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __A ( self ): torch.manual_seed(0 ) A__ : Tuple = VQModel(**self.dummy_movq_kwargs ) return model def __A ( self ): A__ : List[str] = self.dummy_unet A__ : Optional[Any] = self.dummy_movq A__ : Optional[Any] = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="""linear""" , beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , clip_sample=A__ , set_alpha_to_one=A__ , steps_offset=1 , prediction_type="""epsilon""" , thresholding=A__ , ) A__ : Optional[Any] = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def __A ( self , A__ , A__=0 ): A__ : List[str] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(A__ ) ).to(A__ ) A__ : Optional[Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( A__ ) # create hint A__ : Optional[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(A__ ) ).to(A__ ) if str(A__ ).startswith("""mps""" ): A__ : List[Any] = torch.manual_seed(A__ ) else: A__ : str = torch.Generator(device=A__ ).manual_seed(A__ ) A__ : int = { """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """hint""": hint, """generator""": generator, """height""": 64, """width""": 64, """guidance_scale""": 4.0, """num_inference_steps""": 2, """output_type""": """np""", } return inputs def __A ( self ): A__ : Tuple = """cpu""" A__ : Dict = self.get_dummy_components() A__ : Union[str, Any] = self.pipeline_class(**A__ ) A__ : Any = pipe.to(A__ ) pipe.set_progress_bar_config(disable=A__ ) A__ : Dict = pipe(**self.get_dummy_inputs(A__ ) ) A__ : List[Any] = output.images A__ : Dict = pipe( **self.get_dummy_inputs(A__ ) , return_dict=A__ , )[0] A__ : List[Any] = image[0, -3:, -3:, -1] A__ : List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) A__ : str = np.array( [0.6_9_5_9_8_2_6, 0.8_6_8_2_7_9, 0.7_5_5_8_0_9_2, 0.6_8_7_6_9_4_6_7, 0.8_5_8_0_5_8_0_4, 0.6_5_9_7_7_4_9_6, 0.4_4_8_8_5_3_0_2, 0.5_9_5_9_1_1_1, 0.4_2_5_1_5_9_5] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class _a (unittest.TestCase ): '''simple docstring''' def __A ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __A ( self ): A__ : Dict = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_controlnet_robotcat_fp16.npy""" ) A__ : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/hint_image_cat.png""" ) A__ : Dict = torch.from_numpy(np.array(A__ ) ).float() / 2_5_5.0 A__ : int = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) A__ : List[Any] = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(A__ ) A__ : Tuple = KandinskyVaaControlnetPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-controlnet-depth""" , torch_dtype=torch.floataa ) A__ : List[Any] = pipeline.to(A__ ) pipeline.set_progress_bar_config(disable=A__ ) A__ : List[Any] = """A robot, 4k photo""" A__ : Union[str, Any] = torch.Generator(device="""cuda""" ).manual_seed(0 ) A__ , A__ : Dict = pipe_prior( A__ , generator=A__ , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() A__ : Optional[int] = torch.Generator(device="""cuda""" ).manual_seed(0 ) A__ : Any = pipeline( image_embeds=A__ , negative_image_embeds=A__ , hint=A__ , generator=A__ , num_inference_steps=100 , output_type="""np""" , ) A__ : int = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(A__ , A__ )
141
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) A_ : Optional[int] = {'configuration_fnet': ['FNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'FNetConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Optional[Any] = ['FNetTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Tuple = ['FNetTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : List[Any] = [ 'FNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'FNetForMaskedLM', 'FNetForMultipleChoice', 'FNetForNextSentencePrediction', 'FNetForPreTraining', 'FNetForQuestionAnswering', 'FNetForSequenceClassification', 'FNetForTokenClassification', 'FNetLayer', 'FNetModel', 'FNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_fnet import FNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet import FNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet_fast import FNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_fnet import ( FNET_PRETRAINED_MODEL_ARCHIVE_LIST, FNetForMaskedLM, FNetForMultipleChoice, FNetForNextSentencePrediction, FNetForPreTraining, FNetForQuestionAnswering, FNetForSequenceClassification, FNetForTokenClassification, FNetLayer, FNetModel, FNetPreTrainedModel, ) else: import sys A_ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
141
1
import gc import unittest from diffusers import FlaxStableDiffusionInpaintPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) lowerCamelCase : Dict = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) lowerCamelCase : Optional[Any] = """xvjiarui/stable-diffusion-2-inpainting""" lowerCamelCase , lowerCamelCase : List[str] = FlaxStableDiffusionInpaintPipeline.from_pretrained(__magic_name__ , safety_checker=__magic_name__ ) lowerCamelCase : int = """Face of a yellow cat, high resolution, sitting on a park bench""" lowerCamelCase : Optional[Any] = jax.random.PRNGKey(0 ) lowerCamelCase : Tuple = 5_0 lowerCamelCase : Dict = jax.device_count() lowerCamelCase : Union[str, Any] = num_samples * [prompt] lowerCamelCase : Optional[Any] = num_samples * [init_image] lowerCamelCase : Dict = num_samples * [mask_image] lowerCamelCase , lowerCamelCase , lowerCamelCase : Dict = pipeline.prepare_inputs(__magic_name__ , __magic_name__ , __magic_name__ ) # shard inputs and rng lowerCamelCase : List[Any] = replicate(__magic_name__ ) lowerCamelCase : str = jax.random.split(__magic_name__ , jax.device_count() ) lowerCamelCase : Dict = shard(__magic_name__ ) lowerCamelCase : List[Any] = shard(__magic_name__ ) lowerCamelCase : int = shard(__magic_name__ ) lowerCamelCase : Optional[Any] = pipeline( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , jit=__magic_name__ ) lowerCamelCase : List[str] = output.images.reshape(__magic_name__ , 5_1_2 , 5_1_2 , 3 ) lowerCamelCase : Optional[Any] = images[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] lowerCamelCase : Union[str, Any] = jnp.asarray(jax.device_get(image_slice.flatten() ) ) lowerCamelCase : Optional[Any] = jnp.array( [0.3_611_307, 0.37_649_736, 0.3_757_408, 0.38_213_953, 0.39_295_167, 0.3_841_631, 0.41_554_978, 0.4_137_475, 0.4_217_084] ) print(F'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
287
_lowerCamelCase ={ "joule": 1.0, "kilojoule": 1_0_0_0, "megajoule": 1_0_0_0_0_0_0, "gigajoule": 1_0_0_0_0_0_0_0_0_0, "wattsecond": 1.0, "watthour": 3_6_0_0, "kilowatthour": 3_6_0_0_0_0_0, "newtonmeter": 1.0, "calorie_nutr": 4_1_8_6.8, "kilocalorie_nutr": 4_1_8_6_8_0_0.0_0, "electronvolt": 1.6_0_2_1_7_6_6_3_4E-1_9, "britishthermalunit_it": 1_0_5_5.0_5_5_8_5, "footpound": 1.355818, } def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: lowerCamelCase : Dict = ( F'''Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n''' F'''Valid values are: {", ".join(lowerCamelCase )}''' ) raise ValueError(lowerCamelCase ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
287
1
"""simple docstring""" __snake_case = range(2, 20 + 1) __snake_case = [10**k for k in range(ks[-1] + 1)] __snake_case = {} def __lowerCAmelCase ( lowercase : List[Any] , lowercase : Dict , lowercase : int , lowercase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" snake_case : str = sum(a_i[j] for j in range(lowercase , len(lowercase ) ) ) snake_case : int = sum(a_i[j] * base[j] for j in range(min(len(lowercase ) , lowercase ) ) ) snake_case ,snake_case : Any = 0, 0 snake_case : Dict = n - i snake_case : Any = memo.get(lowercase ) if sub_memo is not None: snake_case : int = sub_memo.get(lowercase ) if jumps is not None and len(lowercase ) > 0: # find and make the largest jump without going over snake_case : List[Any] = -1 for _k in range(len(lowercase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: snake_case : Dict = _k break if max_jump >= 0: snake_case ,snake_case ,snake_case : int = jumps[max_jump] # since the difference between jumps is cached, add c snake_case : Optional[Any] = diff + c for j in range(min(lowercase , len(lowercase ) ) ): snake_case ,snake_case : Dict = divmod(lowercase , 10 ) if new_c > 0: add(lowercase , lowercase , lowercase ) else: snake_case : Optional[int] = [] else: snake_case : Tuple = {c: []} snake_case : List[str] = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps snake_case ,snake_case : List[str] = next_term(lowercase , k - 1 , i + dn , lowercase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead snake_case ,snake_case : Tuple = compute(lowercase , lowercase , i + dn , lowercase ) diff += _diff dn += terms_jumped snake_case : Optional[Any] = sub_memo[c] # keep jumps sorted by # of terms skipped snake_case : Any = 0 while j < len(lowercase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(lowercase , (diff, dn, k) ) return (diff, dn) def __lowerCAmelCase ( lowercase : List[str] , lowercase : List[str] , lowercase : int , lowercase : Any ) -> Any: """simple docstring""" if i >= n: return 0, i if k > len(lowercase ): a_i.extend([0 for _ in range(k - len(lowercase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) snake_case : List[Any] = i snake_case ,snake_case ,snake_case : int = 0, 0, 0 for j in range(len(lowercase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 snake_case : Any = ds_c + ds_b diff += addend snake_case : List[Any] = 0 for j in range(lowercase ): snake_case : List[Any] = a_i[j] + addend snake_case ,snake_case : Optional[Any] = divmod(lowercase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(lowercase , lowercase , lowercase ) return diff, i - start_i def __lowerCAmelCase ( lowercase : List[Any] , lowercase : int , lowercase : Union[str, Any] ) -> Dict: """simple docstring""" for j in range(lowercase , len(lowercase ) ): snake_case : Optional[Any] = digits[j] + addend if s >= 10: snake_case ,snake_case : List[str] = divmod(lowercase , 10 ) snake_case : Dict = addend // 10 + quotient else: snake_case : Union[str, Any] = s snake_case : Optional[Any] = addend // 10 if addend == 0: break while addend > 0: snake_case ,snake_case : Optional[int] = divmod(lowercase , 10 ) digits.append(lowercase ) def __lowerCAmelCase ( lowercase : int = 10**15 ) -> int: """simple docstring""" snake_case : Dict = [1] snake_case : List[Any] = 1 snake_case : Dict = 0 while True: snake_case ,snake_case : Optional[Any] = next_term(lowercase , 20 , i + dn , lowercase ) dn += terms_jumped if dn == n - i: break snake_case : List[str] = 0 for j in range(len(lowercase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'''{solution() = }''')
112
"""simple docstring""" from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
112
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __lowercase = {'''configuration_fnet''': ['''FNET_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''FNetConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = ['''FNetTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = ['''FNetTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''FNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FNetForMaskedLM''', '''FNetForMultipleChoice''', '''FNetForNextSentencePrediction''', '''FNetForPreTraining''', '''FNetForQuestionAnswering''', '''FNetForSequenceClassification''', '''FNetForTokenClassification''', '''FNetLayer''', '''FNetModel''', '''FNetPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_fnet import FNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet import FNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet_fast import FNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_fnet import ( FNET_PRETRAINED_MODEL_ARCHIVE_LIST, FNetForMaskedLM, FNetForMultipleChoice, FNetForNextSentencePrediction, FNetForPreTraining, FNetForQuestionAnswering, FNetForSequenceClassification, FNetForTokenClassification, FNetLayer, FNetModel, FNetPreTrainedModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
272
"""simple docstring""" import os import pytest from attr import dataclass lowerCamelCase__ = """us-east-1""" # defaults region @dataclass class A__ : A_ : str A_ : Union[str, Any] = 'arn:aws:iam::558105141721:role/sagemaker_execution_role' A_ : Optional[int] = { 'task_name': 'mnli', 'per_device_train_batch_size': 1_6, 'per_device_eval_batch_size': 1_6, 'do_train': True, 'do_eval': True, 'do_predict': True, 'output_dir': '/opt/ml/model', 'overwrite_output_dir': True, 'max_steps': 5_0_0, 'save_steps': 5_5_0_0, } A_ : List[Any] = {**hyperparameters, 'max_steps': 1_0_0_0} @property def __lowerCamelCase ( self ): if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def __lowerCamelCase ( self ): return f"{self.framework}-transfromers-test" @property def __lowerCamelCase ( self ): return f"./tests/sagemaker/scripts/{self.framework}" @property def __lowerCamelCase ( self ): if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='class' ) def __lowerCAmelCase (_UpperCamelCase ): __lowerCAmelCase : str = SageMakerTestEnvironment(framework=request.cls.framework )
86
0
'''simple docstring''' from __future__ import annotations from collections.abc import Callable lowerCamelCase__ = list[list[float | int]] def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : int = len(__lowerCAmelCase ) _UpperCAmelCase : Matrix = [[0 for _ in range(size + 1 )] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : float for row in range(__lowerCAmelCase ): for col in range(__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = matrix[row][col] _UpperCAmelCase : Optional[int] = vector[row][0] _UpperCAmelCase : int = 0 _UpperCAmelCase : Union[str, Any] = 0 while row < size and col < size: # pivoting _UpperCAmelCase : Optional[Any] = max((abs(augmented[rowa][col] ), rowa) for rowa in range(__lowerCAmelCase , __lowerCAmelCase ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: _UpperCAmelCase : str = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , __lowerCAmelCase ): _UpperCAmelCase : Dict = augmented[rowa][col] / augmented[row][col] _UpperCAmelCase : Optional[Any] = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , __lowerCAmelCase ): for row in range(__lowerCAmelCase ): _UpperCAmelCase : Dict = augmented[row][col] / augmented[col][col] for cola in range(__lowerCAmelCase , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(__lowerCAmelCase ) ] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : int = len(__lowerCAmelCase ) _UpperCAmelCase : Matrix = [[0 for _ in range(__lowerCAmelCase )] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : Matrix = [[0] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : Matrix _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int for x_val, y_val in enumerate(__lowerCAmelCase ): for col in range(__lowerCAmelCase ): _UpperCAmelCase : Dict = (x_val + 1) ** (size - col - 1) _UpperCAmelCase : int = y_val _UpperCAmelCase : List[str] = solve(__lowerCAmelCase , __lowerCAmelCase ) def interpolated_func(__lowerCAmelCase ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(__lowerCAmelCase ) ) return interpolated_func def __lowerCAmelCase (__lowerCAmelCase ): return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def __lowerCAmelCase (__lowerCAmelCase = question_function , __lowerCAmelCase = 10 ): _UpperCAmelCase : list[int] = [func(__lowerCAmelCase ) for x_val in range(1 , order + 1 )] _UpperCAmelCase : list[Callable[[int], int]] = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] _UpperCAmelCase : int = 0 _UpperCAmelCase : Callable[[int], int] _UpperCAmelCase : int for poly in polynomials: _UpperCAmelCase : int = 1 while func(__lowerCAmelCase ) == poly(__lowerCAmelCase ): x_val += 1 ret += poly(__lowerCAmelCase ) return ret if __name__ == "__main__": print(F'''{solution() = }''')
360
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): if number < 0: raise ValueError("number must not be negative" ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
322
0
'''simple docstring''' import unittest from transformers import MobileBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertModel, ) class snake_case : """simple docstring""" def __init__( self : Union[str, Any] , __A : List[Any] , __A : Optional[Any]=1_3 , __A : Optional[int]=7 , __A : Optional[Any]=True , __A : Tuple=True , __A : str=True , __A : Optional[Any]=True , __A : Optional[int]=9_9 , __A : Optional[int]=6_4 , __A : Optional[Any]=3_2 , __A : Tuple=5 , __A : Any=4 , __A : Union[str, Any]=3_7 , __A : List[str]="gelu" , __A : Union[str, Any]=0.1 , __A : List[str]=0.1 , __A : Any=5_1_2 , __A : Optional[Any]=1_6 , __A : Optional[Any]=2 , __A : Union[str, Any]=0.02 , __A : Dict=3 , __A : Any=4 , __A : List[Any]=None , ): __UpperCamelCase = parent __UpperCamelCase = batch_size __UpperCamelCase = seq_length __UpperCamelCase = is_training __UpperCamelCase = use_input_mask __UpperCamelCase = use_token_type_ids __UpperCamelCase = use_labels __UpperCamelCase = vocab_size __UpperCamelCase = hidden_size __UpperCamelCase = embedding_size __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = intermediate_size __UpperCamelCase = hidden_act __UpperCamelCase = hidden_dropout_prob __UpperCamelCase = attention_probs_dropout_prob __UpperCamelCase = max_position_embeddings __UpperCamelCase = type_vocab_size __UpperCamelCase = type_sequence_label_size __UpperCamelCase = initializer_range __UpperCamelCase = num_labels __UpperCamelCase = num_choices __UpperCamelCase = scope def _lowerCamelCase ( self : Any ): __UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCamelCase = None if self.use_input_mask: __UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCamelCase = None if self.use_token_type_ids: __UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None if self.use_labels: __UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __UpperCamelCase = ids_tensor([self.batch_size] , self.num_choices ) __UpperCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : Union[str, Any] ): return MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , embedding_size=self.embedding_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_A , initializer_range=self.initializer_range , ) def _lowerCamelCase ( self : str , __A : List[str] , __A : str , __A : List[Any] , __A : Dict , __A : List[str] , __A : Optional[Any] , __A : Any ): __UpperCamelCase = MobileBertModel(config=_A ) model.to(_A ) model.eval() __UpperCamelCase = model(_A , attention_mask=_A , token_type_ids=_A ) __UpperCamelCase = model(_A , token_type_ids=_A ) __UpperCamelCase = model(_A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _lowerCamelCase ( self : int , __A : Any , __A : int , __A : Tuple , __A : List[str] , __A : str , __A : int , __A : List[str] ): __UpperCamelCase = MobileBertForMaskedLM(config=_A ) model.to(_A ) model.eval() __UpperCamelCase = model(_A , attention_mask=_A , token_type_ids=_A , labels=_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : Optional[Any] , __A : Optional[int] , __A : Union[str, Any] , __A : Union[str, Any] , __A : Optional[int] , __A : str , __A : List[str] , __A : Any ): __UpperCamelCase = MobileBertForNextSentencePrediction(config=_A ) model.to(_A ) model.eval() __UpperCamelCase = model( _A , attention_mask=_A , token_type_ids=_A , labels=_A , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def _lowerCamelCase ( self : Dict , __A : Union[str, Any] , __A : Union[str, Any] , __A : int , __A : List[Any] , __A : Optional[Any] , __A : Dict , __A : List[Any] ): __UpperCamelCase = MobileBertForPreTraining(config=_A ) model.to(_A ) model.eval() __UpperCamelCase = model( _A , attention_mask=_A , token_type_ids=_A , labels=_A , next_sentence_label=_A , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def _lowerCamelCase ( self : List[Any] , __A : int , __A : str , __A : List[Any] , __A : Optional[int] , __A : str , __A : int , __A : Dict ): __UpperCamelCase = MobileBertForQuestionAnswering(config=_A ) model.to(_A ) model.eval() __UpperCamelCase = model( _A , attention_mask=_A , token_type_ids=_A , start_positions=_A , end_positions=_A , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : Tuple , __A : Union[str, Any] , __A : int , __A : Dict , __A : List[str] , __A : Tuple , __A : List[Any] , __A : Union[str, Any] ): __UpperCamelCase = self.num_labels __UpperCamelCase = MobileBertForSequenceClassification(_A ) model.to(_A ) model.eval() __UpperCamelCase = model(_A , attention_mask=_A , token_type_ids=_A , labels=_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : Optional[Any] , __A : Tuple , __A : str , __A : List[Any] , __A : str , __A : Union[str, Any] , __A : Union[str, Any] , __A : List[str] ): __UpperCamelCase = self.num_labels __UpperCamelCase = MobileBertForTokenClassification(config=_A ) model.to(_A ) model.eval() __UpperCamelCase = model(_A , attention_mask=_A , token_type_ids=_A , labels=_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : Tuple , __A : Any , __A : str , __A : List[Any] , __A : Tuple , __A : Optional[int] , __A : Optional[int] , __A : List[str] ): __UpperCamelCase = self.num_choices __UpperCamelCase = MobileBertForMultipleChoice(config=_A ) model.to(_A ) model.eval() __UpperCamelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __UpperCamelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __UpperCamelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __UpperCamelCase = model( _A , attention_mask=_A , token_type_ids=_A , labels=_A , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = self.prepare_config_and_inputs() ( __UpperCamelCase ) = config_and_inputs __UpperCamelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class snake_case ( __a , __a , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] =( ( MobileBertModel, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE_ : Union[str, Any] =( { """feature-extraction""": MobileBertModel, """fill-mask""": MobileBertForMaskedLM, """question-answering""": MobileBertForQuestionAnswering, """text-classification""": MobileBertForSequenceClassification, """token-classification""": MobileBertForTokenClassification, """zero-shot""": MobileBertForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE_ : str =True def _lowerCamelCase ( self : Optional[Any] , __A : Union[str, Any] , __A : Union[str, Any] , __A : Any=False ): __UpperCamelCase = super()._prepare_for_class(_A , _A , return_labels=_A ) if return_labels: if model_class in get_values(_A ): __UpperCamelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=_A ) __UpperCamelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_A ) return inputs_dict def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = MobileBertModelTester(self ) __UpperCamelCase = ConfigTester(self , config_class=_A , hidden_size=3_7 ) def _lowerCamelCase ( self : Union[str, Any] ): self.config_tester.run_common_tests() def _lowerCamelCase ( self : List[str] ): __UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*_A ) def _lowerCamelCase ( self : int ): __UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*_A ) def _lowerCamelCase ( self : str ): __UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*_A ) def _lowerCamelCase ( self : Optional[int] ): __UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*_A ) def _lowerCamelCase ( self : Optional[Any] ): __UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*_A ) def _lowerCamelCase ( self : Any ): __UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*_A ) def _lowerCamelCase ( self : Tuple ): __UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*_A ) def _lowerCamelCase ( self : List[Any] ): __UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*_A ) def lowercase__ ( __lowercase : List[str] ) -> Union[str, Any]: """simple docstring""" return torch.tensor( _lowerCAmelCase , dtype=torch.long , device=_lowerCAmelCase , ) a__ : Union[str, Any] =1E-3 @require_torch @require_sentencepiece @require_tokenizers class snake_case ( unittest.TestCase ): """simple docstring""" @slow def _lowerCamelCase ( self : Any ): __UpperCamelCase = MobileBertModel.from_pretrained('google/mobilebert-uncased' ).to(_A ) __UpperCamelCase = _long_tensor([[1_0_1, 7_1_1_0, 1_0_0_5, 1_0_5_6, 2_0_2_3, 1_1_3_3_3, 1_7_4_1_3, 1_0_2_9, 1_0_2]] ) with torch.no_grad(): __UpperCamelCase = model(_A )[0] __UpperCamelCase = torch.Size((1, 9, 5_1_2) ) self.assertEqual(output.shape , _A ) __UpperCamelCase = torch.tensor( [ [ [-2.473_6526e07, 8.269_1656e04, 1.652_1838e05], [-5.754_1704e-01, 3.905_6022e00, 4.401_1507e00], [2.604_7359e00, 1.567_7652e00, -1.732_4188e-01], ] ] , device=_A , ) # MobileBERT results range from 10e0 to 10e8. Even a 0.0000001% difference with a value of 10e8 results in a # ~1 difference, it's therefore not a good idea to measure using addition. # Here, we instead divide the expected result with the result in order to obtain ~1. We then check that the # result is held between bounds: 1 - TOLERANCE < expected_result / result < 1 + TOLERANCE __UpperCamelCase = torch.all((expected_slice / output[..., :3, :3]) >= 1 - TOLERANCE ) __UpperCamelCase = torch.all((expected_slice / output[..., :3, :3]) <= 1 + TOLERANCE ) self.assertTrue(lower_bound and upper_bound )
53
"""simple docstring""" def UpperCamelCase ( _lowerCAmelCase : int = 3, _lowerCAmelCase : int = 7, _lowerCAmelCase : int = 1000000 ) -> int: _UpperCAmelCase : Dict = 0 _UpperCAmelCase : int = 1 for current_denominator in range(1, limit + 1 ): _UpperCAmelCase : Union[str, Any] = current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: _UpperCAmelCase : Optional[Any] = current_numerator _UpperCAmelCase : str = current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=1_00_00_00))
246
0
'''simple docstring''' def __a ( UpperCAmelCase ) ->int: """simple docstring""" if not isinstance(__a , __a ): raise ValueError("""Input series is not valid, valid series - [2, 4, 6]""" ) if len(__a ) == 0: raise ValueError("""Input list must be a non empty list""" ) if len(__a ) == 1: return True A = series[1] - series[0] for index in range(len(__a ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def __a ( UpperCAmelCase ) ->int: """simple docstring""" if not isinstance(__a , __a ): raise ValueError("""Input series is not valid, valid series - [2, 4, 6]""" ) if len(__a ) == 0: raise ValueError("""Input list must be a non empty list""" ) A = 0 for val in series: answer += val return answer / len(__a ) if __name__ == "__main__": import doctest doctest.testmod()
350
'''simple docstring''' import argparse from copy import deepcopy import numpy as np from datasets import ClassLabel, DatasetDict, load_dataset from evaluate import load from transformers import ( AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, Trainer, TrainerCallback, TrainingArguments, set_seed, ) def __a ( ) ->str: """simple docstring""" A = argparse.ArgumentParser() parser.add_argument("""--model_ckpt""" , type=UpperCAmelCase , default="""microsoft/unixcoder-base-nine""" ) parser.add_argument("""--num_epochs""" , type=UpperCAmelCase , default=5 ) parser.add_argument("""--batch_size""" , type=UpperCAmelCase , default=6 ) parser.add_argument("""--gradient_accumulation_steps""" , type=UpperCAmelCase , default=1 ) parser.add_argument("""--freeze""" , type=UpperCAmelCase , default=UpperCAmelCase ) parser.add_argument("""--learning_rate""" , type=UpperCAmelCase , default=5E-4 ) parser.add_argument("""--seed""" , type=UpperCAmelCase , default=0 ) parser.add_argument("""--lr_scheduler_type""" , type=UpperCAmelCase , default="""cosine""" ) parser.add_argument("""--num_warmup_steps""" , type=UpperCAmelCase , default=10 ) parser.add_argument("""--weight_decay""" , type=UpperCAmelCase , default=0.01 ) parser.add_argument("""--output_dir""" , type=UpperCAmelCase , default="""./results""" ) return parser.parse_args() _lowerCamelCase : Optional[Any] = load('accuracy') def __a ( UpperCAmelCase ) ->Any: """simple docstring""" A , A = eval_pred A = np.argmax(UpperCAmelCase , axis=1 ) return metric.compute(predictions=UpperCAmelCase , references=UpperCAmelCase ) class __UpperCAmelCase ( A__ ): '''simple docstring''' def __init__(self : Union[str, Any] , _lowerCAmelCase : Any ): super().__init__() A = trainer def A (self : Dict , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Any , **_lowerCAmelCase : List[Any] ): if control.should_evaluate: A = deepcopy(_lowerCAmelCase ) self._trainer.evaluate(eval_dataset=self._trainer.train_dataset , metric_key_prefix="""train""" ) return control_copy def __a ( ) ->Optional[int]: """simple docstring""" A = get_args() set_seed(args.seed ) A = load_dataset("""codeparrot/codecomplex""" , split="""train""" ) A = dataset.train_test_split(test_size=0.2 ) A = train_test["""test"""].train_test_split(test_size=0.5 ) A = DatasetDict( { """train""": train_test["""train"""], """test""": test_validation["""train"""], """valid""": test_validation["""test"""], } ) print("""Loading tokenizer and model""" ) A = AutoTokenizer.from_pretrained(args.model_ckpt ) A = tokenizer.eos_token A = AutoModelForSequenceClassification.from_pretrained(args.model_ckpt , num_labels=7 ) A = model.config.eos_token_id if args.freeze: for param in model.roberta.parameters(): A = False A = ClassLabel(num_classes=7 , names=list(set(train_test_validation["""train"""]["""complexity"""] ) ) ) def tokenize(UpperCAmelCase ): A = tokenizer(example["""src"""] , truncation=UpperCAmelCase , max_length=1024 ) A = labels.straint(example["""complexity"""] ) return { "input_ids": inputs["input_ids"], "attention_mask": inputs["attention_mask"], "label": label, } A = train_test_validation.map( UpperCAmelCase , batched=UpperCAmelCase , remove_columns=train_test_validation["""train"""].column_names , ) A = DataCollatorWithPadding(tokenizer=UpperCAmelCase ) A = TrainingArguments( output_dir=args.output_dir , learning_rate=args.learning_rate , lr_scheduler_type=args.lr_scheduler_type , evaluation_strategy="""epoch""" , save_strategy="""epoch""" , logging_strategy="""epoch""" , per_device_train_batch_size=args.batch_size , per_device_eval_batch_size=args.batch_size , num_train_epochs=args.num_epochs , gradient_accumulation_steps=args.gradient_accumulation_steps , weight_decay=0.01 , metric_for_best_model="""accuracy""" , run_name="""complexity-java""" , report_to="""wandb""" , ) A = Trainer( model=UpperCAmelCase , args=UpperCAmelCase , train_dataset=tokenized_datasets["""train"""] , eval_dataset=tokenized_datasets["""valid"""] , tokenizer=UpperCAmelCase , data_collator=UpperCAmelCase , compute_metrics=UpperCAmelCase , ) print("""Training...""" ) trainer.add_callback(CustomCallback(UpperCAmelCase ) ) trainer.train() if __name__ == "__main__": main()
337
0
from tempfile import TemporaryDirectory from unittest import TestCase from unittest.mock import MagicMock, patch from transformers import AutoModel, TFAutoModel from transformers.onnx import FeaturesManager from transformers.testing_utils import SMALL_MODEL_IDENTIFIER, require_tf, require_torch @require_torch @require_tf class _lowerCAmelCase ( __a ): def __a ( self ) -> Any: lowerCAmelCase_ = SMALL_MODEL_IDENTIFIER lowerCAmelCase_ = "pt" lowerCAmelCase_ = "tf" def __a ( self , _UpperCamelCase ) -> Union[str, Any]: lowerCAmelCase_ = AutoModel.from_pretrained(self.test_model ) model_pt.save_pretrained(_UpperCamelCase ) def __a ( self , _UpperCamelCase ) -> Dict: lowerCAmelCase_ = TFAutoModel.from_pretrained(self.test_model , from_pt=_UpperCamelCase ) model_tf.save_pretrained(_UpperCamelCase ) def __a ( self ) -> List[str]: lowerCAmelCase_ = "mock_framework" # Framework provided - return whatever the user provides lowerCAmelCase_ = FeaturesManager.determine_framework(self.test_model , _UpperCamelCase ) self.assertEqual(_UpperCamelCase , _UpperCamelCase ) # Local checkpoint and framework provided - return provided framework # PyTorch checkpoint with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_UpperCamelCase ) lowerCAmelCase_ = FeaturesManager.determine_framework(_UpperCamelCase , _UpperCamelCase ) self.assertEqual(_UpperCamelCase , _UpperCamelCase ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_UpperCamelCase ) lowerCAmelCase_ = FeaturesManager.determine_framework(_UpperCamelCase , _UpperCamelCase ) self.assertEqual(_UpperCamelCase , _UpperCamelCase ) def __a ( self ) -> Union[str, Any]: # PyTorch checkpoint with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(_UpperCamelCase ) lowerCAmelCase_ = FeaturesManager.determine_framework(_UpperCamelCase ) self.assertEqual(_UpperCamelCase , self.framework_pt ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(_UpperCamelCase ) lowerCAmelCase_ = FeaturesManager.determine_framework(_UpperCamelCase ) self.assertEqual(_UpperCamelCase , self.framework_tf ) # Invalid local checkpoint with TemporaryDirectory() as local_invalid_ckpt: with self.assertRaises(_UpperCamelCase ): lowerCAmelCase_ = FeaturesManager.determine_framework(_UpperCamelCase ) def __a ( self ) -> List[str]: lowerCAmelCase_ = MagicMock(return_value=_UpperCamelCase ) with patch("transformers.onnx.features.is_tf_available" , _UpperCamelCase ): lowerCAmelCase_ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_UpperCamelCase , self.framework_pt ) # PyTorch not in environment -> use TensorFlow lowerCAmelCase_ = MagicMock(return_value=_UpperCamelCase ) with patch("transformers.onnx.features.is_torch_available" , _UpperCamelCase ): lowerCAmelCase_ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_UpperCamelCase , self.framework_tf ) # Both in environment -> use PyTorch lowerCAmelCase_ = MagicMock(return_value=_UpperCamelCase ) lowerCAmelCase_ = MagicMock(return_value=_UpperCamelCase ) with patch("transformers.onnx.features.is_tf_available" , _UpperCamelCase ), patch( "transformers.onnx.features.is_torch_available" , _UpperCamelCase ): lowerCAmelCase_ = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(_UpperCamelCase , self.framework_pt ) # Both not in environment -> raise error lowerCAmelCase_ = MagicMock(return_value=_UpperCamelCase ) lowerCAmelCase_ = MagicMock(return_value=_UpperCamelCase ) with patch("transformers.onnx.features.is_tf_available" , _UpperCamelCase ), patch( "transformers.onnx.features.is_torch_available" , _UpperCamelCase ): with self.assertRaises(_UpperCamelCase ): lowerCAmelCase_ = FeaturesManager.determine_framework(self.test_model )
231
import math def lowerCamelCase__ ( __lowerCAmelCase : int ): """simple docstring""" lowerCAmelCase_ = 0 lowerCAmelCase_ = 0 while num > 0: lowerCAmelCase_ = num % 8 lowerCAmelCase_ = octal + (remainder * math.floor(math.pow(10 , __lowerCAmelCase ) )) counter += 1 lowerCAmelCase_ = math.floor(num / 8 ) # basically /= 8 without remainder if any # This formatting removes trailing '.0' from `octal`. return F"""0o{int(__lowerCAmelCase )}""" def lowerCamelCase__ ( ): """simple docstring""" print("\n2 in octal is:" ) print(decimal_to_octal(2 ) ) # = 2 print("\n8 in octal is:" ) print(decimal_to_octal(8 ) ) # = 10 print("\n65 in octal is:" ) print(decimal_to_octal(65 ) ) # = 101 print("\n216 in octal is:" ) print(decimal_to_octal(216 ) ) # = 330 print("\n512 in octal is:" ) print(decimal_to_octal(512 ) ) # = 1000 print("\n" ) if __name__ == "__main__": main()
231
1
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments from transformers.hf_argparser import make_choice_type_function, string_to_bool # Since Python 3.10, we can use the builtin `|` operator for Union types # See PEP 604: https://peps.python.org/pep-0604 A : Union[str, Any] = sys.version_info >= (3, 10) def a__ ( __UpperCamelCase=None , __UpperCamelCase=None ): return field(default_factory=lambda: default , metadata=__UpperCamelCase ) @dataclass class lowerCamelCase : """simple docstring""" lowerCamelCase__ = 4_2 lowerCamelCase__ = 4_2 lowerCamelCase__ = 4_2 lowerCamelCase__ = 4_2 @dataclass class lowerCamelCase : """simple docstring""" lowerCamelCase__ = 4_2 lowerCamelCase__ = field(default='''toto''' , metadata={'''help''': '''help message'''} ) @dataclass class lowerCamelCase : """simple docstring""" lowerCamelCase__ = False lowerCamelCase__ = True lowerCamelCase__ = None class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCamelCase__ = '''titi''' lowerCamelCase__ = '''toto''' class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCamelCase__ = '''titi''' lowerCamelCase__ = '''toto''' lowerCamelCase__ = 4_2 @dataclass class lowerCamelCase : """simple docstring""" lowerCamelCase__ = '''toto''' def __A ( self : str ) -> int: SCREAMING_SNAKE_CASE_ = BasicEnum(self.foo ) @dataclass class lowerCamelCase : """simple docstring""" lowerCamelCase__ = '''toto''' def __A ( self : Optional[Any] ) -> List[Any]: SCREAMING_SNAKE_CASE_ = MixedTypeEnum(self.foo ) @dataclass class lowerCamelCase : """simple docstring""" lowerCamelCase__ = None lowerCamelCase__ = field(default=SCREAMING_SNAKE_CASE__ , metadata={'''help''': '''help message'''} ) lowerCamelCase__ = None lowerCamelCase__ = list_field(default=[] ) lowerCamelCase__ = list_field(default=[] ) @dataclass class lowerCamelCase : """simple docstring""" lowerCamelCase__ = list_field(default=[] ) lowerCamelCase__ = list_field(default=[1, 2, 3] ) lowerCamelCase__ = list_field(default=['''Hallo''', '''Bonjour''', '''Hello'''] ) lowerCamelCase__ = list_field(default=[0.1, 0.2, 0.3] ) @dataclass class lowerCamelCase : """simple docstring""" lowerCamelCase__ = field() lowerCamelCase__ = field() lowerCamelCase__ = field() def __A ( self : Optional[int] ) -> Dict: SCREAMING_SNAKE_CASE_ = BasicEnum(self.required_enum ) @dataclass class lowerCamelCase : """simple docstring""" lowerCamelCase__ = 4_2 lowerCamelCase__ = field() lowerCamelCase__ = None lowerCamelCase__ = field(default='''toto''' , metadata={'''help''': '''help message'''} ) lowerCamelCase__ = list_field(default=['''Hallo''', '''Bonjour''', '''Hello'''] ) if is_python_no_less_than_3_10: @dataclass class lowerCamelCase : """simple docstring""" lowerCamelCase__ = False lowerCamelCase__ = True lowerCamelCase__ = None @dataclass class lowerCamelCase : """simple docstring""" lowerCamelCase__ = None lowerCamelCase__ = field(default=SCREAMING_SNAKE_CASE__ , metadata={'''help''': '''help message'''} ) lowerCamelCase__ = None lowerCamelCase__ = list_field(default=[] ) lowerCamelCase__ = list_field(default=[] ) class lowerCamelCase (unittest.TestCase ): """simple docstring""" def __A ( self : List[Any] , __magic_name__ : argparse.ArgumentParser , __magic_name__ : argparse.ArgumentParser ) -> int: self.assertEqual(len(a._actions ) , len(b._actions ) ) for x, y in zip(a._actions , b._actions ): SCREAMING_SNAKE_CASE_ = {k: v for k, v in vars(__magic_name__ ).items() if k != "container"} SCREAMING_SNAKE_CASE_ = {k: v for k, v in vars(__magic_name__ ).items() if k != "container"} # Choices with mixed type have custom function as "type" # So we need to compare results directly for equality if xx.get("choices" , __magic_name__ ) and yy.get("choices" , __magic_name__ ): for expected_choice in yy["choices"] + xx["choices"]: self.assertEqual(xx["type"](__magic_name__ ) , yy["type"](__magic_name__ ) ) del xx["type"], yy["type"] self.assertEqual(__magic_name__ , __magic_name__ ) def __A ( self : Union[str, Any] ) -> str: SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() expected.add_argument("--foo" , type=__magic_name__ , required=__magic_name__ ) expected.add_argument("--bar" , type=__magic_name__ , required=__magic_name__ ) expected.add_argument("--baz" , type=__magic_name__ , required=__magic_name__ ) expected.add_argument("--flag" , type=__magic_name__ , default=__magic_name__ , const=__magic_name__ , nargs="?" ) self.argparsersEqual(__magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = ["--foo", "1", "--baz", "quux", "--bar", "0.5"] ((SCREAMING_SNAKE_CASE_ ) , ) = parser.parse_args_into_dataclasses(__magic_name__ , look_for_args_file=__magic_name__ ) self.assertFalse(example.flag ) def __A ( self : List[Any] ) -> Any: SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() expected.add_argument("--foo" , default=42 , type=__magic_name__ ) expected.add_argument("--baz" , default="toto" , type=__magic_name__ , help="help message" ) self.argparsersEqual(__magic_name__ , __magic_name__ ) def __A ( self : Optional[int] ) -> List[str]: SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() expected.add_argument("--foo" , type=__magic_name__ , default=__magic_name__ , const=__magic_name__ , nargs="?" ) expected.add_argument("--baz" , type=__magic_name__ , default=__magic_name__ , const=__magic_name__ , nargs="?" ) # A boolean no_* argument always has to come after its "default: True" regular counter-part # and its default must be set to False expected.add_argument("--no_baz" , action="store_false" , default=__magic_name__ , dest="baz" ) expected.add_argument("--opt" , type=__magic_name__ , default=__magic_name__ ) SCREAMING_SNAKE_CASE_ = [WithDefaultBoolExample] if is_python_no_less_than_3_10: dataclass_types.append(__magic_name__ ) for dataclass_type in dataclass_types: SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) self.argparsersEqual(__magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = parser.parse_args([] ) self.assertEqual(__magic_name__ , Namespace(foo=__magic_name__ , baz=__magic_name__ , opt=__magic_name__ ) ) SCREAMING_SNAKE_CASE_ = parser.parse_args(["--foo", "--no_baz"] ) self.assertEqual(__magic_name__ , Namespace(foo=__magic_name__ , baz=__magic_name__ , opt=__magic_name__ ) ) SCREAMING_SNAKE_CASE_ = parser.parse_args(["--foo", "--baz"] ) self.assertEqual(__magic_name__ , Namespace(foo=__magic_name__ , baz=__magic_name__ , opt=__magic_name__ ) ) SCREAMING_SNAKE_CASE_ = parser.parse_args(["--foo", "True", "--baz", "True", "--opt", "True"] ) self.assertEqual(__magic_name__ , Namespace(foo=__magic_name__ , baz=__magic_name__ , opt=__magic_name__ ) ) SCREAMING_SNAKE_CASE_ = parser.parse_args(["--foo", "False", "--baz", "False", "--opt", "False"] ) self.assertEqual(__magic_name__ , Namespace(foo=__magic_name__ , baz=__magic_name__ , opt=__magic_name__ ) ) def __A ( self : Tuple ) -> Optional[int]: SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() expected.add_argument( "--foo" , default="toto" , choices=["titi", "toto", 42] , type=make_choice_type_function(["titi", "toto", 42] ) , ) self.argparsersEqual(__magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = parser.parse_args([] ) self.assertEqual(args.foo , "toto" ) SCREAMING_SNAKE_CASE_ = parser.parse_args_into_dataclasses([] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.toto ) SCREAMING_SNAKE_CASE_ = parser.parse_args(["--foo", "titi"] ) self.assertEqual(args.foo , "titi" ) SCREAMING_SNAKE_CASE_ = parser.parse_args_into_dataclasses(["--foo", "titi"] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.titi ) SCREAMING_SNAKE_CASE_ = parser.parse_args(["--foo", "42"] ) self.assertEqual(args.foo , 42 ) SCREAMING_SNAKE_CASE_ = parser.parse_args_into_dataclasses(["--foo", "42"] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.fourtytwo ) def __A ( self : Optional[Any] ) -> List[Any]: @dataclass class lowerCamelCase : """simple docstring""" lowerCamelCase__ = '''toto''' SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() expected.add_argument( "--foo" , default="toto" , choices=("titi", "toto", 42) , type=make_choice_type_function(["titi", "toto", 42] ) , ) self.argparsersEqual(__magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = parser.parse_args([] ) self.assertEqual(args.foo , "toto" ) SCREAMING_SNAKE_CASE_ = parser.parse_args(["--foo", "titi"] ) self.assertEqual(args.foo , "titi" ) SCREAMING_SNAKE_CASE_ = parser.parse_args(["--foo", "42"] ) self.assertEqual(args.foo , 42 ) def __A ( self : Optional[Any] ) -> List[str]: SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() expected.add_argument("--foo_int" , nargs="+" , default=[] , type=__magic_name__ ) expected.add_argument("--bar_int" , nargs="+" , default=[1, 2, 3] , type=__magic_name__ ) expected.add_argument("--foo_str" , nargs="+" , default=["Hallo", "Bonjour", "Hello"] , type=__magic_name__ ) expected.add_argument("--foo_float" , nargs="+" , default=[0.1, 0.2, 0.3] , type=__magic_name__ ) self.argparsersEqual(__magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = parser.parse_args([] ) self.assertEqual( __magic_name__ , Namespace(foo_int=[] , bar_int=[1, 2, 3] , foo_str=["Hallo", "Bonjour", "Hello"] , foo_float=[0.1, 0.2, 0.3] ) , ) SCREAMING_SNAKE_CASE_ = parser.parse_args("--foo_int 1 --bar_int 2 3 --foo_str a b c --foo_float 0.1 0.7".split() ) self.assertEqual(__magic_name__ , Namespace(foo_int=[1] , bar_int=[2, 3] , foo_str=["a", "b", "c"] , foo_float=[0.1, 0.7] ) ) def __A ( self : Union[str, Any] ) -> str: SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() expected.add_argument("--foo" , default=__magic_name__ , type=__magic_name__ ) expected.add_argument("--bar" , default=__magic_name__ , type=__magic_name__ , help="help message" ) expected.add_argument("--baz" , default=__magic_name__ , type=__magic_name__ ) expected.add_argument("--ces" , nargs="+" , default=[] , type=__magic_name__ ) expected.add_argument("--des" , nargs="+" , default=[] , type=__magic_name__ ) SCREAMING_SNAKE_CASE_ = [OptionalExample] if is_python_no_less_than_3_10: dataclass_types.append(__magic_name__ ) for dataclass_type in dataclass_types: SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) self.argparsersEqual(__magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = parser.parse_args([] ) self.assertEqual(__magic_name__ , Namespace(foo=__magic_name__ , bar=__magic_name__ , baz=__magic_name__ , ces=[] , des=[] ) ) SCREAMING_SNAKE_CASE_ = parser.parse_args("--foo 12 --bar 3.14 --baz 42 --ces a b c --des 1 2 3".split() ) self.assertEqual(__magic_name__ , Namespace(foo=12 , bar=3.14 , baz="42" , ces=["a", "b", "c"] , des=[1, 2, 3] ) ) def __A ( self : int ) -> List[str]: SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() expected.add_argument("--required_list" , nargs="+" , type=__magic_name__ , required=__magic_name__ ) expected.add_argument("--required_str" , type=__magic_name__ , required=__magic_name__ ) expected.add_argument( "--required_enum" , type=make_choice_type_function(["titi", "toto"] ) , choices=["titi", "toto"] , required=__magic_name__ , ) self.argparsersEqual(__magic_name__ , __magic_name__ ) def __A ( self : Any ) -> Optional[int]: SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() expected.add_argument("--foo" , type=__magic_name__ , required=__magic_name__ ) expected.add_argument( "--required_enum" , type=make_choice_type_function(["titi", "toto"] ) , choices=["titi", "toto"] , required=__magic_name__ , ) expected.add_argument("--opt" , type=__magic_name__ , default=__magic_name__ ) expected.add_argument("--baz" , default="toto" , type=__magic_name__ , help="help message" ) expected.add_argument("--foo_str" , nargs="+" , default=["Hallo", "Bonjour", "Hello"] , type=__magic_name__ ) self.argparsersEqual(__magic_name__ , __magic_name__ ) def __A ( self : List[Any] ) -> Any: SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) SCREAMING_SNAKE_CASE_ = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, } SCREAMING_SNAKE_CASE_ = parser.parse_dict(__magic_name__ )[0] SCREAMING_SNAKE_CASE_ = BasicExample(**__magic_name__ ) self.assertEqual(__magic_name__ , __magic_name__ ) def __A ( self : Tuple ) -> Dict: SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) SCREAMING_SNAKE_CASE_ = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, "extra": 42, } self.assertRaises(__magic_name__ , parser.parse_dict , __magic_name__ , allow_extra_keys=__magic_name__ ) def __A ( self : Optional[int] ) -> int: SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) SCREAMING_SNAKE_CASE_ = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, } with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE_ = os.path.join(__magic_name__ , "temp_json" ) os.mkdir(__magic_name__ ) with open(temp_local_path + ".json" , "w+" ) as f: json.dump(__magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = parser.parse_yaml_file(Path(temp_local_path + ".json" ) )[0] SCREAMING_SNAKE_CASE_ = BasicExample(**__magic_name__ ) self.assertEqual(__magic_name__ , __magic_name__ ) def __A ( self : List[Any] ) -> Any: SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) SCREAMING_SNAKE_CASE_ = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, } with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE_ = os.path.join(__magic_name__ , "temp_yaml" ) os.mkdir(__magic_name__ ) with open(temp_local_path + ".yaml" , "w+" ) as f: yaml.dump(__magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = parser.parse_yaml_file(Path(temp_local_path + ".yaml" ) )[0] SCREAMING_SNAKE_CASE_ = BasicExample(**__magic_name__ ) self.assertEqual(__magic_name__ , __magic_name__ ) def __A ( self : Any ) -> Any: SCREAMING_SNAKE_CASE_ = HfArgumentParser(__magic_name__ ) self.assertIsNotNone(__magic_name__ )
369
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowerCamelCase (SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCamelCase__ = ['''image_processor''', '''tokenizer'''] lowerCamelCase__ = '''ViltImageProcessor''' lowerCamelCase__ = ('''BertTokenizer''', '''BertTokenizerFast''') def __init__( self : Optional[int] , __magic_name__ : str=None , __magic_name__ : List[str]=None , **__magic_name__ : Any ) -> str: SCREAMING_SNAKE_CASE_ = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , __magic_name__ , ) SCREAMING_SNAKE_CASE_ = kwargs.pop("feature_extractor" ) SCREAMING_SNAKE_CASE_ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(__magic_name__ , __magic_name__ ) SCREAMING_SNAKE_CASE_ = self.image_processor def __call__( self : List[str] , __magic_name__ : List[str] , __magic_name__ : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , __magic_name__ : bool = True , __magic_name__ : Union[bool, str, PaddingStrategy] = False , __magic_name__ : Union[bool, str, TruncationStrategy] = None , __magic_name__ : Optional[int] = None , __magic_name__ : int = 0 , __magic_name__ : Optional[int] = None , __magic_name__ : Optional[bool] = None , __magic_name__ : Optional[bool] = None , __magic_name__ : bool = False , __magic_name__ : bool = False , __magic_name__ : bool = False , __magic_name__ : bool = False , __magic_name__ : bool = True , __magic_name__ : Optional[Union[str, TensorType]] = None , **__magic_name__ : str , ) -> BatchEncoding: SCREAMING_SNAKE_CASE_ = self.tokenizer( text=__magic_name__ , add_special_tokens=__magic_name__ , padding=__magic_name__ , truncation=__magic_name__ , max_length=__magic_name__ , stride=__magic_name__ , pad_to_multiple_of=__magic_name__ , return_token_type_ids=__magic_name__ , return_attention_mask=__magic_name__ , return_overflowing_tokens=__magic_name__ , return_special_tokens_mask=__magic_name__ , return_offsets_mapping=__magic_name__ , return_length=__magic_name__ , verbose=__magic_name__ , return_tensors=__magic_name__ , **__magic_name__ , ) # add pixel_values + pixel_mask SCREAMING_SNAKE_CASE_ = self.image_processor(__magic_name__ , return_tensors=__magic_name__ ) encoding.update(__magic_name__ ) return encoding def __A ( self : Optional[int] , *__magic_name__ : List[Any] , **__magic_name__ : Optional[Any] ) -> Any: return self.tokenizer.batch_decode(*__magic_name__ , **__magic_name__ ) def __A ( self : Dict , *__magic_name__ : List[Any] , **__magic_name__ : Union[str, Any] ) -> str: return self.tokenizer.decode(*__magic_name__ , **__magic_name__ ) @property def __A ( self : Optional[int] ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ = self.tokenizer.model_input_names SCREAMING_SNAKE_CASE_ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def __A ( self : Dict ) -> List[Any]: warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , __magic_name__ , ) return self.image_processor_class @property def __A ( self : int ) -> List[Any]: warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , __magic_name__ , ) return self.image_processor
305
0
'''simple docstring''' import torch from transformers import AutoModel class snake_case ( torch.nn.Module ): """simple docstring""" def __init__( self : Optional[int] , __A : List[Any]="sayef/fsner-bert-base-uncased" ): super(__A , self ).__init__() __UpperCamelCase = AutoModel.from_pretrained(__A , return_dict=__A ) __UpperCamelCase = torch.nn.CosineSimilarity(3 , 1e-08 ) __UpperCamelCase = torch.nn.Softmax(dim=1 ) def _lowerCamelCase ( self : Tuple , **__A : Optional[int] ): return self.bert(**__A ).last_hidden_state def _lowerCamelCase ( self : Tuple , __A : Tuple ): return token_embeddings.sum(2 , keepdim=__A ) def _lowerCamelCase ( self : List[Any] , __A : str , __A : int , __A : str=1 ): return self.softmax(T * self.cos(__A , __A ) ) def _lowerCamelCase ( self : Optional[int] , __A : str , __A : Any ): __UpperCamelCase = W_supports['sizes'].tolist() __UpperCamelCase = W_supports['start_token_id'].item() __UpperCamelCase = W_supports['end_token_id'].item() del W_supports["sizes"] del W_supports["start_token_id"] del W_supports["end_token_id"] __UpperCamelCase = self.BERT(**__A ) __UpperCamelCase = self.BERT(**__A ) __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = W_supports['input_ids'] == start_token_id __UpperCamelCase = W_supports['input_ids'] == end_token_id for i, size in enumerate(__A ): if i == 0: __UpperCamelCase = 0 else: __UpperCamelCase = support_sizes[i - 1] __UpperCamelCase = S[s : s + size][start_token_masks[s : s + size]] __UpperCamelCase = S[s : s + size][end_token_masks[s : s + size]] __UpperCamelCase = torch.matmul(q[i] , s_start.T ).sum(1 ).softmax(0 ) __UpperCamelCase = torch.matmul(q[i] , s_end.T ).sum(1 ).softmax(0 ) if p_starts is not None: __UpperCamelCase = torch.vstack((p_starts, p_start) ) __UpperCamelCase = torch.vstack((p_ends, p_end) ) else: __UpperCamelCase = p_start __UpperCamelCase = p_end return p_starts, p_ends
53
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ :List[Any] = logging.get_logger(__name__) lowerCAmelCase__ :Tuple = {'''ctrl''': '''https://huggingface.co/ctrl/resolve/main/config.json'''} class __a ( UpperCAmelCase ): _a : str = 'ctrl' _a : Tuple = ['past_key_values'] _a : List[Any] = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self , _SCREAMING_SNAKE_CASE=246534 , _SCREAMING_SNAKE_CASE=256 , _SCREAMING_SNAKE_CASE=1280 , _SCREAMING_SNAKE_CASE=8192 , _SCREAMING_SNAKE_CASE=48 , _SCREAMING_SNAKE_CASE=16 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=1e-6 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=True , **_SCREAMING_SNAKE_CASE , ) -> List[str]: """simple docstring""" _UpperCAmelCase = vocab_size _UpperCAmelCase = n_positions _UpperCAmelCase = n_embd _UpperCAmelCase = n_layer _UpperCAmelCase = n_head _UpperCAmelCase = dff _UpperCAmelCase = resid_pdrop _UpperCAmelCase = embd_pdrop _UpperCAmelCase = layer_norm_epsilon _UpperCAmelCase = initializer_range _UpperCAmelCase = use_cache super().__init__(**_SCREAMING_SNAKE_CASE )
329
0
def lowerCamelCase__ ( UpperCamelCase__ : list ) -> list: '''simple docstring''' def merge(UpperCamelCase__ : list , UpperCamelCase__ : list ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(UpperCamelCase__ ) <= 1: return collection _snake_case = len(UpperCamelCase__ ) // 2 return merge(merge_sort(collection[:mid] ) , merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ = input("""Enter numbers separated by a comma:\n""").strip() UpperCAmelCase_ = [int(item) for item in user_input.split(""",""")] print(*merge_sort(unsorted), sep=""",""")
295
import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class UpperCamelCase_ : @property def lowerCAmelCase ( self ) -> int: return self.get_dummy_input() @property def lowerCAmelCase ( self ) -> Optional[Any]: if self.block_type == "down": return (4, 32, 16, 16) elif self.block_type == "mid": return (4, 32, 32, 32) elif self.block_type == "up": return (4, 32, 64, 64) raise ValueError(F'''\'{self.block_type}\' is not a supported block_type. Set it to \'up\', \'mid\', or \'down\'.''' ) def lowerCAmelCase ( self , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=False , lowerCAmelCase_=False , ) -> List[str]: _snake_case = 4 _snake_case = 32 _snake_case = (32, 32) _snake_case = torch.manual_seed(0 ) _snake_case = torch.device(lowerCAmelCase_ ) _snake_case = (batch_size, num_channels) + sizes _snake_case = randn_tensor(lowerCAmelCase_ , generator=lowerCAmelCase_ , device=lowerCAmelCase_ ) _snake_case = {'hidden_states': hidden_states} if include_temb: _snake_case = 128 _snake_case = randn_tensor((batch_size, temb_channels) , generator=lowerCAmelCase_ , device=lowerCAmelCase_ ) if include_res_hidden_states_tuple: _snake_case = torch.manual_seed(1 ) _snake_case = (randn_tensor(lowerCAmelCase_ , generator=lowerCAmelCase_ , device=lowerCAmelCase_ ),) if include_encoder_hidden_states: _snake_case = floats_tensor((batch_size, 32, 32) ).to(lowerCAmelCase_ ) if include_skip_sample: _snake_case = randn_tensor(((batch_size, 3) + sizes) , generator=lowerCAmelCase_ , device=lowerCAmelCase_ ) return dummy_input def lowerCAmelCase ( self ) -> Tuple: _snake_case = { 'in_channels': 32, 'out_channels': 32, 'temb_channels': 128, } if self.block_type == "up": _snake_case = 32 if self.block_type == "mid": init_dict.pop('out_channels' ) _snake_case = self.dummy_input return init_dict, inputs_dict def lowerCAmelCase ( self , lowerCAmelCase_ ) -> Optional[int]: _snake_case , _snake_case = self.prepare_init_args_and_inputs_for_common() _snake_case = self.block_class(**lowerCAmelCase_ ) unet_block.to(lowerCAmelCase_ ) unet_block.eval() with torch.no_grad(): _snake_case = unet_block(**lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _snake_case = output[0] self.assertEqual(output.shape , self.output_shape ) _snake_case = output[0, -1, -3:, -3:] _snake_case = torch.tensor(lowerCAmelCase_ ).to(lowerCAmelCase_ ) assert torch_all_close(output_slice.flatten() , lowerCAmelCase_ , atol=5E-3 ) @unittest.skipIf(torch_device == 'mps' , 'Training is not supported in mps' ) def lowerCAmelCase ( self ) -> Tuple: _snake_case , _snake_case = self.prepare_init_args_and_inputs_for_common() _snake_case = self.block_class(**lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.train() _snake_case = model(**lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _snake_case = output[0] _snake_case = torch.device(lowerCAmelCase_ ) _snake_case = randn_tensor(output.shape , device=lowerCAmelCase_ ) _snake_case = torch.nn.functional.mse_loss(lowerCAmelCase_ , lowerCAmelCase_ ) loss.backward()
295
1
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType A_ : Tuple = logging.get_logger(__name__) A_ : Optional[int] = { """openai/whisper-base""": """https://huggingface.co/openai/whisper-base/resolve/main/config.json""", } # fmt: off A_ : Tuple = [ 1, 2, 7, 8, 9, 1_0, 1_4, 2_5, 2_6, 2_7, 2_8, 2_9, 3_1, 5_8, 5_9, 6_0, 6_1, 6_2, 6_3, 9_0, 9_1, 9_2, 9_3, 3_5_7, 3_6_6, 4_3_8, 5_3_2, 6_8_5, 7_0_5, 7_9_6, 9_3_0, 1_0_5_8, 1_2_2_0, 1_2_6_7, 1_2_7_9, 1_3_0_3, 1_3_4_3, 1_3_7_7, 1_3_9_1, 1_6_3_5, 1_7_8_2, 1_8_7_5, 2_1_6_2, 2_3_6_1, 2_4_8_8, 3_4_6_7, 4_0_0_8, 4_2_1_1, 4_6_0_0, 4_8_0_8, 5_2_9_9, 5_8_5_5, 6_3_2_9, 7_2_0_3, 9_6_0_9, 9_9_5_9, 1_0_5_6_3, 1_0_7_8_6, 1_1_4_2_0, 1_1_7_0_9, 1_1_9_0_7, 1_3_1_6_3, 1_3_6_9_7, 1_3_7_0_0, 1_4_8_0_8, 1_5_3_0_6, 1_6_4_1_0, 1_6_7_9_1, 1_7_9_9_2, 1_9_2_0_3, 1_9_5_1_0, 2_0_7_2_4, 2_2_3_0_5, 2_2_9_3_5, 2_7_0_0_7, 3_0_1_0_9, 3_0_4_2_0, 3_3_4_0_9, 3_4_9_4_9, 4_0_2_8_3, 4_0_4_9_3, 4_0_5_4_9, 4_7_2_8_2, 4_9_1_4_6, 5_0_2_5_7, 5_0_3_5_9, 5_0_3_6_0, 5_0_3_6_1 ] A_ : Optional[Any] = [ 1, 2, 7, 8, 9, 1_0, 1_4, 2_5, 2_6, 2_7, 2_8, 2_9, 3_1, 5_8, 5_9, 6_0, 6_1, 6_2, 6_3, 9_0, 9_1, 9_2, 9_3, 3_5_9, 5_0_3, 5_2_2, 5_4_2, 8_7_3, 8_9_3, 9_0_2, 9_1_8, 9_2_2, 9_3_1, 1_3_5_0, 1_8_5_3, 1_9_8_2, 2_4_6_0, 2_6_2_7, 3_2_4_6, 3_2_5_3, 3_2_6_8, 3_5_3_6, 3_8_4_6, 3_9_6_1, 4_1_8_3, 4_6_6_7, 6_5_8_5, 6_6_4_7, 7_2_7_3, 9_0_6_1, 9_3_8_3, 1_0_4_2_8, 1_0_9_2_9, 1_1_9_3_8, 1_2_0_3_3, 1_2_3_3_1, 1_2_5_6_2, 1_3_7_9_3, 1_4_1_5_7, 1_4_6_3_5, 1_5_2_6_5, 1_5_6_1_8, 1_6_5_5_3, 1_6_6_0_4, 1_8_3_6_2, 1_8_9_5_6, 2_0_0_7_5, 2_1_6_7_5, 2_2_5_2_0, 2_6_1_3_0, 2_6_1_6_1, 2_6_4_3_5, 2_8_2_7_9, 2_9_4_6_4, 3_1_6_5_0, 3_2_3_0_2, 3_2_4_7_0, 3_6_8_6_5, 4_2_8_6_3, 4_7_4_2_5, 4_9_8_7_0, 5_0_2_5_4, 5_0_2_5_8, 5_0_3_6_0, 5_0_3_6_1, 5_0_3_6_2 ] class lowercase ( _lowerCamelCase ): """simple docstring""" UpperCAmelCase = """whisper""" UpperCAmelCase = ["""past_key_values"""] UpperCAmelCase = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self ,a_=51_865 ,a_=80 ,a_=6 ,a_=4 ,a_=6 ,a_=4 ,a_=1_536 ,a_=1_536 ,a_=0.0 ,a_=0.0 ,a_=50_257 ,a_=True ,a_=True ,a_="gelu" ,a_=256 ,a_=0.0 ,a_=0.0 ,a_=0.0 ,a_=0.02 ,a_=False ,a_=1_500 ,a_=448 ,a_=50_256 ,a_=50_256 ,a_=50_256 ,a_=None ,a_=[220, 50_256] ,a_=False ,a_=256 ,a_=False ,a_=0.05 ,a_=10 ,a_=2 ,a_=0.0 ,a_=10 ,a_=0 ,a_=7 ,**a_ ,) -> List[str]: _UpperCAmelCase : Optional[Any] = vocab_size _UpperCAmelCase : str = num_mel_bins _UpperCAmelCase : Any = d_model _UpperCAmelCase : Optional[Any] = encoder_layers _UpperCAmelCase : List[str] = encoder_attention_heads _UpperCAmelCase : Union[str, Any] = decoder_layers _UpperCAmelCase : Tuple = decoder_attention_heads _UpperCAmelCase : Dict = decoder_ffn_dim _UpperCAmelCase : Optional[Any] = encoder_ffn_dim _UpperCAmelCase : List[Any] = dropout _UpperCAmelCase : List[str] = attention_dropout _UpperCAmelCase : Optional[int] = activation_dropout _UpperCAmelCase : Optional[Any] = activation_function _UpperCAmelCase : Union[str, Any] = init_std _UpperCAmelCase : Optional[Any] = encoder_layerdrop _UpperCAmelCase : List[str] = decoder_layerdrop _UpperCAmelCase : Dict = use_cache _UpperCAmelCase : List[str] = encoder_layers _UpperCAmelCase : List[str] = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCAmelCase : Any = max_source_positions _UpperCAmelCase : List[str] = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. _UpperCAmelCase : str = classifier_proj_size _UpperCAmelCase : Dict = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _UpperCAmelCase : int = apply_spec_augment _UpperCAmelCase : int = mask_time_prob _UpperCAmelCase : List[str] = mask_time_length _UpperCAmelCase : List[Any] = mask_time_min_masks _UpperCAmelCase : Optional[Any] = mask_feature_prob _UpperCAmelCase : int = mask_feature_length _UpperCAmelCase : Dict = mask_feature_min_masks _UpperCAmelCase : Union[str, Any] = median_filter_width super().__init__( pad_token_id=a_ ,bos_token_id=a_ ,eos_token_id=a_ ,is_encoder_decoder=a_ ,decoder_start_token_id=a_ ,suppress_tokens=a_ ,begin_suppress_tokens=a_ ,**a_ ,) class lowercase ( _lowerCamelCase ): """simple docstring""" @property def _snake_case ( self ) -> Mapping[str, Mapping[int, str]]: _UpperCAmelCase : Optional[Any] = OrderedDict( [ ("""input_features""", {0: """batch""", 1: """feature_size""", 2: """encoder_sequence"""}), ] ) if self.use_past: _UpperCAmelCase : Any = {0: """batch"""} else: _UpperCAmelCase : List[str] = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(a_ ,direction="""inputs""" ) return common_inputs def _snake_case ( self ,a_ ,a_ = -1 ,a_ = -1 ,a_ = False ,a_ = None ,a_ = 22_050 ,a_ = 5.0 ,a_ = 220 ,) -> Mapping[str, Any]: _UpperCAmelCase : Optional[int] = OrderedDict() _UpperCAmelCase : Any = OnnxConfig.generate_dummy_inputs( self ,preprocessor=preprocessor.feature_extractor ,batch_size=a_ ,framework=a_ ,sampling_rate=a_ ,time_duration=a_ ,frequency=a_ ,) _UpperCAmelCase : Union[str, Any] = encoder_inputs["""input_features"""].shape[2] _UpperCAmelCase : Optional[Any] = encoder_sequence_length // 2 if self.use_past else seq_length _UpperCAmelCase : Dict = super().generate_dummy_inputs( preprocessor.tokenizer ,a_ ,a_ ,a_ ,a_ ) _UpperCAmelCase : Union[str, Any] = encoder_inputs.pop("""input_features""" ) _UpperCAmelCase : Union[str, Any] = decoder_inputs.pop("""decoder_input_ids""" ) if "past_key_values" in decoder_inputs: _UpperCAmelCase : Tuple = decoder_inputs.pop("""past_key_values""" ) return dummy_inputs @property def _snake_case ( self ) -> float: return 1E-3
215
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import _LazyModule A_ : int = {"""tokenization_bertweet""": ["""BertweetTokenizer"""]} if TYPE_CHECKING: from .tokenization_bertweet import BertweetTokenizer else: import sys A_ : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
215
1
"""simple docstring""" import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoFeatureExtractor, WavaVecaFeatureExtractor from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 _A = get_tests_dir("""fixtures""") class _lowerCamelCase ( unittest.TestCase ): def _lowerCAmelCase ( self : str ) -> Tuple: """simple docstring""" # A mock response for an HTTP head request to emulate server down lowerCAmelCase__ : Any = mock.Mock() lowerCAmelCase__ : List[str] = 5_00 lowerCAmelCase__ : Tuple = {} lowerCAmelCase__ : Union[str, Any] = HTTPError lowerCAmelCase__ : Union[str, Any] = {} # Download this model to make sure it's in the cache. lowerCAmelCase__ : int = WavaVecaFeatureExtractor.from_pretrained("""hf-internal-testing/tiny-random-wav2vec2""" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("""requests.Session.request""" , return_value=UpperCamelCase ) as mock_head: lowerCAmelCase__ : str = WavaVecaFeatureExtractor.from_pretrained("""hf-internal-testing/tiny-random-wav2vec2""" ) # This check we did call the fake head request mock_head.assert_called() def _lowerCAmelCase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" # This test is for deprecated behavior and can be removed in v5 lowerCAmelCase__ : str = WavaVecaFeatureExtractor.from_pretrained( """https://huggingface.co/hf-internal-testing/tiny-random-wav2vec2/resolve/main/preprocessor_config.json""" ) @is_staging_test class _lowerCamelCase ( unittest.TestCase ): @classmethod def _lowerCAmelCase ( cls : str ) -> str: """simple docstring""" lowerCAmelCase__ : Optional[Any] = TOKEN HfFolder.save_token(UpperCamelCase ) @classmethod def _lowerCAmelCase ( cls : Any ) -> List[Any]: """simple docstring""" try: delete_repo(token=cls._token , repo_id="""test-feature-extractor""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-feature-extractor-org""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""test-dynamic-feature-extractor""" ) except HTTPError: pass def _lowerCAmelCase ( self : Optional[Any] ) -> str: """simple docstring""" lowerCAmelCase__ : Tuple = WavaVecaFeatureExtractor.from_pretrained(UpperCamelCase ) feature_extractor.push_to_hub("""test-feature-extractor""" , use_auth_token=self._token ) lowerCAmelCase__ : List[Any] = WavaVecaFeatureExtractor.from_pretrained(f"""{USER}/test-feature-extractor""" ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(UpperCamelCase , getattr(UpperCamelCase , UpperCamelCase ) ) # Reset repo delete_repo(token=self._token , repo_id="""test-feature-extractor""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained( UpperCamelCase , repo_id="""test-feature-extractor""" , push_to_hub=UpperCamelCase , use_auth_token=self._token ) lowerCAmelCase__ : Optional[Any] = WavaVecaFeatureExtractor.from_pretrained(f"""{USER}/test-feature-extractor""" ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(UpperCamelCase , getattr(UpperCamelCase , UpperCamelCase ) ) def _lowerCAmelCase ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" lowerCAmelCase__ : int = WavaVecaFeatureExtractor.from_pretrained(UpperCamelCase ) feature_extractor.push_to_hub("""valid_org/test-feature-extractor""" , use_auth_token=self._token ) lowerCAmelCase__ : Optional[Any] = WavaVecaFeatureExtractor.from_pretrained("""valid_org/test-feature-extractor""" ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(UpperCamelCase , getattr(UpperCamelCase , UpperCamelCase ) ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-feature-extractor""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained( UpperCamelCase , repo_id="""valid_org/test-feature-extractor-org""" , push_to_hub=UpperCamelCase , use_auth_token=self._token ) lowerCAmelCase__ : str = WavaVecaFeatureExtractor.from_pretrained("""valid_org/test-feature-extractor-org""" ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(UpperCamelCase , getattr(UpperCamelCase , UpperCamelCase ) ) def _lowerCAmelCase ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" CustomFeatureExtractor.register_for_auto_class() lowerCAmelCase__ : Tuple = CustomFeatureExtractor.from_pretrained(UpperCamelCase ) feature_extractor.push_to_hub("""test-dynamic-feature-extractor""" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual( feature_extractor.auto_map , {"""AutoFeatureExtractor""": """custom_feature_extraction.CustomFeatureExtractor"""} , ) lowerCAmelCase__ : Tuple = AutoFeatureExtractor.from_pretrained( f"""{USER}/test-dynamic-feature-extractor""" , trust_remote_code=UpperCamelCase ) # Can't make an isinstance check because the new_feature_extractor is from the CustomFeatureExtractor class of a dynamic module self.assertEqual(new_feature_extractor.__class__.__name__ , """CustomFeatureExtractor""" )
212
"""simple docstring""" from argparse import ArgumentParser from . import BaseTransformersCLICommand def lowercase_ ( __UpperCAmelCase ) -> List[str]: return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class _lowerCamelCase ( a_ ): @staticmethod def _lowerCAmelCase ( UpperCamelCase : ArgumentParser ) -> List[Any]: """simple docstring""" lowerCAmelCase__ : int = parser.add_parser("""download""" ) download_parser.add_argument( """--cache-dir""" , type=UpperCamelCase , default=UpperCamelCase , help="""Path to location to store the models""" ) download_parser.add_argument( """--force""" , action="""store_true""" , help="""Force the model to be download even if already in cache-dir""" ) download_parser.add_argument( """--trust-remote-code""" , action="""store_true""" , help="""Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you've reviewed the code as it will execute on your local machine""" , ) download_parser.add_argument("""model""" , type=UpperCamelCase , help="""Name of the model to download""" ) download_parser.set_defaults(func=UpperCamelCase ) def __init__( self : Optional[int] , UpperCamelCase : str , UpperCamelCase : str , UpperCamelCase : bool , UpperCamelCase : bool ) -> Any: """simple docstring""" lowerCAmelCase__ : int = model lowerCAmelCase__ : Union[str, Any] = cache lowerCAmelCase__ : Optional[int] = force lowerCAmelCase__ : Dict = trust_remote_code def _lowerCAmelCase ( self : int ) -> Any: """simple docstring""" from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
212
1
'''simple docstring''' from __future__ import annotations lowerCamelCase_ = [ [-1, 0], # left [0, -1], # down [1, 0], # right [0, 1], # up ] def __lowercase ( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> tuple[list[list[int]], list[list[int]]]: '''simple docstring''' _A = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__lowercase ) ) ] # the reference grid _A = 1 _A = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__lowercase ) ) ] # the action grid _A = init[0] _A = init[1] _A = 0 _A = g + heuristic[x][y] # cost from starting cell to destination cell _A = [[f, g, x, y]] _A = False # flag that is set when search is complete _A = False # flag set if we can't find expand while not found and not resign: if len(__lowercase ) == 0: raise ValueError("Algorithm is unable to find solution" ) else: # to choose the least costliest action so as to move closer to the goal cell.sort() cell.reverse() _A = cell.pop() _A = next_cell[2] _A = next_cell[3] _A = next_cell[1] if x == goal[0] and y == goal[1]: _A = True else: for i in range(len(__lowercase ) ): # to try out different valid actions _A = x + DIRECTIONS[i][0] _A = y + DIRECTIONS[i][1] if xa >= 0 and xa < len(__lowercase ) and ya >= 0 and ya < len(grid[0] ): if closed[xa][ya] == 0 and grid[xa][ya] == 0: _A = g + cost _A = ga + heuristic[xa][ya] cell.append([fa, ga, xa, ya] ) _A = 1 _A = i _A = [] _A = goal[0] _A = goal[1] invpath.append([x, y] ) # we get the reverse path from here while x != init[0] or y != init[1]: _A = x - DIRECTIONS[action[x][y]][0] _A = y - DIRECTIONS[action[x][y]][1] _A = xa _A = ya invpath.append([x, y] ) _A = [] for i in range(len(__lowercase ) ): path.append(invpath[len(__lowercase ) - 1 - i] ) return path, action if __name__ == "__main__": lowerCamelCase_ = [ [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0], [0, 0, 0, 0, 1, 0], ] lowerCamelCase_ = [0, 0] # all coordinates are given in format [y,x] lowerCamelCase_ = [len(grid) - 1, len(grid[0]) - 1] lowerCamelCase_ = 1 # the cost map which pushes the path closer to the goal lowerCamelCase_ = [[0 for row in range(len(grid[0]))] for col in range(len(grid))] for i in range(len(grid)): for j in range(len(grid[0])): lowerCamelCase_ = abs(i - goal[0]) + abs(j - goal[1]) if grid[i][j] == 1: # added extra penalty in the heuristic map lowerCamelCase_ = 99 lowerCamelCase_ , lowerCamelCase_ = search(grid, init, goal, cost, heuristic) print('''ACTION MAP''') for i in range(len(action)): print(action[i]) for i in range(len(path)): print(path[i])
79
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'MIT/ast-finetuned-audioset-10-10-0.4593': ( 'https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json' ), } class UpperCAmelCase_ ( snake_case ): UpperCamelCase ="audio-spectrogram-transformer" def __init__( self , UpperCamelCase_=7_68 , UpperCamelCase_=12 , UpperCamelCase_=12 , UpperCamelCase_=30_72 , UpperCamelCase_="gelu" , UpperCamelCase_=0.0 , UpperCamelCase_=0.0 , UpperCamelCase_=0.0_2 , UpperCamelCase_=1E-12 , UpperCamelCase_=16 , UpperCamelCase_=True , UpperCamelCase_=10 , UpperCamelCase_=10 , UpperCamelCase_=10_24 , UpperCamelCase_=1_28 , **UpperCamelCase_ , ) -> Optional[int]: super().__init__(**UpperCamelCase_ ) __lowercase : Optional[Any] = hidden_size __lowercase : List[str] = num_hidden_layers __lowercase : List[str] = num_attention_heads __lowercase : Dict = intermediate_size __lowercase : List[str] = hidden_act __lowercase : Union[str, Any] = hidden_dropout_prob __lowercase : Optional[Any] = attention_probs_dropout_prob __lowercase : Dict = initializer_range __lowercase : Optional[int] = layer_norm_eps __lowercase : Optional[int] = patch_size __lowercase : List[str] = qkv_bias __lowercase : Union[str, Any] = frequency_stride __lowercase : List[Any] = time_stride __lowercase : Tuple = max_length __lowercase : int = num_mel_bins
249
0
'''simple docstring''' def __lowercase ( __lowercase , __lowercase ) -> float: '''simple docstring''' def get_matched_characters(__lowercase , __lowercase ) -> str: _A = [] _A = min(len(_stra ) , len(_stra ) ) // 2 for i, l in enumerate(_stra ): _A = int(max(0 , i - limit ) ) _A = int(min(i + limit + 1 , len(_stra ) ) ) if l in _stra[left:right]: matched.append(__lowercase ) _A = F'''{_stra[0:_stra.index(__lowercase )]} {_stra[_stra.index(__lowercase ) + 1:]}''' return "".join(__lowercase ) # matching characters _A = get_matched_characters(__lowercase , __lowercase ) _A = get_matched_characters(__lowercase , __lowercase ) _A = len(__lowercase ) # transposition _A = ( len([(ca, ca) for ca, ca in zip(__lowercase , __lowercase ) if ca != ca] ) // 2 ) if not match_count: _A = 0.0 else: _A = ( 1 / 3 * ( match_count / len(__lowercase ) + match_count / len(__lowercase ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters _A = 0 for ca, ca in zip(stra[:4] , stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('''hello''', '''world'''))
369
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) lowerCamelCase_ = { '''configuration_trocr''': ['''TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TrOCRConfig'''], '''processing_trocr''': ['''TrOCRProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ '''TROCR_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TrOCRForCausalLM''', '''TrOCRPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys lowerCamelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
174
0
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to properly calculate the metrics on the # validation dataset when in a distributed system, and builds off the # `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## lowerCAmelCase__ = 1_6 lowerCAmelCase__ = 3_2 def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__ = 16 ): """simple docstring""" lowercase__ : Optional[int] = AutoTokenizer.from_pretrained("bert-base-cased" ) lowercase__ : str = load_dataset("glue" , "mrpc" ) def tokenize_function(lowerCamelCase__ ): # max_length=None => use the model max length (it's actually the default) lowercase__ : Dict = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=__a , max_length=__a ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowercase__ : Dict = datasets.map( __a , batched=__a , remove_columns=["idx", "sentence1", "sentence2"] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowercase__ : Tuple = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(lowerCamelCase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. lowercase__ : Dict = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowercase__ : Tuple = 16 elif accelerator.mixed_precision != "no": lowercase__ : Optional[Any] = 8 else: lowercase__ : Any = None return tokenizer.pad( __a , padding="longest" , max_length=__a , pad_to_multiple_of=__a , return_tensors="pt" , ) # Instantiate dataloaders. lowercase__ : List[Any] = DataLoader( tokenized_datasets["train"] , shuffle=__a , collate_fn=__a , batch_size=__a ) lowercase__ : List[Any] = DataLoader( tokenized_datasets["validation"] , shuffle=__a , collate_fn=__a , batch_size=__a ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''', None) == "1": from accelerate.test_utils.training import mocked_dataloaders lowerCAmelCase__ = mocked_dataloaders # noqa: F811 def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" if os.environ.get("TESTING_MOCKED_DATALOADERS" , __a ) == "1": lowercase__ : Optional[int] = 2 # Initialize accelerator lowercase__ : Optional[Any] = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowercase__ : str = config['''lr'''] lowercase__ : Any = int(config["num_epochs"] ) lowercase__ : str = int(config["seed"] ) lowercase__ : Union[str, Any] = int(config["batch_size"] ) lowercase__ : List[str] = evaluate.load("glue" , "mrpc" ) # If the batch size is too big we use gradient accumulation lowercase__ : Any = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: lowercase__ : Optional[Any] = batch_size // MAX_GPU_BATCH_SIZE lowercase__ : Any = MAX_GPU_BATCH_SIZE set_seed(__a ) lowercase__ : Tuple = get_dataloaders(__a , __a ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowercase__ : Optional[int] = AutoModelForSequenceClassification.from_pretrained("bert-base-cased" , return_dict=__a ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowercase__ : Union[str, Any] = model.to(accelerator.device ) # Instantiate optimizer lowercase__ : Tuple = AdamW(params=model.parameters() , lr=__a ) # Instantiate scheduler lowercase__ : Dict = get_linear_schedule_with_warmup( optimizer=__a , num_warmup_steps=100 , num_training_steps=(len(__a ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowercase__ : Union[str, Any] = accelerator.prepare( __a , __a , __a , __a , __a ) # Now we train the model for epoch in range(__a ): model.train() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) lowercase__ : List[str] = model(**__a ) lowercase__ : str = outputs.loss lowercase__ : int = loss / gradient_accumulation_steps accelerator.backward(__a ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() lowercase__ : Any = 0 for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowercase__ : Optional[Any] = model(**__a ) lowercase__ : Tuple = outputs.logits.argmax(dim=-1 ) lowercase__ : List[Any] = accelerator.gather((predictions, batch["labels"]) ) # New Code # # First we check if it's a distributed system if accelerator.use_distributed: # Then see if we're on the last batch of our eval dataloader if step == len(__a ) - 1: # Last batch needs to be truncated on distributed systems as it contains additional samples lowercase__ : Any = predictions[: len(eval_dataloader.dataset ) - samples_seen] lowercase__ : int = references[: len(eval_dataloader.dataset ) - samples_seen] else: # Otherwise we add the number of samples seen samples_seen += references.shape[0] # All of this can be avoided if you use `Accelerator.gather_for_metrics` instead of `Accelerator.gather`: # accelerator.gather_for_metrics((predictions, batch["labels"])) metric.add_batch( predictions=__a , references=__a , ) lowercase__ : Union[str, Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"""epoch {epoch}:""" , __a ) def __lowerCamelCase ( ): """simple docstring""" lowercase__ : int = argparse.ArgumentParser(description="Simple example of training script." ) parser.add_argument( "--mixed_precision" , type=__a , default=__a , choices=["no", "fp16", "bf16", "fp8"] , help="Whether to use mixed precision. Choose" "between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10." "and an Nvidia Ampere GPU." , ) parser.add_argument("--cpu" , action="store_true" , help="If passed, will train on the CPU." ) lowercase__ : int = parser.parse_args() lowercase__ : Optional[Any] = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(__a , __a ) if __name__ == "__main__": main()
130
"""simple docstring""" import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets UpperCAmelCase_ : Optional[Any] = datasets.logging.get_logger(__name__) UpperCAmelCase_ : List[str] = """\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = \"Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric\", author = \"Moosavi, Nafise Sadat and Strube, Michael\", booktitle = \"Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)\", month = aug, year = \"2016\", address = \"Berlin, Germany\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/P16-1060\", doi = \"10.18653/v1/P16-1060\", pages = \"632--642\", } """ UpperCAmelCase_ : Tuple = """\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the \"([pos] [word])\" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a \"-\" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section \"*_conll File Format\"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. """ UpperCAmelCase_ : Union[str, Any] = """ Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting 'keep_singletons=False', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the 'NP_only' option, the scorer will only evaluate the resolution of NPs. min_span: By setting 'min_span', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: 'mentions': mentions 'muc': MUC metric [Vilain et al, 1995] 'bcub': B-cubed [Bagga and Baldwin, 1998] 'ceafe': CEAFe [Luo et al., 2005] 'lea': LEA [Moosavi and Strube, 2016] 'conll_score': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric('coval') >>> words = ['bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -', ... 'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)', ... 'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)', ... 'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -', ... 'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -', ... 'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {'mentions/recall': 1.0,[...] 'conll_score': 100.0} """ def _A (__a , __a , __a=False , __a=False , __a=True , __a=False , __a="dummy_doc" ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE_ : int = {doc: key_lines} SCREAMING_SNAKE_CASE_ : List[str] = {doc: sys_lines} SCREAMING_SNAKE_CASE_ : Dict = {} SCREAMING_SNAKE_CASE_ : Dict = 0 SCREAMING_SNAKE_CASE_ : List[str] = 0 SCREAMING_SNAKE_CASE_ : Tuple = 0 SCREAMING_SNAKE_CASE_ : int = 0 SCREAMING_SNAKE_CASE_ : List[str] = 0 SCREAMING_SNAKE_CASE_ : Any = 0 SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Tuple = reader.get_doc_mentions(__a , key_doc_lines[doc] , __a ) key_singletons_num += singletons_num if NP_only or min_span: SCREAMING_SNAKE_CASE_ : Optional[Any] = reader.set_annotated_parse_trees(__a , key_doc_lines[doc] , __a , __a ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : int = reader.get_doc_mentions(__a , sys_doc_lines[doc] , __a ) sys_singletons_num += singletons_num if NP_only or min_span: SCREAMING_SNAKE_CASE_ : Union[str, Any] = reader.set_annotated_parse_trees(__a , key_doc_lines[doc] , __a , __a ) if remove_nested: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : List[str] = reader.remove_nested_coref_mentions(__a , __a ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : List[str] = reader.remove_nested_coref_mentions(__a , __a ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters SCREAMING_SNAKE_CASE_ : Optional[Any] = reader.get_mention_assignments(__a , __a ) SCREAMING_SNAKE_CASE_ : Optional[Any] = reader.get_mention_assignments(__a , __a ) SCREAMING_SNAKE_CASE_ : Optional[Any] = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( '''Number of resulting singleton clusters in the key ''' f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' '''files, respectively''' ) return doc_coref_infos def _A (__a , __a , __a , __a , __a , __a , __a ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = get_coref_infos(__a , __a , __a , __a , __a , __a ) SCREAMING_SNAKE_CASE_ : str = {} SCREAMING_SNAKE_CASE_ : Union[str, Any] = 0 SCREAMING_SNAKE_CASE_ : str = 0 for name, metric in metrics: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Tuple = evaluator.evaluate_documents(__a , __a , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(10 ) , f'Recall: {recall * 1_00:.2f}' , f' Precision: {precision * 1_00:.2f}' , f' F1: {fa * 1_00:.2f}' , ) if conll_subparts_num == 3: SCREAMING_SNAKE_CASE_ : Tuple = (conll / 3) * 1_00 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({'''conll_score''': conll} ) return output_scores def _A (__a ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: SCREAMING_SNAKE_CASE_ : Any = line.split()[5] if not parse_col == "-": SCREAMING_SNAKE_CASE_ : Any = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''')), '''references''': datasets.Sequence(datasets.Value('''string''')), }) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] , lowercase_ : List[str] , lowercase_ : Optional[int] , lowercase_ : Dict=True , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=False , lowercase_ : Dict=False): '''simple docstring''' SCREAMING_SNAKE_CASE_ : int = [ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: SCREAMING_SNAKE_CASE_ : Union[str, Any] = util.check_gold_parse_annotation(lowercase_) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''') # util.parse_key_file(key_file) # key_file = key_file + ".parsed" SCREAMING_SNAKE_CASE_ : Optional[Any] = evaluate( key_lines=lowercase_ , sys_lines=lowercase_ , metrics=lowercase_ , NP_only=lowercase_ , remove_nested=lowercase_ , keep_singletons=lowercase_ , min_span=lowercase_ , ) return score
91
0
import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : str = (PNDMScheduler,) __SCREAMING_SNAKE_CASE : Union[str, Any] = (('''num_inference_steps''', 50),) def a ( self , **snake_case ): snake_case_ = { 'num_train_timesteps': 1000, 'beta_start': 0.00_01, 'beta_end': 0.02, 'beta_schedule': 'linear', } config.update(**snake_case ) return config def a ( self , snake_case=0 , **snake_case ): snake_case_ = dict(self.forward_default_kwargs ) snake_case_ = kwargs.pop('num_inference_steps' , snake_case ) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config(**snake_case ) snake_case_ = scheduler_class(**snake_case ) scheduler.set_timesteps(snake_case ) # copy over dummy past residuals snake_case_ = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(snake_case ) snake_case_ = scheduler_class.from_pretrained(snake_case ) new_scheduler.set_timesteps(snake_case ) # copy over dummy past residuals snake_case_ = dummy_past_residuals[:] snake_case_ = scheduler.step_prk(snake_case , snake_case , snake_case , **snake_case ).prev_sample snake_case_ = new_scheduler.step_prk(snake_case , snake_case , snake_case , **snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" snake_case_ = scheduler.step_plms(snake_case , snake_case , snake_case , **snake_case ).prev_sample snake_case_ = new_scheduler.step_plms(snake_case , snake_case , snake_case , **snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def a ( self ): pass def a ( self , snake_case=0 , **snake_case ): snake_case_ = dict(self.forward_default_kwargs ) snake_case_ = kwargs.pop('num_inference_steps' , snake_case ) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config() snake_case_ = scheduler_class(**snake_case ) scheduler.set_timesteps(snake_case ) # copy over dummy past residuals (must be after setting timesteps) snake_case_ = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(snake_case ) snake_case_ = scheduler_class.from_pretrained(snake_case ) # copy over dummy past residuals new_scheduler.set_timesteps(snake_case ) # copy over dummy past residual (must be after setting timesteps) snake_case_ = dummy_past_residuals[:] snake_case_ = scheduler.step_prk(snake_case , snake_case , snake_case , **snake_case ).prev_sample snake_case_ = new_scheduler.step_prk(snake_case , snake_case , snake_case , **snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" snake_case_ = scheduler.step_plms(snake_case , snake_case , snake_case , **snake_case ).prev_sample snake_case_ = new_scheduler.step_plms(snake_case , snake_case , snake_case , **snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def a ( self , **snake_case ): snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(**snake_case ) snake_case_ = scheduler_class(**snake_case ) snake_case_ = 10 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter scheduler.set_timesteps(snake_case ) for i, t in enumerate(scheduler.prk_timesteps ): snake_case_ = model(snake_case , snake_case ) snake_case_ = scheduler.step_prk(snake_case , snake_case , snake_case ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): snake_case_ = model(snake_case , snake_case ) snake_case_ = scheduler.step_plms(snake_case , snake_case , snake_case ).prev_sample return sample def a ( self ): snake_case_ = dict(self.forward_default_kwargs ) snake_case_ = kwargs.pop('num_inference_steps' , snake_case ) for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config() snake_case_ = scheduler_class(**snake_case ) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample if num_inference_steps is not None and hasattr(snake_case , 'set_timesteps' ): scheduler.set_timesteps(snake_case ) elif num_inference_steps is not None and not hasattr(snake_case , 'set_timesteps' ): snake_case_ = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] snake_case_ = dummy_past_residuals[:] snake_case_ = scheduler.step_prk(snake_case , 0 , snake_case , **snake_case ).prev_sample snake_case_ = scheduler.step_prk(snake_case , 1 , snake_case , **snake_case ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) snake_case_ = scheduler.step_plms(snake_case , 0 , snake_case , **snake_case ).prev_sample snake_case_ = scheduler.step_plms(snake_case , 1 , snake_case , **snake_case ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def a ( self ): for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=snake_case ) def a ( self ): for steps_offset in [0, 1]: self.check_over_configs(steps_offset=snake_case ) snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(steps_offset=1 ) snake_case_ = scheduler_class(**snake_case ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ) , ) def a ( self ): for beta_start, beta_end in zip([0.00_01, 0.0_01] , [0.0_02, 0.02] ): self.check_over_configs(beta_start=snake_case , beta_end=snake_case ) def a ( self ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=snake_case ) def a ( self ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=snake_case ) def a ( self ): for t in [1, 5, 10]: self.check_over_forward(time_step=snake_case ) def a ( self ): for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=snake_case ) def a ( self ): # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 snake_case_ = 27 for scheduler_class in self.scheduler_classes: snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = self.get_scheduler_config() snake_case_ = scheduler_class(**snake_case ) scheduler.set_timesteps(snake_case ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): snake_case_ = scheduler.step_prk(snake_case , snake_case , snake_case ).prev_sample def a ( self ): with self.assertRaises(snake_case ): snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config() snake_case_ = scheduler_class(**snake_case ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def a ( self ): snake_case_ = self.full_loop() snake_case_ = torch.sum(torch.abs(snake_case ) ) snake_case_ = torch.mean(torch.abs(snake_case ) ) assert abs(result_sum.item() - 1_98.13_18 ) < 1e-2 assert abs(result_mean.item() - 0.25_80 ) < 1e-3 def a ( self ): snake_case_ = self.full_loop(prediction_type='v_prediction' ) snake_case_ = torch.sum(torch.abs(snake_case ) ) snake_case_ = torch.mean(torch.abs(snake_case ) ) assert abs(result_sum.item() - 67.39_86 ) < 1e-2 assert abs(result_mean.item() - 0.08_78 ) < 1e-3 def a ( self ): # We specify different beta, so that the first alpha is 0.99 snake_case_ = self.full_loop(set_alpha_to_one=snake_case , beta_start=0.01 ) snake_case_ = torch.sum(torch.abs(snake_case ) ) snake_case_ = torch.mean(torch.abs(snake_case ) ) assert abs(result_sum.item() - 2_30.03_99 ) < 1e-2 assert abs(result_mean.item() - 0.29_95 ) < 1e-3 def a ( self ): # We specify different beta, so that the first alpha is 0.99 snake_case_ = self.full_loop(set_alpha_to_one=snake_case , beta_start=0.01 ) snake_case_ = torch.sum(torch.abs(snake_case ) ) snake_case_ = torch.mean(torch.abs(snake_case ) ) assert abs(result_sum.item() - 1_86.94_82 ) < 1e-2 assert abs(result_mean.item() - 0.24_34 ) < 1e-3
200
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
200
1
import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency UpperCAmelCase_ = { 'E': 12.70, 'T': 9.06, 'A': 8.17, 'O': 7.51, 'I': 6.97, 'N': 6.75, 'S': 6.33, 'H': 6.09, 'R': 5.99, 'D': 4.25, 'L': 4.03, 'C': 2.78, 'U': 2.76, 'M': 2.41, 'W': 2.36, 'F': 2.23, 'G': 2.02, 'Y': 1.97, 'P': 1.93, 'B': 1.29, 'V': 0.98, 'K': 0.77, 'J': 0.15, 'X': 0.15, 'Q': 0.10, 'Z': 0.07, } UpperCAmelCase_ = 'ETAOINSHRDLCUMWFGYPBVKJXQZ' UpperCAmelCase_ = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def lowerCamelCase__ ( A__ : str ): '''simple docstring''' __lowerCamelCase = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def lowerCamelCase__ ( A__ : tuple ): '''simple docstring''' return x[0] def lowerCamelCase__ ( A__ : str ): '''simple docstring''' __lowerCamelCase = get_letter_count(A__ ) __lowerCamelCase = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(A__ ) __lowerCamelCase = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find , reverse=A__ ) __lowerCamelCase = """""".join(freq_to_letter[freq] ) __lowerCamelCase = list(freq_to_letter_str.items() ) freq_pairs.sort(key=A__ , reverse=A__ ) __lowerCamelCase = [freq_pair[1] for freq_pair in freq_pairs] return "".join(A__ ) def lowerCamelCase__ ( A__ : str ): '''simple docstring''' __lowerCamelCase = get_frequency_order(A__ ) __lowerCamelCase = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
12
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging UpperCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Tuple = ['pixel_values'] def __init__( self: Any , UpperCamelCase_: bool = True , UpperCamelCase_: Union[int, float] = 1 / 2_55 , UpperCamelCase_: bool = True , UpperCamelCase_: int = 8 , **UpperCamelCase_: Tuple , ): super().__init__(**UpperCamelCase_ ) __lowerCamelCase = do_rescale __lowerCamelCase = rescale_factor __lowerCamelCase = do_pad __lowerCamelCase = pad_size def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: np.ndarray , UpperCamelCase_: float , UpperCamelCase_: Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase_: Tuple ): return rescale(UpperCamelCase_ , scale=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: np.ndarray , UpperCamelCase_: int , UpperCamelCase_: Optional[Union[str, ChannelDimension]] = None ): __lowerCamelCase, __lowerCamelCase = get_image_size(UpperCamelCase_ ) __lowerCamelCase = (old_height // size + 1) * size - old_height __lowerCamelCase = (old_width // size + 1) * size - old_width return pad(UpperCamelCase_ , ((0, pad_height), (0, pad_width)) , mode="""symmetric""" , data_format=UpperCamelCase_ ) def lowerCAmelCase__ ( self: str , UpperCamelCase_: ImageInput , UpperCamelCase_: Optional[bool] = None , UpperCamelCase_: Optional[float] = None , UpperCamelCase_: Optional[bool] = None , UpperCamelCase_: Optional[int] = None , UpperCamelCase_: Optional[Union[str, TensorType]] = None , UpperCamelCase_: Union[str, ChannelDimension] = ChannelDimension.FIRST , **UpperCamelCase_: Any , ): __lowerCamelCase = do_rescale if do_rescale is not None else self.do_rescale __lowerCamelCase = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCamelCase = do_pad if do_pad is not None else self.do_pad __lowerCamelCase = pad_size if pad_size is not None else self.pad_size __lowerCamelCase = make_list_of_images(UpperCamelCase_ ) if not valid_images(UpperCamelCase_ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) # All transformations expect numpy arrays. __lowerCamelCase = [to_numpy_array(UpperCamelCase_ ) for image in images] if do_rescale: __lowerCamelCase = [self.rescale(image=UpperCamelCase_ , scale=UpperCamelCase_ ) for image in images] if do_pad: __lowerCamelCase = [self.pad(UpperCamelCase_ , size=UpperCamelCase_ ) for image in images] __lowerCamelCase = [to_channel_dimension_format(UpperCamelCase_ , UpperCamelCase_ ) for image in images] __lowerCamelCase = {"""pixel_values""": images} return BatchFeature(data=UpperCamelCase_ , tensor_type=UpperCamelCase_ )
12
1
"""simple docstring""" from argparse import ArgumentParser from .env import EnvironmentCommand def _snake_case ( ): lowerCAmelCase : int = ArgumentParser('''Diffusers CLI tool''' , usage='''diffusers-cli <command> [<args>]''' ) lowerCAmelCase : Any = parser.add_subparsers(help='''diffusers-cli command helpers''' ) # Register commands EnvironmentCommand.register_subcommand(_snake_case ) # Let's go lowerCAmelCase : str = parser.parse_args() if not hasattr(_snake_case , '''func''' ): parser.print_help() exit(1 ) # Run lowerCAmelCase : Optional[Any] = args.func(_snake_case ) service.run() if __name__ == "__main__": main()
314
"""simple docstring""" import argparse import json import os import torch from torch import nn from transformers import NllbMoeConfig, NllbMoeModel from transformers.modeling_utils import dtype_byte_size from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME def _snake_case ( _snake_case : Optional[int] ): lowerCAmelCase : List[str] = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(_snake_case , _snake_case ) def _snake_case ( _snake_case : List[str] ): lowerCAmelCase, lowerCAmelCase : str = emb.weight.shape lowerCAmelCase : Optional[Any] = nn.Linear(_snake_case , _snake_case , bias=_snake_case ) lowerCAmelCase : Tuple = emb.weight.data return lin_layer def _snake_case ( _snake_case : Union[str, Any] , _snake_case : Dict=None ): lowerCAmelCase : Union[str, Any] = {} for old_key in state_dict.keys(): lowerCAmelCase : Union[str, Any] = old_key if "moe_layer.experts." in key: if expert_idx is not None: lowerCAmelCase : str = key.replace('''moe_layer.experts.0''' , f'''ffn.experts.expert_{expert_idx}''' ) else: lowerCAmelCase : Optional[Any] = key.replace('''moe_layer.experts.''' , '''ffn.experts.expert_''' ) if "gate" in key: lowerCAmelCase : Any = key.replace('''.moe_layer.gate.wg''' , '''.ffn.router.classifier''' ) if "fc2" and "experts" not in key: lowerCAmelCase : Tuple = key.replace('''.fc2.''' , '''.ffn.fc2.''' ) if "fc1" and "experts" not in key: lowerCAmelCase : int = key.replace('''.fc1.''' , '''.ffn.fc1.''' ) if ".encoder_attn." in key: lowerCAmelCase : List[str] = key.replace('''.encoder_attn.''' , '''.cross_attention.''' ) if "encoder_attn_layer_norm" in key: lowerCAmelCase : int = key.replace('''encoder_attn_layer_norm''' , '''cross_attention_layer_norm''' ) if "final_layer_norm" in key: lowerCAmelCase : List[str] = key.replace('''final_layer_norm''' , '''ff_layer_norm''' ) lowerCAmelCase : Tuple = state_dict[old_key] return new_dict def _snake_case ( _snake_case : Optional[int] , _snake_case : Optional[int] , _snake_case : Optional[int] , _snake_case : Union[str, Any] , _snake_case : str = WEIGHTS_NAME ): lowerCAmelCase : Optional[Any] = [] lowerCAmelCase : Tuple = 0 os.makedirs(_snake_case , exist_ok=_snake_case ) for expert in range(_snake_case ): lowerCAmelCase : Any = switch_checkpoint_path + f'''-rank-{expert}.pt''' if os.path.isfile(_snake_case ): lowerCAmelCase : List[str] = torch.load(_snake_case )['''model'''] remove_ignore_keys_(_snake_case ) lowerCAmelCase : Any = rename_fairseq_keys(_snake_case , _snake_case ) lowerCAmelCase : Any = os.path.join( _snake_case , weights_name.replace('''.bin''' , f'''-{len(_snake_case )+1:05d}-of-???.bin''' ) ) torch.save(_snake_case , _snake_case ) sharded_state_dicts.append(expert_state.keys() ) total_size += sum([value.numel() for key, value in expert_state.items()] ) * dtype_byte_size( expert_state[list(_snake_case )[0]].dtype ) # Add the last block lowerCAmelCase : List[str] = os.path.join(_snake_case , weights_name.replace('''.bin''' , f'''-{len(_snake_case )+1:05d}-of-???.bin''' ) ) lowerCAmelCase : str = torch.load(switch_checkpoint_path + '''-shared.pt''' )['''model'''] remove_ignore_keys_(_snake_case ) lowerCAmelCase : Union[str, Any] = rename_fairseq_keys(_snake_case , _snake_case ) lowerCAmelCase : Dict = shared_weights['''decoder.embed_tokens.weight'''] sharded_state_dicts.append(shared_weights.keys() ) # If we only have the shared weights (dummy model/experts saved on the same file) if len(_snake_case ) == 1: lowerCAmelCase : List[str] = os.path.join(_snake_case , _snake_case ) torch.save(_snake_case , _snake_case ) return {weights_name: sharded_state_dicts[0]}, None else: torch.save(_snake_case , _snake_case ) # Otherwise, let's build the index lowerCAmelCase : Dict = {} for idx, shard in enumerate(_snake_case ): lowerCAmelCase : Union[str, Any] = weights_name.replace('''.bin''' , f'''-{idx+1:05d}-of-{len(_snake_case ):05d}.bin''' ) lowerCAmelCase : Any = os.path.join(_snake_case , weights_name.replace('''.bin''' , f'''-{idx+1:05d}-of-???.bin''' ) ) os.rename(_snake_case , os.path.join(_snake_case , _snake_case ) ) for key in shard: lowerCAmelCase : List[Any] = shard_file # Add the metadata lowerCAmelCase : Dict = {'''total_size''': total_size} lowerCAmelCase : int = {'''metadata''': metadata, '''weight_map''': weight_map} with open(os.path.join(_snake_case , _snake_case ) , '''w''' , encoding='''utf-8''' ) as f: lowerCAmelCase : Union[str, Any] = json.dumps(_snake_case , indent=2 , sort_keys=_snake_case ) + '''\n''' f.write(_snake_case ) return metadata, index if __name__ == "__main__": snake_case__ : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--nllb_moe_checkpoint_path''', default='''/home/arthur_huggingface_co/fairseq/weights/checkpoints/model_moe_54b/checkpoint_2_300000''', type=str, required=False, help='''Path to a directory containing a folder per layer. Follows the original Google format.''', ) parser.add_argument('''--dtype''', default='''float32''', type=str, required=False, help='''dtype of the saved model''') parser.add_argument( '''--pytorch_dump_folder_path''', default='''/home/arthur_huggingface_co/fairseq/weights/checkpoints/hf-converted-moe-54b''', type=str, required=False, help='''Path to the output pytorch model.''', ) snake_case__ : List[str] = parser.parse_args() snake_case__ , snake_case__ : Tuple = shard_on_the_fly( args.nllb_moe_checkpoint_path, args.pytorch_dump_folder_path, 128, args.dtype, ) snake_case__ : str = NllbMoeConfig.from_pretrained( '''facebook/nllb-200-3.3B''', encoder_sparse_step=4, decoder_sparse_step=4, num_experts=128 ) config.save_pretrained(args.pytorch_dump_folder_path) snake_case__ : Any = NllbMoeModel.from_pretrained(args.pytorch_dump_folder_path) print('''Done''') model.save_pretrained(args.pytorch_dump_folder_path)
314
1
import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def UpperCamelCase ( _A, _A=0.999, _A="cosine", ): """simple docstring""" if alpha_transform_type == "cosine": def alpha_bar_fn(_A ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(_A ): return math.exp(t * -12.0 ) else: raise ValueError(f'Unsupported alpha_tranform_type: {alpha_transform_type}' ) __magic_name__ : Union[str, Any] = [] for i in range(_A ): __magic_name__ : Optional[int] = i / num_diffusion_timesteps __magic_name__ : List[str] = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(_A ) / alpha_bar_fn(_A ), _A ) ) return torch.tensor(_A, dtype=torch.floataa ) class snake_case__ ( _lowerCAmelCase , _lowerCAmelCase ): lowercase__ : List[str] = [e.name for e in KarrasDiffusionSchedulers] lowercase__ : Tuple = 2 @register_to_config def __init__( self , lowerCAmelCase__ = 10_00 , lowerCAmelCase__ = 0.0_0_0_8_5 , lowerCAmelCase__ = 0.0_1_2 , lowerCAmelCase__ = "linear" , lowerCAmelCase__ = None , lowerCAmelCase__ = "epsilon" , lowerCAmelCase__ = "linspace" , lowerCAmelCase__ = 0 , ) -> List[str]: if trained_betas is not None: __magic_name__ : Tuple = torch.tensor(lowerCAmelCase__ , dtype=torch.floataa ) elif beta_schedule == "linear": __magic_name__ : Optional[Any] = torch.linspace(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. __magic_name__ : List[str] = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , lowerCAmelCase__ , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule __magic_name__ : Optional[int] = betas_for_alpha_bar(lowerCAmelCase__ ) else: raise NotImplementedError(F'{beta_schedule} does is not implemented for {self.__class__}' ) __magic_name__ : List[str] = 1.0 - self.betas __magic_name__ : Tuple = torch.cumprod(self.alphas , dim=0 ) # set all values self.set_timesteps(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__=None ) -> List[Any]: if schedule_timesteps is None: __magic_name__ : Union[str, Any] = self.timesteps __magic_name__ : Dict = (schedule_timesteps == timestep).nonzero() # The sigma index that is taken for the **very** first `step` # is always the second index (or the last index if there is only 1) # This way we can ensure we don't accidentally skip a sigma in # case we start in the middle of the denoising schedule (e.g. for image-to-image) if len(self._index_counter ) == 0: __magic_name__ : Tuple = 1 if len(lowerCAmelCase__ ) > 1 else 0 else: __magic_name__ : Union[str, Any] = timestep.cpu().item() if torch.is_tensor(lowerCAmelCase__ ) else timestep __magic_name__ : Dict = self._index_counter[timestep_int] return indices[pos].item() @property def __magic_name__ ( self ) -> Optional[int]: # standard deviation of the initial noise distribution if self.config.timestep_spacing in ["linspace", "trailing"]: return self.sigmas.max() return (self.sigmas.max() ** 2 + 1) ** 0.5 def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , ) -> torch.FloatTensor: __magic_name__ : Union[str, Any] = self.index_for_timestep(lowerCAmelCase__ ) if self.state_in_first_order: __magic_name__ : Dict = self.sigmas[step_index] else: __magic_name__ : List[Any] = self.sigmas_interpol[step_index] __magic_name__ : str = sample / ((sigma**2 + 1) ** 0.5) return sample def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None , ) -> int: __magic_name__ : str = num_inference_steps __magic_name__ : int = num_train_timesteps or self.config.num_train_timesteps # "linspace", "leading", "trailing" corresponds to annotation of Table 2. of https://arxiv.org/abs/2305.08891 if self.config.timestep_spacing == "linspace": __magic_name__ : str = np.linspace(0 , num_train_timesteps - 1 , lowerCAmelCase__ , dtype=lowerCAmelCase__ )[::-1].copy() elif self.config.timestep_spacing == "leading": __magic_name__ : Dict = num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 __magic_name__ : Tuple = (np.arange(0 , lowerCAmelCase__ ) * step_ratio).round()[::-1].copy().astype(lowerCAmelCase__ ) timesteps += self.config.steps_offset elif self.config.timestep_spacing == "trailing": __magic_name__ : Union[str, Any] = num_train_timesteps / self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 __magic_name__ : Tuple = (np.arange(lowerCAmelCase__ , 0 , -step_ratio )).round().copy().astype(lowerCAmelCase__ ) timesteps -= 1 else: raise ValueError( F'{self.config.timestep_spacing} is not supported. Please make sure to choose one of \'linspace\', \'leading\' or \'trailing\'.' ) __magic_name__ : Dict = np.array(((1 - self.alphas_cumprod) / self.alphas_cumprod) ** 0.5 ) __magic_name__ : Union[str, Any] = torch.from_numpy(np.log(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) __magic_name__ : Tuple = np.interp(lowerCAmelCase__ , np.arange(0 , len(lowerCAmelCase__ ) ) , lowerCAmelCase__ ) __magic_name__ : Any = np.concatenate([sigmas, [0.0]] ).astype(np.floataa ) __magic_name__ : str = torch.from_numpy(lowerCAmelCase__ ).to(device=lowerCAmelCase__ ) # interpolate sigmas __magic_name__ : Any = sigmas.log().lerp(sigmas.roll(1 ).log() , 0.5 ).exp() __magic_name__ : Union[str, Any] = torch.cat([sigmas[:1], sigmas[1:].repeat_interleave(2 ), sigmas[-1:]] ) __magic_name__ : Union[str, Any] = torch.cat( [sigmas_interpol[:1], sigmas_interpol[1:].repeat_interleave(2 ), sigmas_interpol[-1:]] ) if str(lowerCAmelCase__ ).startswith("""mps""" ): # mps does not support float64 __magic_name__ : Any = torch.from_numpy(lowerCAmelCase__ ).to(lowerCAmelCase__ , dtype=torch.floataa ) else: __magic_name__ : Union[str, Any] = torch.from_numpy(lowerCAmelCase__ ).to(lowerCAmelCase__ ) # interpolate timesteps __magic_name__ : List[Any] = self.sigma_to_t(lowerCAmelCase__ ).to(lowerCAmelCase__ , dtype=timesteps.dtype ) __magic_name__ : Any = torch.stack((timesteps_interpol[1:-1, None], timesteps[1:, None]) , dim=-1 ).flatten() __magic_name__ : Optional[int] = torch.cat([timesteps[:1], interleaved_timesteps] ) __magic_name__ : Any = None # for exp beta schedules, such as the one for `pipeline_shap_e.py` # we need an index counter __magic_name__ : List[str] = defaultdict(lowerCAmelCase__ ) def __magic_name__ ( self , lowerCAmelCase__ ) -> Dict: # get log sigma __magic_name__ : str = sigma.log() # get distribution __magic_name__ : str = log_sigma - self.log_sigmas[:, None] # get sigmas range __magic_name__ : Union[str, Any] = dists.ge(0 ).cumsum(dim=0 ).argmax(dim=0 ).clamp(max=self.log_sigmas.shape[0] - 2 ) __magic_name__ : List[str] = low_idx + 1 __magic_name__ : Optional[int] = self.log_sigmas[low_idx] __magic_name__ : List[Any] = self.log_sigmas[high_idx] # interpolate sigmas __magic_name__ : Union[str, Any] = (low - log_sigma) / (low - high) __magic_name__ : str = w.clamp(0 , 1 ) # transform interpolation to time range __magic_name__ : Any = (1 - w) * low_idx + w * high_idx __magic_name__ : List[Any] = t.view(sigma.shape ) return t @property def __magic_name__ ( self ) -> List[str]: return self.sample is None def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = True , ) -> Union[SchedulerOutput, Tuple]: __magic_name__ : Optional[int] = self.index_for_timestep(lowerCAmelCase__ ) # advance index counter by 1 __magic_name__ : Optional[int] = timestep.cpu().item() if torch.is_tensor(lowerCAmelCase__ ) else timestep self._index_counter[timestep_int] += 1 if self.state_in_first_order: __magic_name__ : List[Any] = self.sigmas[step_index] __magic_name__ : List[Any] = self.sigmas_interpol[step_index + 1] __magic_name__ : Tuple = self.sigmas[step_index + 1] else: # 2nd order / KDPM2's method __magic_name__ : int = self.sigmas[step_index - 1] __magic_name__ : Tuple = self.sigmas_interpol[step_index] __magic_name__ : Optional[Any] = self.sigmas[step_index] # currently only gamma=0 is supported. This usually works best anyways. # We can support gamma in the future but then need to scale the timestep before # passing it to the model which requires a change in API __magic_name__ : Optional[Any] = 0 __magic_name__ : str = sigma * (gamma + 1) # Note: sigma_hat == sigma for now # 1. compute predicted original sample (x_0) from sigma-scaled predicted noise if self.config.prediction_type == "epsilon": __magic_name__ : Any = sigma_hat if self.state_in_first_order else sigma_interpol __magic_name__ : Any = sample - sigma_input * model_output elif self.config.prediction_type == "v_prediction": __magic_name__ : Union[str, Any] = sigma_hat if self.state_in_first_order else sigma_interpol __magic_name__ : int = model_output * (-sigma_input / (sigma_input**2 + 1) ** 0.5) + ( sample / (sigma_input**2 + 1) ) elif self.config.prediction_type == "sample": raise NotImplementedError("""prediction_type not implemented yet: sample""" ) else: raise ValueError( F'prediction_type given as {self.config.prediction_type} must be one of `epsilon`, or `v_prediction`' ) if self.state_in_first_order: # 2. Convert to an ODE derivative for 1st order __magic_name__ : int = (sample - pred_original_sample) / sigma_hat # 3. delta timestep __magic_name__ : str = sigma_interpol - sigma_hat # store for 2nd order step __magic_name__ : Tuple = sample else: # DPM-Solver-2 # 2. Convert to an ODE derivative for 2nd order __magic_name__ : Optional[int] = (sample - pred_original_sample) / sigma_interpol # 3. delta timestep __magic_name__ : List[str] = sigma_next - sigma_hat __magic_name__ : Tuple = self.sample __magic_name__ : Tuple = None __magic_name__ : Tuple = sample + derivative * dt if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=lowerCAmelCase__ ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ) -> torch.FloatTensor: # Make sure sigmas and timesteps have the same device and dtype as original_samples __magic_name__ : List[Any] = self.sigmas.to(device=original_samples.device , dtype=original_samples.dtype ) if original_samples.device.type == "mps" and torch.is_floating_point(lowerCAmelCase__ ): # mps does not support float64 __magic_name__ : List[str] = self.timesteps.to(original_samples.device , dtype=torch.floataa ) __magic_name__ : Any = timesteps.to(original_samples.device , dtype=torch.floataa ) else: __magic_name__ : int = self.timesteps.to(original_samples.device ) __magic_name__ : Optional[int] = timesteps.to(original_samples.device ) __magic_name__ : Dict = [self.index_for_timestep(lowerCAmelCase__ , lowerCAmelCase__ ) for t in timesteps] __magic_name__ : List[Any] = sigmas[step_indices].flatten() while len(sigma.shape ) < len(original_samples.shape ): __magic_name__ : int = sigma.unsqueeze(-1 ) __magic_name__ : Any = original_samples + noise * sigma return noisy_samples def __len__( self ) -> Any: return self.config.num_train_timesteps
342
import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class snake_case__ ( unittest.TestCase ): def __magic_name__ ( self ) -> str: __magic_name__ : Tuple = """ylacombe/bark-small""" __magic_name__ : List[str] = tempfile.mkdtemp() __magic_name__ : Optional[Any] = """en_speaker_1""" __magic_name__ : Union[str, Any] = """This is a test string""" __magic_name__ : Optional[int] = """speaker_embeddings_path.json""" __magic_name__ : Any = """speaker_embeddings""" def __magic_name__ ( self , **lowerCAmelCase__ ) -> List[Any]: return AutoTokenizer.from_pretrained(self.checkpoint , **lowerCAmelCase__ ) def __magic_name__ ( self ) -> Optional[Any]: shutil.rmtree(self.tmpdirname ) def __magic_name__ ( self ) -> Tuple: __magic_name__ : Optional[Any] = self.get_tokenizer() __magic_name__ : int = BarkProcessor(tokenizer=lowerCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) __magic_name__ : Union[str, Any] = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def __magic_name__ ( self ) -> Optional[int]: __magic_name__ : Optional[int] = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) __magic_name__ : Optional[Any] = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) __magic_name__ : str = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token="""(BOS)""" , eos_token="""(EOS)""" , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def __magic_name__ ( self ) -> Any: __magic_name__ : List[str] = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) __magic_name__ : Union[str, Any] = 35 __magic_name__ : List[Any] = 2 __magic_name__ : Dict = 8 __magic_name__ : Tuple = { """semantic_prompt""": np.ones(lowerCAmelCase__ ), """coarse_prompt""": np.ones((nb_codebooks_coarse, seq_len) ), """fine_prompt""": np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset __magic_name__ : Optional[int] = processor(text=self.input_string , voice_preset=lowerCAmelCase__ ) __magic_name__ : Union[str, Any] = inputs["""history_prompt"""] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowerCAmelCase__ , np.array([] ) ).tolist() ) # test loading voice preset from npz file __magic_name__ : Dict = os.path.join(self.tmpdirname , """file.npz""" ) np.savez(lowerCAmelCase__ , **lowerCAmelCase__ ) __magic_name__ : Optional[Any] = processor(text=self.input_string , voice_preset=lowerCAmelCase__ ) __magic_name__ : List[Any] = inputs["""history_prompt"""] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowerCAmelCase__ , np.array([] ) ).tolist() ) # test loading voice preset from the hub __magic_name__ : Tuple = processor(text=self.input_string , voice_preset=self.voice_preset ) def __magic_name__ ( self ) -> Optional[Any]: __magic_name__ : str = self.get_tokenizer() __magic_name__ : Dict = BarkProcessor(tokenizer=lowerCAmelCase__ ) __magic_name__ : Optional[Any] = processor(text=self.input_string ) __magic_name__ : List[Any] = tokenizer( self.input_string , padding="""max_length""" , max_length=2_56 , add_special_tokens=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__ , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
342
1
import argparse from collections import defaultdict import yaml __lowerCAmelCase = 'docs/source/en/_toctree.yml' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = defaultdict(__a ) for doc in model_doc: counts[doc["local"]] += 1 _snake_case = [key for key, value in counts.items() if value > 1] _snake_case = [] for duplicate_key in duplicates: _snake_case = list({doc["""title"""] for doc in model_doc if doc["""local"""] == duplicate_key} ) if len(__a ) > 1: raise ValueError( f"""{duplicate_key} is present several times in the documentation table of content at """ """`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the """ """others.""" ) # Only add this once new_doc.append({"""local""": duplicate_key, """title""": titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in model_doc if counts[doc["""local"""]] == 1] ) # Sort return sorted(__a , key=lambda _SCREAMING_SNAKE_CASE : s["title"].lower() ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE=False ): with open(__a , encoding="""utf-8""" ) as f: _snake_case = yaml.safe_load(f.read() ) # Get to the API doc _snake_case = 0 while content[api_idx]["title"] != "API": api_idx += 1 _snake_case = content[api_idx]['sections'] # Then to the model doc _snake_case = 0 while api_doc[model_idx]["title"] != "Models": model_idx += 1 _snake_case = api_doc[model_idx]['sections'] _snake_case = [(idx, section) for idx, section in enumerate(__a ) if 'sections' in section] _snake_case = False for idx, modality_doc in modalities_docs: _snake_case = modality_doc['sections'] _snake_case = clean_model_doc_toc(__a ) if old_modality_doc != new_modality_doc: _snake_case = True if overwrite: _snake_case = new_modality_doc if diff: if overwrite: _snake_case = model_doc _snake_case = api_doc with open(__a , """w""" , encoding="""utf-8""" ) as f: f.write(yaml.dump(__a , allow_unicode=__a ) ) else: raise ValueError( """The model doc part of the table of content is not properly sorted, run `make style` to fix this.""" ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') __lowerCAmelCase = parser.parse_args() check_model_doc(args.fix_and_overwrite)
355
'''simple docstring''' import inspect import tempfile import unittest from huggingface_hub import hf_hub_download from transformers import is_torch_available from transformers.testing_utils import is_flaky, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin __lowerCAmelCase = 1E-4 if is_torch_available(): import torch from transformers import AutoformerConfig, AutoformerForPrediction, AutoformerModel from transformers.models.autoformer.modeling_autoformer import AutoformerDecoder, AutoformerEncoder @require_torch class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=16 , UpperCAmelCase=13 , UpperCAmelCase=7 , UpperCAmelCase=14 , UpperCAmelCase=10 , UpperCAmelCase=19 , UpperCAmelCase=5 , UpperCAmelCase=4 , UpperCAmelCase=True , UpperCAmelCase=16 , UpperCAmelCase=2 , UpperCAmelCase=4 , UpperCAmelCase=4 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=[1, 2, 3, 4, 5] , UpperCAmelCase=25 , UpperCAmelCase=5 , ) -> int: _snake_case = d_model _snake_case = parent _snake_case = batch_size _snake_case = prediction_length _snake_case = context_length _snake_case = cardinality _snake_case = num_time_features _snake_case = lags_sequence _snake_case = embedding_dimension _snake_case = is_training _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = context_length _snake_case = prediction_length + label_length _snake_case = label_length _snake_case = moving_average _snake_case = autocorrelation_factor def lowercase (self ) -> str: return AutoformerConfig( d_model=self.d_model , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , prediction_length=self.prediction_length , context_length=self.context_length , label_length=self.label_length , lags_sequence=self.lags_sequence , num_time_features=self.num_time_features , num_static_categorical_features=1 , cardinality=[self.cardinality] , embedding_dimension=[self.embedding_dimension] , moving_average=self.moving_average , ) def lowercase (self , UpperCAmelCase ) -> Tuple: _snake_case = config.context_length + max(config.lags_sequence ) _snake_case = ids_tensor([self.batch_size, 1] , config.cardinality[0] ) _snake_case = floats_tensor([self.batch_size, _past_length, config.num_time_features] ) _snake_case = floats_tensor([self.batch_size, _past_length] ) _snake_case = floats_tensor([self.batch_size, _past_length] ) > 0.5 # decoder inputs _snake_case = floats_tensor([self.batch_size, config.prediction_length, config.num_time_features] ) _snake_case = floats_tensor([self.batch_size, config.prediction_length] ) _snake_case = { """past_values""": past_values, """static_categorical_features""": static_categorical_features, """past_time_features""": past_time_features, """past_observed_mask""": past_observed_mask, """future_time_features""": future_time_features, """future_values""": future_values, } return inputs_dict def lowercase (self ) -> int: _snake_case = self.get_config() _snake_case = self.prepare_autoformer_inputs_dict(UpperCAmelCase ) return config, inputs_dict def lowercase (self ) -> List[Any]: _snake_case, _snake_case = self.prepare_config_and_inputs() return config, inputs_dict def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> Tuple: _snake_case = AutoformerModel(config=UpperCAmelCase ).to(UpperCAmelCase ).eval() _snake_case = model(**UpperCAmelCase ) _snake_case = outputs.encoder_last_hidden_state _snake_case = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: _snake_case = model.get_encoder() encoder.save_pretrained(UpperCAmelCase ) _snake_case = AutoformerEncoder.from_pretrained(UpperCAmelCase ).to(UpperCAmelCase ) _snake_case, _snake_case, _snake_case, _snake_case, _snake_case = model.create_network_inputs(**UpperCAmelCase ) _snake_case, _snake_case = model.decomposition_layer(transformer_inputs[:, : config.context_length, ...] ) _snake_case = torch.cat( (transformer_inputs[:, : config.context_length, ...], feature[:, : config.context_length, ...]) , dim=-1 , ) _snake_case = encoder(inputs_embeds=UpperCAmelCase )[0] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1e-3 ) _snake_case = ( torch.mean(transformer_inputs[:, : config.context_length, ...] , dim=1 ) .unsqueeze(1 ) .repeat(1 , config.prediction_length , 1 ) ) _snake_case = torch.zeros( [transformer_inputs.shape[0], config.prediction_length, transformer_inputs.shape[2]] , device=enc_input.device , ) _snake_case = torch.cat( ( torch.cat((seasonal_input[:, -config.label_length :, ...], zeros) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) _snake_case = torch.cat( ( torch.cat((trend_input[:, -config.label_length :, ...], mean) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) with tempfile.TemporaryDirectory() as tmpdirname: _snake_case = model.get_decoder() decoder.save_pretrained(UpperCAmelCase ) _snake_case = AutoformerDecoder.from_pretrained(UpperCAmelCase ).to(UpperCAmelCase ) _snake_case = decoder( trend=UpperCAmelCase , inputs_embeds=UpperCAmelCase , encoder_hidden_states=UpperCAmelCase , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1e-3 ) @require_torch class _lowerCAmelCase ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = (AutoformerModel, AutoformerForPrediction) if is_torch_available() else () lowerCAmelCase_ = (AutoformerForPrediction,) if is_torch_available() else () lowerCAmelCase_ = {"feature-extraction": AutoformerModel} if is_torch_available() else {} lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def lowercase (self ) -> List[Any]: _snake_case = AutoformerModelTester(self ) _snake_case = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase ) def lowercase (self ) -> List[Any]: self.config_tester.run_common_tests() def lowercase (self ) -> Any: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(UpperCAmelCase ) _snake_case, _snake_case = model_class.from_pretrained(UpperCAmelCase , output_loading_info=UpperCAmelCase ) self.assertEqual(info["""missing_keys"""] , [] ) def lowercase (self ) -> List[Any]: _snake_case = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*UpperCAmelCase ) @unittest.skip(reason="""Model has no tokens embeddings""" ) def lowercase (self ) -> Tuple: pass def lowercase (self ) -> Any: _snake_case = inspect.signature(getattr(UpperCAmelCase , """forward""" ) ) # The main input is the name of the argument after `self` _snake_case = list(model_signature.parameters.keys() )[1] self.assertEqual(AutoformerModel.main_input_name , UpperCAmelCase ) def lowercase (self ) -> List[Any]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = [ """past_values""", """past_time_features""", """past_observed_mask""", """static_categorical_features""", """static_real_features""", """future_values""", """future_time_features""", ] if model.__class__.__name__ in ["AutoformerForPrediction"]: expected_arg_names.append("""future_observed_mask""" ) expected_arg_names.extend( [ """decoder_attention_mask""", """head_mask""", """decoder_head_mask""", """cross_attn_head_mask""", """encoder_outputs""", """past_key_values""", """output_hidden_states""", """output_attentions""", """use_cache""", """return_dict""", ] ) self.assertListEqual(arg_names[: len(UpperCAmelCase )] , UpperCAmelCase ) def lowercase (self ) -> List[Any]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = True _snake_case = getattr(self.model_tester , """seq_length""" , UpperCAmelCase ) _snake_case = getattr(self.model_tester , """decoder_seq_length""" , UpperCAmelCase ) _snake_case = getattr(self.model_tester , """encoder_seq_length""" , UpperCAmelCase ) _snake_case = getattr(self.model_tester , """d_model""" , UpperCAmelCase ) _snake_case = getattr(self.model_tester , """num_attention_heads""" , UpperCAmelCase ) _snake_case = d_model // num_attention_heads for model_class in self.all_model_classes: _snake_case = True _snake_case = False _snake_case = True _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _snake_case = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(UpperCAmelCase ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _snake_case = True _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _snake_case = outputs.encoder_attentions self.assertEqual(len(UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) _snake_case = len(UpperCAmelCase ) _snake_case = 7 if "last_hidden_state" in outputs: correct_outlen += 1 if "trend" in outputs: correct_outlen += 1 if "past_key_values" in outputs: correct_outlen += 1 # past_key_values have been returned if "loss" in outputs: correct_outlen += 1 if "params" in outputs: correct_outlen += 1 self.assertEqual(UpperCAmelCase , UpperCAmelCase ) # decoder attentions _snake_case = outputs.decoder_attentions self.assertIsInstance(UpperCAmelCase , (list, tuple) ) self.assertEqual(len(UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # cross attentions _snake_case = outputs.cross_attentions self.assertIsInstance(UpperCAmelCase , (list, tuple) ) self.assertEqual(len(UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(cross_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # Check attention is always last and order is fine _snake_case = True _snake_case = True _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) self.assertEqual(out_len + 2 , len(UpperCAmelCase ) ) _snake_case = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) @is_flaky() def lowercase (self ) -> List[Any]: super().test_retain_grad_hidden_states_attentions() def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE="train-batch.pt" ): _snake_case = hf_hub_download(repo_id="""hf-internal-testing/tourism-monthly-batch""" , filename=_SCREAMING_SNAKE_CASE , repo_type="""dataset""" ) _snake_case = torch.load(_SCREAMING_SNAKE_CASE , map_location=_SCREAMING_SNAKE_CASE ) return batch @require_torch @slow class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowercase (self ) -> Union[str, Any]: _snake_case = AutoformerModel.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(UpperCAmelCase ) _snake_case = prepare_batch() with torch.no_grad(): _snake_case = model( past_values=batch["""past_values"""] , past_time_features=batch["""past_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , static_categorical_features=batch["""static_categorical_features"""] , future_values=batch["""future_values"""] , future_time_features=batch["""future_time_features"""] , )[0] _snake_case = torch.Size( (64, model.config.prediction_length + model.config.label_length, model.config.feature_size) ) self.assertEqual(output.shape , UpperCAmelCase ) _snake_case = torch.tensor( [[0.3593, -1.3398, 0.6330], [0.2279, 1.5396, -0.1792], [0.0450, 1.3225, -0.2335]] , device=UpperCAmelCase ) self.assertTrue(torch.allclose(output[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> str: _snake_case = AutoformerForPrediction.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(UpperCAmelCase ) _snake_case = prepare_batch("""val-batch.pt""" ) with torch.no_grad(): _snake_case = model( past_values=batch["""past_values"""] , past_time_features=batch["""past_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , static_categorical_features=batch["""static_categorical_features"""] , ).encoder_last_hidden_state _snake_case = torch.Size((64, model.config.context_length, model.config.d_model) ) self.assertEqual(output.shape , UpperCAmelCase ) _snake_case = torch.tensor( [[-0.0734, -0.9036, 0.8358], [4.7186, 2.4113, 1.9581], [1.7953, 2.3558, 1.2970]] , device=UpperCAmelCase ) self.assertTrue(torch.allclose(output[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> Optional[int]: _snake_case = AutoformerForPrediction.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(UpperCAmelCase ) _snake_case = prepare_batch("""val-batch.pt""" ) with torch.no_grad(): _snake_case = model.generate( static_categorical_features=batch["""static_categorical_features"""] , past_time_features=batch["""past_time_features"""] , past_values=batch["""past_values"""] , future_time_features=batch["""future_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , ) _snake_case = torch.Size((64, model.config.num_parallel_samples, model.config.prediction_length) ) self.assertEqual(outputs.sequences.shape , UpperCAmelCase ) _snake_case = torch.tensor([3130.6763, 4056.5293, 7053.0786] , device=UpperCAmelCase ) _snake_case = outputs.sequences.mean(dim=1 ) self.assertTrue(torch.allclose(mean_prediction[0, -3:] , UpperCAmelCase , rtol=1e-1 ) )
270
0
'''simple docstring''' import argparse import struct import unittest class UpperCAmelCase : '''simple docstring''' def __init__( self , __lowerCAmelCase ) -> None: lowercase__ : Tuple = data # Initialize hash values lowercase__ : Tuple = [ 0X6A09E667, 0XBB67AE85, 0X3C6EF372, 0XA54FF53A, 0X510E527F, 0X9B05688C, 0X1F83D9AB, 0X5BE0CD19, ] # Initialize round constants lowercase__ : Tuple = [ 0X428A2F98, 0X71374491, 0XB5C0FBCF, 0XE9B5DBA5, 0X3956C25B, 0X59F111F1, 0X923F82A4, 0XAB1C5ED5, 0XD807AA98, 0X12835B01, 0X243185BE, 0X550C7DC3, 0X72BE5D74, 0X80DEB1FE, 0X9BDC06A7, 0XC19BF174, 0XE49B69C1, 0XEFBE4786, 0X0FC19DC6, 0X240CA1CC, 0X2DE92C6F, 0X4A7484AA, 0X5CB0A9DC, 0X76F988DA, 0X983E5152, 0XA831C66D, 0XB00327C8, 0XBF597FC7, 0XC6E00BF3, 0XD5A79147, 0X06CA6351, 0X14292967, 0X27B70A85, 0X2E1B2138, 0X4D2C6DFC, 0X53380D13, 0X650A7354, 0X766A0ABB, 0X81C2C92E, 0X92722C85, 0XA2BFE8A1, 0XA81A664B, 0XC24B8B70, 0XC76C51A3, 0XD192E819, 0XD6990624, 0XF40E3585, 0X106AA070, 0X19A4C116, 0X1E376C08, 0X2748774C, 0X34B0BCB5, 0X391C0CB3, 0X4ED8AA4A, 0X5B9CCA4F, 0X682E6FF3, 0X748F82EE, 0X78A5636F, 0X84C87814, 0X8CC70208, 0X90BEFFFA, 0XA4506CEB, 0XBEF9A3F7, 0XC67178F2, ] lowercase__ : int = self.preprocessing(self.data ) self.final_hash() @staticmethod def _lowerCAmelCase( __lowerCAmelCase ) -> bytes: lowercase__ : Any = b'\x80' + (b'\x00' * (63 - (len(SCREAMING_SNAKE_CASE__ ) + 8) % 64)) lowercase__ : List[str] = struct.pack('''>Q''' , (len(SCREAMING_SNAKE_CASE__ ) * 8) ) return data + padding + big_endian_integer def _lowerCAmelCase( self ) -> None: # Convert into blocks of 64 bytes lowercase__ : int = [ self.preprocessed_data[x : x + 64] for x in range(0 , len(self.preprocessed_data ) , 64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers lowercase__ : List[str] = list(struct.unpack('''>16L''' , SCREAMING_SNAKE_CASE__ ) ) # add 48 0-ed integers words += [0] * 48 lowercase__ : Tuple = self.hashes for index in range(0 , 64 ): if index > 15: # modify the zero-ed indexes at the end of the array lowercase__ : Optional[int] = ( self.ror(words[index - 15] , 7 ) ^ self.ror(words[index - 15] , 18 ) ^ (words[index - 15] >> 3) ) lowercase__ : List[str] = ( self.ror(words[index - 2] , 17 ) ^ self.ror(words[index - 2] , 19 ) ^ (words[index - 2] >> 10) ) lowercase__ : Union[str, Any] = ( words[index - 16] + sa + words[index - 7] + sa ) % 0X100000000 # Compression lowercase__ : Any = self.ror(SCREAMING_SNAKE_CASE__ , 6 ) ^ self.ror(SCREAMING_SNAKE_CASE__ , 11 ) ^ self.ror(SCREAMING_SNAKE_CASE__ , 25 ) lowercase__ : List[str] = (e & f) ^ ((~e & 0XFFFFFFFF) & g) lowercase__ : Tuple = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0X100000000 lowercase__ : int = self.ror(SCREAMING_SNAKE_CASE__ , 2 ) ^ self.ror(SCREAMING_SNAKE_CASE__ , 13 ) ^ self.ror(SCREAMING_SNAKE_CASE__ , 22 ) lowercase__ : Optional[Any] = (a & b) ^ (a & c) ^ (b & c) lowercase__ : Tuple = (sa + maj) % 0X100000000 lowercase__ : List[str] = ( g, f, e, ((d + tempa) % 0X100000000), c, b, a, ((tempa + tempa) % 0X100000000), ) lowercase__ : Tuple = [a, b, c, d, e, f, g, h] # Modify final values lowercase__ : Dict = [ ((element + mutated_hash_values[index]) % 0X100000000) for index, element in enumerate(self.hashes ) ] lowercase__ : Any = ''.join([hex(SCREAMING_SNAKE_CASE__ )[2:].zfill(8 ) for value in self.hashes] ) def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase ) -> int: return 0XFFFFFFFF & (value << (32 - rotations)) | (value >> rotations) class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase( self ) -> None: import hashlib lowercase__ : Union[str, Any] = bytes('''Test String''' , '''utf-8''' ) self.assertEqual(SHAaaa(SCREAMING_SNAKE_CASE__ ).hash , hashlib.shaaaa(SCREAMING_SNAKE_CASE__ ).hexdigest() ) def __UpperCamelCase ( ): import doctest doctest.testmod() lowercase__ : Dict = argparse.ArgumentParser() parser.add_argument( '''-s''' , '''--string''' , dest='''input_string''' , default='''Hello World!! Welcome to Cryptography''' , help='''Hash the string''' , ) parser.add_argument( '''-f''' , '''--file''' , dest='''input_file''' , help='''Hash contents of a file''' ) lowercase__ : Optional[int] = parser.parse_args() lowercase__ : List[Any] = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , '''rb''' ) as f: lowercase__ : Optional[Any] = f.read() else: lowercase__ : Optional[Any] = bytes(__A , '''utf-8''' ) print(SHAaaa(__A ).hash ) if __name__ == "__main__": main()
198
from __future__ import annotations UpperCAmelCase_ : Tuple = [] def SCREAMING_SNAKE_CASE_ ( __A : list[list[int]] , __A : int , __A : int ) -> bool: """simple docstring""" for i in range(len(__A ) ): if board[row][i] == 1: return False for i in range(len(__A ) ): if board[i][column] == 1: return False for i, j in zip(range(__A , -1 , -1 ) , range(__A , -1 , -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(__A , -1 , -1 ) , range(__A , len(__A ) ) ): if board[i][j] == 1: return False return True def SCREAMING_SNAKE_CASE_ ( __A : list[list[int]] , __A : int ) -> bool: """simple docstring""" if row >= len(__A ): solution.append(__A ) printboard(__A ) print() return True for i in range(len(__A ) ): if is_safe(__A , __A , __A ): a_ : Any = 1 solve(__A , row + 1 ) a_ : Tuple = 0 return False def SCREAMING_SNAKE_CASE_ ( __A : list[list[int]] ) -> None: """simple docstring""" for i in range(len(__A ) ): for j in range(len(__A ) ): if board[i][j] == 1: print('Q' , end=' ' ) else: print('.' , end=' ' ) print() # n=int(input("The no. of queens")) UpperCAmelCase_ : List[str] = 8 UpperCAmelCase_ : str = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print('The total no. of solutions are :', len(solution))
32
0
"""simple docstring""" import re from filelock import FileLock try: import nltk a__ : Tuple = True except (ImportError, ModuleNotFoundError): a__ : str = False if NLTK_AVAILABLE: with FileLock('''.lock''') as lock: nltk.download('''punkt''', quiet=True) def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' re.sub("<n>" , "" , lowerCAmelCase_ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(lowerCAmelCase_ ) )
195
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
195
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowercase : Any = { 'configuration_vision_text_dual_encoder': ['VisionTextDualEncoderConfig'], 'processing_vision_text_dual_encoder': ['VisionTextDualEncoderProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = ['VisionTextDualEncoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Any = ['FlaxVisionTextDualEncoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : List[Any] = ['TFVisionTextDualEncoderModel'] if TYPE_CHECKING: from .configuration_vision_text_dual_encoder import VisionTextDualEncoderConfig from .processing_vision_text_dual_encoder import VisionTextDualEncoderProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_text_dual_encoder import VisionTextDualEncoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_text_dual_encoder import FlaxVisionTextDualEncoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_text_dual_encoder import TFVisionTextDualEncoderModel else: import sys lowercase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
20
'''simple docstring''' from scipy.stats import pearsonr import datasets lowercase : Optional[int] = '\nPearson correlation coefficient and p-value for testing non-correlation.\nThe Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases.\nThe p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets.\n' lowercase : Optional[Any] = '\nArgs:\n predictions (`list` of `int`): Predicted class labels, as returned by a model.\n references (`list` of `int`): Ground truth labels.\n return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`.\n\nReturns:\n pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation.\n p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities.\n\nExamples:\n\n Example 1-A simple example using only predictions and references.\n >>> pearsonr_metric = datasets.load_metric("pearsonr")\n >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5])\n >>> print(round(results[\'pearsonr\'], 2))\n -0.74\n\n Example 2-The same as Example 1, but that also returns the `p-value`.\n >>> pearsonr_metric = datasets.load_metric("pearsonr")\n >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True)\n >>> print(sorted(list(results.keys())))\n [\'p-value\', \'pearsonr\']\n >>> print(round(results[\'pearsonr\'], 2))\n -0.74\n >>> print(round(results[\'p-value\'], 2))\n 0.15\n' lowercase : str = '\n@article{2020SciPy-NMeth,\nauthor = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, Ilhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Antonio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\ntitle = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\njournal = {Nature Methods},\nyear = {2020},\nvolume = {17},\npages = {261--272},\nadsurl = {https://rdcu.be/b08Wh},\ndoi = {10.1038/s41592-019-0686-2},\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A ( datasets.Metric ): def __lowerCAmelCase ( self ) -> List[str]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) , reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html'''] , ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=False ) -> Union[str, Any]: """simple docstring""" if return_pvalue: A : Union[str, Any] = pearsonr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )[0] )}
3
0
"""simple docstring""" import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print("Googling.....") __magic_name__ = 'https://www.google.com/search?q=' + ' '.join(sys.argv[1:]) __magic_name__ = requests.get(url, headers={"UserAgent": UserAgent().random}) # res.raise_for_status() with open("project1a.html", "wb") as out_file: # only for knowing the class for data in res.iter_content(10000): out_file.write(data) __magic_name__ = BeautifulSoup(res.text, "html.parser") __magic_name__ = list(soup.select(".eZt8xd"))[:5] print(len(links)) for link in links: if link.text == "Maps": webbrowser.open(link.get("href")) else: webbrowser.open(F"""https://google.com{link.get("href")}""")
368
"""simple docstring""" def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ = False ): if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 10 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3_3170_4406_4679_8873_8596_1981 and not allow_probable: raise ValueError( """Warning: upper bound of deterministic test is exceeded. """ """Pass allow_probable=True to allow probabilistic test. """ """A return value of True indicates a probable prime.""" ) # array bounds provided by analysis __SCREAMING_SNAKE_CASE = [ 2047, 137_3653, 2532_6001, 32_1503_1751, 2_1523_0289_8747, 3_4747_4966_0383, 341_5500_7172_8321, 1, 382_5123_0565_4641_3051, 1, 1, 3186_6585_7834_0311_5116_7461, 3_3170_4406_4679_8873_8596_1981, ] __SCREAMING_SNAKE_CASE = [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41] for idx, _p in enumerate(UpperCamelCase_ , 1 ): if n < _p: # then we have our last prime to check __SCREAMING_SNAKE_CASE = primes[:idx] break __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = n - 1, 0 # break up n -1 into a power of 2 (s) and # remaining odd component # essentially, solve for d * 2 ** s == n - 1 while d % 2 == 0: d //= 2 s += 1 for prime in plist: __SCREAMING_SNAKE_CASE = False for r in range(UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = pow(UpperCamelCase_ , d * 2**r , UpperCamelCase_ ) # see article for analysis explanation for m if (r == 0 and m == 1) or ((m + 1) % n == 0): __SCREAMING_SNAKE_CASE = True # this loop will not determine compositeness break if pr: continue # if pr is False, then the above loop never evaluated to true, # and the n MUST be composite return False return True def _lowerCAmelCase ( ): assert not miller_rabin(561 ) assert miller_rabin(563 ) # 2047 assert not miller_rabin(83_8201 ) assert miller_rabin(83_8207 ) # 1_373_653 assert not miller_rabin(1731_6001 ) assert miller_rabin(1731_6017 ) # 25_326_001 assert not miller_rabin(30_7838_6641 ) assert miller_rabin(30_7838_6653 ) # 3_215_031_751 assert not miller_rabin(1_7130_4557_4801 ) assert miller_rabin(1_7130_4557_4819 ) # 2_152_302_898_747 assert not miller_rabin(2_7797_9972_8307 ) assert miller_rabin(2_7797_9972_8327 ) # 3_474_749_660_383 assert not miller_rabin(113_8500_2390_9441 ) assert miller_rabin(113_8500_2390_9527 ) # 341_550_071_728_321 assert not miller_rabin(127_5041_0188_4880_4351 ) assert miller_rabin(127_5041_0188_4880_4391 ) # 3_825_123_056_546_413_051 assert not miller_rabin(796_6646_4458_5077_8779_1867 ) assert miller_rabin(796_6646_4458_5077_8779_1951 ) # 318_665_857_834_031_151_167_461 assert not miller_rabin(5528_4067_7446_6478_9766_0333 ) assert miller_rabin(5528_4067_7446_6478_9766_0359 ) # 3_317_044_064_679_887_385_961_981 # upper limit for probabilistic test if __name__ == "__main__": test_miller_rabin()
255
0
'''simple docstring''' import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _UpperCamelCase : '''simple docstring''' def __init__( self : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int]=1_3 , _lowerCAmelCase : str=3_2 , _lowerCAmelCase : Tuple=3 , _lowerCAmelCase : str=4 , _lowerCAmelCase : Optional[Any]=[1_0, 2_0, 3_0, 4_0] , _lowerCAmelCase : Dict=[2, 2, 3, 2] , _lowerCAmelCase : str=True , _lowerCAmelCase : List[Any]=True , _lowerCAmelCase : Optional[Any]=3_7 , _lowerCAmelCase : str="gelu" , _lowerCAmelCase : Union[str, Any]=1_0 , _lowerCAmelCase : Optional[int]=0.02 , _lowerCAmelCase : Any=["stage2", "stage3", "stage4"] , _lowerCAmelCase : List[Any]=3 , _lowerCAmelCase : Tuple=None , ): '''simple docstring''' __lowercase =parent __lowercase =batch_size __lowercase =image_size __lowercase =num_channels __lowercase =num_stages __lowercase =hidden_sizes __lowercase =depths __lowercase =is_training __lowercase =use_labels __lowercase =intermediate_size __lowercase =hidden_act __lowercase =type_sequence_label_size __lowercase =initializer_range __lowercase =out_features __lowercase =num_labels __lowercase =scope __lowercase =num_stages def __lowerCamelCase ( self : Optional[Any]): '''simple docstring''' __lowercase =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) __lowercase =None if self.use_labels: __lowercase =ids_tensor([self.batch_size] , self.type_sequence_label_size) __lowercase =self.get_config() return config, pixel_values, labels def __lowerCamelCase ( self : Dict): '''simple docstring''' return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def __lowerCamelCase ( self : Tuple): '''simple docstring''' return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=5_1_2 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=_lowerCAmelCase , auxiliary_loss_weight=0.4 , auxiliary_in_channels=4_0 , auxiliary_channels=2_5_6 , auxiliary_num_convs=1 , auxiliary_concat_input=_lowerCAmelCase , loss_ignore_index=2_5_5 , num_labels=self.num_labels , ) def __lowerCamelCase ( self : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[Any]): '''simple docstring''' __lowercase =UperNetForSemanticSegmentation(config=_lowerCAmelCase) model.to(_lowerCAmelCase) model.eval() __lowercase =model(_lowerCAmelCase) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size)) def __lowerCamelCase ( self : Dict): '''simple docstring''' __lowercase =self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) =config_and_inputs __lowercase ={'pixel_values': pixel_values} return config, inputs_dict @require_torch class _UpperCamelCase ( A , A , unittest.TestCase ): '''simple docstring''' lowerCAmelCase__ = (UperNetForSemanticSegmentation,) if is_torch_available() else () lowerCAmelCase__ = {"""image-segmentation""": UperNetForSemanticSegmentation} if is_torch_available() else {} lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False def __lowerCamelCase ( self : Tuple): '''simple docstring''' __lowercase =UperNetModelTester(self) __lowercase =ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=3_7) def __lowerCamelCase ( self : Tuple): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowerCamelCase ( self : Tuple): '''simple docstring''' return def __lowerCamelCase ( self : Optional[int]): '''simple docstring''' __lowercase , __lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase =model_class(_lowerCAmelCase) __lowercase =inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase =[*signature.parameters.keys()] __lowercase =['pixel_values'] self.assertListEqual(arg_names[:1] , _lowerCAmelCase) def __lowerCamelCase ( self : List[str]): '''simple docstring''' __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*_lowerCAmelCase) @unittest.skip(reason='UperNet does not use inputs_embeds') def __lowerCamelCase ( self : Any): '''simple docstring''' pass @unittest.skip(reason='UperNet does not support input and output embeddings') def __lowerCamelCase ( self : Any): '''simple docstring''' pass @unittest.skip(reason='UperNet does not have a base model') def __lowerCamelCase ( self : int): '''simple docstring''' pass @unittest.skip(reason='UperNet does not have a base model') def __lowerCamelCase ( self : List[Any]): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip(reason='UperNet has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`') def __lowerCamelCase ( self : Dict): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.') def __lowerCamelCase ( self : str): '''simple docstring''' pass def __lowerCamelCase ( self : Optional[int]): '''simple docstring''' def check_hidden_states_output(_lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int): __lowercase =model_class(_lowerCAmelCase) model.to(_lowerCAmelCase) model.eval() with torch.no_grad(): __lowercase =model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase)) __lowercase =outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase =self.model_tester.num_stages self.assertEqual(len(_lowerCAmelCase) , expected_num_stages + 1) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:]) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __lowercase , __lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase =True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase =True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase) def __lowerCamelCase ( self : List[Any]): '''simple docstring''' __lowercase , __lowercase =self.model_tester.prepare_config_and_inputs_for_common() __lowercase =_config_zero_init(_lowerCAmelCase) __lowercase =_config_zero_init(configs_no_init.backbone_config) for model_class in self.all_model_classes: __lowercase =model_class(config=_lowerCAmelCase) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip(reason='UperNet does not have tied weights') def __lowerCamelCase ( self : int): '''simple docstring''' pass @slow def __lowerCamelCase ( self : List[Any]): '''simple docstring''' for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase =UperNetForSemanticSegmentation.from_pretrained(_lowerCAmelCase) self.assertIsNotNone(_lowerCAmelCase) def _A ( ): """simple docstring""" __lowercase =hf_hub_download( repo_id='hf-internal-testing/fixtures_ade20k' , repo_type='dataset' , filename='ADE_val_00000001.jpg' ) __lowercase =Image.open(_lowerCAmelCase ).convert('RGB' ) return image @require_torch @require_vision @slow class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __lowerCamelCase ( self : str): '''simple docstring''' __lowercase =AutoImageProcessor.from_pretrained('openmmlab/upernet-swin-tiny') __lowercase =UperNetForSemanticSegmentation.from_pretrained('openmmlab/upernet-swin-tiny').to(_lowerCAmelCase) __lowercase =prepare_img() __lowercase =processor(images=_lowerCAmelCase , return_tensors='pt').to(_lowerCAmelCase) with torch.no_grad(): __lowercase =model(**_lowerCAmelCase) __lowercase =torch.Size((1, model.config.num_labels, 5_1_2, 5_1_2)) self.assertEqual(outputs.logits.shape , _lowerCAmelCase) __lowercase =torch.tensor( [[-7.5958, -7.5958, -7.4302], [-7.5958, -7.5958, -7.4302], [-7.4797, -7.4797, -7.3068]]).to(_lowerCAmelCase) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , _lowerCAmelCase , atol=1e-4)) def __lowerCamelCase ( self : Any): '''simple docstring''' __lowercase =AutoImageProcessor.from_pretrained('openmmlab/upernet-convnext-tiny') __lowercase =UperNetForSemanticSegmentation.from_pretrained('openmmlab/upernet-convnext-tiny').to(_lowerCAmelCase) __lowercase =prepare_img() __lowercase =processor(images=_lowerCAmelCase , return_tensors='pt').to(_lowerCAmelCase) with torch.no_grad(): __lowercase =model(**_lowerCAmelCase) __lowercase =torch.Size((1, model.config.num_labels, 5_1_2, 5_1_2)) self.assertEqual(outputs.logits.shape , _lowerCAmelCase) __lowercase =torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]]).to(_lowerCAmelCase) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , _lowerCAmelCase , atol=1e-4))
166
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = { """alibaba-damo/mgp-str-base""": """https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json""", } class _UpperCamelCase ( A ): '''simple docstring''' lowerCAmelCase__ = """mgp-str""" def __init__( self : int , _lowerCAmelCase : str=[3_2, 1_2_8] , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : int=3 , _lowerCAmelCase : str=2_7 , _lowerCAmelCase : List[str]=3_8 , _lowerCAmelCase : Tuple=5_0_2_5_7 , _lowerCAmelCase : str=3_0_5_2_2 , _lowerCAmelCase : Optional[int]=7_6_8 , _lowerCAmelCase : Optional[int]=1_2 , _lowerCAmelCase : Optional[Any]=1_2 , _lowerCAmelCase : Optional[int]=4.0 , _lowerCAmelCase : List[Any]=True , _lowerCAmelCase : Tuple=False , _lowerCAmelCase : List[Any]=1e-5 , _lowerCAmelCase : List[Any]=0.0 , _lowerCAmelCase : str=0.0 , _lowerCAmelCase : Tuple=0.0 , _lowerCAmelCase : str=False , _lowerCAmelCase : List[Any]=0.02 , **_lowerCAmelCase : Optional[Any] , ): '''simple docstring''' super().__init__(**_lowerCAmelCase) __lowercase =image_size __lowercase =patch_size __lowercase =num_channels __lowercase =max_token_length __lowercase =num_character_labels __lowercase =num_bpe_labels __lowercase =num_wordpiece_labels __lowercase =hidden_size __lowercase =num_hidden_layers __lowercase =num_attention_heads __lowercase =mlp_ratio __lowercase =distilled __lowercase =layer_norm_eps __lowercase =drop_rate __lowercase =qkv_bias __lowercase =attn_drop_rate __lowercase =drop_path_rate __lowercase =output_aa_attentions __lowercase =initializer_range
166
1
lowerCAmelCase__ = 'Tobias Carryer' from time import time class a_ : '''simple docstring''' def __init__( self : int , lowercase__ : str , lowercase__ : Optional[int] , lowercase__ : int , lowercase__ : List[str]=int(time())): # noqa: B008 '''simple docstring''' lowerCAmelCase__ = multiplier lowerCAmelCase__ = increment lowerCAmelCase__ = modulo lowerCAmelCase__ = seed def __snake_case ( self : Optional[int]): '''simple docstring''' lowerCAmelCase__ = (self.multiplier * self.seed + self.increment) % self.modulo return self.seed if __name__ == "__main__": # Show the LCG in action. lowerCAmelCase__ = LinearCongruentialGenerator(1664525, 1013904223, 2 << 31) while True: print(lcg.next_number())
355
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class a_ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : str , lowercase__ : pyspark.sql.DataFrame , lowercase__ : Optional[NamedSplit] = None , lowercase__ : Optional[Features] = None , lowercase__ : bool = True , lowercase__ : str = None , lowercase__ : bool = False , lowercase__ : str = None , lowercase__ : bool = True , lowercase__ : str = "arrow" , **lowercase__ : Any , ): '''simple docstring''' super().__init__( split=lowercase__ , features=lowercase__ , cache_dir=lowercase__ , keep_in_memory=lowercase__ , streaming=lowercase__ , **lowercase__ , ) lowerCAmelCase__ = load_from_cache_file lowerCAmelCase__ = file_format lowerCAmelCase__ = Spark( df=lowercase__ , features=lowercase__ , cache_dir=lowercase__ , working_dir=lowercase__ , **lowercase__ , ) def __snake_case ( self : Tuple): '''simple docstring''' if self.streaming: return self.builder.as_streaming_dataset(split=self.split) lowerCAmelCase__ = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=lowercase__ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split)
119
0
"""simple docstring""" import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __lowerCamelCase : '''simple docstring''' def __init__( self : Union[str, Any] , a_ : int , a_ : List[Any]=13 , a_ : Dict=32 , a_ : int=2 , a_ : Dict=3 , a_ : List[str]=16 , a_ : Tuple=[1, 2, 1] , a_ : Dict=[2, 2, 4] , a_ : str=2 , a_ : Optional[Any]=2.0 , a_ : Any=True , a_ : Union[str, Any]=0.0 , a_ : Tuple=0.0 , a_ : Any=0.1 , a_ : List[Any]="gelu" , a_ : Union[str, Any]=False , a_ : Optional[Any]=True , a_ : Tuple=0.02 , a_ : Any=1e-5 , a_ : List[Any]=True , a_ : List[Any]=None , a_ : List[Any]=True , a_ : Union[str, Any]=10 , a_ : Optional[int]=8 , ): lowerCAmelCase_ : int = parent lowerCAmelCase_ : Dict = batch_size lowerCAmelCase_ : Dict = image_size lowerCAmelCase_ : str = patch_size lowerCAmelCase_ : List[Any] = num_channels lowerCAmelCase_ : Optional[int] = embed_dim lowerCAmelCase_ : Tuple = depths lowerCAmelCase_ : int = num_heads lowerCAmelCase_ : Tuple = window_size lowerCAmelCase_ : Tuple = mlp_ratio lowerCAmelCase_ : List[str] = qkv_bias lowerCAmelCase_ : Dict = hidden_dropout_prob lowerCAmelCase_ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase_ : Any = drop_path_rate lowerCAmelCase_ : Dict = hidden_act lowerCAmelCase_ : int = use_absolute_embeddings lowerCAmelCase_ : Union[str, Any] = patch_norm lowerCAmelCase_ : Optional[Any] = layer_norm_eps lowerCAmelCase_ : Optional[int] = initializer_range lowerCAmelCase_ : Tuple = is_training lowerCAmelCase_ : str = scope lowerCAmelCase_ : Dict = use_labels lowerCAmelCase_ : int = type_sequence_label_size lowerCAmelCase_ : Optional[Any] = encoder_stride def lowerCamelCase ( self : Optional[int] ): lowerCAmelCase_ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase_ : Tuple = None if self.use_labels: lowerCAmelCase_ : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase_ : List[str] = self.get_config() return config, pixel_values, labels def lowerCamelCase ( self : Tuple ): return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def lowerCamelCase ( self : int , a_ : List[str] , a_ : Tuple , a_ : Tuple ): lowerCAmelCase_ : Any = SwinvaModel(config=a_ ) model.to(a_ ) model.eval() lowerCAmelCase_ : Tuple = model(a_ ) lowerCAmelCase_ : List[str] = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) lowerCAmelCase_ : int = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def lowerCamelCase ( self : str , a_ : Optional[int] , a_ : Optional[Any] , a_ : int ): lowerCAmelCase_ : Optional[int] = SwinvaForMaskedImageModeling(config=a_ ) model.to(a_ ) model.eval() lowerCAmelCase_ : Any = model(a_ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase_ : Tuple = 1 lowerCAmelCase_ : int = SwinvaForMaskedImageModeling(a_ ) model.to(a_ ) model.eval() lowerCAmelCase_ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase_ : Optional[Any] = model(a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def lowerCamelCase ( self : List[Any] , a_ : Optional[Any] , a_ : Optional[int] , a_ : Union[str, Any] ): lowerCAmelCase_ : Tuple = self.type_sequence_label_size lowerCAmelCase_ : Any = SwinvaForImageClassification(a_ ) model.to(a_ ) model.eval() lowerCAmelCase_ : Optional[int] = model(a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowerCamelCase ( self : List[Any] ): lowerCAmelCase_ : Optional[int] = self.prepare_config_and_inputs() lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : int = config_and_inputs lowerCAmelCase_ : List[Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __lowerCamelCase ( A__ , A__ , unittest.TestCase ): '''simple docstring''' a_ : Any = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) a_ : Optional[Any] = ( {"""feature-extraction""": SwinvaModel, """image-classification""": SwinvaForImageClassification} if is_torch_available() else {} ) a_ : List[str] = False a_ : Any = False a_ : Optional[Any] = False a_ : List[Any] = False def lowerCamelCase ( self : List[str] ): lowerCAmelCase_ : str = SwinvaModelTester(self ) lowerCAmelCase_ : str = ConfigTester(self , config_class=a_ , embed_dim=37 ) def lowerCamelCase ( self : str ): self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCamelCase ( self : str ): lowerCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) @unittest.skip(reason="Got `CUDA error: misaligned address` with PyTorch 2.0.0." ) def lowerCamelCase ( self : int ): pass @unittest.skip(reason="Swinv2 does not use inputs_embeds" ) def lowerCamelCase ( self : Dict ): pass def lowerCamelCase ( self : Optional[int] ): lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase_ : List[str] = model_class(a_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase_ : Optional[Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a_ , nn.Linear ) ) def lowerCamelCase ( self : Optional[Any] ): lowerCAmelCase_ , lowerCAmelCase_ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase_ : int = model_class(a_ ) lowerCAmelCase_ : Tuple = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase_ : List[str] = [*signature.parameters.keys()] lowerCAmelCase_ : Optional[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , a_ ) def lowerCamelCase ( self : Dict ): lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase_ : List[str] = True for model_class in self.all_model_classes: lowerCAmelCase_ : Optional[int] = True lowerCAmelCase_ : Union[str, Any] = False lowerCAmelCase_ : Optional[Any] = True lowerCAmelCase_ : Tuple = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): lowerCAmelCase_ : Dict = model(**self._prepare_for_class(a_ , a_ ) ) lowerCAmelCase_ : Tuple = outputs.attentions lowerCAmelCase_ : str = len(self.model_tester.depths ) self.assertEqual(len(a_ ) , a_ ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCAmelCase_ : List[str] = True lowerCAmelCase_ : int = config.window_size**2 lowerCAmelCase_ : Any = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): lowerCAmelCase_ : str = model(**self._prepare_for_class(a_ , a_ ) ) lowerCAmelCase_ : Any = outputs.attentions self.assertEqual(len(a_ ) , a_ ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) lowerCAmelCase_ : Optional[int] = len(a_ ) # Check attention is always last and order is fine lowerCAmelCase_ : Optional[int] = True lowerCAmelCase_ : Optional[Any] = True lowerCAmelCase_ : str = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): lowerCAmelCase_ : List[Any] = model(**self._prepare_for_class(a_ , a_ ) ) if hasattr(self.model_tester , "num_hidden_states_types" ): lowerCAmelCase_ : Any = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states lowerCAmelCase_ : Tuple = 2 self.assertEqual(out_len + added_hidden_states , len(a_ ) ) lowerCAmelCase_ : Optional[Any] = outputs.attentions self.assertEqual(len(a_ ) , a_ ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def lowerCamelCase ( self : int , a_ : Any , a_ : int , a_ : str , a_ : int ): lowerCAmelCase_ : int = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): lowerCAmelCase_ : Optional[int] = model(**self._prepare_for_class(a_ , a_ ) ) lowerCAmelCase_ : List[str] = outputs.hidden_states lowerCAmelCase_ : Optional[Any] = getattr( self.model_tester , "expected_num_hidden_layers" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(a_ ) , a_ ) # Swinv2 has a different seq_length lowerCAmelCase_ : int = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase_ : Optional[Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) lowerCAmelCase_ : Optional[Any] = outputs.reshaped_hidden_states self.assertEqual(len(a_ ) , a_ ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = reshaped_hidden_states[0].shape lowerCAmelCase_ : Optional[Any] = ( reshaped_hidden_states[0].view(a_ , a_ , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def lowerCamelCase ( self : Optional[Any] ): lowerCAmelCase_ , lowerCAmelCase_ : Any = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase_ : Dict = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: lowerCAmelCase_ : Tuple = True self.check_hidden_states_output(a_ , a_ , a_ , a_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase_ : Union[str, Any] = True self.check_hidden_states_output(a_ , a_ , a_ , a_ ) def lowerCamelCase ( self : Dict ): lowerCAmelCase_ , lowerCAmelCase_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase_ : Union[str, Any] = 3 lowerCAmelCase_ : int = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) lowerCAmelCase_ : int = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase_ : List[Any] = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) lowerCAmelCase_ : Optional[Any] = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: lowerCAmelCase_ : Dict = True self.check_hidden_states_output(a_ , a_ , a_ , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase_ : int = True self.check_hidden_states_output(a_ , a_ , a_ , (padded_height, padded_width) ) def lowerCamelCase ( self : Union[str, Any] ): lowerCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*a_ ) def lowerCamelCase ( self : Any ): lowerCAmelCase_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a_ ) @slow def lowerCamelCase ( self : List[str] ): for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : List[Any] = SwinvaModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) def lowerCamelCase ( self : Dict ): lowerCAmelCase_ , lowerCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase_ : List[Any] = _config_zero_init(a_ ) for model_class in self.all_model_classes: lowerCAmelCase_ : Any = model_class(config=a_ ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def lowerCamelCase ( self : Optional[Any] ): return ( AutoImageProcessor.from_pretrained("microsoft/swinv2-tiny-patch4-window8-256" ) if is_vision_available() else None ) @slow def lowerCamelCase ( self : int ): lowerCAmelCase_ : Dict = SwinvaForImageClassification.from_pretrained("microsoft/swinv2-tiny-patch4-window8-256" ).to( a_ ) lowerCAmelCase_ : Optional[Any] = self.default_image_processor lowerCAmelCase_ : List[str] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowerCAmelCase_ : Tuple = image_processor(images=a_ , return_tensors="pt" ).to(a_ ) # forward pass with torch.no_grad(): lowerCAmelCase_ : Union[str, Any] = model(**a_ ) # verify the logits lowerCAmelCase_ : Tuple = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , a_ ) lowerCAmelCase_ : Tuple = torch.tensor([-0.3947, -0.4306, 0.0026] ).to(a_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a_ , atol=1e-4 ) )
241
"""simple docstring""" # coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import platform import sys lowercase__ = """3""" print("""Python version:""", sys.version) print("""OS platform:""", platform.platform()) print("""OS architecture:""", platform.machine()) try: import torch print("""Torch version:""", torch.__version__) print("""Cuda available:""", torch.cuda.is_available()) print("""Cuda version:""", torch.version.cuda) print("""CuDNN version:""", torch.backends.cudnn.version()) print("""Number of GPUs available:""", torch.cuda.device_count()) except ImportError: print("""Torch version:""", None) try: import transformers print("""transformers version:""", transformers.__version__) except ImportError: print("""transformers version:""", None)
241
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ : Optional[int] =logging.get_logger(__name__) lowerCAmelCase__ : List[str] ={ 'EleutherAI/gpt-neox-20b': 'https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json', # See all GPTNeoX models at https://huggingface.co/models?filter=gpt_neox } class __lowercase (__SCREAMING_SNAKE_CASE ): """simple docstring""" _UpperCAmelCase = """gpt_neox""" def __init__( self , lowerCAmelCase__=5_0_4_3_2 , lowerCAmelCase__=6_1_4_4 , lowerCAmelCase__=4_4 , lowerCAmelCase__=6_4 , lowerCAmelCase__=2_4_5_7_6 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.25 , lowerCAmelCase__=1_0_0_0_0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.1 , lowerCAmelCase__=2_0_4_8 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-5 , lowerCAmelCase__=True , lowerCAmelCase__=0 , lowerCAmelCase__=2 , lowerCAmelCase__=False , lowerCAmelCase__=True , lowerCAmelCase__=None , **lowerCAmelCase__ , ): """simple docstring""" super().__init__(bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Any = vocab_size SCREAMING_SNAKE_CASE_ : Dict = max_position_embeddings SCREAMING_SNAKE_CASE_ : List[str] = hidden_size SCREAMING_SNAKE_CASE_ : Dict = num_hidden_layers SCREAMING_SNAKE_CASE_ : Optional[int] = num_attention_heads SCREAMING_SNAKE_CASE_ : Optional[int] = intermediate_size SCREAMING_SNAKE_CASE_ : Optional[int] = hidden_act SCREAMING_SNAKE_CASE_ : Union[str, Any] = rotary_pct SCREAMING_SNAKE_CASE_ : Optional[Any] = rotary_emb_base SCREAMING_SNAKE_CASE_ : Optional[int] = attention_dropout SCREAMING_SNAKE_CASE_ : List[str] = hidden_dropout SCREAMING_SNAKE_CASE_ : Union[str, Any] = classifier_dropout SCREAMING_SNAKE_CASE_ : Optional[int] = initializer_range SCREAMING_SNAKE_CASE_ : Tuple = layer_norm_eps SCREAMING_SNAKE_CASE_ : int = use_cache SCREAMING_SNAKE_CASE_ : List[Any] = tie_word_embeddings SCREAMING_SNAKE_CASE_ : Dict = use_parallel_residual SCREAMING_SNAKE_CASE_ : Any = rope_scaling self._rope_scaling_validation() if self.hidden_size % self.num_attention_heads != 0: raise ValueError( 'The hidden size is not divisble by the number of attention heads! Make sure to update them!' ) def UpperCamelCase__ ( self ): """simple docstring""" if self.rope_scaling is None: return if not isinstance(self.rope_scaling , lowerCAmelCase__ ) or len(self.rope_scaling ) != 2: raise ValueError( '`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ' F'''got {self.rope_scaling}''' ) SCREAMING_SNAKE_CASE_ : int = self.rope_scaling.get('type' , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Dict = self.rope_scaling.get('factor' , lowerCAmelCase__ ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F'''`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}''' ) if rope_scaling_factor is None or not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) or rope_scaling_factor <= 1.0: raise ValueError(F'''`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}''' )
162
from collections.abc import Sequence def a__ ( A__, A__ = False ): if not arr: return 0 SCREAMING_SNAKE_CASE_ : str = 0 if allow_empty_subarrays else float('-inf' ) SCREAMING_SNAKE_CASE_ : Tuple = 0.0 for num in arr: SCREAMING_SNAKE_CASE_ : int = max(0 if allow_empty_subarrays else num, curr_sum + num ) SCREAMING_SNAKE_CASE_ : List[Any] = max(A__, A__ ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() lowerCAmelCase__ : Union[str, Any] =[-2, 1, -3, 4, -1, 2, 1, -5, 4] print(F"""{max_subarray_sum(nums) = }""")
162
1
"""simple docstring""" import inspect import unittest import numpy as np from transformers import ViTConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax from transformers.models.vit.modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self : List[Any] ,A_ : Any ,A_ : str=13 ,A_ : List[str]=30 ,A_ : Any=2 ,A_ : Union[str, Any]=3 ,A_ : List[str]=True ,A_ : Any=True ,A_ : List[Any]=32 ,A_ : List[Any]=5 ,A_ : List[Any]=4 ,A_ : Optional[int]=37 ,A_ : List[str]="gelu" ,A_ : Optional[int]=0.1 ,A_ : Optional[int]=0.1 ,A_ : Tuple=10 ,A_ : Any=0.02 ,) -> Union[str, Any]: A = parent A = batch_size A = image_size A = patch_size A = num_channels A = is_training A = use_labels A = hidden_size A = num_hidden_layers A = num_attention_heads A = intermediate_size A = hidden_act A = hidden_dropout_prob A = attention_probs_dropout_prob A = type_sequence_label_size A = initializer_range # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) A = (image_size // patch_size) ** 2 A = num_patches + 1 def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> List[Any]: A = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A = ViTConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=A_ ,initializer_range=self.initializer_range ,) return config, pixel_values def _SCREAMING_SNAKE_CASE ( self : int ,A_ : Optional[Any] ,A_ : Union[str, Any] ) -> List[str]: A = FlaxViTModel(config=A_ ) A = model(A_ ) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) A = (self.image_size, self.image_size) A = (self.patch_size, self.patch_size) A = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, num_patches + 1, self.hidden_size) ) def _SCREAMING_SNAKE_CASE ( self : int ,A_ : Dict ,A_ : Tuple ) -> Tuple: A = self.type_sequence_label_size A = FlaxViTForImageClassification(config=A_ ) A = model(A_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) # test greyscale images A = 1 A = FlaxViTForImageClassification(A_ ) A = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A = model(A_ ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Tuple: A = self.prepare_config_and_inputs() ( ( A ) , ( A ) , ) = config_and_inputs A = {'pixel_values': pixel_values} return config, inputs_dict @require_flax class lowerCAmelCase_ ( _lowercase , unittest.TestCase ): '''simple docstring''' _lowerCamelCase: Optional[int] = (FlaxViTModel, FlaxViTForImageClassification) if is_flax_available() else () def _SCREAMING_SNAKE_CASE ( self : Any ) -> None: A = FlaxViTModelTester(self ) A = ConfigTester(self ,config_class=A_ ,has_text_modality=A_ ,hidden_size=37 ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Dict: self.config_tester.run_common_tests() def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Dict: A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def _SCREAMING_SNAKE_CASE ( self : Any ) -> Optional[int]: A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> int: A , A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A = model_class(A_ ) A = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A = [*signature.parameters.keys()] A = ['pixel_values'] self.assertListEqual(arg_names[:1] ,A_ ) def _SCREAMING_SNAKE_CASE ( self : str ) -> Union[str, Any]: A , A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): A = self._prepare_for_class(A_ ,A_ ) A = model_class(A_ ) @jax.jit def model_jitted(A_ : List[Any] ,**A_ : List[Any] ): return model(pixel_values=A_ ,**A_ ) with self.subTest('JIT Enabled' ): A = model_jitted(**A_ ).to_tuple() with self.subTest('JIT Disabled' ): with jax.disable_jit(): A = model_jitted(**A_ ).to_tuple() self.assertEqual(len(A_ ) ,len(A_ ) ) for jitted_output, output in zip(A_ ,A_ ): self.assertEqual(jitted_output.shape ,output.shape ) @slow def _SCREAMING_SNAKE_CASE ( self : Dict ) -> Dict: for model_class_name in self.all_model_classes: A = model_class_name.from_pretrained('google/vit-base-patch16-224' ) A = model(np.ones((1, 3, 224, 224) ) ) self.assertIsNotNone(A_ )
74
"""simple docstring""" import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _lowercase = logging.get_logger(__name__) class lowerCAmelCase_ ( _lowercase ): '''simple docstring''' def __init__( self : Union[str, Any] ,*A_ : List[str] ,**A_ : int ) -> None: warnings.warn( 'The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use YolosImageProcessor instead.' ,A_ ,) super().__init__(*A_ ,**A_ )
74
1
"""simple docstring""" import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, ) from transformers.models.deberta.modeling_deberta import DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST class _lowerCAmelCase ( a ): """simple docstring""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase=1_3 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=9_9 , __UpperCAmelCase=3_2 , __UpperCAmelCase=5 , __UpperCAmelCase=4 , __UpperCAmelCase=3_7 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=5_1_2 , __UpperCAmelCase=1_6 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase="None" , __UpperCAmelCase=3 , __UpperCAmelCase=4 , __UpperCAmelCase=None , ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = parent lowerCAmelCase__ :int = batch_size lowerCAmelCase__ :List[str] = seq_length lowerCAmelCase__ :Tuple = is_training lowerCAmelCase__ :Tuple = use_input_mask lowerCAmelCase__ :Dict = use_token_type_ids lowerCAmelCase__ :Union[str, Any] = use_labels lowerCAmelCase__ :Tuple = vocab_size lowerCAmelCase__ :List[Any] = hidden_size lowerCAmelCase__ :Tuple = num_hidden_layers lowerCAmelCase__ :str = num_attention_heads lowerCAmelCase__ :List[str] = intermediate_size lowerCAmelCase__ :Optional[Any] = hidden_act lowerCAmelCase__ :Union[str, Any] = hidden_dropout_prob lowerCAmelCase__ :Any = attention_probs_dropout_prob lowerCAmelCase__ :Dict = max_position_embeddings lowerCAmelCase__ :Tuple = type_vocab_size lowerCAmelCase__ :List[str] = type_sequence_label_size lowerCAmelCase__ :Tuple = initializer_range lowerCAmelCase__ :Optional[Any] = num_labels lowerCAmelCase__ :int = num_choices lowerCAmelCase__ :Union[str, Any] = relative_attention lowerCAmelCase__ :int = position_biased_input lowerCAmelCase__ :Optional[int] = pos_att_type lowerCAmelCase__ :Dict = scope def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ :int = None if self.use_input_mask: lowerCAmelCase__ :int = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) lowerCAmelCase__ :Optional[Any] = None if self.use_token_type_ids: lowerCAmelCase__ :Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase__ :Dict = None lowerCAmelCase__ :Union[str, Any] = None lowerCAmelCase__ :Dict = None if self.use_labels: lowerCAmelCase__ :Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ :List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ :Optional[int] = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ :Dict = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case ( self ): '''simple docstring''' return DebertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[Any] = self.get_config() lowerCAmelCase__ :Optional[Any] = 3_0_0 return config def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' self.parent.assertListEqual(list(result.loss.size() ) , [] ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Dict = DebertaModel(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ :Any = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase )[0] lowerCAmelCase__ :List[Any] = model(__UpperCAmelCase , token_type_ids=__UpperCAmelCase )[0] lowerCAmelCase__ :Dict = model(__UpperCAmelCase )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :str = DebertaForMaskedLM(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ :List[str] = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Tuple = self.num_labels lowerCAmelCase__ :int = DebertaForSequenceClassification(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ :Optional[int] = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(__UpperCAmelCase ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = self.num_labels lowerCAmelCase__ :Any = DebertaForTokenClassification(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ :List[Any] = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :int = DebertaForQuestionAnswering(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ :str = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , start_positions=__UpperCAmelCase , end_positions=__UpperCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = self.prepare_config_and_inputs() ( lowerCAmelCase__ ) :Tuple = config_and_inputs lowerCAmelCase__ :int = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class _lowerCAmelCase ( a , a , unittest.TestCase ): """simple docstring""" __magic_name__ :List[str] = ( ( DebertaModel, DebertaForMaskedLM, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaForQuestionAnswering, ) if is_torch_available() else () ) __magic_name__ :Optional[Any] = ( { """feature-extraction""": DebertaModel, """fill-mask""": DebertaForMaskedLM, """question-answering""": DebertaForQuestionAnswering, """text-classification""": DebertaForSequenceClassification, """token-classification""": DebertaForTokenClassification, """zero-shot""": DebertaForSequenceClassification, } if is_torch_available() else {} ) __magic_name__ :Tuple = True __magic_name__ :List[Any] = False __magic_name__ :Optional[Any] = False __magic_name__ :str = False __magic_name__ :int = False def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[str] = DebertaModelTester(self ) lowerCAmelCase__ :List[Any] = ConfigTester(self , config_class=__UpperCAmelCase , hidden_size=3_7 ) def snake_case ( self ): '''simple docstring''' self.config_tester.run_common_tests() def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*__UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*__UpperCAmelCase ) @slow def snake_case ( self ): '''simple docstring''' for model_name in DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ :int = DebertaModel.from_pretrained(__UpperCAmelCase ) self.assertIsNotNone(__UpperCAmelCase ) @require_torch @require_sentencepiece @require_tokenizers class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @unittest.skip(reason='Model not available yet' ) def snake_case ( self ): '''simple docstring''' pass @slow def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :str = DebertaModel.from_pretrained('microsoft/deberta-base' ) lowerCAmelCase__ :str = torch.tensor([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] ) lowerCAmelCase__ :Tuple = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): lowerCAmelCase__ :int = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase )[0] # compare the actual values for a slice. lowerCAmelCase__ :str = torch.tensor( [[[-0.59_86, -0.80_55, -0.84_62], [1.44_84, -0.93_48, -0.80_59], [0.31_23, 0.00_32, -1.41_31]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __UpperCAmelCase , atol=1E-4 ) , F"{output[:, 1:4, 1:4]}" )
370
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer __A = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast __A = TaTokenizerFast __A = {"""configuration_mt5""": ["""MT5Config""", """MT5OnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ """MT5EncoderModel""", """MT5ForConditionalGeneration""", """MT5ForQuestionAnswering""", """MT5Model""", """MT5PreTrainedModel""", """MT5Stack""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = ["""TFMT5EncoderModel""", """TFMT5ForConditionalGeneration""", """TFMT5Model"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = ["""FlaxMT5EncoderModel""", """FlaxMT5ForConditionalGeneration""", """FlaxMT5Model"""] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys __A = _LazyModule( __name__, globals()["""__file__"""], _import_structure, extra_objects={"""MT5Tokenizer""": MTaTokenizer, """MT5TokenizerFast""": MTaTokenizerFast}, module_spec=__spec__, )
254
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available lowerCAmelCase__ = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['GPTSw3Tokenizer'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
11
"""simple docstring""" import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, ) from transformers.models.deberta.modeling_deberta import DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCAmelCase_ ( _lowercase): def __init__( self : Tuple , __UpperCamelCase : Optional[Any] , __UpperCamelCase : List[str]=13 , __UpperCamelCase : str=7 , __UpperCamelCase : List[Any]=True , __UpperCamelCase : List[str]=True , __UpperCamelCase : int=True , __UpperCamelCase : Optional[int]=True , __UpperCamelCase : List[Any]=99 , __UpperCamelCase : Dict=32 , __UpperCamelCase : int=5 , __UpperCamelCase : str=4 , __UpperCamelCase : Any=37 , __UpperCamelCase : Tuple="gelu" , __UpperCamelCase : List[str]=0.1 , __UpperCamelCase : Any=0.1 , __UpperCamelCase : Any=512 , __UpperCamelCase : List[str]=16 , __UpperCamelCase : Optional[Any]=2 , __UpperCamelCase : List[str]=0.0_2 , __UpperCamelCase : str=False , __UpperCamelCase : Dict=True , __UpperCamelCase : Tuple="None" , __UpperCamelCase : Dict=3 , __UpperCamelCase : Dict=4 , __UpperCamelCase : Any=None , ) -> Tuple: _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_input_mask _UpperCamelCase = use_token_type_ids _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = type_sequence_label_size _UpperCamelCase = initializer_range _UpperCamelCase = num_labels _UpperCamelCase = num_choices _UpperCamelCase = relative_attention _UpperCamelCase = position_biased_input _UpperCamelCase = pos_att_type _UpperCamelCase = scope def _UpperCamelCase ( self : Union[str, Any] ) -> Union[str, Any]: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCamelCase = None if self.use_input_mask: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _UpperCamelCase = None if self.use_token_type_ids: _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _UpperCamelCase = ids_tensor([self.batch_size] , self.num_choices ) _UpperCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCamelCase ( self : Optional[int] ) -> Optional[Any]: return DebertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def _UpperCamelCase ( self : Optional[int] ) -> List[Any]: _UpperCamelCase = self.get_config() _UpperCamelCase = 300 return config def _UpperCamelCase ( self : int , __UpperCamelCase : List[Any] ) -> str: self.parent.assertListEqual(list(result.loss.size() ) , [] ) def _UpperCamelCase ( self : Any , __UpperCamelCase : Any , __UpperCamelCase : Optional[int] , __UpperCamelCase : List[Any] , __UpperCamelCase : Optional[int] , __UpperCamelCase : Tuple , __UpperCamelCase : List[Any] , __UpperCamelCase : Optional[Any] ) -> List[str]: _UpperCamelCase = DebertaModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() _UpperCamelCase = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase )[0] _UpperCamelCase = model(__UpperCamelCase , token_type_ids=__UpperCamelCase )[0] _UpperCamelCase = model(__UpperCamelCase )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def _UpperCamelCase ( self : Tuple , __UpperCamelCase : Dict , __UpperCamelCase : List[str] , __UpperCamelCase : Optional[int] , __UpperCamelCase : List[Any] , __UpperCamelCase : Any , __UpperCamelCase : str , __UpperCamelCase : Optional[int] ) -> Tuple: _UpperCamelCase = DebertaForMaskedLM(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() _UpperCamelCase = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : Optional[int] , __UpperCamelCase : str , __UpperCamelCase : Any , __UpperCamelCase : List[Any] , __UpperCamelCase : int , __UpperCamelCase : Optional[Any] , __UpperCamelCase : str , __UpperCamelCase : Tuple ) -> List[Any]: _UpperCamelCase = self.num_labels _UpperCamelCase = DebertaForSequenceClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() _UpperCamelCase = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(__UpperCamelCase ) def _UpperCamelCase ( self : Optional[Any] , __UpperCamelCase : Optional[Any] , __UpperCamelCase : int , __UpperCamelCase : List[Any] , __UpperCamelCase : Dict , __UpperCamelCase : List[Any] , __UpperCamelCase : Optional[int] , __UpperCamelCase : str ) -> Dict: _UpperCamelCase = self.num_labels _UpperCamelCase = DebertaForTokenClassification(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() _UpperCamelCase = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : List[Any] , __UpperCamelCase : str , __UpperCamelCase : List[str] , __UpperCamelCase : Union[str, Any] , __UpperCamelCase : int , __UpperCamelCase : str , __UpperCamelCase : List[Any] , __UpperCamelCase : List[Any] ) -> List[Any]: _UpperCamelCase = DebertaForQuestionAnswering(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() _UpperCamelCase = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , start_positions=__UpperCamelCase , end_positions=__UpperCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _UpperCamelCase ( self : Any ) -> Union[str, Any]: _UpperCamelCase = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) = config_and_inputs _UpperCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class UpperCAmelCase_ ( _lowercase , _lowercase , unittest.TestCase): snake_case__ = ( ( DebertaModel, DebertaForMaskedLM, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaForQuestionAnswering, ) if is_torch_available() else () ) snake_case__ = ( { '''feature-extraction''': DebertaModel, '''fill-mask''': DebertaForMaskedLM, '''question-answering''': DebertaForQuestionAnswering, '''text-classification''': DebertaForSequenceClassification, '''token-classification''': DebertaForTokenClassification, '''zero-shot''': DebertaForSequenceClassification, } if is_torch_available() else {} ) snake_case__ = True snake_case__ = False snake_case__ = False snake_case__ = False snake_case__ = False def _UpperCamelCase ( self : Union[str, Any] ) -> Tuple: _UpperCamelCase = DebertaModelTester(self ) _UpperCamelCase = ConfigTester(self , config_class=__UpperCamelCase , hidden_size=37 ) def _UpperCamelCase ( self : Optional[int] ) -> int: self.config_tester.run_common_tests() def _UpperCamelCase ( self : Any ) -> List[str]: _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*__UpperCamelCase ) def _UpperCamelCase ( self : Optional[Any] ) -> Optional[int]: _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*__UpperCamelCase ) def _UpperCamelCase ( self : List[str] ) -> Union[str, Any]: _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*__UpperCamelCase ) def _UpperCamelCase ( self : Optional[int] ) -> Union[str, Any]: _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*__UpperCamelCase ) def _UpperCamelCase ( self : Dict ) -> Tuple: _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*__UpperCamelCase ) @slow def _UpperCamelCase ( self : Any ) -> Optional[Any]: for model_name in DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase = DebertaModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) @require_torch @require_sentencepiece @require_tokenizers class UpperCAmelCase_ ( unittest.TestCase): @unittest.skip(reason='''Model not available yet''' ) def _UpperCamelCase ( self : Tuple ) -> Union[str, Any]: pass @slow def _UpperCamelCase ( self : Tuple ) -> Union[str, Any]: _UpperCamelCase = DebertaModel.from_pretrained('''microsoft/deberta-base''' ) _UpperCamelCase = torch.tensor([[0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2]] ) _UpperCamelCase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _UpperCamelCase = model(__UpperCamelCase , attention_mask=__UpperCamelCase )[0] # compare the actual values for a slice. _UpperCamelCase = torch.tensor( [[[-0.5_9_8_6, -0.8_0_5_5, -0.8_4_6_2], [1.4_4_8_4, -0.9_3_4_8, -0.8_0_5_9], [0.3_1_2_3, 0.0_0_3_2, -1.4_1_3_1]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __UpperCamelCase , atol=1E-4 ) , F'''{output[:, 1:4, 1:4]}''' )
256
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __A : int = { 'configuration_whisper': ['WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'WhisperConfig', 'WhisperOnnxConfig'], 'feature_extraction_whisper': ['WhisperFeatureExtractor'], 'processing_whisper': ['WhisperProcessor'], 'tokenization_whisper': ['WhisperTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Union[str, Any] = ['WhisperTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Any = [ 'WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'WhisperForConditionalGeneration', 'WhisperModel', 'WhisperPreTrainedModel', 'WhisperForAudioClassification', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[Any] = [ 'TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFWhisperForConditionalGeneration', 'TFWhisperModel', 'TFWhisperPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Union[str, Any] = [ 'FlaxWhisperForConditionalGeneration', 'FlaxWhisperModel', 'FlaxWhisperPreTrainedModel', 'FlaxWhisperForAudioClassification', ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys __A : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
8
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __A : int = { 'configuration_whisper': ['WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'WhisperConfig', 'WhisperOnnxConfig'], 'feature_extraction_whisper': ['WhisperFeatureExtractor'], 'processing_whisper': ['WhisperProcessor'], 'tokenization_whisper': ['WhisperTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Union[str, Any] = ['WhisperTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Any = [ 'WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'WhisperForConditionalGeneration', 'WhisperModel', 'WhisperPreTrainedModel', 'WhisperForAudioClassification', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[Any] = [ 'TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFWhisperForConditionalGeneration', 'TFWhisperModel', 'TFWhisperPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Union[str, Any] = [ 'FlaxWhisperForConditionalGeneration', 'FlaxWhisperModel', 'FlaxWhisperPreTrainedModel', 'FlaxWhisperForAudioClassification', ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys __A : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
8
1
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import MvpTokenizer __lowercase = logging.get_logger(__name__) __lowercase = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} # See all MVP models at https://huggingface.co/models?filter=mvp __lowercase = { '''vocab_file''': { '''RUCAIBox/mvp''': '''https://huggingface.co/RUCAIBox/mvp/resolve/main/vocab.json''', }, '''added_tokens.json''': { '''RUCAIBox/mvp''': '''https://huggingface.co/RUCAIBox/mvp/resolve/main/added_tokens.json''', }, '''merges_file''': { '''RUCAIBox/mvp''': '''https://huggingface.co/RUCAIBox/mvp/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''RUCAIBox/mvp''': '''https://huggingface.co/RUCAIBox/mvp/resolve/main/tokenizer.json''', }, } __lowercase = { '''RUCAIBox/mvp''': 1_0_2_4, } class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : Optional[int] = VOCAB_FILES_NAMES UpperCAmelCase_ : Tuple = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase_ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase_ : Tuple = ['''input_ids''', '''attention_mask'''] UpperCAmelCase_ : int = MvpTokenizer def __init__( self , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase="replace" , __lowerCAmelCase="<s>" , __lowerCAmelCase="</s>" , __lowerCAmelCase="</s>" , __lowerCAmelCase="<s>" , __lowerCAmelCase="<unk>" , __lowerCAmelCase="<pad>" , __lowerCAmelCase="<mask>" , __lowerCAmelCase=False , __lowerCAmelCase=True , **__lowerCAmelCase , ): """simple docstring""" super().__init__( __lowerCAmelCase , __lowerCAmelCase , tokenizer_file=__lowerCAmelCase , errors=__lowerCAmelCase , bos_token=__lowerCAmelCase , eos_token=__lowerCAmelCase , sep_token=__lowerCAmelCase , cls_token=__lowerCAmelCase , unk_token=__lowerCAmelCase , pad_token=__lowerCAmelCase , mask_token=__lowerCAmelCase , add_prefix_space=__lowerCAmelCase , trim_offsets=__lowerCAmelCase , **__lowerCAmelCase , ) lowerCAmelCase = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get("""add_prefix_space""" , __lowerCAmelCase) != add_prefix_space: lowerCAmelCase = getattr(__lowerCAmelCase , pre_tok_state.pop("""type""")) lowerCAmelCase = add_prefix_space lowerCAmelCase = pre_tok_class(**__lowerCAmelCase) lowerCAmelCase = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowerCAmelCase = """post_processor""" lowerCAmelCase = getattr(self.backend_tokenizer , __lowerCAmelCase , __lowerCAmelCase) if tokenizer_component_instance: lowerCAmelCase = json.loads(tokenizer_component_instance.__getstate__()) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowerCAmelCase = tuple(state["""sep"""]) if "cls" in state: lowerCAmelCase = tuple(state["""cls"""]) lowerCAmelCase = False if state.get("""add_prefix_space""" , __lowerCAmelCase) != add_prefix_space: lowerCAmelCase = add_prefix_space lowerCAmelCase = True if state.get("""trim_offsets""" , __lowerCAmelCase) != trim_offsets: lowerCAmelCase = trim_offsets lowerCAmelCase = True if changes_to_apply: lowerCAmelCase = getattr(__lowerCAmelCase , state.pop("""type""")) lowerCAmelCase = component_class(**__lowerCAmelCase) setattr(self.backend_tokenizer , __lowerCAmelCase , __lowerCAmelCase) @property def a_ ( self): """simple docstring""" if self._mask_token is None: if self.verbose: logger.error("""Using mask_token, but it is not set yet.""") return None return str(self._mask_token) @mask_token.setter def a_ ( self , __lowerCAmelCase): """simple docstring""" lowerCAmelCase = AddedToken(__lowerCAmelCase , lstrip=__lowerCAmelCase , rstrip=__lowerCAmelCase) if isinstance(__lowerCAmelCase , __lowerCAmelCase) else value lowerCAmelCase = value def a_ ( self , *__lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" lowerCAmelCase = kwargs.get("""is_split_into_words""" , __lowerCAmelCase) if is_split_into_words and not self.add_prefix_space: raise ValueError( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " """to use it with pretokenized inputs.""") return super()._batch_encode_plus(*__lowerCAmelCase , **__lowerCAmelCase) def a_ ( self , *__lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" lowerCAmelCase = kwargs.get("""is_split_into_words""" , __lowerCAmelCase) if is_split_into_words and not self.add_prefix_space: raise ValueError( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " """to use it with pretokenized inputs.""") return super()._encode_plus(*__lowerCAmelCase , **__lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase = None): """simple docstring""" lowerCAmelCase = self._tokenizer.model.save(__lowerCAmelCase , name=__lowerCAmelCase) return tuple(__lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase=None): """simple docstring""" lowerCAmelCase = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def a_ ( self , __lowerCAmelCase , __lowerCAmelCase = None): """simple docstring""" lowerCAmelCase = [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0]
272
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowercase = { '''configuration_nezha''': ['''NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''NezhaConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''NezhaForNextSentencePrediction''', '''NezhaForMaskedLM''', '''NezhaForPreTraining''', '''NezhaForMultipleChoice''', '''NezhaForQuestionAnswering''', '''NezhaForSequenceClassification''', '''NezhaForTokenClassification''', '''NezhaModel''', '''NezhaPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
272
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __a = logging.get_logger(__name__) __a = '▁' __a = {'vocab_file': 'sentencepiece.bpe.model'} __a = { 'vocab_file': { 'xlm-roberta-base': 'https://huggingface.co/xlm-roberta-base/resolve/main/sentencepiece.bpe.model', 'xlm-roberta-large': 'https://huggingface.co/xlm-roberta-large/resolve/main/sentencepiece.bpe.model', 'xlm-roberta-large-finetuned-conll02-dutch': ( 'https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/sentencepiece.bpe.model' ), 'xlm-roberta-large-finetuned-conll02-spanish': ( 'https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/sentencepiece.bpe.model' ), 'xlm-roberta-large-finetuned-conll03-english': ( 'https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/sentencepiece.bpe.model' ), 'xlm-roberta-large-finetuned-conll03-german': ( 'https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/sentencepiece.bpe.model' ), } } __a = { 'xlm-roberta-base': 512, 'xlm-roberta-large': 512, 'xlm-roberta-large-finetuned-conll02-dutch': 512, 'xlm-roberta-large-finetuned-conll02-spanish': 512, 'xlm-roberta-large-finetuned-conll03-english': 512, 'xlm-roberta-large-finetuned-conll03-german': 512, } class __a( lowercase__ ): """simple docstring""" lowerCAmelCase = VOCAB_FILES_NAMES lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase = ['''input_ids''', '''attention_mask'''] def __init__( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE="<s>" ,_SCREAMING_SNAKE_CASE="</s>" ,_SCREAMING_SNAKE_CASE="</s>" ,_SCREAMING_SNAKE_CASE="<s>" ,_SCREAMING_SNAKE_CASE="<unk>" ,_SCREAMING_SNAKE_CASE="<pad>" ,_SCREAMING_SNAKE_CASE="<mask>" ,_SCREAMING_SNAKE_CASE = None ,**_SCREAMING_SNAKE_CASE ,) -> Dict: UpperCAmelCase_ : Optional[int] = AddedToken(_a ,lstrip=_a ,rstrip=_a ) if isinstance(_a ,_a ) else mask_token UpperCAmelCase_ : Dict = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_a ,eos_token=_a ,unk_token=_a ,sep_token=_a ,cls_token=_a ,pad_token=_a ,mask_token=_a ,sp_model_kwargs=self.sp_model_kwargs ,**_a ,) UpperCAmelCase_ : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_a ) ) UpperCAmelCase_ : Optional[int] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token UpperCAmelCase_ : List[str] = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab UpperCAmelCase_ : List[str] = 1 UpperCAmelCase_ : Tuple = len(self.sp_model ) + self.fairseq_offset UpperCAmelCase_ : Dict = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ) -> Union[str, Any]: UpperCAmelCase_ : List[Any] = self.__dict__.copy() UpperCAmelCase_ : Optional[Any] = None UpperCAmelCase_ : Dict = self.sp_model.serialized_model_proto() return state def __setstate__( self ,_SCREAMING_SNAKE_CASE ) -> Dict: UpperCAmelCase_ : Optional[Any] = d # for backward compatibility if not hasattr(self ,'''sp_model_kwargs''' ): UpperCAmelCase_ : str = {} UpperCAmelCase_ : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ) -> Tuple: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase_ : List[str] = [self.cls_token_id] UpperCAmelCase_ : Any = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = False ) -> Optional[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a ,token_ids_a=_a ,already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ) -> str: UpperCAmelCase_ : List[Any] = [self.sep_token_id] UpperCAmelCase_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def a__ ( self ) -> List[Any]: return len(self.sp_model ) + self.fairseq_offset + 1 # Add the <mask> token def a__ ( self ) -> Union[str, Any]: UpperCAmelCase_ : Union[str, Any] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def a__ ( self ,_SCREAMING_SNAKE_CASE ) -> int: return self.sp_model.encode(_a ,out_type=_a ) def a__ ( self ,_SCREAMING_SNAKE_CASE ) -> Optional[int]: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCAmelCase_ : Optional[Any] = self.sp_model.PieceToId(_a ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def a__ ( self ,_SCREAMING_SNAKE_CASE ) -> int: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def a__ ( self ,_SCREAMING_SNAKE_CASE ) -> Optional[Any]: UpperCAmelCase_ : Dict = ''.join(_a ).replace(_a ,''' ''' ).strip() return out_string def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ) -> int: if not os.path.isdir(_a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase_ : Any = os.path.join( _a ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,_a ) elif not os.path.isfile(self.vocab_file ): with open(_a ,'''wb''' ) as fi: UpperCAmelCase_ : Union[str, Any] = self.sp_model.serialized_model_proto() fi.write(_a ) return (out_vocab_file,)
355
import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging __a = logging.get_logger(__name__) def lowerCamelCase__ ( _lowercase ): '''simple docstring''' UpperCAmelCase_ : str = r'''\w+[.]\d+''' UpperCAmelCase_ : int = re.findall(_lowercase , _lowercase ) for pat in pats: UpperCAmelCase_ : List[Any] = key.replace(_lowercase , '''_'''.join(pat.split('''.''' ) ) ) return key def lowerCamelCase__ ( _lowercase , _lowercase , _lowercase ): '''simple docstring''' UpperCAmelCase_ : Dict = pt_tuple_key[:-1] + ('''scale''',) if ( any('''norm''' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): UpperCAmelCase_ : List[Any] = pt_tuple_key[:-1] + ('''scale''',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: UpperCAmelCase_ : Any = pt_tuple_key[:-1] + ('''scale''',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: UpperCAmelCase_ : Tuple = pt_tuple_key[:-1] + ('''embedding''',) return renamed_pt_tuple_key, pt_tensor # conv layer UpperCAmelCase_ : List[Any] = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: UpperCAmelCase_ : List[Any] = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer UpperCAmelCase_ : Dict = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight": UpperCAmelCase_ : Tuple = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight UpperCAmelCase_ : Any = pt_tuple_key[:-1] + ('''weight''',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias UpperCAmelCase_ : Tuple = pt_tuple_key[:-1] + ('''bias''',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def lowerCamelCase__ ( _lowercase , _lowercase , _lowercase=42 ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params UpperCAmelCase_ : str = flax_model.init_weights(PRNGKey(_lowercase ) ) UpperCAmelCase_ : List[Any] = flatten_dict(_lowercase ) UpperCAmelCase_ : int = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): UpperCAmelCase_ : Optional[int] = rename_key(_lowercase ) UpperCAmelCase_ : List[str] = tuple(renamed_pt_key.split('''.''' ) ) # Correctly rename weight parameters UpperCAmelCase_, UpperCAmelCase_ : Any = rename_key_and_reshape_tensor(_lowercase , _lowercase , _lowercase ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f'''PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ''' f'''{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.''' ) # also add unexpected weight so that warning is thrown UpperCAmelCase_ : int = jnp.asarray(_lowercase ) return unflatten_dict(_lowercase )
235
0
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class lowercase ( unittest.TestCase ): @slow def _snake_case ( self ) -> Optional[Any]: lowerCAmelCase = FlaxMTaForConditionalGeneration.from_pretrained("""google/mt5-small""" ) lowerCAmelCase = AutoTokenizer.from_pretrained("""google/mt5-small""" ) lowerCAmelCase = tokenizer("""Hello there""" , return_tensors="""np""" ).input_ids lowerCAmelCase = tokenizer("""Hi I am""" , return_tensors="""np""" ).input_ids lowerCAmelCase = shift_tokens_right(lowercase , model.config.pad_token_id , model.config.decoder_start_token_id ) lowerCAmelCase = model(lowercase , decoder_input_ids=lowercase ).logits lowerCAmelCase = optax.softmax_cross_entropy(lowercase , onehot(lowercase , logits.shape[-1] ) ).mean() lowerCAmelCase = -(labels.shape[-1] * loss.item()) lowerCAmelCase = -84.9_127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1e-4 )
46
"""simple docstring""" from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool from google.protobuf import symbol_database as _symbol_database from google.protobuf.internal import builder as _builder # @@protoc_insertion_point(imports) SCREAMING_SNAKE_CASE__ = _symbol_database.Default() SCREAMING_SNAKE_CASE__ = _descriptor_pool.Default().AddSerializedFile( b"\n\x19sentencepiece_model.proto\x12\rsentencepiece\"\x80\x0c\n\x0bTrainerSpec\x12\r\n\x05input\x18\x01 \x03(\t\x12\x14\n\x0cinput_format\x18\x07 \x01(\t\x12\x14\n\x0cmodel_prefix\x18\x02 \x01(\t\x12\x41\n\nmodel_type\x18\x03 \x01(\x0e\x32$.sentencepiece.TrainerSpec.ModelType:\x07UNIGRAM\x12\x18\n\nvocab_size\x18\x04 \x01(\x05:\x04\x38\x30\x30\x30\x12\x17\n\x0f\x61\x63\x63\x65pt_language\x18\x05 \x03(\t\x12 \n\x15self_test_sample_size\x18\x06 \x01(\x05:\x01\x30\x12*\n\x1b\x65nable_differential_privacy\x18\x32 \x01(\x08:\x05\x66\x61lse\x12+\n differential_privacy_noise_level\x18\x33 \x01(\x02:\x01\x30\x12\x32\n\'differential_privacy_clipping_threshold\x18\x34 \x01(\x04:\x01\x30\x12\"\n\x12\x63haracter_coverage\x18\n \x01(\x02:\x06\x30.9995\x12\x1e\n\x13input_sentence_size\x18\x0b \x01(\x04:\x01\x30\x12$\n\x16shuffle_input_sentence\x18\x13 \x01(\x08:\x04true\x12 \n\x14mining_sentence_size\x18\x0c \x01(\x05\x42\x02\x18\x01\x12\"\n\x16training_sentence_size\x18\r \x01(\x05\x42\x02\x18\x01\x12(\n\x17seed_sentencepiece_size\x18\x0e \x01(\x05:\x07\x31\x30\x30\x30\x30\x30\x30\x12\x1e\n\x10shrinking_factor\x18\x0f \x01(\x02:\x04\x30.75\x12!\n\x13max_sentence_length\x18\x12 \x01(\x05:\x04\x34\x31\x39\x32\x12\x17\n\x0bnum_threads\x18\x10 \x01(\x05:\x02\x31\x36\x12\x1d\n\x12num_sub_iterations\x18\x11 \x01(\x05:\x01\x32\x12$\n\x18max_sentencepiece_length\x18\x14 \x01(\x05:\x02\x31\x36\x12%\n\x17split_by_unicode_script\x18\x15 \x01(\x08:\x04true\x12\x1d\n\x0fsplit_by_number\x18\x17 \x01(\x08:\x04true\x12!\n\x13split_by_whitespace\x18\x16 \x01(\x08:\x04true\x12)\n\x1atreat_whitespace_as_suffix\x18\x18 \x01(\x08:\x05\x66\x61lse\x12+\n\x1c\x61llow_whitespace_only_pieces\x18\x1a \x01(\x08:\x05\x66\x61lse\x12\x1b\n\x0csplit_digits\x18\x19 \x01(\x08:\x05\x66\x61lse\x12#\n\x19pretokenization_delimiter\x18\x35 \x01(\t:\x00\x12\x17\n\x0f\x63ontrol_symbols\x18\x1e \x03(\t\x12\x1c\n\x14user_defined_symbols\x18\x1f \x03(\t\x12\x16\n\x0erequired_chars\x18$ \x01(\t\x12\x1c\n\rbyte_fallback\x18# \x01(\x08:\x05\x66\x61lse\x12+\n\x1dvocabulary_output_piece_score\x18 \x01(\x08:\x04true\x12\x1e\n\x10hard_vocab_limit\x18! \x01(\x08:\x04true\x12\x1c\n\ruse_all_vocab\x18\" \x01(\x08:\x05\x66\x61lse\x12\x11\n\x06unk_id\x18( \x01(\x05:\x01\x30\x12\x11\n\x06\x62os_id\x18) \x01(\x05:\x01\x31\x12\x11\n\x06\x65os_id\x18* \x01(\x05:\x01\x32\x12\x12\n\x06pad_id\x18+ \x01(\x05:\x02-1\x12\x18\n\tunk_piece\x18- \x01(\t:\x05<unk>\x12\x16\n\tbos_piece\x18. \x01(\t:\x03<s>\x12\x17\n\teos_piece\x18/ \x01(\t:\x04</s>\x12\x18\n\tpad_piece\x18\x30 \x01(\t:\x05<pad>\x12\x1a\n\x0bunk_surface\x18, \x01(\t:\x05 \xe2\x81\x87 \x12+\n\x1ctrain_extremely_large_corpus\x18\x31 \x01(\x08:\x05\x66\x61lse\"5\n\tModelType\x12\x0b\n\x07UNIGRAM\x10\x01\x12\x07\n\x03\x42PE\x10\x02\x12\x08\n\x04WORD\x10\x03\x12\x08\n\x04\x43HAR\x10\x04*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\"\xd1\x01\n\x0eNormalizerSpec\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x1c\n\x14precompiled_charsmap\x18\x02 \x01(\x0c\x12\x1e\n\x10\x61\x64\x64_dummy_prefix\x18\x03 \x01(\x08:\x04true\x12&\n\x18remove_extra_whitespaces\x18\x04 \x01(\x08:\x04true\x12 \n\x12\x65scape_whitespaces\x18\x05 \x01(\x08:\x04true\x12\x1e\n\x16normalization_rule_tsv\x18\x06 \x01(\t*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\"y\n\x0cSelfTestData\x12\x33\n\x07samples\x18\x01 \x03(\x0b\x32\".sentencepiece.SelfTestData.Sample\x1a)\n\x06Sample\x12\r\n\x05input\x18\x01 \x01(\t\x12\x10\n\x08\x65xpected\x18\x02 \x01(\t*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\"\xfe\x03\n\nModelProto\x12\x37\n\x06pieces\x18\x01 \x03(\x0b\x32\'.sentencepiece.ModelProto.SentencePiece\x12\x30\n\x0ctrainer_spec\x18\x02 \x01(\x0b\x32\x1a.sentencepiece.TrainerSpec\x12\x36\n\x0fnormalizer_spec\x18\x03 \x01(\x0b\x32\x1d.sentencepiece.NormalizerSpec\x12\x33\n\x0eself_test_data\x18\x04 \x01(\x0b\x32\x1b.sentencepiece.SelfTestData\x12\x38\n\x11\x64\x65normalizer_spec\x18\x05 \x01(\x0b\x32\x1d.sentencepiece.NormalizerSpec\x1a\xd2\x01\n\rSentencePiece\x12\r\n\x05piece\x18\x01 \x01(\t\x12\r\n\x05score\x18\x02 \x01(\x02\x12\x42\n\x04type\x18\x03 \x01(\x0e\x32,.sentencepiece.ModelProto.SentencePiece.Type:\x06NORMAL\"T\n\x04Type\x12\n\n\x06NORMAL\x10\x01\x12\x0b\n\x07UNKNOWN\x10\x02\x12\x0b\n\x07\x43ONTROL\x10\x03\x12\x10\n\x0cUSER_DEFINED\x10\x04\x12\x08\n\x04\x42YTE\x10\x06\x12\n\n\x06UNUSED\x10\x05*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\x42\x02H\x03" ) SCREAMING_SNAKE_CASE__ = globals() _builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals) _builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, "sentencepiece_model_pb2", _globals) if _descriptor._USE_C_DESCRIPTORS is False: SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = b"H\003" # (generated by protobuf compiler, but `_TRAINERSPEC` is not defined) # _TRAINERSPEC.fields_by_name["mining_sentence_size"]._options = None # _TRAINERSPEC.fields_by_name["mining_sentence_size"]._serialized_options = b"\030\001" # _TRAINERSPEC.fields_by_name["training_sentence_size"]._options = None # _TRAINERSPEC.fields_by_name["training_sentence_size"]._serialized_options = b"\030\001" SCREAMING_SNAKE_CASE__ = 45 SCREAMING_SNAKE_CASE__ = 1_581 SCREAMING_SNAKE_CASE__ = 1_517 SCREAMING_SNAKE_CASE__ = 1_570 SCREAMING_SNAKE_CASE__ = 1_584 SCREAMING_SNAKE_CASE__ = 1_793 SCREAMING_SNAKE_CASE__ = 1_795 SCREAMING_SNAKE_CASE__ = 1_916 SCREAMING_SNAKE_CASE__ = 1_864 SCREAMING_SNAKE_CASE__ = 1_905 SCREAMING_SNAKE_CASE__ = 1_919 SCREAMING_SNAKE_CASE__ = 2_429 SCREAMING_SNAKE_CASE__ = 2_208 SCREAMING_SNAKE_CASE__ = 2_418 SCREAMING_SNAKE_CASE__ = 2_323 SCREAMING_SNAKE_CASE__ = 2_407 # @@protoc_insertion_point(module_scope)
150
0
import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class lowerCAmelCase : '''simple docstring''' def __init__( self : Tuple , __a : List[str] , __a : str=2 , __a : Tuple=3 , __a : Optional[Any]=4 , __a : Optional[Any]=2 , __a : List[str]=7 , __a : Any=True , __a : str=True , __a : str=True , __a : Optional[int]=True , __a : Optional[int]=99 , __a : Optional[int]=36 , __a : List[Any]=3 , __a : List[Any]=4 , __a : int=37 , __a : str="gelu" , __a : Optional[Any]=0.1 , __a : List[str]=0.1 , __a : Tuple=512 , __a : Tuple=16 , __a : Dict=2 , __a : Tuple=0.02 , __a : Optional[Any]=6 , __a : int=6 , __a : Optional[Any]=3 , __a : Optional[Any]=4 , __a : List[str]=None , __a : str=1000 , ) -> Tuple: """simple docstring""" __lowercase : Tuple = parent __lowercase : int = batch_size __lowercase : Dict = num_channels __lowercase : List[Any] = image_size __lowercase : Tuple = patch_size __lowercase : Optional[int] = text_seq_length __lowercase : Optional[Any] = is_training __lowercase : str = use_input_mask __lowercase : List[Any] = use_token_type_ids __lowercase : List[str] = use_labels __lowercase : int = vocab_size __lowercase : Dict = hidden_size __lowercase : Tuple = num_hidden_layers __lowercase : int = num_attention_heads __lowercase : Tuple = intermediate_size __lowercase : List[str] = hidden_act __lowercase : int = hidden_dropout_prob __lowercase : List[str] = attention_probs_dropout_prob __lowercase : int = max_position_embeddings __lowercase : Tuple = type_vocab_size __lowercase : Any = type_sequence_label_size __lowercase : List[Any] = initializer_range __lowercase : List[Any] = coordinate_size __lowercase : Any = shape_size __lowercase : Optional[Any] = num_labels __lowercase : Optional[Any] = num_choices __lowercase : str = scope __lowercase : Union[str, Any] = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) __lowercase : str = text_seq_length __lowercase : List[Any] = (image_size // patch_size) ** 2 + 1 __lowercase : Union[str, Any] = self.text_seq_length + self.image_seq_length def lowerCAmelCase ( self : Any ) -> int: """simple docstring""" __lowercase : Any = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) __lowercase : List[Any] = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: __lowercase : Optional[int] = bbox[i, j, 3] __lowercase : Union[str, Any] = bbox[i, j, 1] __lowercase : Dict = t if bbox[i, j, 2] < bbox[i, j, 0]: __lowercase : Dict = bbox[i, j, 2] __lowercase : List[str] = bbox[i, j, 0] __lowercase : Optional[Any] = t __lowercase : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase : Optional[int] = None if self.use_input_mask: __lowercase : List[str] = random_attention_mask([self.batch_size, self.text_seq_length] ) __lowercase : Any = None if self.use_token_type_ids: __lowercase : List[str] = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) __lowercase : List[Any] = None __lowercase : Dict = None if self.use_labels: __lowercase : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase : List[str] = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) __lowercase : Optional[Any] = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def lowerCAmelCase ( self : Dict , __a : int , __a : Optional[Any] , __a : Any , __a : List[Any] , __a : Optional[Any] , __a : Union[str, Any] , __a : int , __a : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" __lowercase : Dict = LayoutLMvaModel(config=__a ) model.to(__a ) model.eval() # text + image __lowercase : Tuple = model(__a , pixel_values=__a ) __lowercase : List[Any] = model( __a , bbox=__a , pixel_values=__a , attention_mask=__a , token_type_ids=__a ) __lowercase : Any = model(__a , bbox=__a , pixel_values=__a , token_type_ids=__a ) __lowercase : Tuple = model(__a , bbox=__a , pixel_values=__a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only __lowercase : int = model(__a ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only __lowercase : Tuple = model(pixel_values=__a ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def lowerCAmelCase ( self : Optional[Any] , __a : Union[str, Any] , __a : Dict , __a : Optional[int] , __a : int , __a : List[str] , __a : Dict , __a : int , __a : Dict ) -> Optional[int]: """simple docstring""" __lowercase : Tuple = self.num_labels __lowercase : List[Any] = LayoutLMvaForSequenceClassification(__a ) model.to(__a ) model.eval() __lowercase : Dict = model( __a , bbox=__a , pixel_values=__a , attention_mask=__a , token_type_ids=__a , labels=__a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase ( self : int , __a : str , __a : Union[str, Any] , __a : Tuple , __a : Tuple , __a : Dict , __a : Optional[int] , __a : Optional[Any] , __a : Any ) -> Dict: """simple docstring""" __lowercase : List[str] = self.num_labels __lowercase : int = LayoutLMvaForTokenClassification(config=__a ) model.to(__a ) model.eval() __lowercase : Union[str, Any] = model( __a , bbox=__a , pixel_values=__a , attention_mask=__a , token_type_ids=__a , labels=__a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def lowerCAmelCase ( self : str , __a : int , __a : List[str] , __a : Dict , __a : Any , __a : Any , __a : List[str] , __a : Tuple , __a : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase : Union[str, Any] = LayoutLMvaForQuestionAnswering(config=__a ) model.to(__a ) model.eval() __lowercase : Optional[int] = model( __a , bbox=__a , pixel_values=__a , attention_mask=__a , token_type_ids=__a , start_positions=__a , end_positions=__a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCAmelCase ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" __lowercase : Dict = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) : Any = config_and_inputs __lowercase : Union[str, Any] = { """input_ids""": input_ids, """bbox""": bbox, """pixel_values""": pixel_values, """token_type_ids""": token_type_ids, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class lowerCAmelCase ( __a , __a , unittest.TestCase ): '''simple docstring''' _A : Optional[int] = False _A : List[str] = False _A : Optional[Any] = False _A : Union[str, Any] = ( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) _A : Any = ( {'''document-question-answering''': LayoutLMvaForQuestionAnswering, '''feature-extraction''': LayoutLMvaModel} if is_torch_available() else {} ) def lowerCAmelCase ( self : Optional[Any] , __a : Optional[Any] , __a : Optional[int] , __a : List[Any] , __a : int , __a : List[str] ) -> Union[str, Any]: """simple docstring""" return True def lowerCAmelCase ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase : int = LayoutLMvaModelTester(self ) __lowercase : List[str] = ConfigTester(self , config_class=__a , hidden_size=37 ) def lowerCAmelCase ( self : Tuple , __a : Union[str, Any] , __a : List[Any] , __a : List[str]=False ) -> List[str]: """simple docstring""" __lowercase : Tuple = copy.deepcopy(__a ) if model_class in get_values(__a ): __lowercase : Optional[int] = { k: v.unsqueeze(1 ).expand(-1 , self.model_tester.num_choices , -1 ).contiguous() if isinstance(__a , torch.Tensor ) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(__a ): __lowercase : List[str] = torch.ones(self.model_tester.batch_size , dtype=torch.long , device=__a ) elif model_class in get_values(__a ): __lowercase : Optional[Any] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__a ) __lowercase : List[str] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__a ) elif model_class in [ *get_values(__a ), ]: __lowercase : Dict = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__a ) elif model_class in [ *get_values(__a ), ]: __lowercase : Any = torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=torch.long , device=__a , ) return inputs_dict def lowerCAmelCase ( self : Dict ) -> Optional[int]: """simple docstring""" self.config_tester.run_common_tests() def lowerCAmelCase ( self : int ) -> List[Any]: """simple docstring""" __lowercase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a ) def lowerCAmelCase ( self : Any ) -> int: """simple docstring""" __lowercase : Optional[int] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __lowercase : Tuple = type self.model_tester.create_and_check_model(*__a ) def lowerCAmelCase ( self : Dict ) -> int: """simple docstring""" __lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__a ) def lowerCAmelCase ( self : Union[str, Any] ) -> Any: """simple docstring""" __lowercase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__a ) def lowerCAmelCase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" __lowercase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__a ) @slow def lowerCAmelCase ( self : Dict ) -> List[str]: """simple docstring""" for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase : Union[str, Any] = LayoutLMvaModel.from_pretrained(__a ) self.assertIsNotNone(__a ) def snake_case_ ( ): __lowercase : Dict = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch class lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def lowerCAmelCase ( self : int ) -> List[Any]: """simple docstring""" return LayoutLMvaImageProcessor(apply_ocr=__a ) if is_vision_available() else None @slow def lowerCAmelCase ( self : List[str] ) -> List[Any]: """simple docstring""" __lowercase : Dict = LayoutLMvaModel.from_pretrained("""microsoft/layoutlmv3-base""" ).to(__a ) __lowercase : int = self.default_image_processor __lowercase : Dict = prepare_img() __lowercase : List[str] = image_processor(images=__a , return_tensors="""pt""" ).pixel_values.to(__a ) __lowercase : Any = torch.tensor([[1, 2]] ) __lowercase : Any = torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]] ).unsqueeze(0 ) # forward pass __lowercase : Union[str, Any] = model( input_ids=input_ids.to(__a ) , bbox=bbox.to(__a ) , pixel_values=pixel_values.to(__a ) , ) # verify the logits __lowercase : Any = torch.Size((1, 199, 768) ) self.assertEqual(outputs.last_hidden_state.shape , __a ) __lowercase : Optional[Any] = torch.tensor( [[-0.0529, 0.3618, 0.1632], [-0.1587, -0.1667, -0.0400], [-0.1557, -0.1671, -0.0505]] ).to(__a ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , __a , atol=1E-4 ) )
306
import argparse import logging import os import time import timeit import datasets import numpy as np import pycuda.autoinit # noqa: F401 import pycuda.driver as cuda import tensorrt as trt import torch from absl import logging as absl_logging from accelerate import Accelerator from datasets import load_dataset, load_metric from torch.utils.data import DataLoader from utils_qa import postprocess_qa_predictions import transformers from transformers import AutoTokenizer, EvalPrediction, default_data_collator, set_seed from transformers.trainer_pt_utils import nested_concat, nested_truncate lowerCamelCase : str = trt.Logger(trt.Logger.WARNING) lowerCamelCase : Any = absl_logging.get_absl_logger() absl_logger.setLevel(logging.WARNING) lowerCamelCase : Optional[Any] = logging.getLogger(__name__) lowerCamelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--onnx_model_path''', default=None, type=str, required=True, help='''Path to ONNX model: ''', ) parser.add_argument( '''--output_dir''', default=None, type=str, required=True, help='''The output directory where the model checkpoints and predictions will be written.''', ) # Other parameters parser.add_argument( '''--tokenizer_name''', default='''''', type=str, required=True, help='''Pretrained tokenizer name or path if not the same as model_name''', ) parser.add_argument( '''--version_2_with_negative''', action='''store_true''', help='''If true, the SQuAD examples contain some that do not have an answer.''', ) parser.add_argument( '''--null_score_diff_threshold''', type=float, default=0.0, help='''If null_score - best_non_null is greater than the threshold predict null.''', ) parser.add_argument( '''--max_seq_length''', default=3_84, type=int, help=( '''The maximum total input sequence length after WordPiece tokenization. Sequences ''' '''longer than this will be truncated, and sequences shorter than this will be padded.''' ), ) parser.add_argument( '''--doc_stride''', default=1_28, type=int, help='''When splitting up a long document into chunks, how much stride to take between chunks.''', ) parser.add_argument('''--per_device_eval_batch_size''', default=8, type=int, help='''Batch size per GPU/CPU for evaluation.''') parser.add_argument( '''--n_best_size''', default=20, type=int, help='''The total number of n-best predictions to generate in the nbest_predictions.json output file.''', ) parser.add_argument( '''--max_answer_length''', default=30, type=int, help=( '''The maximum length of an answer that can be generated. This is needed because the start ''' '''and end predictions are not conditioned on one another.''' ), ) parser.add_argument('''--seed''', type=int, default=42, help='''random seed for initialization''') parser.add_argument( '''--dataset_name''', type=str, default=None, required=True, help='''The name of the dataset to use (via the datasets library).''', ) parser.add_argument( '''--dataset_config_name''', type=str, default=None, help='''The configuration name of the dataset to use (via the datasets library).''', ) parser.add_argument( '''--preprocessing_num_workers''', type=int, default=4, help='''A csv or a json file containing the training data.''' ) parser.add_argument('''--overwrite_cache''', action='''store_true''', help='''Overwrite the cached training and evaluation sets''') parser.add_argument( '''--fp16''', action='''store_true''', help='''Whether to use 16-bit (mixed) precision instead of 32-bit''', ) parser.add_argument( '''--int8''', action='''store_true''', help='''Whether to use INT8''', ) lowerCamelCase : Dict = parser.parse_args() if args.tokenizer_name: lowerCamelCase : str = AutoTokenizer.from_pretrained(args.tokenizer_name, use_fast=True) else: raise ValueError( '''You are instantiating a new tokenizer from scratch. This is not supported by this script.''' '''You can do it from another script, save it, and load it from here, using --tokenizer_name.''' ) logger.info('''Training/evaluation parameters %s''', args) lowerCamelCase : List[str] = args.per_device_eval_batch_size lowerCamelCase : Any = (args.eval_batch_size, args.max_seq_length) # TRT Engine properties lowerCamelCase : List[str] = True lowerCamelCase : List[Any] = '''temp_engine/bert-fp32.engine''' if args.fpaa: lowerCamelCase : Optional[Any] = '''temp_engine/bert-fp16.engine''' if args.inta: lowerCamelCase : int = '''temp_engine/bert-int8.engine''' # import ONNX file if not os.path.exists('''temp_engine'''): os.makedirs('''temp_engine''') lowerCamelCase : int = 1 << (int)(trt.NetworkDefinitionCreationFlag.EXPLICIT_BATCH) with trt.Builder(TRT_LOGGER) as builder, builder.create_network(EXPLICIT_BATCH) as network, trt.OnnxParser( network, TRT_LOGGER ) as parser: with open(args.onnx_model_path, '''rb''') as model: if not parser.parse(model.read()): for error in range(parser.num_errors): print(parser.get_error(error)) # Query input names and shapes from parsed TensorRT network lowerCamelCase : Union[str, Any] = [network.get_input(i) for i in range(network.num_inputs)] lowerCamelCase : Dict = [_input.name for _input in network_inputs] # ex: ["actual_input1"] with builder.create_builder_config() as config: lowerCamelCase : List[str] = 1 << 50 if STRICT_TYPES: config.set_flag(trt.BuilderFlag.STRICT_TYPES) if args.fpaa: config.set_flag(trt.BuilderFlag.FPaa) if args.inta: config.set_flag(trt.BuilderFlag.INTa) lowerCamelCase : Optional[int] = builder.create_optimization_profile() config.add_optimization_profile(profile) for i in range(len(input_names)): profile.set_shape(input_names[i], INPUT_SHAPE, INPUT_SHAPE, INPUT_SHAPE) lowerCamelCase : Optional[Any] = builder.build_engine(network, config) # serialize_engine and store in file (can be directly loaded and deserialized): with open(engine_name, '''wb''') as f: f.write(engine.serialize()) def snake_case_ ( lowerCAmelCase_ : Any , lowerCAmelCase_ : int , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : str , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Tuple ): __lowercase : List[str] = np.asarray(inputs["""input_ids"""] , dtype=np.intaa ) __lowercase : Union[str, Any] = np.asarray(inputs["""attention_mask"""] , dtype=np.intaa ) __lowercase : int = np.asarray(inputs["""token_type_ids"""] , dtype=np.intaa ) # Copy inputs cuda.memcpy_htod_async(d_inputs[0] , input_ids.ravel() , lowerCAmelCase_ ) cuda.memcpy_htod_async(d_inputs[1] , attention_mask.ravel() , lowerCAmelCase_ ) cuda.memcpy_htod_async(d_inputs[2] , token_type_ids.ravel() , lowerCAmelCase_ ) # start time __lowercase : Optional[Any] = time.time() # Run inference context.execute_async( bindings=[int(lowerCAmelCase_ ) for d_inp in d_inputs] + [int(lowerCAmelCase_ ), int(lowerCAmelCase_ )] , stream_handle=stream.handle ) # Transfer predictions back from GPU cuda.memcpy_dtoh_async(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) cuda.memcpy_dtoh_async(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # Synchronize the stream and take time stream.synchronize() # end time __lowercase : int = time.time() __lowercase : Union[str, Any] = end_time - start_time __lowercase : Any = (h_outputa, h_outputa) # print(outputs) return outputs, infer_time # Initialize the accelerator. We will let the accelerator handle device placement for us in this example. lowerCamelCase : Tuple = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO, ) # Setup logging, we only want one process per machine to log things on the screen. # accelerator.is_local_main_process is only True for one process per machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). if args.dataset_name is not None: # Downloading and loading a dataset from the hub. lowerCamelCase : List[Any] = load_dataset(args.dataset_name, args.dataset_config_name) else: raise ValueError('''Evaluation requires a dataset name''') # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Preprocessing the datasets. # Preprocessing is slighlty different for training and evaluation. lowerCamelCase : Optional[Any] = raw_datasets['''validation'''].column_names lowerCamelCase : Union[str, Any] = '''question''' if '''question''' in column_names else column_names[0] lowerCamelCase : str = '''context''' if '''context''' in column_names else column_names[1] lowerCamelCase : Dict = '''answers''' if '''answers''' in column_names else column_names[2] # Padding side determines if we do (question|context) or (context|question). lowerCamelCase : Dict = tokenizer.padding_side == '''right''' if args.max_seq_length > tokenizer.model_max_length: logger.warning( f'''The max_seq_length passed ({args.max_seq_length}) is larger than the maximum length for the''' f'''model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.''' ) lowerCamelCase : Tuple = min(args.max_seq_length, tokenizer.model_max_length) def snake_case_ ( lowerCAmelCase_ : int ): # Some of the questions have lots of whitespace on the left, which is not useful and will make the # truncation of the context fail (the tokenized question will take a lots of space). So we remove that # left whitespace __lowercase : str = [q.lstrip() for q in examples[question_column_name]] # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results # in one example possible giving several features when a context is long, each of those features having a # context that overlaps a bit the context of the previous feature. __lowercase : List[str] = tokenizer( examples[question_column_name if pad_on_right else context_column_name] , examples[context_column_name if pad_on_right else question_column_name] , truncation="""only_second""" if pad_on_right else """only_first""" , max_length=lowerCAmelCase_ , stride=args.doc_stride , return_overflowing_tokens=lowerCAmelCase_ , return_offsets_mapping=lowerCAmelCase_ , padding="""max_length""" , ) # Since one example might give us several features if it has a long context, we need a map from a feature to # its corresponding example. This key gives us just that. __lowercase : List[str] = tokenized_examples.pop("""overflow_to_sample_mapping""" ) # For evaluation, we will need to convert our predictions to substrings of the context, so we keep the # corresponding example_id and we will store the offset mappings. __lowercase : Any = [] for i in range(len(tokenized_examples["""input_ids"""] ) ): # Grab the sequence corresponding to that example (to know what is the context and what is the question). __lowercase : Dict = tokenized_examples.sequence_ids(lowerCAmelCase_ ) __lowercase : List[Any] = 1 if pad_on_right else 0 # One example can give several spans, this is the index of the example containing this span of text. __lowercase : List[str] = sample_mapping[i] tokenized_examples["example_id"].append(examples["""id"""][sample_index] ) # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token # position is part of the context or not. __lowercase : Dict = [ (o if sequence_ids[k] == context_index else None) for k, o in enumerate(tokenized_examples["""offset_mapping"""][i] ) ] return tokenized_examples lowerCamelCase : Tuple = raw_datasets['''validation'''] # Validation Feature Creation lowerCamelCase : Optional[int] = eval_examples.map( prepare_validation_features, batched=True, num_proc=args.preprocessing_num_workers, remove_columns=column_names, load_from_cache_file=not args.overwrite_cache, desc='''Running tokenizer on validation dataset''', ) lowerCamelCase : Union[str, Any] = default_data_collator lowerCamelCase : Optional[Any] = eval_dataset.remove_columns(['''example_id''', '''offset_mapping''']) lowerCamelCase : List[str] = DataLoader( eval_dataset_for_model, collate_fn=data_collator, batch_size=args.per_device_eval_batch_size ) def snake_case_ ( lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Dict="eval" ): # Post-processing: we match the start logits and end logits to answers in the original context. __lowercase : int = postprocess_qa_predictions( examples=lowerCAmelCase_ , features=lowerCAmelCase_ , predictions=lowerCAmelCase_ , version_2_with_negative=args.version_2_with_negative , n_best_size=args.n_best_size , max_answer_length=args.max_answer_length , null_score_diff_threshold=args.null_score_diff_threshold , output_dir=args.output_dir , prefix=lowerCAmelCase_ , ) # Format the result to the format the metric expects. if args.version_2_with_negative: __lowercase : Optional[int] = [ {"""id""": k, """prediction_text""": v, """no_answer_probability""": 0.0} for k, v in predictions.items() ] else: __lowercase : List[Any] = [{"""id""": k, """prediction_text""": v} for k, v in predictions.items()] __lowercase : Optional[int] = [{"""id""": ex["""id"""], """answers""": ex[answer_column_name]} for ex in examples] return EvalPrediction(predictions=lowerCAmelCase_ , label_ids=lowerCAmelCase_ ) lowerCamelCase : Dict = load_metric('''squad_v2''' if args.version_2_with_negative else '''squad''') # Evaluation! logger.info('''Loading ONNX model %s for evaluation''', args.onnx_model_path) with open(engine_name, '''rb''') as f, trt.Runtime(TRT_LOGGER) as runtime, runtime.deserialize_cuda_engine( f.read() ) as engine, engine.create_execution_context() as context: # setup for TRT inferrence for i in range(len(input_names)): context.set_binding_shape(i, INPUT_SHAPE) assert context.all_binding_shapes_specified def snake_case_ ( lowerCAmelCase_ : str ): return trt.volume(engine.get_binding_shape(lowerCAmelCase_ ) ) * engine.get_binding_dtype(lowerCAmelCase_ ).itemsize # Allocate device memory for inputs and outputs. lowerCamelCase : int = [cuda.mem_alloc(binding_nbytes(binding)) for binding in engine if engine.binding_is_input(binding)] # Allocate output buffer lowerCamelCase : Dict = cuda.pagelocked_empty(tuple(context.get_binding_shape(3)), dtype=np.floataa) lowerCamelCase : str = cuda.pagelocked_empty(tuple(context.get_binding_shape(4)), dtype=np.floataa) lowerCamelCase : Dict = cuda.mem_alloc(h_outputa.nbytes) lowerCamelCase : Optional[Any] = cuda.mem_alloc(h_outputa.nbytes) # Create a stream in which to copy inputs/outputs and run inference. lowerCamelCase : Optional[int] = cuda.Stream() # Evaluation logger.info('''***** Running Evaluation *****''') logger.info(f''' Num examples = {len(eval_dataset)}''') logger.info(f''' Batch size = {args.per_device_eval_batch_size}''') lowerCamelCase : int = 0.0 lowerCamelCase : List[str] = 0 lowerCamelCase : List[str] = timeit.default_timer() lowerCamelCase : List[Any] = None for step, batch in enumerate(eval_dataloader): lowerCamelCase ,lowerCamelCase : str = model_infer(batch, context, d_inputs, h_outputa, h_outputa, d_outputa, d_outputa, stream) total_time += infer_time niter += 1 lowerCamelCase ,lowerCamelCase : Union[str, Any] = outputs lowerCamelCase : Optional[Any] = torch.tensor(start_logits) lowerCamelCase : List[str] = torch.tensor(end_logits) # necessary to pad predictions and labels for being gathered lowerCamelCase : Optional[int] = accelerator.pad_across_processes(start_logits, dim=1, pad_index=-1_00) lowerCamelCase : Dict = accelerator.pad_across_processes(end_logits, dim=1, pad_index=-1_00) lowerCamelCase : List[Any] = (accelerator.gather(start_logits).cpu().numpy(), accelerator.gather(end_logits).cpu().numpy()) lowerCamelCase : Dict = logits if all_preds is None else nested_concat(all_preds, logits, padding_index=-1_00) if all_preds is not None: lowerCamelCase : Tuple = nested_truncate(all_preds, len(eval_dataset)) lowerCamelCase : Dict = timeit.default_timer() - start_time logger.info(''' Evaluation done in total %f secs (%f sec per example)''', evalTime, evalTime / len(eval_dataset)) # Inference time from TRT logger.info('''Average Inference Time = {:.3f} ms'''.format(total_time * 10_00 / niter)) logger.info('''Total Inference Time = {:.3f} ms'''.format(total_time * 10_00)) logger.info('''Total Number of Inference = %d''', niter) lowerCamelCase : str = post_processing_function(eval_examples, eval_dataset, all_preds) lowerCamelCase : Optional[Any] = metric.compute(predictions=prediction.predictions, references=prediction.label_ids) logger.info(f'''Evaluation metrics: {eval_metric}''')
306
1
'''simple docstring''' import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() A__ : Optional[int] = logging.get_logger(__name__) A__ : List[Any] = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn.grep_linear""": """encoder.layers.*.attention.gru_rel_pos_linear""", """self_attn.relative_attention_bias""": """encoder.layers.*.attention.rel_attn_embed""", """self_attn.grep_a""": """encoder.layers.*.attention.gru_rel_pos_const""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """ctc_proj""", """mask_emb""": """masked_spec_embed""", } A__ : Tuple = [ """ctc_proj""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def UpperCAmelCase__ ( UpperCAmelCase_ : int , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[int] ) -> Tuple: for attribute in key.split('.' ): __lowerCamelCase : List[Any] = getattr(UpperCAmelCase_ , UpperCAmelCase_ ) if weight_type is not None: __lowerCamelCase : Any = getattr(UpperCAmelCase_ , UpperCAmelCase_ ).shape else: __lowerCamelCase : Any = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": __lowerCamelCase : Tuple = value elif weight_type == "weight_g": __lowerCamelCase : Optional[int] = value elif weight_type == "weight_v": __lowerCamelCase : str = value elif weight_type == "bias": __lowerCamelCase : List[Any] = value else: __lowerCamelCase : List[str] = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def UpperCAmelCase__ ( UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Union[str, Any] ) -> Optional[int]: __lowerCamelCase : Optional[Any] = [] __lowerCamelCase : Dict = fairseq_model.state_dict() __lowerCamelCase : List[str] = hf_model.feature_extractor for name, value in fairseq_dict.items(): __lowerCamelCase : Union[str, Any] = False if "conv_layers" in name: load_conv_layer( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , hf_model.config.feat_extract_norm == 'group' , ) __lowerCamelCase : str = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: __lowerCamelCase : int = True if "*" in mapped_key: __lowerCamelCase : Optional[int] = name.split(UpperCAmelCase_ )[0].split('.' )[-2] __lowerCamelCase : List[str] = mapped_key.replace('*' , UpperCAmelCase_ ) if "weight_g" in name: __lowerCamelCase : Dict = 'weight_g' elif "weight_v" in name: __lowerCamelCase : Any = 'weight_v' elif "bias" in name and "relative_attention_bias" not in name: __lowerCamelCase : Any = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj __lowerCamelCase : List[Any] = 'weight' else: __lowerCamelCase : str = None set_recursively(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) continue if not is_used: unused_weights.append(UpperCAmelCase_ ) logger.warning(F'Unused weights: {unused_weights}' ) def UpperCAmelCase__ ( UpperCAmelCase_ : Any , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : List[Any] ) -> Tuple: __lowerCamelCase : List[str] = full_name.split('conv_layers.' )[-1] __lowerCamelCase : List[Any] = name.split('.' ) __lowerCamelCase : Any = int(items[0] ) __lowerCamelCase : Optional[Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) __lowerCamelCase : Union[str, Any] = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) __lowerCamelCase : Tuple = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) __lowerCamelCase : str = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) __lowerCamelCase : List[Any] = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(UpperCAmelCase_ ) @torch.no_grad() def UpperCAmelCase__ ( UpperCAmelCase_ : int , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Optional[Any]=None ) -> Optional[int]: # load the pre-trained checkpoints __lowerCamelCase : Optional[Any] = torch.load(UpperCAmelCase_ ) __lowerCamelCase : Optional[Any] = WavLMConfigOrig(checkpoint['cfg'] ) __lowerCamelCase : Union[str, Any] = WavLMOrig(UpperCAmelCase_ ) model.load_state_dict(checkpoint['model'] ) model.eval() if config_path is not None: __lowerCamelCase : Optional[int] = WavLMConfig.from_pretrained(UpperCAmelCase_ ) else: __lowerCamelCase : Any = WavLMConfig() __lowerCamelCase : Optional[int] = WavLMModel(UpperCAmelCase_ ) recursively_load_weights(UpperCAmelCase_ , UpperCAmelCase_ ) hf_wavlm.save_pretrained(UpperCAmelCase_ ) if __name__ == "__main__": A__ : List[Any] = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") A__ : List[str] = parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
185
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class UpperCAmelCase_ (_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase : Union[str, Any] = StableDiffusionInstructPixaPixPipeline lowerCamelCase : List[Any] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width', 'cross_attention_kwargs'} lowerCamelCase : List[str] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS lowerCamelCase : Optional[int] = IMAGE_TO_IMAGE_IMAGE_PARAMS lowerCamelCase : List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase_ ( self ) -> int: torch.manual_seed(0 ) __lowerCamelCase : Any = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=8 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) __lowerCamelCase : Union[str, Any] = PNDMScheduler(skip_prk_steps=SCREAMING_SNAKE_CASE_ ) torch.manual_seed(0 ) __lowerCamelCase : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCamelCase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) __lowerCamelCase : int = CLIPTextModel(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Optional[int] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __lowerCamelCase : Optional[Any] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowercase_ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=0 ) -> Dict: __lowerCamelCase : int = floats_tensor((1, 3, 32, 32) , rng=random.Random(SCREAMING_SNAKE_CASE_ ) ).to(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCamelCase : int = Image.fromarray(np.uinta(SCREAMING_SNAKE_CASE_ ) ).convert('RGB' ) if str(SCREAMING_SNAKE_CASE_ ).startswith('mps' ): __lowerCamelCase : Union[str, Any] = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) else: __lowerCamelCase : Optional[int] = torch.Generator(device=SCREAMING_SNAKE_CASE_ ).manual_seed(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Tuple = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'image_guidance_scale': 1, 'output_type': 'numpy', } return inputs def lowercase_ ( self ) -> Union[str, Any]: __lowerCamelCase : Optional[int] = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase : Any = self.get_dummy_components() __lowerCamelCase : Tuple = StableDiffusionInstructPixaPixPipeline(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Optional[int] = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = self.get_dummy_inputs(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = sd_pipe(**SCREAMING_SNAKE_CASE_ ).images __lowerCamelCase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCamelCase : Optional[int] = np.array([0.7_5_2_6, 0.3_7_5_0, 0.4_5_4_7, 0.6_1_1_7, 0.5_8_6_6, 0.5_0_1_6, 0.4_3_2_7, 0.5_6_4_2, 0.4_8_1_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowercase_ ( self ) -> Union[str, Any]: __lowerCamelCase : List[str] = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase : Union[str, Any] = self.get_dummy_components() __lowerCamelCase : List[Any] = StableDiffusionInstructPixaPixPipeline(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Tuple = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Union[str, Any] = self.get_dummy_inputs(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Union[str, Any] = 'french fries' __lowerCamelCase : List[Any] = sd_pipe(**SCREAMING_SNAKE_CASE_ , negative_prompt=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Tuple = output.images __lowerCamelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCamelCase : Optional[Any] = np.array([0.7_5_1_1, 0.3_6_4_2, 0.4_5_5_3, 0.6_2_3_6, 0.5_7_9_7, 0.5_0_1_3, 0.4_3_4_3, 0.5_6_1_1, 0.4_8_3_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowercase_ ( self ) -> Tuple: __lowerCamelCase : List[str] = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase : Union[str, Any] = self.get_dummy_components() __lowerCamelCase : Any = StableDiffusionInstructPixaPixPipeline(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : List[str] = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = self.get_dummy_inputs(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = [inputs['prompt']] * 2 __lowerCamelCase : Tuple = np.array(inputs['image'] ).astype(np.floataa ) / 2_5_5.0 __lowerCamelCase : List[str] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ).unsqueeze(0 ).to(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Dict = image / 2 + 0.5 __lowerCamelCase : Optional[Any] = image.permute(0 , 3 , 1 , 2 ) __lowerCamelCase : Dict = image.repeat(2 , 1 , 1 , 1 ) __lowerCamelCase : int = sd_pipe(**SCREAMING_SNAKE_CASE_ ).images __lowerCamelCase : Tuple = image[-1, -3:, -3:, -1] assert image.shape == (2, 32, 32, 3) __lowerCamelCase : Union[str, Any] = np.array([0.5_8_1_2, 0.5_7_4_8, 0.5_2_2_2, 0.5_9_0_8, 0.5_6_9_5, 0.7_1_7_4, 0.6_8_0_4, 0.5_5_2_3, 0.5_5_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowercase_ ( self ) -> Tuple: __lowerCamelCase : str = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase : int = self.get_dummy_components() __lowerCamelCase : Union[str, Any] = EulerAncestralDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule='scaled_linear' ) __lowerCamelCase : str = StableDiffusionInstructPixaPixPipeline(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Dict = self.get_dummy_inputs(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = sd_pipe(**SCREAMING_SNAKE_CASE_ ).images __lowerCamelCase : Optional[Any] = image[0, -3:, -3:, -1] __lowerCamelCase : Tuple = [round(SCREAMING_SNAKE_CASE_ , 4 ) for x in image_slice.flatten().tolist()] print(','.join([str(SCREAMING_SNAKE_CASE_ ) for x in slice] ) ) assert image.shape == (1, 32, 32, 3) __lowerCamelCase : List[str] = np.array([0.7_4_1_7, 0.3_8_4_2, 0.4_7_3_2, 0.5_7_7_6, 0.5_8_9_1, 0.5_1_3_9, 0.4_0_5_2, 0.5_6_7_3, 0.4_9_8_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowercase_ ( self ) -> List[Any]: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def lowercase_ ( self ) -> Optional[int]: __lowerCamelCase : Union[str, Any] = self.get_dummy_components() __lowerCamelCase : Tuple = StableDiffusionInstructPixaPixPipeline(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : int = VaeImageProcessor(do_resize=SCREAMING_SNAKE_CASE_ , do_normalize=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : int = pipe(**self.get_dummy_inputs_by_type(SCREAMING_SNAKE_CASE_ , input_image_type='pt' ) )[0] __lowerCamelCase : Optional[Any] = components['vae'] __lowerCamelCase : Dict = self.get_dummy_inputs_by_type(SCREAMING_SNAKE_CASE_ , input_image_type='pt' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): __lowerCamelCase : str = vae.encode(inputs[image_param] ).latent_dist.mode() __lowerCamelCase : str = pipe(**SCREAMING_SNAKE_CASE_ )[0] __lowerCamelCase : Optional[Any] = np.abs(out - out_latents_inputs ).max() self.assertLess(SCREAMING_SNAKE_CASE_ , 1E-4 , 'passing latents as image input generate different result from passing image' ) @slow @require_torch_gpu class UpperCAmelCase_ (unittest.TestCase ): """simple docstring""" def lowercase_ ( self ) -> int: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ ( self , SCREAMING_SNAKE_CASE_=0 ) -> str: __lowerCamelCase : str = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Any = load_image( 'https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg' ) __lowerCamelCase : Any = { 'prompt': 'turn him into a cyborg', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'image_guidance_scale': 1.0, 'output_type': 'numpy', } return inputs def lowercase_ ( self ) -> str: __lowerCamelCase : int = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=SCREAMING_SNAKE_CASE_ ) pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) pipe.enable_attention_slicing() __lowerCamelCase : Optional[Any] = self.get_inputs() __lowerCamelCase : List[str] = pipe(**SCREAMING_SNAKE_CASE_ ).images __lowerCamelCase : Optional[int] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) __lowerCamelCase : Any = np.array([0.5_9_0_2, 0.6_0_1_5, 0.6_0_2_7, 0.5_9_8_3, 0.6_0_9_2, 0.6_0_6_1, 0.5_7_6_5, 0.5_7_8_5, 0.5_5_5_5] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def lowercase_ ( self ) -> Any: __lowerCamelCase : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Any = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) pipe.enable_attention_slicing() __lowerCamelCase : Optional[Any] = self.get_inputs() __lowerCamelCase : Optional[int] = pipe(**SCREAMING_SNAKE_CASE_ ).images __lowerCamelCase : str = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) __lowerCamelCase : Optional[Any] = np.array([0.6_5_7_8, 0.6_8_1_7, 0.6_9_7_2, 0.6_7_6_1, 0.6_8_5_6, 0.6_9_1_6, 0.6_4_2_8, 0.6_5_1_6, 0.6_3_0_1] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def lowercase_ ( self ) -> Optional[int]: __lowerCamelCase : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Dict = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) pipe.enable_attention_slicing() __lowerCamelCase : Union[str, Any] = self.get_inputs() __lowerCamelCase : str = pipe(**SCREAMING_SNAKE_CASE_ ).images __lowerCamelCase : Dict = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) __lowerCamelCase : Union[str, Any] = np.array([0.3_8_2_8, 0.3_8_3_4, 0.3_8_1_8, 0.3_7_9_2, 0.3_8_6_5, 0.3_7_5_2, 0.3_7_9_2, 0.3_8_4_7, 0.3_7_5_3] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def lowercase_ ( self ) -> Optional[int]: __lowerCamelCase : Union[str, Any] = 0 def callback_fn(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> None: __lowerCamelCase : List[Any] = True nonlocal number_of_steps number_of_steps += 1 if step == 1: __lowerCamelCase : Any = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) __lowerCamelCase : Union[str, Any] = latents[0, -3:, -3:, -1] __lowerCamelCase : str = np.array([-0.2_4_6_3, -0.4_6_4_4, -0.9_7_5_6, 1.5_1_7_6, 1.4_4_1_4, 0.7_8_6_6, 0.9_8_9_7, 0.8_5_2_1, 0.7_9_8_3] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 elif step == 2: __lowerCamelCase : Optional[int] = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) __lowerCamelCase : List[Any] = latents[0, -3:, -3:, -1] __lowerCamelCase : Any = np.array([-0.2_6_4_4, -0.4_6_2_6, -0.9_6_5_3, 1.5_1_7_6, 1.4_5_5_1, 0.7_6_8_6, 0.9_8_0_5, 0.8_4_5_2, 0.8_1_1_5] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 __lowerCamelCase : int = False __lowerCamelCase : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=SCREAMING_SNAKE_CASE_ , torch_dtype=torch.floataa ) __lowerCamelCase : Optional[Any] = pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) pipe.enable_attention_slicing() __lowerCamelCase : Optional[int] = self.get_inputs() pipe(**SCREAMING_SNAKE_CASE_ , callback=SCREAMING_SNAKE_CASE_ , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def lowercase_ ( self ) -> Union[str, Any]: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCamelCase : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=SCREAMING_SNAKE_CASE_ , torch_dtype=torch.floataa ) __lowerCamelCase : List[Any] = pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __lowerCamelCase : List[str] = self.get_inputs() __lowerCamelCase : Tuple = pipe(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Optional[Any] = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 10**9 def lowercase_ ( self ) -> List[Any]: __lowerCamelCase : Optional[int] = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 __lowerCamelCase : Union[str, Any] = inputs['image'].resize((5_04, 5_04) ) __lowerCamelCase : int = 'timbrooks/instruct-pix2pix' __lowerCamelCase : Any = StableDiffusionInstructPixaPixPipeline.from_pretrained( SCREAMING_SNAKE_CASE_ , safety_checker=SCREAMING_SNAKE_CASE_ , ) pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) pipe.enable_attention_slicing() __lowerCamelCase : Dict = pipe(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Any = output.images[0] __lowerCamelCase : Optional[int] = image[2_55:2_58, 3_83:3_86, -1] assert image.shape == (5_04, 5_04, 3) __lowerCamelCase : List[str] = np.array([0.2_7_2_6, 0.2_5_2_9, 0.2_6_6_4, 0.2_6_5_5, 0.2_6_4_1, 0.2_6_4_2, 0.2_5_9_1, 0.2_6_4_9, 0.2_5_9_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3
185
1
import argparse import requests import torch from PIL import Image from torchvision.transforms import Compose, Normalize, Resize, ToTensor from transformers import SwinaSRConfig, SwinaSRForImageSuperResolution, SwinaSRImageProcessor def UpperCAmelCase__ ( lowerCamelCase ): lowercase :Dict = SwinaSRConfig() if "Swin2SR_ClassicalSR_X4_64" in checkpoint_url: lowercase :Optional[int] = 4 elif "Swin2SR_CompressedSR_X4_48" in checkpoint_url: lowercase :int = 4 lowercase :int = 48 lowercase :Optional[int] = "pixelshuffle_aux" elif "Swin2SR_Lightweight_X2_64" in checkpoint_url: lowercase :List[str] = [6, 6, 6, 6] lowercase :List[str] = 60 lowercase :str = [6, 6, 6, 6] lowercase :Optional[int] = "pixelshuffledirect" elif "Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR" in checkpoint_url: lowercase :List[str] = 4 lowercase :Tuple = "nearest+conv" elif "Swin2SR_Jpeg_dynamic" in checkpoint_url: lowercase :Optional[Any] = 1 lowercase :Optional[int] = 1 lowercase :int = 126 lowercase :Optional[int] = 7 lowercase :Optional[Any] = 255.0 lowercase :Tuple = "" return config def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase ): if "patch_embed.proj" in name and "layers" not in name: lowercase :str = name.replace("patch_embed.proj", "embeddings.patch_embeddings.projection" ) if "patch_embed.norm" in name: lowercase :List[str] = name.replace("patch_embed.norm", "embeddings.patch_embeddings.layernorm" ) if "layers" in name: lowercase :Optional[Any] = name.replace("layers", "encoder.stages" ) if "residual_group.blocks" in name: lowercase :List[str] = name.replace("residual_group.blocks", "layers" ) if "attn.proj" in name: lowercase :Optional[Any] = name.replace("attn.proj", "attention.output.dense" ) if "attn" in name: lowercase :List[Any] = name.replace("attn", "attention.self" ) if "norm1" in name: lowercase :Union[str, Any] = name.replace("norm1", "layernorm_before" ) if "norm2" in name: lowercase :Dict = name.replace("norm2", "layernorm_after" ) if "mlp.fc1" in name: lowercase :str = name.replace("mlp.fc1", "intermediate.dense" ) if "mlp.fc2" in name: lowercase :Optional[Any] = name.replace("mlp.fc2", "output.dense" ) if "q_bias" in name: lowercase :Any = name.replace("q_bias", "query.bias" ) if "k_bias" in name: lowercase :Optional[Any] = name.replace("k_bias", "key.bias" ) if "v_bias" in name: lowercase :Tuple = name.replace("v_bias", "value.bias" ) if "cpb_mlp" in name: lowercase :List[str] = name.replace("cpb_mlp", "continuous_position_bias_mlp" ) if "patch_embed.proj" in name: lowercase :Tuple = name.replace("patch_embed.proj", "patch_embed.projection" ) if name == "norm.weight": lowercase :List[Any] = "layernorm.weight" if name == "norm.bias": lowercase :Dict = "layernorm.bias" if "conv_first" in name: lowercase :Union[str, Any] = name.replace("conv_first", "first_convolution" ) if ( "upsample" in name or "conv_before_upsample" in name or "conv_bicubic" in name or "conv_up" in name or "conv_hr" in name or "conv_last" in name or "aux" in name ): # heads if "conv_last" in name: lowercase :Dict = name.replace("conv_last", "final_convolution" ) if config.upsampler in ["pixelshuffle", "pixelshuffle_aux", "nearest+conv"]: if "conv_before_upsample.0" in name: lowercase :Union[str, Any] = name.replace("conv_before_upsample.0", "conv_before_upsample" ) if "upsample.0" in name: lowercase :Union[str, Any] = name.replace("upsample.0", "upsample.convolution_0" ) if "upsample.2" in name: lowercase :Union[str, Any] = name.replace("upsample.2", "upsample.convolution_1" ) lowercase :List[Any] = "upsample." + name elif config.upsampler == "pixelshuffledirect": lowercase :Tuple = name.replace("upsample.0.weight", "upsample.conv.weight" ) lowercase :List[str] = name.replace("upsample.0.bias", "upsample.conv.bias" ) else: pass else: lowercase :Any = "swin2sr." + name return name def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase ): for key in orig_state_dict.copy().keys(): lowercase :List[Any] = orig_state_dict.pop(lowerCamelCase ) if "qkv" in key: lowercase :Dict = key.split("." ) lowercase :Tuple = int(key_split[1] ) lowercase :Dict = int(key_split[4] ) lowercase :List[str] = config.embed_dim if "weight" in key: lowercase :int = val[:dim, :] lowercase :str = val[dim : dim * 2, :] lowercase :List[Any] = val[-dim:, :] else: lowercase :List[str] = val[:dim] lowercase :str = val[dim : dim * 2] lowercase :List[Any] = val[-dim:] pass else: lowercase :Optional[Any] = val return orig_state_dict def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): lowercase :Tuple = get_config(lowerCamelCase ) lowercase :List[str] = SwinaSRForImageSuperResolution(lowerCamelCase ) model.eval() lowercase :Optional[int] = torch.hub.load_state_dict_from_url(lowerCamelCase, map_location="cpu" ) lowercase :int = convert_state_dict(lowerCamelCase, lowerCamelCase ) lowercase , lowercase :Optional[int] = model.load_state_dict(lowerCamelCase, strict=lowerCamelCase ) if len(lowerCamelCase ) > 0: raise ValueError("Missing keys when converting: {}".format(lowerCamelCase ) ) for key in unexpected_keys: if not ("relative_position_index" in key or "relative_coords_table" in key or "self_mask" in key): raise ValueError(F"Unexpected key {key} in state_dict" ) # verify values lowercase :str = "https://github.com/mv-lab/swin2sr/blob/main/testsets/real-inputs/shanghai.jpg?raw=true" lowercase :Union[str, Any] = Image.open(requests.get(lowerCamelCase, stream=lowerCamelCase ).raw ).convert("RGB" ) lowercase :List[str] = SwinaSRImageProcessor() # pixel_values = processor(image, return_tensors="pt").pixel_values lowercase :Any = 126 if "Jpeg" in checkpoint_url else 256 lowercase :Optional[Any] = Compose( [ Resize((image_size, image_size) ), ToTensor(), Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225] ), ] ) lowercase :Any = transforms(lowerCamelCase ).unsqueeze(0 ) if config.num_channels == 1: lowercase :Any = pixel_values[:, 0, :, :].unsqueeze(1 ) lowercase :int = model(lowerCamelCase ) # assert values if "Swin2SR_ClassicalSR_X2_64" in checkpoint_url: lowercase :List[str] = torch.Size([1, 3, 512, 512] ) lowercase :Any = torch.tensor( [[-0.7_087, -0.7_138, -0.6_721], [-0.8_340, -0.8_095, -0.7_298], [-0.9_149, -0.8_414, -0.7_940]] ) elif "Swin2SR_ClassicalSR_X4_64" in checkpoint_url: lowercase :Optional[int] = torch.Size([1, 3, 1024, 1024] ) lowercase :List[Any] = torch.tensor( [[-0.7_775, -0.8_105, -0.8_933], [-0.7_764, -0.8_356, -0.9_225], [-0.7_976, -0.8_686, -0.9_579]] ) elif "Swin2SR_CompressedSR_X4_48" in checkpoint_url: # TODO values didn't match exactly here lowercase :Dict = torch.Size([1, 3, 1024, 1024] ) lowercase :List[Any] = torch.tensor( [[-0.8_035, -0.7_504, -0.7_491], [-0.8_538, -0.8_124, -0.7_782], [-0.8_804, -0.8_651, -0.8_493]] ) elif "Swin2SR_Lightweight_X2_64" in checkpoint_url: lowercase :Optional[Any] = torch.Size([1, 3, 512, 512] ) lowercase :Union[str, Any] = torch.tensor( [[-0.7_669, -0.8_662, -0.8_767], [-0.8_810, -0.9_962, -0.9_820], [-0.9_340, -1.0_322, -1.1_149]] ) elif "Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR" in checkpoint_url: lowercase :List[Any] = torch.Size([1, 3, 1024, 1024] ) lowercase :Union[str, Any] = torch.tensor( [[-0.5_238, -0.5_557, -0.6_321], [-0.6_016, -0.5_903, -0.6_391], [-0.6_244, -0.6_334, -0.6_889]] ) assert ( outputs.reconstruction.shape == expected_shape ), F"Shape of reconstruction should be {expected_shape}, but is {outputs.reconstruction.shape}" assert torch.allclose(outputs.reconstruction[0, 0, :3, :3], lowerCamelCase, atol=1e-3 ) print("Looks ok!" ) lowercase :Tuple = { "https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X2_64.pth": ( "swin2SR-classical-sr-x2-64" ), "https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X4_64.pth": ( "swin2SR-classical-sr-x4-64" ), "https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_CompressedSR_X4_48.pth": ( "swin2SR-compressed-sr-x4-48" ), "https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_Lightweight_X2_64.pth": ( "swin2SR-lightweight-x2-64" ), "https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR.pth": ( "swin2SR-realworld-sr-x4-64-bsrgan-psnr" ), } lowercase :Tuple = url_to_name[checkpoint_url] if pytorch_dump_folder_path is not None: print(F"Saving model {model_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(lowerCamelCase ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) processor.save_pretrained(lowerCamelCase ) if push_to_hub: model.push_to_hub(F"caidas/{model_name}" ) processor.push_to_hub(F"caidas/{model_name}" ) if __name__ == "__main__": _UpperCAmelCase : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint_url", default="https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X2_64.pth", type=str, help="URL of the original Swin2SR checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument("--push_to_hub", action="store_true", help="Whether to push the converted model to the hub.") _UpperCAmelCase : Dict = parser.parse_args() convert_swinasr_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
158
import pytest _UpperCAmelCase : List[Any] = "__dummy_dataset1__" _UpperCAmelCase : Union[str, Any] = "\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = \"https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/\"\nURLS = {\"train\": REPO_URL + \"wikiann-bn-train.jsonl\", \"validation\": REPO_URL + \"wikiann-bn-validation.jsonl\"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\n \"ner_tags\": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n \"O\",\n \"B-PER\",\n \"I-PER\",\n \"B-ORG\",\n \"I-ORG\",\n \"B-LOC\",\n \"I-LOC\",\n ]\n )\n ),\n \"langs\": datasets.Sequence(datasets.Value(\"string\")),\n \"spans\": datasets.Sequence(datasets.Value(\"string\")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={\"filepath\": dl_path[\"train\"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={\"filepath\": dl_path[\"validation\"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n" @pytest.fixture def UpperCAmelCase__ ( ): return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def UpperCAmelCase__ ( ): return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): lowercase :Tuple = dataset_loading_script_name lowercase :Dict = tmp_path / "datasets" / script_name script_dir.mkdir(parents=lowerCamelCase ) lowercase :int = script_dir / F"{script_name}.py" with open(lowerCamelCase, "w" ) as f: f.write(lowerCamelCase ) return str(lowerCamelCase )
158
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class a__ ( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): lowerCamelCase : Optional[int] =StableDiffusionInpaintPipeline lowerCamelCase : str =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS lowerCamelCase : str =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS lowerCamelCase : Tuple =frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess lowerCamelCase : Optional[Any] =frozenset([] ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" torch.manual_seed(0 ) __lowerCamelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=9 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=_SCREAMING_SNAKE_CASE , ) __lowerCamelCase = PNDMScheduler(skip_prk_steps=_SCREAMING_SNAKE_CASE ) torch.manual_seed(0 ) __lowerCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) __lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act='''gelu''' , projection_dim=5_12 , ) __lowerCamelCase = CLIPTextModel(_SCREAMING_SNAKE_CASE ) __lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) __lowerCamelCase = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def SCREAMING_SNAKE_CASE__ ( self : Any , a : List[str] , a : Dict=0 ): """simple docstring""" __lowerCamelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(_SCREAMING_SNAKE_CASE ) ).to(_SCREAMING_SNAKE_CASE ) __lowerCamelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCamelCase = Image.fromarray(np.uinta(_SCREAMING_SNAKE_CASE ) ).convert('''RGB''' ).resize((64, 64) ) __lowerCamelCase = Image.fromarray(np.uinta(image + 4 ) ).convert('''RGB''' ).resize((64, 64) ) if str(_SCREAMING_SNAKE_CASE ).startswith('''mps''' ): __lowerCamelCase = torch.manual_seed(_SCREAMING_SNAKE_CASE ) else: __lowerCamelCase = torch.Generator(device=_SCREAMING_SNAKE_CASE ).manual_seed(_SCREAMING_SNAKE_CASE ) __lowerCamelCase = { """prompt""": """A painting of a squirrel eating a burger""", """image""": init_image, """mask_image""": mask_image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def SCREAMING_SNAKE_CASE__ ( self : Tuple ): """simple docstring""" __lowerCamelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = StableDiffusionInpaintPipeline(**_SCREAMING_SNAKE_CASE ) __lowerCamelCase = sd_pipe.to(_SCREAMING_SNAKE_CASE ) sd_pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) __lowerCamelCase = self.get_dummy_inputs(_SCREAMING_SNAKE_CASE ) __lowerCamelCase = sd_pipe(**_SCREAMING_SNAKE_CASE ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __lowerCamelCase = np.array([0.47_27, 0.57_35, 0.39_41, 0.54_46, 0.59_26, 0.43_94, 0.50_62, 0.46_54, 0.44_76] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class a__ ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) __lowerCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench.npy''' ) __lowerCamelCase = """stabilityai/stable-diffusion-2-inpainting""" __lowerCamelCase = StableDiffusionInpaintPipeline.from_pretrained(_SCREAMING_SNAKE_CASE , safety_checker=_SCREAMING_SNAKE_CASE ) pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() __lowerCamelCase = """Face of a yellow cat, high resolution, sitting on a park bench""" __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe( prompt=_SCREAMING_SNAKE_CASE , image=_SCREAMING_SNAKE_CASE , mask_image=_SCREAMING_SNAKE_CASE , generator=_SCREAMING_SNAKE_CASE , output_type='''np''' , ) __lowerCamelCase = output.images[0] assert image.shape == (5_12, 5_12, 3) assert np.abs(expected_image - image ).max() < 9e-3 def SCREAMING_SNAKE_CASE__ ( self : Tuple ): """simple docstring""" __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) __lowerCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench_fp16.npy''' ) __lowerCamelCase = """stabilityai/stable-diffusion-2-inpainting""" __lowerCamelCase = StableDiffusionInpaintPipeline.from_pretrained( _SCREAMING_SNAKE_CASE , torch_dtype=torch.floataa , safety_checker=_SCREAMING_SNAKE_CASE , ) pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() __lowerCamelCase = """Face of a yellow cat, high resolution, sitting on a park bench""" __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe( prompt=_SCREAMING_SNAKE_CASE , image=_SCREAMING_SNAKE_CASE , mask_image=_SCREAMING_SNAKE_CASE , generator=_SCREAMING_SNAKE_CASE , output_type='''np''' , ) __lowerCamelCase = output.images[0] assert image.shape == (5_12, 5_12, 3) assert np.abs(expected_image - image ).max() < 5e-1 def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) __lowerCamelCase = """stabilityai/stable-diffusion-2-inpainting""" __lowerCamelCase = PNDMScheduler.from_pretrained(_SCREAMING_SNAKE_CASE , subfolder='''scheduler''' ) __lowerCamelCase = StableDiffusionInpaintPipeline.from_pretrained( _SCREAMING_SNAKE_CASE , safety_checker=_SCREAMING_SNAKE_CASE , scheduler=_SCREAMING_SNAKE_CASE , torch_dtype=torch.floataa , ) pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __lowerCamelCase = """Face of a yellow cat, high resolution, sitting on a park bench""" __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe( prompt=_SCREAMING_SNAKE_CASE , image=_SCREAMING_SNAKE_CASE , mask_image=_SCREAMING_SNAKE_CASE , generator=_SCREAMING_SNAKE_CASE , num_inference_steps=2 , output_type='''np''' , ) __lowerCamelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 10**9
67
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() A: Optional[int] = logging.get_logger(__name__) A: Optional[int] = torch.device("cpu") def _snake_case ( ): UpperCAmelCase : List[Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase : Tuple = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return im def _snake_case ( UpperCamelCase : int ): if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.1_7_0_3e0_0, 2.1_1_0_7e0_0, -2.0_8_1_1e0_0, 8.8_6_8_5e-0_1, 2.4_3_6_0e-0_1] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.9_6_3_6e-0_1, 2.3_4_7_8e-0_1, -1.6_9_6_3e0_0, -1.7_3_8_1e0_0, -8.6_3_3_7e-0_1] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.2_7_6_8e-0_1, -4.7_4_2_9e-0_1, -1.0_8_9_7e0_0, -1.0_2_4_8e0_0, 3.5_5_2_3e-0_2] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.5_3_3_0e-0_1, 2.4_2_1_1e-0_1, -6.0_1_8_5e-0_1, -8.2_7_8_9e-0_1, -6.0_4_4_6e-0_2] ) def _snake_case ( UpperCamelCase : Optional[Any] , UpperCamelCase : List[str] , UpperCamelCase : str ): UpperCAmelCase : int = dct.pop(UpperCamelCase ) UpperCAmelCase : Any = val def _snake_case ( UpperCamelCase : Union[str, Any] ): UpperCAmelCase : Optional[int] = [] for k in state_dict.keys(): UpperCAmelCase : Optional[Any] = k if ".pwconv" in k: UpperCAmelCase : int = k_new.replace(""".pwconv""" , """.point_wise_conv""" ) if ".dwconv" in k: UpperCAmelCase : Tuple = k_new.replace(""".dwconv""" , """.depth_wise_conv""" ) if ".Proj." in k: UpperCAmelCase : List[Any] = k_new.replace(""".Proj.""" , """.proj.""" ) if "patch_embed" in k_new: UpperCAmelCase : Any = k_new.replace("""patch_embed""" , """swiftformer.patch_embed.patch_embedding""" ) if "network" in k_new: UpperCAmelCase : int = k_new.split(""".""" ) if ls[2].isdigit(): UpperCAmelCase : List[Any] = """swiftformer.encoder.network.""" + ls[1] + """.blocks.""" + ls[2] + """.""" + """.""".join(ls[3:] ) else: UpperCAmelCase : Any = k_new.replace("""network""" , """swiftformer.encoder.network""" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def _snake_case ( UpperCamelCase : int , UpperCamelCase : int , UpperCamelCase : Optional[int] ): UpperCAmelCase : List[Any] = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase : Optional[Any] = 1000 UpperCAmelCase : Tuple = """huggingface/label-files""" UpperCAmelCase : List[str] = """imagenet-1k-id2label.json""" UpperCAmelCase : Dict = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase : Dict = {int(UpperCamelCase ): v for k, v in idalabel.items()} UpperCAmelCase : str = idalabel UpperCAmelCase : Dict = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": UpperCAmelCase : Any = [3, 3, 6, 4] UpperCAmelCase : List[str] = [48, 56, 112, 220] elif swiftformer_name == "swiftformer_s": UpperCAmelCase : Dict = [3, 3, 9, 6] UpperCAmelCase : Union[str, Any] = [48, 64, 168, 224] elif swiftformer_name == "swiftformer_l1": UpperCAmelCase : int = [4, 3, 10, 5] UpperCAmelCase : Optional[int] = [48, 96, 192, 384] elif swiftformer_name == "swiftformer_l3": UpperCAmelCase : Union[str, Any] = [4, 4, 12, 6] UpperCAmelCase : List[Any] = [64, 128, 320, 512] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("""https""" ): UpperCAmelCase : List[str] = torch.hub.load_state_dict_from_url(UpperCamelCase , map_location="""cpu""" , check_hash=UpperCamelCase ) else: UpperCAmelCase : Any = torch.load(UpperCamelCase , map_location="""cpu""" ) UpperCAmelCase : Optional[Any] = checkpoint UpperCAmelCase : Dict = create_rename_keys(UpperCamelCase ) for rename_key_src, rename_key_dest in rename_keys: rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) # load HuggingFace model UpperCAmelCase : List[Any] = SwiftFormerForImageClassification(UpperCamelCase ).eval() hf_model.load_state_dict(UpperCamelCase ) # prepare test inputs UpperCAmelCase : Dict = prepare_img() UpperCAmelCase : Tuple = ViTImageProcessor.from_pretrained("""preprocessor_config""" ) UpperCAmelCase : Optional[int] = processor(images=UpperCamelCase , return_tensors="""pt""" ) # compare outputs from both models UpperCAmelCase : Optional[int] = get_expected_output(UpperCamelCase ) UpperCAmelCase : List[str] = hf_model(inputs["""pixel_values"""] ).logits assert hf_logits.shape == torch.Size([1, 1000] ) assert torch.allclose(hf_logits[0, 0:5] , UpperCamelCase , atol=1e-3 ) Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) print(F"Saving model {swiftformer_name} to {pytorch_dump_folder_path}" ) hf_model.save_pretrained(UpperCamelCase ) if __name__ == "__main__": A: List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--swiftformer_name", default="swiftformer_xs", choices=["swiftformer_xs", "swiftformer_s", "swiftformer_l1", "swiftformer_l3"], type=str, help="Name of the SwiftFormer model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default="./converted_outputs/", type=str, help="Path to the output PyTorch model directory.", ) parser.add_argument("--original_ckpt", default=None, type=str, help="Path to the original model checkpoint.") A: str = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
109
0
import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _snake_case : str = mock.Mock() _snake_case : List[str] = 5_00 _snake_case : Any = {} _snake_case : str = HTTPError _snake_case : List[Any] = {} # Download this model to make sure it's in the cache. _snake_case : Optional[Any] = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=UpperCamelCase ) as mock_head: _snake_case : Any = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def UpperCamelCase_ ( self : Any ): '''simple docstring''' _snake_case : Union[str, Any] = mock.Mock() _snake_case : Union[str, Any] = 5_00 _snake_case : Any = {} _snake_case : Any = HTTPError _snake_case : Optional[int] = {} # Download this model to make sure it's in the cache. _snake_case : Optional[int] = GPTaTokenizerFast.from_pretrained('gpt2' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=UpperCamelCase ) as mock_head: _snake_case : int = GPTaTokenizerFast.from_pretrained('gpt2' ) # This check we did call the fake head request mock_head.assert_called() def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' try: _snake_case : Optional[int] = tempfile.mktemp() with open(UpperCamelCase , 'wb' ) as f: http_get('https://huggingface.co/albert-base-v1/resolve/main/spiece.model' , UpperCamelCase ) _snake_case : Optional[Any] = AlbertTokenizer.from_pretrained(UpperCamelCase ) finally: os.remove(UpperCamelCase ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile('tokenizer.json' ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open('tokenizer.json' , 'wb' ) as f: http_get('https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json' , UpperCamelCase ) _snake_case : List[str] = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 10_00 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove('tokenizer.json' ) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _snake_case : Optional[Any] = AlbertTokenizer.from_pretrained('https://huggingface.co/albert-base-v1/resolve/main/spiece.model' ) @is_staging_test class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' a_ : Dict =["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """bla""", """blou"""] @classmethod def UpperCamelCase_ ( cls : Tuple ): '''simple docstring''' _snake_case : Optional[Any] = TOKEN HfFolder.save_token(UpperCamelCase ) @classmethod def UpperCamelCase_ ( cls : Tuple ): '''simple docstring''' try: delete_repo(token=cls._token , repo_id='test-tokenizer' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-tokenizer-org' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-tokenizer' ) except HTTPError: pass def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: _snake_case : Tuple = os.path.join(UpperCamelCase , 'vocab.txt' ) with open(UpperCamelCase , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) _snake_case : int = BertTokenizer(UpperCamelCase ) tokenizer.push_to_hub('test-tokenizer' , use_auth_token=self._token ) _snake_case : Any = BertTokenizer.from_pretrained(f"""{USER}/test-tokenizer""" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='test-tokenizer' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(UpperCamelCase , repo_id='test-tokenizer' , push_to_hub=UpperCamelCase , use_auth_token=self._token ) _snake_case : Optional[Any] = BertTokenizer.from_pretrained(f"""{USER}/test-tokenizer""" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: _snake_case : int = os.path.join(UpperCamelCase , 'vocab.txt' ) with open(UpperCamelCase , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) _snake_case : Tuple = BertTokenizer(UpperCamelCase ) tokenizer.push_to_hub('valid_org/test-tokenizer-org' , use_auth_token=self._token ) _snake_case : Optional[int] = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='valid_org/test-tokenizer-org' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( UpperCamelCase , repo_id='valid_org/test-tokenizer-org' , push_to_hub=UpperCamelCase , use_auth_token=self._token ) _snake_case : int = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: _snake_case : Dict = os.path.join(UpperCamelCase , 'vocab.txt' ) with open(UpperCamelCase , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) _snake_case : Union[str, Any] = CustomTokenizer(UpperCamelCase ) # No fast custom tokenizer tokenizer.push_to_hub('test-dynamic-tokenizer' , use_auth_token=self._token ) _snake_case : int = AutoTokenizer.from_pretrained(f"""{USER}/test-dynamic-tokenizer""" , trust_remote_code=UpperCamelCase ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizer' ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: _snake_case : Tuple = os.path.join(UpperCamelCase , 'vocab.txt' ) with open(UpperCamelCase , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) _snake_case : Union[str, Any] = BertTokenizerFast.from_pretrained(UpperCamelCase ) bert_tokenizer.save_pretrained(UpperCamelCase ) _snake_case : List[str] = CustomTokenizerFast.from_pretrained(UpperCamelCase ) tokenizer.push_to_hub('test-dynamic-tokenizer' , use_auth_token=self._token ) _snake_case : Optional[int] = AutoTokenizer.from_pretrained(f"""{USER}/test-dynamic-tokenizer""" , trust_remote_code=UpperCamelCase ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizerFast' ) _snake_case : List[str] = AutoTokenizer.from_pretrained( f"""{USER}/test-dynamic-tokenizer""" , use_fast=UpperCamelCase , trust_remote_code=UpperCamelCase ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizer' ) class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _snake_case : Tuple = Trie() trie.add('Hello 友達' ) self.assertEqual(trie.data , {'H': {'e': {'l': {'l': {'o': {' ': {'友': {'達': {'': 1}}}}}}}}} ) trie.add('Hello' ) trie.data self.assertEqual(trie.data , {'H': {'e': {'l': {'l': {'o': {'': 1, ' ': {'友': {'達': {'': 1}}}}}}}}} ) def UpperCamelCase_ ( self : str ): '''simple docstring''' _snake_case : int = Trie() self.assertEqual(trie.split('[CLS] This is a extra_id_100' ) , ['[CLS] This is a extra_id_100'] ) trie.add('[CLS]' ) trie.add('extra_id_1' ) trie.add('extra_id_100' ) self.assertEqual(trie.split('[CLS] This is a extra_id_100' ) , ['[CLS]', ' This is a ', 'extra_id_100'] ) def UpperCamelCase_ ( self : int ): '''simple docstring''' _snake_case : Optional[Any] = Trie() trie.add('A' ) self.assertEqual(trie.split('ABC' ) , ['A', 'BC'] ) self.assertEqual(trie.split('BCA' ) , ['BC', 'A'] ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _snake_case : List[str] = Trie() trie.add('TOKEN]' ) trie.add('[SPECIAL_TOKEN]' ) self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]' ) , ['This is something ', '[SPECIAL_TOKEN]'] ) def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' _snake_case : List[Any] = Trie() trie.add('A' ) trie.add('P' ) trie.add('[SPECIAL_TOKEN]' ) self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]' ) , ['This is something ', '[SPECIAL_TOKEN]'] ) def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _snake_case : List[str] = Trie() trie.add('AB' ) trie.add('B' ) trie.add('C' ) self.assertEqual(trie.split('ABC' ) , ['AB', 'C'] ) def UpperCamelCase_ ( self : int ): '''simple docstring''' _snake_case : int = Trie() trie.add('ABC' ) trie.add('B' ) trie.add('CD' ) self.assertEqual(trie.split('ABCD' ) , ['ABC', 'D'] ) def UpperCamelCase_ ( self : Any ): '''simple docstring''' _snake_case : Dict = Trie() _snake_case : Any = trie.cut_text('ABC' , [0, 0, 2, 1, 2, 3] ) self.assertEqual(UpperCamelCase , ['AB', 'C'] )
369
import argparse import torch from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert from transformers.utils import logging logging.set_verbosity_info() def lowerCamelCase_ ( lowerCAmelCase: List[str] , lowerCAmelCase: int , lowerCAmelCase: List[Any] )-> Dict: # Initialise PyTorch model _snake_case : Dict = RemBertConfig.from_json_file(lowerCAmelCase ) print('Building PyTorch model from configuration: {}'.format(str(lowerCAmelCase ) ) ) _snake_case : Optional[Any] = RemBertModel(lowerCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_rembert(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # Save pytorch-model print('Save PyTorch model to {}'.format(lowerCAmelCase ) ) torch.save(model.state_dict() , lowerCAmelCase ) if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--rembert_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained RemBERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) lowerCAmelCase_ = parser.parse_args() convert_rembert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.rembert_config_file, args.pytorch_dump_path)
260
0
def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = current_set.copy() for row_index, row in enumerate(__lowercase ): lowercase = row[0] for column_index, column in enumerate(__lowercase ): if magnitude == 0: lowercase = column continue lowercase = column / magnitude # Subtract to cancel term lowercase = current_set[0] lowercase = [first_row] lowercase = current_set[1::] for row in current_set: lowercase = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(__lowercase ) continue for column_index in range(len(__lowercase ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(__lowercase ) # Create next recursion iteration set if len(final_set[0] ) != 3: lowercase = final_set[0] lowercase = [] lowercase = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) lowercase = simplify(__lowercase ) for i in range(len(__lowercase ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , __lowercase ) lowercase = resultant return final_set def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): if len(__lowercase ) == 0: raise IndexError('solve_simultaneous() requires n lists of length n+1' ) lowercase = len(__lowercase ) + 1 if any(len(__lowercase ) != _length for item in equations ): raise IndexError('solve_simultaneous() requires n lists of length n+1' ) for row in equations: if any(not isinstance(__lowercase , (int, float) ) for column in row ): raise ValueError('solve_simultaneous() requires lists of integers' ) if len(__lowercase ) == 1: return [equations[0][-1] / equations[0][0]] lowercase = equations.copy() if any(0 in row for row in data_set ): lowercase = data_set.copy() lowercase = [] for row_index, row in enumerate(__lowercase ): if 0 not in row: lowercase = data_set.pop(__lowercase ) break if not full_row: raise ValueError('solve_simultaneous() requires at least 1 full equation' ) data_set.insert(0 , __lowercase ) lowercase = data_set.copy() lowercase = simplify(__lowercase ) lowercase = simplified[::-1] lowercase = [] for row in simplified: lowercase = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue lowercase = row.copy()[: len(__lowercase ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(__lowercase ) == 0: solutions.append(0 ) continue lowercase = temp_row[1::] lowercase = temp_row[::-1] for column_index, column in enumerate(__lowercase ): current_solution -= column * solutions[column_index] solutions.append(__lowercase ) lowercase = [] for item in solutions: final.append(float(round(__lowercase , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
195
'''simple docstring''' import math def UpperCAmelCase_ ( __lowercase : int ) -> bool: '''simple docstring''' return math.sqrt(__lowercase ) * math.sqrt(__lowercase ) == num def UpperCAmelCase_ ( __lowercase : int ) -> bool: '''simple docstring''' _UpperCAmelCase = 0 _UpperCAmelCase = n while left <= right: _UpperCAmelCase = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: _UpperCAmelCase = mid - 1 else: _UpperCAmelCase = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
22
0
from __future__ import annotations from collections.abc import Iterable, Iterator from dataclasses import dataclass a : Any = (3, 9, -11, 0, 7, 5, 1, -1) a : List[str] = (4, 6, 2, 0, 8, 10, 3, -2) @dataclass class _a : A = 42 A = 42 class _a : def __init__(self, SCREAMING_SNAKE_CASE_ ) -> None: UpperCAmelCase_: Node | None = None for i in sorted(SCREAMING_SNAKE_CASE_, reverse=SCREAMING_SNAKE_CASE_ ): UpperCAmelCase_: Any = Node(SCREAMING_SNAKE_CASE_, self.head ) def __iter__(self ) -> Iterator[int]: UpperCAmelCase_: List[Any] = self.head while node: yield node.data UpperCAmelCase_: Optional[Any] = node.next_node def __len__(self ) -> int: return sum(1 for _ in self ) def __str__(self ) -> str: return " -> ".join([str(SCREAMING_SNAKE_CASE_ ) for node in self] ) def lowerCAmelCase_ (lowerCAmelCase__: SortedLinkedList , lowerCAmelCase__: SortedLinkedList ): """simple docstring""" return SortedLinkedList(list(lowerCAmelCase__ ) + list(lowerCAmelCase__ ) ) if __name__ == "__main__": import doctest doctest.testmod() a : str = SortedLinkedList print(merge_lists(SSL(test_data_odd), SSL(test_data_even)))
82
from __future__ import annotations def lowerCAmelCase_ (lowerCAmelCase__: list[float] ): """simple docstring""" UpperCAmelCase_: Union[str, Any] = 0.00 UpperCAmelCase_: List[str] = 0 for resistor in resistors: if resistor <= 0: UpperCAmelCase_: Dict = F'Resistor at index {index} has a negative or zero value!' raise ValueError(lowerCAmelCase__ ) first_sum += 1 / float(lowerCAmelCase__ ) index += 1 return 1 / first_sum def lowerCAmelCase_ (lowerCAmelCase__: list[float] ): """simple docstring""" UpperCAmelCase_: Any = 0.00 UpperCAmelCase_: int = 0 for resistor in resistors: sum_r += resistor if resistor < 0: UpperCAmelCase_: int = F'Resistor at index {index} has a negative value!' raise ValueError(lowerCAmelCase__ ) index += 1 return sum_r if __name__ == "__main__": import doctest doctest.testmod()
82
1
"""simple docstring""" import logging import math from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Sequence, Tuple, Union import torch from .tensor_utils import tensor_tree_map, tree_map def lowercase_ ( _snake_case ): SCREAMING_SNAKE_CASE__ : str = [] if isinstance(_snake_case ,_snake_case ): for v in tree.values(): shapes.extend(_fetch_dims(_snake_case ) ) elif isinstance(_snake_case ,(list, tuple) ): for t in tree: shapes.extend(_fetch_dims(_snake_case ) ) elif isinstance(_snake_case ,torch.Tensor ): shapes.append(tree.shape ) else: raise ValueError("""Not supported""" ) return shapes @torch.jit.ignore def lowercase_ ( _snake_case ,_snake_case ): SCREAMING_SNAKE_CASE__ : Dict = [] for d in reversed(_snake_case ): idx.append(flat_idx % d ) SCREAMING_SNAKE_CASE__ : List[str] = flat_idx // d return tuple(reversed(_snake_case ) ) @torch.jit.ignore def lowercase_ ( _snake_case ,_snake_case ,_snake_case ,_snake_case = None ,_snake_case = None ,): # start_edges and end_edges both indicate whether, starting from any given # dimension, the start/end index is at the top/bottom edge of the # corresponding tensor, modeled as a tree def reduce_edge_list(_snake_case ) -> None: SCREAMING_SNAKE_CASE__ : List[str] = True for i in range(len(_snake_case ) ): SCREAMING_SNAKE_CASE__ : Dict = -1 * (i + 1) l[reversed_idx] &= tally SCREAMING_SNAKE_CASE__ : Optional[Any] = l[reversed_idx] if start_edges is None: SCREAMING_SNAKE_CASE__ : List[str] = [s == 0 for s in start] reduce_edge_list(_snake_case ) if end_edges is None: SCREAMING_SNAKE_CASE__ : List[str] = [e == (d - 1) for e, d in zip(_snake_case ,_snake_case )] reduce_edge_list(_snake_case ) # Base cases. Either start/end are empty and we're done, or the final, # one-dimensional tensor can be simply sliced if len(_snake_case ) == 0: return [()] elif len(_snake_case ) == 1: return [(slice(start[0] ,end[0] + 1 ),)] SCREAMING_SNAKE_CASE__ : List[Tuple[slice, ...]] = [] SCREAMING_SNAKE_CASE__ : List[slice] = [] # Dimensions common to start and end can be selected directly for s, e in zip(_snake_case ,_snake_case ): if s == e: path_list.append(slice(_snake_case ,s + 1 ) ) else: break SCREAMING_SNAKE_CASE__ : Tuple[slice, ...] = tuple(_snake_case ) SCREAMING_SNAKE_CASE__ : List[Any] = len(_snake_case ) # start == end, and we're done if divergence_idx == len(_snake_case ): return [path] def upper() -> Tuple[Tuple[slice, ...], ...]: assert start_edges is not None assert end_edges is not None SCREAMING_SNAKE_CASE__ : Optional[Any] = start[divergence_idx] return tuple( path + (slice(_snake_case ,sdi + 1 ),) + s for s in _get_minimal_slice_set( start[divergence_idx + 1 :] ,[d - 1 for d in dims[divergence_idx + 1 :]] ,dims[divergence_idx + 1 :] ,start_edges=start_edges[divergence_idx + 1 :] ,end_edges=[True for _ in end_edges[divergence_idx + 1 :]] ,) ) def lower() -> Tuple[Tuple[slice, ...], ...]: assert start_edges is not None assert end_edges is not None SCREAMING_SNAKE_CASE__ : List[Any] = end[divergence_idx] return tuple( path + (slice(_snake_case ,edi + 1 ),) + s for s in _get_minimal_slice_set( [0 for _ in start[divergence_idx + 1 :]] ,end[divergence_idx + 1 :] ,dims[divergence_idx + 1 :] ,start_edges=[True for _ in start_edges[divergence_idx + 1 :]] ,end_edges=end_edges[divergence_idx + 1 :] ,) ) # If both start and end are at the edges of the subtree rooted at # divergence_idx, we can just select the whole subtree at once if start_edges[divergence_idx] and end_edges[divergence_idx]: slices.append(path + (slice(start[divergence_idx] ,end[divergence_idx] + 1 ),) ) # If just start is at the edge, we can grab almost all of the subtree, # treating only the ragged bottom edge as an edge case elif start_edges[divergence_idx]: slices.append(path + (slice(start[divergence_idx] ,end[divergence_idx] ),) ) slices.extend(lower() ) # Analogous to the previous case, but the top is ragged this time elif end_edges[divergence_idx]: slices.extend(upper() ) slices.append(path + (slice(start[divergence_idx] + 1 ,end[divergence_idx] + 1 ),) ) # If both sides of the range are ragged, we need to handle both sides # separately. If there's contiguous meat in between them, we can index it # in one big chunk else: slices.extend(upper() ) SCREAMING_SNAKE_CASE__ : Dict = end[divergence_idx] - start[divergence_idx] if middle_ground > 1: slices.append(path + (slice(start[divergence_idx] + 1 ,end[divergence_idx] ),) ) slices.extend(lower() ) return slices @torch.jit.ignore def lowercase_ ( _snake_case ,_snake_case ,_snake_case ,_snake_case ): SCREAMING_SNAKE_CASE__ : Tuple = t.shape[:no_batch_dims] SCREAMING_SNAKE_CASE__ : Dict = list(_flat_idx_to_idx(_snake_case ,_snake_case ) ) # _get_minimal_slice_set is inclusive SCREAMING_SNAKE_CASE__ : Any = list(_flat_idx_to_idx(flat_end - 1 ,_snake_case ) ) # Get an ordered list of slices to perform SCREAMING_SNAKE_CASE__ : List[str] = _get_minimal_slice_set( _snake_case ,_snake_case ,_snake_case ,) SCREAMING_SNAKE_CASE__ : Any = [t[s] for s in slices] return torch.cat([s.view((-1,) + t.shape[no_batch_dims:] ) for s in sliced_tensors] ) def lowercase_ ( _snake_case ,_snake_case ,_snake_case ,_snake_case ,_snake_case = False ,_snake_case = None ,_snake_case = False ,): if not (len(_snake_case ) > 0): raise ValueError("""Must provide at least one input""" ) SCREAMING_SNAKE_CASE__ : str = [shape[:no_batch_dims] for shape in _fetch_dims(_snake_case )] SCREAMING_SNAKE_CASE__ : Dict = tuple([max(_snake_case ) for s in zip(*_snake_case )] ) def _prep_inputs(_snake_case ) -> torch.Tensor: if not low_mem: if not sum(t.shape[:no_batch_dims] ) == no_batch_dims: SCREAMING_SNAKE_CASE__ : Any = t.expand(orig_batch_dims + t.shape[no_batch_dims:] ) SCREAMING_SNAKE_CASE__ : Optional[int] = t.reshape(-1 ,*t.shape[no_batch_dims:] ) else: SCREAMING_SNAKE_CASE__ : Optional[Any] = t.expand(orig_batch_dims + t.shape[no_batch_dims:] ) return t SCREAMING_SNAKE_CASE__ : Dict[str, Any] = tensor_tree_map(_prep_inputs ,_snake_case ) SCREAMING_SNAKE_CASE__ : int = None if _out is not None: SCREAMING_SNAKE_CASE__ : Dict = tensor_tree_map(lambda _snake_case : t.view([-1] + list(t.shape[no_batch_dims:] ) ) ,_out ) SCREAMING_SNAKE_CASE__ : Optional[Any] = 1 for d in orig_batch_dims: flat_batch_dim *= d SCREAMING_SNAKE_CASE__ : Optional[Any] = flat_batch_dim // chunk_size + (flat_batch_dim % chunk_size != 0) def _select_chunk(_snake_case ) -> torch.Tensor: return t[i : i + chunk_size] if t.shape[0] != 1 else t SCREAMING_SNAKE_CASE__ : int = 0 SCREAMING_SNAKE_CASE__ : Dict = prepped_outputs for _ in range(_snake_case ): # Chunk the input if not low_mem: SCREAMING_SNAKE_CASE__ : List[str] = _select_chunk else: SCREAMING_SNAKE_CASE__ : int = partial( _chunk_slice ,flat_start=_snake_case ,flat_end=min(_snake_case ,i + chunk_size ) ,no_batch_dims=len(_snake_case ) ,) SCREAMING_SNAKE_CASE__ : Dict[str, Any] = tensor_tree_map(_snake_case ,_snake_case ) # Run the layer on the chunk SCREAMING_SNAKE_CASE__ : str = layer(**_snake_case ) # Allocate space for the output if out is None: SCREAMING_SNAKE_CASE__ : Dict = tensor_tree_map(lambda _snake_case : t.new_zeros((flat_batch_dim,) + t.shape[1:] ) ,_snake_case ) # Put the chunk in its pre-allocated space if isinstance(_snake_case ,_snake_case ): def assign(_snake_case ,_snake_case ) -> None: for k, v in da.items(): if isinstance(_snake_case ,_snake_case ): assign(_snake_case ,da[k] ) else: if _add_into_out: v[i : i + chunk_size] += da[k] else: SCREAMING_SNAKE_CASE__ : Union[str, Any] = da[k] assign(_snake_case ,_snake_case ) elif isinstance(_snake_case ,_snake_case ): for xa, xa in zip(_snake_case ,_snake_case ): if _add_into_out: xa[i : i + chunk_size] += xa else: SCREAMING_SNAKE_CASE__ : List[str] = xa elif isinstance(_snake_case ,torch.Tensor ): if _add_into_out: out[i : i + chunk_size] += output_chunk else: SCREAMING_SNAKE_CASE__ : Dict = output_chunk else: raise ValueError("""Not supported""" ) i += chunk_size SCREAMING_SNAKE_CASE__ : Dict = tensor_tree_map(lambda _snake_case : t.view(orig_batch_dims + t.shape[1:] ) ,_snake_case ) return out class lowerCAmelCase_ : """simple docstring""" def __init__(self , SCREAMING_SNAKE_CASE__ = 5_12 , ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = max_chunk_size SCREAMING_SNAKE_CASE__ : Optional[int] = None SCREAMING_SNAKE_CASE__ : Optional[tuple] = None def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> int: """simple docstring""" logging.info("""Tuning chunk size...""" ) if min_chunk_size >= self.max_chunk_size: return min_chunk_size SCREAMING_SNAKE_CASE__ : List[int] = [2**l for l in range(int(math.log(self.max_chunk_size , 2 ) ) + 1 )] SCREAMING_SNAKE_CASE__ : Any = [c for c in candidates if c > min_chunk_size] SCREAMING_SNAKE_CASE__ : Dict = [min_chunk_size] + candidates candidates[-1] += 4 def test_chunk_size(SCREAMING_SNAKE_CASE__ ) -> bool: try: with torch.no_grad(): fn(*SCREAMING_SNAKE_CASE__ , chunk_size=SCREAMING_SNAKE_CASE__ ) return True except RuntimeError: return False SCREAMING_SNAKE_CASE__ : Any = 0 SCREAMING_SNAKE_CASE__ : Dict = len(SCREAMING_SNAKE_CASE__ ) - 1 while i > min_viable_chunk_size_index: SCREAMING_SNAKE_CASE__ : int = test_chunk_size(candidates[i] ) if not viable: SCREAMING_SNAKE_CASE__ : Dict = (min_viable_chunk_size_index + i) // 2 else: SCREAMING_SNAKE_CASE__ : Any = i SCREAMING_SNAKE_CASE__ : Dict = (i + len(SCREAMING_SNAKE_CASE__ ) - 1) // 2 return candidates[min_viable_chunk_size_index] def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> bool: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = True for aa, aa in zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): assert type(SCREAMING_SNAKE_CASE__ ) == type(SCREAMING_SNAKE_CASE__ ) if isinstance(SCREAMING_SNAKE_CASE__ , (list, tuple) ): consistent &= self._compare_arg_caches(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) elif isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): SCREAMING_SNAKE_CASE__ : str = [v for _, v in sorted(aa.items() , key=lambda SCREAMING_SNAKE_CASE__ : x[0] )] SCREAMING_SNAKE_CASE__ : Tuple = [v for _, v in sorted(aa.items() , key=lambda SCREAMING_SNAKE_CASE__ : x[0] )] consistent &= self._compare_arg_caches(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else: consistent &= aa == aa return consistent def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = True SCREAMING_SNAKE_CASE__ : tuple = tree_map(lambda SCREAMING_SNAKE_CASE__ : a.shape if isinstance(SCREAMING_SNAKE_CASE__ , torch.Tensor ) else a , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if self.cached_arg_data is not None: # If args have changed shape/value, we need to re-tune assert len(self.cached_arg_data ) == len(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Any = self._compare_arg_caches(self.cached_arg_data , SCREAMING_SNAKE_CASE__ ) else: # Otherwise, we can reuse the precomputed value SCREAMING_SNAKE_CASE__ : Dict = False if not consistent: SCREAMING_SNAKE_CASE__ : Dict = self._determine_favorable_chunk_size( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ) SCREAMING_SNAKE_CASE__ : Any = arg_data assert self.cached_chunk_size is not None return self.cached_chunk_size
25
"""simple docstring""" import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase_ (a__ , a__ , unittest.TestCase ): """simple docstring""" __UpperCamelCase : Optional[Any] = IFPipeline __UpperCamelCase : Dict = TEXT_TO_IMAGE_PARAMS - {'''width''', '''height''', '''latents'''} __UpperCamelCase : Any = TEXT_TO_IMAGE_BATCH_PARAMS __UpperCamelCase : Union[str, Any] = PipelineTesterMixin.required_optional_params - {'''latents'''} def __magic_name__ (self ) -> Union[str, Any]: """simple docstring""" return self._get_dummy_components() def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=0 ) -> List[Any]: """simple docstring""" if str(SCREAMING_SNAKE_CASE__ ).startswith("""mps""" ): SCREAMING_SNAKE_CASE__ : Dict = torch.manual_seed(SCREAMING_SNAKE_CASE__ ) else: SCREAMING_SNAKE_CASE__ : Any = torch.Generator(device=SCREAMING_SNAKE_CASE__ ).manual_seed(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : int = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs def __magic_name__ (self ) -> Tuple: """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != """cuda""" , reason="""float16 requires CUDA""" ) def __magic_name__ (self ) -> List[str]: """simple docstring""" super().test_save_load_floataa(expected_max_diff=1E-1 ) def __magic_name__ (self ) -> List[Any]: """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __magic_name__ (self ) -> Tuple: """simple docstring""" self._test_save_load_local() def __magic_name__ (self ) -> Dict: """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def __magic_name__ (self ) -> Optional[int]: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @slow @require_torch_gpu class lowerCAmelCase_ (unittest.TestCase ): """simple docstring""" def __magic_name__ (self ) -> Dict: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __magic_name__ (self ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = IFPipeline.from_pretrained("""DeepFloyd/IF-I-XL-v1.0""" , variant="""fp16""" , torch_dtype=torch.floataa ) SCREAMING_SNAKE_CASE__ : Dict = IFSuperResolutionPipeline.from_pretrained( """DeepFloyd/IF-II-L-v1.0""" , variant="""fp16""" , torch_dtype=torch.floataa , text_encoder=SCREAMING_SNAKE_CASE__ , tokenizer=SCREAMING_SNAKE_CASE__ ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to("""cuda""" ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Union[str, Any] = pipe_a.encode_prompt("""anime turtle""" , device="""cuda""" ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() SCREAMING_SNAKE_CASE__ : List[str] = None SCREAMING_SNAKE_CASE__ : Union[str, Any] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img SCREAMING_SNAKE_CASE__ : Union[str, Any] = IFImgaImgPipeline(**pipe_a.components ) SCREAMING_SNAKE_CASE__ : Optional[Any] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting SCREAMING_SNAKE_CASE__ : Optional[Any] = IFInpaintingPipeline(**pipe_a.components ) SCREAMING_SNAKE_CASE__ : int = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Optional[int]: """simple docstring""" _start_torch_memory_measurement() SCREAMING_SNAKE_CASE__ : Tuple = torch.Generator(device="""cpu""" ).manual_seed(0 ) SCREAMING_SNAKE_CASE__ : Dict = pipe_a( prompt_embeds=SCREAMING_SNAKE_CASE__ , negative_prompt_embeds=SCREAMING_SNAKE_CASE__ , num_inference_steps=2 , generator=SCREAMING_SNAKE_CASE__ , output_type="""np""" , ) SCREAMING_SNAKE_CASE__ : int = output.images[0] assert image.shape == (64, 64, 3) SCREAMING_SNAKE_CASE__ : Any = torch.cuda.max_memory_allocated() assert mem_bytes < 13 * 10**9 SCREAMING_SNAKE_CASE__ : Optional[int] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy""" ) assert_mean_pixel_difference(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # pipeline 2 _start_torch_memory_measurement() SCREAMING_SNAKE_CASE__ : int = torch.Generator(device="""cpu""" ).manual_seed(0 ) SCREAMING_SNAKE_CASE__ : Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Optional[Any] = pipe_a( prompt_embeds=SCREAMING_SNAKE_CASE__ , negative_prompt_embeds=SCREAMING_SNAKE_CASE__ , image=SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , num_inference_steps=2 , output_type="""np""" , ) SCREAMING_SNAKE_CASE__ : List[str] = output.images[0] assert image.shape == (2_56, 2_56, 3) SCREAMING_SNAKE_CASE__ : List[Any] = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 SCREAMING_SNAKE_CASE__ : int = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy""" ) assert_mean_pixel_difference(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> str: """simple docstring""" _start_torch_memory_measurement() SCREAMING_SNAKE_CASE__ : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Tuple = torch.Generator(device="""cpu""" ).manual_seed(0 ) SCREAMING_SNAKE_CASE__ : Any = pipe_a( prompt_embeds=SCREAMING_SNAKE_CASE__ , negative_prompt_embeds=SCREAMING_SNAKE_CASE__ , image=SCREAMING_SNAKE_CASE__ , num_inference_steps=2 , generator=SCREAMING_SNAKE_CASE__ , output_type="""np""" , ) SCREAMING_SNAKE_CASE__ : Dict = output.images[0] assert image.shape == (64, 64, 3) SCREAMING_SNAKE_CASE__ : Optional[Any] = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 SCREAMING_SNAKE_CASE__ : str = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy""" ) assert_mean_pixel_difference(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # pipeline 2 _start_torch_memory_measurement() SCREAMING_SNAKE_CASE__ : Union[str, Any] = torch.Generator(device="""cpu""" ).manual_seed(0 ) SCREAMING_SNAKE_CASE__ : List[Any] = floats_tensor((1, 3, 2_56, 2_56) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Optional[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Optional[Any] = pipe_a( prompt_embeds=SCREAMING_SNAKE_CASE__ , negative_prompt_embeds=SCREAMING_SNAKE_CASE__ , image=SCREAMING_SNAKE_CASE__ , original_image=SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , num_inference_steps=2 , output_type="""np""" , ) SCREAMING_SNAKE_CASE__ : List[Any] = output.images[0] assert image.shape == (2_56, 2_56, 3) SCREAMING_SNAKE_CASE__ : Optional[Any] = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 SCREAMING_SNAKE_CASE__ : str = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy""" ) assert_mean_pixel_difference(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: """simple docstring""" _start_torch_memory_measurement() SCREAMING_SNAKE_CASE__ : Dict = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Dict = floats_tensor((1, 3, 64, 64) , rng=random.Random(1 ) ).to(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Dict = torch.Generator(device="""cpu""" ).manual_seed(0 ) SCREAMING_SNAKE_CASE__ : List[str] = pipe_a( prompt_embeds=SCREAMING_SNAKE_CASE__ , negative_prompt_embeds=SCREAMING_SNAKE_CASE__ , image=SCREAMING_SNAKE_CASE__ , mask_image=SCREAMING_SNAKE_CASE__ , num_inference_steps=2 , generator=SCREAMING_SNAKE_CASE__ , output_type="""np""" , ) SCREAMING_SNAKE_CASE__ : List[Any] = output.images[0] assert image.shape == (64, 64, 3) SCREAMING_SNAKE_CASE__ : Tuple = torch.cuda.max_memory_allocated() assert mem_bytes < 10 * 10**9 SCREAMING_SNAKE_CASE__ : Optional[Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy""" ) assert_mean_pixel_difference(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # pipeline 2 _start_torch_memory_measurement() SCREAMING_SNAKE_CASE__ : int = torch.Generator(device="""cpu""" ).manual_seed(0 ) SCREAMING_SNAKE_CASE__ : Dict = floats_tensor((1, 3, 64, 64) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : int = floats_tensor((1, 3, 2_56, 2_56) , rng=random.Random(0 ) ).to(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : int = floats_tensor((1, 3, 2_56, 2_56) , rng=random.Random(1 ) ).to(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : int = pipe_a( prompt_embeds=SCREAMING_SNAKE_CASE__ , negative_prompt_embeds=SCREAMING_SNAKE_CASE__ , image=SCREAMING_SNAKE_CASE__ , mask_image=SCREAMING_SNAKE_CASE__ , original_image=SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , num_inference_steps=2 , output_type="""np""" , ) SCREAMING_SNAKE_CASE__ : Dict = output.images[0] assert image.shape == (2_56, 2_56, 3) SCREAMING_SNAKE_CASE__ : List[str] = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 10**9 SCREAMING_SNAKE_CASE__ : Any = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy""" ) assert_mean_pixel_difference(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def lowercase_ ( ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
25
1
"""simple docstring""" import argparse import json import os import numpy as np import PIL import requests import tensorflow.keras.applications.efficientnet as efficientnet import torch from huggingface_hub import hf_hub_download from PIL import Image from tensorflow.keras.preprocessing import image from transformers import ( EfficientNetConfig, EfficientNetForImageClassification, EfficientNetImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) A_ = { '''b0''': efficientnet.EfficientNetBa, '''b1''': efficientnet.EfficientNetBa, '''b2''': efficientnet.EfficientNetBa, '''b3''': efficientnet.EfficientNetBa, '''b4''': efficientnet.EfficientNetBa, '''b5''': efficientnet.EfficientNetBa, '''b6''': efficientnet.EfficientNetBa, '''b7''': efficientnet.EfficientNetBa, } A_ = { '''b0''': { '''hidden_dim''': 1280, '''width_coef''': 1.0, '''depth_coef''': 1.0, '''image_size''': 224, '''dropout_rate''': 0.2, '''dw_padding''': [], }, '''b1''': { '''hidden_dim''': 1280, '''width_coef''': 1.0, '''depth_coef''': 1.1, '''image_size''': 240, '''dropout_rate''': 0.2, '''dw_padding''': [16], }, '''b2''': { '''hidden_dim''': 1408, '''width_coef''': 1.1, '''depth_coef''': 1.2, '''image_size''': 260, '''dropout_rate''': 0.3, '''dw_padding''': [5, 8, 16], }, '''b3''': { '''hidden_dim''': 1536, '''width_coef''': 1.2, '''depth_coef''': 1.4, '''image_size''': 300, '''dropout_rate''': 0.3, '''dw_padding''': [5, 18], }, '''b4''': { '''hidden_dim''': 1792, '''width_coef''': 1.4, '''depth_coef''': 1.8, '''image_size''': 380, '''dropout_rate''': 0.4, '''dw_padding''': [6], }, '''b5''': { '''hidden_dim''': 2048, '''width_coef''': 1.6, '''depth_coef''': 2.2, '''image_size''': 456, '''dropout_rate''': 0.4, '''dw_padding''': [13, 27], }, '''b6''': { '''hidden_dim''': 2304, '''width_coef''': 1.8, '''depth_coef''': 2.6, '''image_size''': 528, '''dropout_rate''': 0.5, '''dw_padding''': [31], }, '''b7''': { '''hidden_dim''': 2560, '''width_coef''': 2.0, '''depth_coef''': 3.1, '''image_size''': 600, '''dropout_rate''': 0.5, '''dw_padding''': [18], }, } def _lowerCAmelCase ( UpperCAmelCase__ : List[str] ) ->Any: A__ : Optional[int] = EfficientNetConfig() A__ : str = CONFIG_MAP[model_name]["""hidden_dim"""] A__ : Union[str, Any] = CONFIG_MAP[model_name]["""width_coef"""] A__ : List[str] = CONFIG_MAP[model_name]["""depth_coef"""] A__ : Dict = CONFIG_MAP[model_name]["""image_size"""] A__ : Optional[Any] = CONFIG_MAP[model_name]["""dropout_rate"""] A__ : Union[str, Any] = CONFIG_MAP[model_name]["""dw_padding"""] A__ : List[str] = """huggingface/label-files""" A__ : Optional[int] = """imagenet-1k-id2label.json""" A__ : List[str] = 1_0_0_0 A__ : Union[str, Any] = json.load(open(hf_hub_download(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ), """r""" ) ) A__ : str = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : Optional[Any] = idalabel A__ : List[str] = {v: k for k, v in idalabel.items()} return config def _lowerCAmelCase ( ) ->Tuple: A__ : Tuple = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : Optional[Any] = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) return im def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int] ) ->List[Any]: A__ : List[str] = CONFIG_MAP[model_name]["""image_size"""] A__ : str = EfficientNetImageProcessor( size={"""height""": size, """width""": size}, image_mean=[0.485, 0.456, 0.406], image_std=[0.4785_3944, 0.473_2864, 0.4743_4163], do_center_crop=UpperCAmelCase__, ) return preprocessor def _lowerCAmelCase ( UpperCAmelCase__ : Tuple ) ->Any: A__ : List[Any] = [v.split("""_""" )[0].split("""block""" )[1] for v in original_param_names if v.startswith("""block""" )] A__ : Optional[Any] = sorted(set(UpperCAmelCase__ ) ) A__ : List[str] = len(UpperCAmelCase__ ) A__ : Dict = {b: str(UpperCAmelCase__ ) for b, i in zip(UpperCAmelCase__, range(UpperCAmelCase__ ) )} A__ : Union[str, Any] = [] rename_keys.append(("""stem_conv/kernel:0""", """embeddings.convolution.weight""") ) rename_keys.append(("""stem_bn/gamma:0""", """embeddings.batchnorm.weight""") ) rename_keys.append(("""stem_bn/beta:0""", """embeddings.batchnorm.bias""") ) rename_keys.append(("""stem_bn/moving_mean:0""", """embeddings.batchnorm.running_mean""") ) rename_keys.append(("""stem_bn/moving_variance:0""", """embeddings.batchnorm.running_var""") ) for b in block_names: A__ : Tuple = block_name_mapping[b] rename_keys.append((f'block{b}_expand_conv/kernel:0', f'encoder.blocks.{hf_b}.expansion.expand_conv.weight') ) rename_keys.append((f'block{b}_expand_bn/gamma:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.weight') ) rename_keys.append((f'block{b}_expand_bn/beta:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.bias') ) rename_keys.append( (f'block{b}_expand_bn/moving_mean:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.running_mean') ) rename_keys.append( (f'block{b}_expand_bn/moving_variance:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.running_var') ) rename_keys.append( (f'block{b}_dwconv/depthwise_kernel:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_conv.weight') ) rename_keys.append((f'block{b}_bn/gamma:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.weight') ) rename_keys.append((f'block{b}_bn/beta:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.bias') ) rename_keys.append( (f'block{b}_bn/moving_mean:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_mean') ) rename_keys.append( (f'block{b}_bn/moving_variance:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_var') ) rename_keys.append((f'block{b}_se_reduce/kernel:0', f'encoder.blocks.{hf_b}.squeeze_excite.reduce.weight') ) rename_keys.append((f'block{b}_se_reduce/bias:0', f'encoder.blocks.{hf_b}.squeeze_excite.reduce.bias') ) rename_keys.append((f'block{b}_se_expand/kernel:0', f'encoder.blocks.{hf_b}.squeeze_excite.expand.weight') ) rename_keys.append((f'block{b}_se_expand/bias:0', f'encoder.blocks.{hf_b}.squeeze_excite.expand.bias') ) rename_keys.append( (f'block{b}_project_conv/kernel:0', f'encoder.blocks.{hf_b}.projection.project_conv.weight') ) rename_keys.append((f'block{b}_project_bn/gamma:0', f'encoder.blocks.{hf_b}.projection.project_bn.weight') ) rename_keys.append((f'block{b}_project_bn/beta:0', f'encoder.blocks.{hf_b}.projection.project_bn.bias') ) rename_keys.append( (f'block{b}_project_bn/moving_mean:0', f'encoder.blocks.{hf_b}.projection.project_bn.running_mean') ) rename_keys.append( (f'block{b}_project_bn/moving_variance:0', f'encoder.blocks.{hf_b}.projection.project_bn.running_var') ) rename_keys.append(("""top_conv/kernel:0""", """encoder.top_conv.weight""") ) rename_keys.append(("""top_bn/gamma:0""", """encoder.top_bn.weight""") ) rename_keys.append(("""top_bn/beta:0""", """encoder.top_bn.bias""") ) rename_keys.append(("""top_bn/moving_mean:0""", """encoder.top_bn.running_mean""") ) rename_keys.append(("""top_bn/moving_variance:0""", """encoder.top_bn.running_var""") ) A__ : Dict = {} for item in rename_keys: if item[0] in original_param_names: A__ : Union[str, Any] = """efficientnet.""" + item[1] A__ : str = """classifier.weight""" A__ : Dict = """classifier.bias""" return key_mapping def _lowerCAmelCase ( UpperCAmelCase__ : List[Any], UpperCAmelCase__ : str, UpperCAmelCase__ : List[str] ) ->List[str]: for key, value in tf_params.items(): if "normalization" in key: continue A__ : Union[str, Any] = key_mapping[key] if "_conv" in key and "kernel" in key: A__ : Optional[int] = torch.from_numpy(UpperCAmelCase__ ).permute(3, 2, 0, 1 ) elif "depthwise_kernel" in key: A__ : int = torch.from_numpy(UpperCAmelCase__ ).permute(2, 3, 0, 1 ) elif "kernel" in key: A__ : Optional[Any] = torch.from_numpy(np.transpose(UpperCAmelCase__ ) ) else: A__ : List[str] = torch.from_numpy(UpperCAmelCase__ ) # Replace HF parameters with original TF model parameters assert hf_params[hf_key].shape == new_hf_value.shape hf_params[hf_key].copy_(UpperCAmelCase__ ) @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Dict ) ->Union[str, Any]: A__ : Dict = model_classes[model_name]( include_top=UpperCAmelCase__, weights="""imagenet""", input_tensor=UpperCAmelCase__, input_shape=UpperCAmelCase__, pooling=UpperCAmelCase__, classes=1_0_0_0, classifier_activation="""softmax""", ) A__ : str = original_model.trainable_variables A__ : str = original_model.non_trainable_variables A__ : str = {param.name: param.numpy() for param in tf_params} for param in tf_non_train_params: A__ : Tuple = param.numpy() A__ : int = list(tf_params.keys() ) # Load HuggingFace model A__ : Optional[Any] = get_efficientnet_config(UpperCAmelCase__ ) A__ : Union[str, Any] = EfficientNetForImageClassification(UpperCAmelCase__ ).eval() A__ : Optional[int] = hf_model.state_dict() # Create src-to-dst parameter name mapping dictionary print("""Converting parameters...""" ) A__ : List[Any] = rename_keys(UpperCAmelCase__ ) replace_params(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) # Initialize preprocessor and preprocess input image A__ : Optional[int] = convert_image_processor(UpperCAmelCase__ ) A__ : str = preprocessor(images=prepare_img(), return_tensors="""pt""" ) # HF model inference hf_model.eval() with torch.no_grad(): A__ : Optional[Any] = hf_model(**UpperCAmelCase__ ) A__ : Union[str, Any] = outputs.logits.detach().numpy() # Original model inference A__ : Dict = False A__ : Tuple = CONFIG_MAP[model_name]["""image_size"""] A__ : Optional[Any] = prepare_img().resize((image_size, image_size), resample=PIL.Image.NEAREST ) A__ : int = image.img_to_array(UpperCAmelCase__ ) A__ : List[Any] = np.expand_dims(UpperCAmelCase__, axis=0 ) A__ : Dict = original_model.predict(UpperCAmelCase__ ) # Check whether original and HF model outputs match -> np.allclose assert np.allclose(UpperCAmelCase__, UpperCAmelCase__, atol=1e-3 ), "The predicted logits are not the same." print("""Model outputs match!""" ) if save_model: # Create folder to save model if not os.path.isdir(UpperCAmelCase__ ): os.mkdir(UpperCAmelCase__ ) # Save converted model and image processor hf_model.save_pretrained(UpperCAmelCase__ ) preprocessor.save_pretrained(UpperCAmelCase__ ) if push_to_hub: # Push model and image processor to hub print(f'Pushing converted {model_name} to the hub...' ) A__ : Union[str, Any] = f'efficientnet-{model_name}' preprocessor.push_to_hub(UpperCAmelCase__ ) hf_model.push_to_hub(UpperCAmelCase__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''b0''', type=str, help='''Version name of the EfficientNet model you want to convert, select from [b0, b1, b2, b3, b4, b5, b6, b7].''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''hf_model''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--save_model''', action='''store_true''', help='''Save model to local''') parser.add_argument('''--push_to_hub''', action='''store_true''', help='''Push model and image processor to the hub''') A_ = parser.parse_args() convert_efficientnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.save_model, args.push_to_hub)
296
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'Salesforce/blip-image-captioning-base' snake_case_ = ( 'This is a tool that generates a description of an image. It takes an input named `image` which should be the ' 'image to caption, and returns a text that contains the description in English.' ) snake_case_ = 'image_captioner' snake_case_ = AutoModelForVisionaSeq snake_case_ = ['image'] snake_case_ = ['text'] def __init__( self : int , *snake_case : Optional[int] , **snake_case : Optional[int] ): '''simple docstring''' requires_backends(self , ["""vision"""] ) super().__init__(*snake_case , **snake_case ) def _UpperCamelCase ( self : int , snake_case : "Image" ): '''simple docstring''' return self.pre_processor(images=snake_case , return_tensors="""pt""" ) def _UpperCamelCase ( self : int , snake_case : List[Any] ): '''simple docstring''' return self.model.generate(**snake_case ) def _UpperCamelCase ( self : Optional[int] , snake_case : Any ): '''simple docstring''' return self.pre_processor.batch_decode(snake_case , skip_special_tokens=snake_case )[0].strip()
296
1
import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin UpperCAmelCase : Optional[int] =random.Random() def _lowerCAmelCase (_lowerCAmelCase , _lowerCAmelCase=1.0 , _lowerCAmelCase=None , _lowerCAmelCase=None): if rng is None: UpperCamelCase_ = global_rng UpperCamelCase_ = [] for batch_idx in range(shape[0]): values.append([]) for _ in range(shape[1]): values[-1].append(rng.random() * scale) return values class _lowercase (unittest.TestCase ): '''simple docstring''' def __init__( self , snake_case__ , snake_case__=7 , snake_case__=400 , snake_case__=2000 , snake_case__=1 , snake_case__=0.0 , snake_case__=1_6000 , snake_case__=True , snake_case__=True , ): '''simple docstring''' UpperCamelCase_ = parent UpperCamelCase_ = batch_size UpperCamelCase_ = min_seq_length UpperCamelCase_ = max_seq_length UpperCamelCase_ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) UpperCamelCase_ = feature_size UpperCamelCase_ = padding_value UpperCamelCase_ = sampling_rate UpperCamelCase_ = return_attention_mask UpperCamelCase_ = do_normalize def _lowerCamelCase ( self ): '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _lowerCamelCase ( self , snake_case__=False , snake_case__=False ): '''simple docstring''' def _flatten(snake_case__ ): return list(itertools.chain(*__snake_case ) ) if equal_length: UpperCamelCase_ = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size UpperCamelCase_ = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: UpperCamelCase_ = [np.asarray(__snake_case ) for x in speech_inputs] return speech_inputs class _lowercase (lowerCamelCase_ , unittest.TestCase ): '''simple docstring''' lowercase__ = WavaVecaFeatureExtractor def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = WavaVecaFeatureExtractionTester(self ) def _lowerCamelCase ( self , snake_case__ ): '''simple docstring''' self.assertTrue(np.all(np.mean(__snake_case , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(__snake_case , axis=0 ) - 1 ) < 1e-3 ) ) def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 UpperCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCamelCase_ = [np.asarray(__snake_case ) for speech_input in speech_inputs] # Test not batched input UpperCamelCase_ = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values UpperCamelCase_ = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(__snake_case , __snake_case , atol=1e-3 ) ) # Test batched UpperCamelCase_ = feat_extract(__snake_case , return_tensors="np" ).input_values UpperCamelCase_ = feat_extract(__snake_case , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(__snake_case , __snake_case ): self.assertTrue(np.allclose(__snake_case , __snake_case , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. UpperCamelCase_ = [floats_list((1, x) )[0] for x in (800, 800, 800)] UpperCamelCase_ = np.asarray(__snake_case ) UpperCamelCase_ = feat_extract(__snake_case , return_tensors="np" ).input_values UpperCamelCase_ = feat_extract(__snake_case , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(__snake_case , __snake_case ): self.assertTrue(np.allclose(__snake_case , __snake_case , atol=1e-3 ) ) def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCamelCase_ = ["longest", "max_length", "do_not_pad"] UpperCamelCase_ = [None, 1600, None] for max_length, padding in zip(__snake_case , __snake_case ): UpperCamelCase_ = feat_extract(__snake_case , padding=__snake_case , max_length=__snake_case , return_tensors="np" ) UpperCamelCase_ = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase_ = range(800 , 1400 , 200 ) UpperCamelCase_ = [floats_list((1, x) )[0] for x in lengths] UpperCamelCase_ = ["longest", "max_length", "do_not_pad"] UpperCamelCase_ = [None, 1600, None] for max_length, padding in zip(__snake_case , __snake_case ): UpperCamelCase_ = feat_extract(__snake_case , max_length=__snake_case , padding=__snake_case ) UpperCamelCase_ = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCamelCase_ = feat_extract( __snake_case , truncation=__snake_case , max_length=1000 , padding="max_length" , return_tensors="np" ) UpperCamelCase_ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCamelCase_ = feat_extract( __snake_case , truncation=__snake_case , max_length=1000 , padding="longest" , return_tensors="np" ) UpperCamelCase_ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) UpperCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] UpperCamelCase_ = feat_extract( __snake_case , truncation=__snake_case , max_length=2000 , padding="longest" , return_tensors="np" ) UpperCamelCase_ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def _lowerCamelCase ( self ): '''simple docstring''' import torch UpperCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase_ = np.random.rand(100 ).astype(np.floataa ) UpperCamelCase_ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: UpperCamelCase_ = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) UpperCamelCase_ = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def _lowerCamelCase ( self ): '''simple docstring''' for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: UpperCamelCase_ = WavaVecaConfig.from_pretrained(__snake_case ) UpperCamelCase_ = WavaVecaFeatureExtractor.from_pretrained(__snake_case ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == "layer" )
128
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class snake_case_ (lowerCamelCase_ , unittest.TestCase ): UpperCAmelCase__ : Tuple = XLMTokenizer UpperCAmelCase__ : List[str] = False def lowerCamelCase__( self :Any ) -> Optional[Any]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt a__ = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] a__ = dict(zip(__snake_case ,range(len(__snake_case ) ) ) ) a__ = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] a__ = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['vocab_file'] ) a__ = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file ,'w' ) as fp: fp.write(json.dumps(__snake_case ) ) with open(self.merges_file ,'w' ) as fp: fp.write('\n'.join(__snake_case ) ) def lowerCamelCase__( self :Any ,__snake_case :int ) -> Optional[int]: a__ = 'lower newer' a__ = 'lower newer' return input_text, output_text def lowerCamelCase__( self :Tuple ) -> Tuple: a__ = XLMTokenizer(self.vocab_file ,self.merges_file ) a__ = 'lower' a__ = ['low', 'er</w>'] a__ = tokenizer.tokenize(__snake_case ) self.assertListEqual(__snake_case ,__snake_case ) a__ = tokens + ['<unk>'] a__ = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__snake_case ) ,__snake_case ) @slow def lowerCamelCase__( self :List[str] ) -> Optional[Any]: a__ = XLMTokenizer.from_pretrained('xlm-mlm-en-2048' ) a__ = tokenizer.encode('sequence builders' ,add_special_tokens=__snake_case ) a__ = tokenizer.encode('multi-sequence build' ,add_special_tokens=__snake_case ) a__ = tokenizer.build_inputs_with_special_tokens(__snake_case ) a__ = tokenizer.build_inputs_with_special_tokens(__snake_case ,__snake_case ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
240
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) __snake_case :str = { '''configuration_efficientformer''': [ '''EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''EfficientFormerConfig''', ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case :Tuple = ['''EfficientFormerImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case :int = [ '''EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''EfficientFormerForImageClassification''', '''EfficientFormerForImageClassificationWithTeacher''', '''EfficientFormerModel''', '''EfficientFormerPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case :Any = [ '''TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFEfficientFormerForImageClassification''', '''TFEfficientFormerForImageClassificationWithTeacher''', '''TFEfficientFormerModel''', '''TFEfficientFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys __snake_case :List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
357
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _A : def __init__( self : str , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : Tuple=13 , __SCREAMING_SNAKE_CASE : Union[str, Any]=32 , __SCREAMING_SNAKE_CASE : List[Any]=2 , __SCREAMING_SNAKE_CASE : Dict=3 , __SCREAMING_SNAKE_CASE : Any=16 , __SCREAMING_SNAKE_CASE : Optional[int]=[1, 2, 1] , __SCREAMING_SNAKE_CASE : Optional[int]=[2, 2, 4] , __SCREAMING_SNAKE_CASE : Any=2 , __SCREAMING_SNAKE_CASE : Tuple=2.0 , __SCREAMING_SNAKE_CASE : List[str]=True , __SCREAMING_SNAKE_CASE : Dict=0.0 , __SCREAMING_SNAKE_CASE : List[str]=0.0 , __SCREAMING_SNAKE_CASE : str=0.1 , __SCREAMING_SNAKE_CASE : int="gelu" , __SCREAMING_SNAKE_CASE : Tuple=False , __SCREAMING_SNAKE_CASE : List[Any]=True , __SCREAMING_SNAKE_CASE : Dict=0.02 , __SCREAMING_SNAKE_CASE : Dict=1E-5 , __SCREAMING_SNAKE_CASE : Optional[Any]=True , __SCREAMING_SNAKE_CASE : int=None , __SCREAMING_SNAKE_CASE : str=True , __SCREAMING_SNAKE_CASE : Optional[int]=10 , __SCREAMING_SNAKE_CASE : int=8 , ): '''simple docstring''' __a = parent __a = batch_size __a = image_size __a = patch_size __a = num_channels __a = embed_dim __a = depths __a = num_heads __a = window_size __a = mlp_ratio __a = qkv_bias __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = drop_path_rate __a = hidden_act __a = use_absolute_embeddings __a = patch_norm __a = layer_norm_eps __a = initializer_range __a = is_training __a = scope __a = use_labels __a = type_sequence_label_size __a = encoder_stride def _lowerCamelCase ( self : Dict): '''simple docstring''' __a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.type_sequence_label_size) __a = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : List[Any]): '''simple docstring''' return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCamelCase ( self : str , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : Optional[Any]): '''simple docstring''' __a = SwinvaModel(config=__SCREAMING_SNAKE_CASE) model.to(__SCREAMING_SNAKE_CASE) model.eval() __a = model(__SCREAMING_SNAKE_CASE) __a = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths) - 1)) __a = int(config.embed_dim * 2 ** (len(config.depths) - 1)) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim)) def _lowerCamelCase ( self : Dict , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : Any): '''simple docstring''' __a = SwinvaForMaskedImageModeling(config=__SCREAMING_SNAKE_CASE) model.to(__SCREAMING_SNAKE_CASE) model.eval() __a = model(__SCREAMING_SNAKE_CASE) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size)) # test greyscale images __a = 1 __a = SwinvaForMaskedImageModeling(__SCREAMING_SNAKE_CASE) model.to(__SCREAMING_SNAKE_CASE) model.eval() __a = floats_tensor([self.batch_size, 1, self.image_size, self.image_size]) __a = model(__SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size)) def _lowerCamelCase ( self : Optional[int] , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : List[Any]): '''simple docstring''' __a = self.type_sequence_label_size __a = SwinvaForImageClassification(__SCREAMING_SNAKE_CASE) model.to(__SCREAMING_SNAKE_CASE) model.eval() __a = model(__SCREAMING_SNAKE_CASE , labels=__SCREAMING_SNAKE_CASE) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size)) def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' __a = self.prepare_config_and_inputs() __a , __a , __a = config_and_inputs __a = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class _A ( __UpperCAmelCase ,__UpperCAmelCase ,unittest.TestCase ): UpperCamelCase__ : Dict = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) UpperCamelCase__ : Optional[int] = ( {'''feature-extraction''': SwinvaModel, '''image-classification''': SwinvaForImageClassification} if is_torch_available() else {} ) UpperCamelCase__ : int = False UpperCamelCase__ : Tuple = False UpperCamelCase__ : Optional[Any] = False UpperCamelCase__ : Optional[Any] = False def _lowerCamelCase ( self : List[Any]): '''simple docstring''' __a = SwinvaModelTester(self) __a = ConfigTester(self , config_class=__SCREAMING_SNAKE_CASE , embed_dim=37) def _lowerCamelCase ( self : List[Any]): '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__SCREAMING_SNAKE_CASE) @unittest.skip(reason='''Got `CUDA error: misaligned address` with PyTorch 2.0.0.''') def _lowerCamelCase ( self : List[Any]): '''simple docstring''' pass @unittest.skip(reason='''Swinv2 does not use inputs_embeds''') def _lowerCamelCase ( self : List[str]): '''simple docstring''' pass def _lowerCamelCase ( self : Any): '''simple docstring''' __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(__SCREAMING_SNAKE_CASE) self.assertIsInstance(model.get_input_embeddings() , (nn.Module)) __a = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__SCREAMING_SNAKE_CASE , nn.Linear)) def _lowerCamelCase ( self : Optional[int]): '''simple docstring''' __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(__SCREAMING_SNAKE_CASE) __a = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a = [*signature.parameters.keys()] __a = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' __a , __a = self.model_tester.prepare_config_and_inputs_for_common() __a = True for model_class in self.all_model_classes: __a = True __a = False __a = True __a = model_class(__SCREAMING_SNAKE_CASE) model.to(__SCREAMING_SNAKE_CASE) model.eval() with torch.no_grad(): __a = model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE)) __a = outputs.attentions __a = len(self.model_tester.depths) self.assertEqual(len(__SCREAMING_SNAKE_CASE) , __SCREAMING_SNAKE_CASE) # check that output_attentions also work using config del inputs_dict["output_attentions"] __a = True __a = config.window_size**2 __a = model_class(__SCREAMING_SNAKE_CASE) model.to(__SCREAMING_SNAKE_CASE) model.eval() with torch.no_grad(): __a = model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE)) __a = outputs.attentions self.assertEqual(len(__SCREAMING_SNAKE_CASE) , __SCREAMING_SNAKE_CASE) self.assertListEqual( list(attentions[0].shape[-3:]) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) __a = len(__SCREAMING_SNAKE_CASE) # Check attention is always last and order is fine __a = True __a = True __a = model_class(__SCREAMING_SNAKE_CASE) model.to(__SCREAMING_SNAKE_CASE) model.eval() with torch.no_grad(): __a = model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE)) if hasattr(self.model_tester , '''num_hidden_states_types'''): __a = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states __a = 2 self.assertEqual(out_len + added_hidden_states , len(__SCREAMING_SNAKE_CASE)) __a = outputs.attentions self.assertEqual(len(__SCREAMING_SNAKE_CASE) , __SCREAMING_SNAKE_CASE) self.assertListEqual( list(self_attentions[0].shape[-3:]) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def _lowerCamelCase ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Dict , __SCREAMING_SNAKE_CASE : List[str]): '''simple docstring''' __a = model_class(__SCREAMING_SNAKE_CASE) model.to(__SCREAMING_SNAKE_CASE) model.eval() with torch.no_grad(): __a = model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE)) __a = outputs.hidden_states __a = getattr( self.model_tester , '''expected_num_hidden_layers''' , len(self.model_tester.depths) + 1) self.assertEqual(len(__SCREAMING_SNAKE_CASE) , __SCREAMING_SNAKE_CASE) # Swinv2 has a different seq_length __a = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable) else (config.patch_size, config.patch_size) ) __a = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:]) , [num_patches, self.model_tester.embed_dim] , ) __a = outputs.reshaped_hidden_states self.assertEqual(len(__SCREAMING_SNAKE_CASE) , __SCREAMING_SNAKE_CASE) __a , __a , __a , __a = reshaped_hidden_states[0].shape __a = ( reshaped_hidden_states[0].view(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , height * width).permute(0 , 2 , 1) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:]) , [num_patches, self.model_tester.embed_dim] , ) def _lowerCamelCase ( self : Optional[int]): '''simple docstring''' __a , __a = self.model_tester.prepare_config_and_inputs_for_common() __a = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __a = True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __a = True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Optional[int]): '''simple docstring''' __a , __a = self.model_tester.prepare_config_and_inputs_for_common() __a = 3 __a = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable) else (self.model_tester.image_size, self.model_tester.image_size) ) __a = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable) else (config.patch_size, config.patch_size) ) __a = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __a = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __a = True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , (padded_height, padded_width)) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __a = True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , (padded_height, padded_width)) def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : List[Any]): '''simple docstring''' __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__SCREAMING_SNAKE_CASE) @slow def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = SwinvaModel.from_pretrained(__SCREAMING_SNAKE_CASE) self.assertIsNotNone(__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Any): '''simple docstring''' __a , __a = self.model_tester.prepare_config_and_inputs_for_common() __a = _config_zero_init(__SCREAMING_SNAKE_CASE) for model_class in self.all_model_classes: __a = model_class(config=__SCREAMING_SNAKE_CASE) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) @require_vision @require_torch class _A ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' return ( AutoImageProcessor.from_pretrained('''microsoft/swinv2-tiny-patch4-window8-256''') if is_vision_available() else None ) @slow def _lowerCamelCase ( self : str): '''simple docstring''' __a = SwinvaForImageClassification.from_pretrained('''microsoft/swinv2-tiny-patch4-window8-256''').to( __SCREAMING_SNAKE_CASE) __a = self.default_image_processor __a = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''') __a = image_processor(images=__SCREAMING_SNAKE_CASE , return_tensors='''pt''').to(__SCREAMING_SNAKE_CASE) # forward pass with torch.no_grad(): __a = model(**__SCREAMING_SNAKE_CASE) # verify the logits __a = torch.Size((1, 1_000)) self.assertEqual(outputs.logits.shape , __SCREAMING_SNAKE_CASE) __a = torch.tensor([-0.39_47, -0.43_06, 0.00_26]).to(__SCREAMING_SNAKE_CASE) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __SCREAMING_SNAKE_CASE , atol=1E-4))
131
0
"""simple docstring""" import inspect import os import sys import unittest import accelerate from accelerate.test_utils import execute_subprocess_async, require_tpu class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def lowerCamelCase ( self : Optional[int] ): lowerCAmelCase_ : int = inspect.getfile(accelerate.test_utils ) lowerCAmelCase_ : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_script.py"] ) lowerCAmelCase_ : Dict = os.path.sep.join(inspect.getfile(self.__class__ ).split(os.path.sep )[:-1] ) @require_tpu def lowerCamelCase ( self : Optional[int] ): lowerCAmelCase_ : List[Any] = f''' {self.test_dir}/xla_spawn.py --num_cores 8 {self.test_file_path} '''.split() lowerCAmelCase_ : Dict = [sys.executable] + distributed_args execute_subprocess_async(a_ , env=os.environ.copy() )
241
"""simple docstring""" from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) # TODO Update this lowercase__ = { """facebook/esm-1b""": """https://huggingface.co/facebook/esm-1b/resolve/main/config.json""", # See all ESM models at https://huggingface.co/models?filter=esm } class __lowerCamelCase ( A__ ): '''simple docstring''' a_ : str = """esm""" def __init__( self : Union[str, Any] , a_ : int=None , a_ : List[str]=None , a_ : Optional[int]=None , a_ : Optional[int]=7_68 , a_ : List[Any]=12 , a_ : List[str]=12 , a_ : Optional[Any]=30_72 , a_ : Optional[Any]=0.1 , a_ : Tuple=0.1 , a_ : Union[str, Any]=10_26 , a_ : List[str]=0.02 , a_ : Optional[int]=1e-1_2 , a_ : int="absolute" , a_ : Union[str, Any]=True , a_ : int=None , a_ : int=False , a_ : Optional[Any]=False , a_ : Any=None , a_ : List[str]=None , **a_ : int , ): super().__init__(pad_token_id=a_ , mask_token_id=a_ , **a_ ) lowerCAmelCase_ : str = vocab_size lowerCAmelCase_ : List[str] = hidden_size lowerCAmelCase_ : str = num_hidden_layers lowerCAmelCase_ : str = num_attention_heads lowerCAmelCase_ : Tuple = intermediate_size lowerCAmelCase_ : List[str] = hidden_dropout_prob lowerCAmelCase_ : Dict = attention_probs_dropout_prob lowerCAmelCase_ : Dict = max_position_embeddings lowerCAmelCase_ : Optional[Any] = initializer_range lowerCAmelCase_ : Tuple = layer_norm_eps lowerCAmelCase_ : Dict = position_embedding_type lowerCAmelCase_ : str = use_cache lowerCAmelCase_ : str = emb_layer_norm_before lowerCAmelCase_ : Any = token_dropout lowerCAmelCase_ : str = is_folding_model if is_folding_model: if esmfold_config is None: logger.info("No esmfold_config supplied for folding model, using default values." ) lowerCAmelCase_ : int = EsmFoldConfig() elif isinstance(a_ , a_ ): lowerCAmelCase_ : int = EsmFoldConfig(**a_ ) lowerCAmelCase_ : Tuple = esmfold_config if vocab_list is None: logger.warning("No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!" ) lowerCAmelCase_ : Any = get_default_vocab_list() else: lowerCAmelCase_ : Optional[Any] = vocab_list else: lowerCAmelCase_ : List[str] = None lowerCAmelCase_ : Tuple = None if self.esmfold_config is not None and getattr(self.esmfold_config , "use_esm_attn_map" , a_ ): raise ValueError("The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!" ) def lowerCamelCase ( self : Dict ): lowerCAmelCase_ : List[str] = super().to_dict() if isinstance(self.esmfold_config , a_ ): lowerCAmelCase_ : int = self.esmfold_config.to_dict() return output @dataclass class __lowerCamelCase : '''simple docstring''' a_ : str = None a_ : bool = True a_ : bool = False a_ : bool = False a_ : bool = False a_ : float = 0 a_ : bool = True a_ : bool = False a_ : int = 128 a_ : "TrunkConfig" = None def lowerCamelCase ( self : str ): if self.trunk is None: lowerCAmelCase_ : List[Any] = TrunkConfig() elif isinstance(self.trunk , a_ ): lowerCAmelCase_ : str = TrunkConfig(**self.trunk ) def lowerCamelCase ( self : Any ): lowerCAmelCase_ : Any = asdict(self ) lowerCAmelCase_ : int = self.trunk.to_dict() return output @dataclass class __lowerCamelCase : '''simple docstring''' a_ : int = 48 a_ : int = 1024 a_ : int = 128 a_ : int = 32 a_ : int = 32 a_ : int = 32 a_ : float = 0 a_ : float = 0 a_ : bool = False a_ : int = 4 a_ : Optional[int] = 128 a_ : "StructureModuleConfig" = None def lowerCamelCase ( self : Optional[Any] ): if self.structure_module is None: lowerCAmelCase_ : Any = StructureModuleConfig() elif isinstance(self.structure_module , a_ ): lowerCAmelCase_ : Union[str, Any] = StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(f'''`max_recycles` should be positive, got {self.max_recycles}.''' ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( "`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got" f''' {self.sequence_state_dim} and {self.sequence_state_dim}.''' ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( "`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got" f''' {self.pairwise_state_dim} and {self.pairwise_state_dim}.''' ) lowerCAmelCase_ : List[str] = self.sequence_state_dim // self.sequence_head_width lowerCAmelCase_ : str = self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( "`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got" f''' {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}.''' ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( "`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got" f''' {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}.''' ) if self.pairwise_state_dim % 2 != 0: raise ValueError(f'''`pairwise_state_dim` should be even, got {self.pairwise_state_dim}.''' ) if self.dropout >= 0.4: raise ValueError(f'''`dropout` should not be greater than 0.4, got {self.dropout}.''' ) def lowerCamelCase ( self : Optional[int] ): lowerCAmelCase_ : Union[str, Any] = asdict(self ) lowerCAmelCase_ : str = self.structure_module.to_dict() return output @dataclass class __lowerCamelCase : '''simple docstring''' a_ : int = 384 a_ : int = 128 a_ : int = 16 a_ : int = 128 a_ : int = 12 a_ : int = 4 a_ : int = 8 a_ : float = 0.1 a_ : int = 8 a_ : int = 1 a_ : int = 2 a_ : int = 7 a_ : int = 10 a_ : float = 1E-8 a_ : float = 1E5 def lowerCamelCase ( self : Optional[int] ): return asdict(self ) def __lowerCamelCase ( ) -> Tuple: """simple docstring""" return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
241
1
"""simple docstring""" import json import os import shutil import warnings from argparse import ArgumentParser, Namespace from pathlib import Path from typing import List from ..utils import logging from . import BaseTransformersCLICommand try: from cookiecutter.main import cookiecutter lowerCamelCase : Any = True except ImportError: lowerCamelCase : Any = False lowerCamelCase : int = logging.get_logger(__name__) # pylint: disable=invalid-name def _SCREAMING_SNAKE_CASE ( lowercase : Namespace ): '''simple docstring''' return AddNewModelCommand(args.testing , args.testing_file , path=args.path ) class A( UpperCamelCase ): '''simple docstring''' @staticmethod def a__ ( A_ : ArgumentParser ) -> int: """simple docstring""" lowerCamelCase_ = parser.add_parser('add-new-model' ) add_new_model_parser.add_argument('--testing' , action='store_true' , help='If in testing mode.' ) add_new_model_parser.add_argument('--testing_file' , type=A_ , help='Configuration file on which to run.' ) add_new_model_parser.add_argument( '--path' , type=A_ , help='Path to cookiecutter. Should only be used for testing purposes.' ) add_new_model_parser.set_defaults(func=A_ ) def __init__( self : Tuple , A_ : bool , A_ : str , A_ : str=None , *A_ : int ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = testing lowerCamelCase_ = testing_file lowerCamelCase_ = path def a__ ( self : Dict ) -> Dict: """simple docstring""" warnings.warn( 'The command `transformers-cli add-new-model` is deprecated and will be removed in v5 of Transformers. ' 'It is not actively maintained anymore, so might give a result that won\'t pass all tests and quality ' 'checks, you should use `transformers-cli add-new-model-like` instead.' ) if not _has_cookiecutter: raise ImportError( 'Model creation dependencies are required to use the `add_new_model` command. Install them by running ' 'the following at the root of your `transformers` clone:\n\n\t$ pip install -e .[modelcreation]\n' ) # Ensure that there is no other `cookiecutter-template-xxx` directory in the current working directory lowerCamelCase_ = [directory for directory in os.listdir() if 'cookiecutter-template-' == directory[:22]] if len(A_ ) > 0: raise ValueError( 'Several directories starting with `cookiecutter-template-` in current working directory. ' 'Please clean your directory by removing all folders starting with `cookiecutter-template-` or ' 'change your working directory.' ) lowerCamelCase_ = ( Path(A_ ).parent.parent.parent.parent if self._path is None else Path(self._path ).parent.parent ) lowerCamelCase_ = path_to_transformer_root / 'templates' / 'adding_a_new_model' # Execute cookiecutter if not self._testing: cookiecutter(str(A_ ) ) else: with open(self._testing_file , 'r' ) as configuration_file: lowerCamelCase_ = json.load(A_ ) cookiecutter( str(path_to_cookiecutter if self._path is None else self._path ) , no_input=A_ , extra_context=A_ , ) lowerCamelCase_ = [directory for directory in os.listdir() if 'cookiecutter-template-' in directory[:22]][0] # Retrieve configuration with open(directory + '/configuration.json' , 'r' ) as configuration_file: lowerCamelCase_ = json.load(A_ ) lowerCamelCase_ = configuration['lowercase_modelname'] lowerCamelCase_ = configuration['generate_tensorflow_pytorch_and_flax'] os.remove(f"""{directory}/configuration.json""" ) lowerCamelCase_ = 'PyTorch' in generate_tensorflow_pytorch_and_flax lowerCamelCase_ = 'TensorFlow' in generate_tensorflow_pytorch_and_flax lowerCamelCase_ = 'Flax' in generate_tensorflow_pytorch_and_flax lowerCamelCase_ = f"""{path_to_transformer_root}/src/transformers/models/{lowercase_model_name}""" os.makedirs(A_ , exist_ok=A_ ) os.makedirs(f"""{path_to_transformer_root}/tests/models/{lowercase_model_name}""" , exist_ok=A_ ) # Tests require submodules as they have parent imports with open(f"""{path_to_transformer_root}/tests/models/{lowercase_model_name}/__init__.py""" , 'w' ): pass shutil.move( f"""{directory}/__init__.py""" , f"""{model_dir}/__init__.py""" , ) shutil.move( f"""{directory}/configuration_{lowercase_model_name}.py""" , f"""{model_dir}/configuration_{lowercase_model_name}.py""" , ) def remove_copy_lines(A_ : Tuple ): with open(A_ , 'r' ) as f: lowerCamelCase_ = f.readlines() with open(A_ , 'w' ) as f: for line in lines: if "# Copied from transformers." not in line: f.write(A_ ) if output_pytorch: if not self._testing: remove_copy_lines(f"""{directory}/modeling_{lowercase_model_name}.py""" ) shutil.move( f"""{directory}/modeling_{lowercase_model_name}.py""" , f"""{model_dir}/modeling_{lowercase_model_name}.py""" , ) shutil.move( f"""{directory}/test_modeling_{lowercase_model_name}.py""" , f"""{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_{lowercase_model_name}.py""" , ) else: os.remove(f"""{directory}/modeling_{lowercase_model_name}.py""" ) os.remove(f"""{directory}/test_modeling_{lowercase_model_name}.py""" ) if output_tensorflow: if not self._testing: remove_copy_lines(f"""{directory}/modeling_tf_{lowercase_model_name}.py""" ) shutil.move( f"""{directory}/modeling_tf_{lowercase_model_name}.py""" , f"""{model_dir}/modeling_tf_{lowercase_model_name}.py""" , ) shutil.move( f"""{directory}/test_modeling_tf_{lowercase_model_name}.py""" , f"""{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_tf_{lowercase_model_name}.py""" , ) else: os.remove(f"""{directory}/modeling_tf_{lowercase_model_name}.py""" ) os.remove(f"""{directory}/test_modeling_tf_{lowercase_model_name}.py""" ) if output_flax: if not self._testing: remove_copy_lines(f"""{directory}/modeling_flax_{lowercase_model_name}.py""" ) shutil.move( f"""{directory}/modeling_flax_{lowercase_model_name}.py""" , f"""{model_dir}/modeling_flax_{lowercase_model_name}.py""" , ) shutil.move( f"""{directory}/test_modeling_flax_{lowercase_model_name}.py""" , f"""{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_flax_{lowercase_model_name}.py""" , ) else: os.remove(f"""{directory}/modeling_flax_{lowercase_model_name}.py""" ) os.remove(f"""{directory}/test_modeling_flax_{lowercase_model_name}.py""" ) shutil.move( f"""{directory}/{lowercase_model_name}.md""" , f"""{path_to_transformer_root}/docs/source/en/model_doc/{lowercase_model_name}.md""" , ) shutil.move( f"""{directory}/tokenization_{lowercase_model_name}.py""" , f"""{model_dir}/tokenization_{lowercase_model_name}.py""" , ) shutil.move( f"""{directory}/tokenization_fast_{lowercase_model_name}.py""" , f"""{model_dir}/tokenization_{lowercase_model_name}_fast.py""" , ) from os import fdopen, remove from shutil import copymode, move from tempfile import mkstemp def replace(A_ : str , A_ : str , A_ : List[str] ): # Create temp file lowerCamelCase_ , lowerCamelCase_ = mkstemp() lowerCamelCase_ = False with fdopen(A_ , 'w' ) as new_file: with open(A_ ) as old_file: for line in old_file: new_file.write(A_ ) if line_to_copy_below in line: lowerCamelCase_ = True for line_to_copy in lines_to_copy: new_file.write(A_ ) if not line_found: raise ValueError(f"""Line {line_to_copy_below} was not found in file.""" ) # Copy the file permissions from the old file to the new file copymode(A_ , A_ ) # Remove original file remove(A_ ) # Move new file move(A_ , A_ ) def skip_units(A_ : Union[str, Any] ): return ( ("generating PyTorch" in line and not output_pytorch) or ("generating TensorFlow" in line and not output_tensorflow) or ("generating Flax" in line and not output_flax) ) def replace_in_files(A_ : str ): with open(A_ ) as datafile: lowerCamelCase_ = [] lowerCamelCase_ = False lowerCamelCase_ = False for line in datafile: if "# To replace in: " in line and "##" not in line: lowerCamelCase_ = line.split('"' )[1] lowerCamelCase_ = skip_units(A_ ) elif "# Below: " in line and "##" not in line: lowerCamelCase_ = line.split('"' )[1] lowerCamelCase_ = skip_units(A_ ) elif "# End." in line and "##" not in line: if not skip_file and not skip_snippet: replace(A_ , A_ , A_ ) lowerCamelCase_ = [] elif "# Replace with" in line and "##" not in line: lowerCamelCase_ = [] elif "##" not in line: lines_to_copy.append(A_ ) remove(A_ ) replace_in_files(f"""{directory}/to_replace_{lowercase_model_name}.py""" ) os.rmdir(A_ )
359
import math def _SCREAMING_SNAKE_CASE ( lowercase : float , lowercase : float ): '''simple docstring''' return math.pow(lowercase , 2 ) - a def _SCREAMING_SNAKE_CASE ( lowercase : float ): '''simple docstring''' return 2 * x def _SCREAMING_SNAKE_CASE ( lowercase : float ): '''simple docstring''' lowerCamelCase_ = 2.0 while start <= a: lowerCamelCase_ = math.pow(lowercase , 2 ) return start def _SCREAMING_SNAKE_CASE ( lowercase : float , lowercase : int = 99_99 , lowercase : float = 0.00_0000_0000_0001 ): '''simple docstring''' if a < 0: raise ValueError('math domain error' ) lowerCamelCase_ = get_initial_point(lowercase ) for _ in range(lowercase ): lowerCamelCase_ = value lowerCamelCase_ = value - fx(lowercase , lowercase ) / fx_derivative(lowercase ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
208
0
'''simple docstring''' import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class A_ ( _snake_case , _snake_case , _snake_case ): '''simple docstring''' @register_to_config def __init__( self : Union[str, Any] , lowercase_ : int , lowercase_ : int , lowercase_ : int , lowercase_ : float , lowercase_ : int , lowercase_ : int , lowercase_ : int , lowercase_ : int , lowercase_ : str , lowercase_ : bool = False , ) -> Any: super().__init__() UpperCAmelCase : Dict = nn.Embedding(lowercase_ , lowercase_ ) UpperCAmelCase : List[Any] = nn.Embedding(lowercase_ , lowercase_ ) UpperCAmelCase : int = False UpperCAmelCase : Optional[int] = nn.Dropout(p=lowercase_ ) UpperCAmelCase : Union[str, Any] = TaConfig( vocab_size=lowercase_ , d_model=lowercase_ , num_heads=lowercase_ , d_kv=lowercase_ , d_ff=lowercase_ , dropout_rate=lowercase_ , feed_forward_proj=lowercase_ , is_decoder=lowercase_ , is_encoder_decoder=lowercase_ , ) UpperCAmelCase : Optional[Any] = nn.ModuleList() for lyr_num in range(lowercase_ ): UpperCAmelCase : Union[str, Any] = TaBlock(lowercase_ ) self.encoders.append(lowercase_ ) UpperCAmelCase : Optional[int] = TaLayerNorm(lowercase_ ) UpperCAmelCase : List[str] = nn.Dropout(p=lowercase_ ) def UpperCAmelCase_ ( self : List[str] , lowercase_ : Tuple , lowercase_ : List[Any] ) -> Dict: UpperCAmelCase : Dict = self.token_embedder(lowercase_ ) UpperCAmelCase : List[Any] = encoder_input_tokens.shape[1] UpperCAmelCase : List[str] = torch.arange(lowercase_ , device=encoder_input_tokens.device ) x += self.position_encoding(lowercase_ ) UpperCAmelCase : int = self.dropout_pre(lowercase_ ) # inverted the attention mask UpperCAmelCase : int = encoder_input_tokens.size() UpperCAmelCase : int = self.get_extended_attention_mask(lowercase_ , lowercase_ ) for lyr in self.encoders: UpperCAmelCase : Optional[int] = lyr(lowercase_ , lowercase_ )[0] UpperCAmelCase : Tuple = self.layer_norm(lowercase_ ) return self.dropout_post(lowercase_ ), encoder_inputs_mask
151
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() lowercase__ = logging.get_logger(__name__) lowercase__ = { "post_extract_proj": "feature_projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.upsample.0": "encoder.upsample.projection", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "layer_norm", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): for attribute in key.split('.' ): UpperCAmelCase : Any = getattr(UpperCAmelCase_ , UpperCAmelCase_ ) if weight_type is not None: UpperCAmelCase : Tuple = getattr(UpperCAmelCase_ , UpperCAmelCase_ ).shape else: UpperCAmelCase : str = hf_pointer.shape assert hf_shape == value.shape, ( F"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": UpperCAmelCase : Union[str, Any] = value elif weight_type == "weight_g": UpperCAmelCase : int = value elif weight_type == "weight_v": UpperCAmelCase : str = value elif weight_type == "bias": UpperCAmelCase : List[Any] = value else: UpperCAmelCase : List[str] = value logger.info(F"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): UpperCAmelCase : str = [] UpperCAmelCase : Optional[int] = fairseq_model.state_dict() UpperCAmelCase : Dict = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase : Dict = False if "conv_layers" in name: load_conv_layer( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , hf_model.config.feat_extract_norm == 'group' , ) UpperCAmelCase : Optional[Any] = True else: for key, mapped_key in MAPPING.items(): UpperCAmelCase : int = 'sew.' + mapped_key if (is_finetuned and mapped_key != 'lm_head') else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: UpperCAmelCase : Any = True if "*" in mapped_key: UpperCAmelCase : Any = name.split(UpperCAmelCase_ )[0].split('.' )[-2] UpperCAmelCase : int = mapped_key.replace('*' , UpperCAmelCase_ ) if "weight_g" in name: UpperCAmelCase : Optional[Any] = 'weight_g' elif "weight_v" in name: UpperCAmelCase : Optional[int] = 'weight_v' elif "weight" in name: UpperCAmelCase : Optional[Any] = 'weight' elif "bias" in name: UpperCAmelCase : str = 'bias' else: UpperCAmelCase : str = None set_recursively(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) continue if not is_used: unused_weights.append(UpperCAmelCase_ ) logger.warning(F"""Unused weights: {unused_weights}""" ) def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): UpperCAmelCase : List[Any] = full_name.split('conv_layers.' )[-1] UpperCAmelCase : Optional[Any] = name.split('.' ) UpperCAmelCase : Optional[Any] = int(items[0] ) UpperCAmelCase : List[str] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) UpperCAmelCase : Union[str, Any] = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) UpperCAmelCase : str = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) UpperCAmelCase : Union[str, Any] = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) UpperCAmelCase : str = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(UpperCAmelCase_ ) def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ ): UpperCAmelCase : Optional[Any] = SEWConfig() if is_finetuned: UpperCAmelCase : List[str] = model.wav_encoder.wav_model.cfg else: UpperCAmelCase : Optional[Any] = model.cfg UpperCAmelCase : str = fs_config.conv_bias UpperCAmelCase : Optional[Any] = eval(fs_config.conv_feature_layers ) UpperCAmelCase : Optional[Any] = [x[0] for x in conv_layers] UpperCAmelCase : str = [x[1] for x in conv_layers] UpperCAmelCase : str = [x[2] for x in conv_layers] UpperCAmelCase : Tuple = 'gelu' UpperCAmelCase : List[str] = 'layer' if fs_config.extractor_mode == 'layer_norm' else 'group' UpperCAmelCase : List[Any] = 0.0 UpperCAmelCase : Optional[int] = fs_config.activation_fn.name UpperCAmelCase : Tuple = fs_config.encoder_embed_dim UpperCAmelCase : List[str] = 0.02 UpperCAmelCase : Any = fs_config.encoder_ffn_embed_dim UpperCAmelCase : Any = 1E-5 UpperCAmelCase : Any = fs_config.encoder_layerdrop UpperCAmelCase : List[str] = fs_config.encoder_attention_heads UpperCAmelCase : Union[str, Any] = fs_config.conv_pos_groups UpperCAmelCase : str = fs_config.conv_pos UpperCAmelCase : Union[str, Any] = len(UpperCAmelCase_ ) UpperCAmelCase : List[str] = fs_config.encoder_layers UpperCAmelCase : Any = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: UpperCAmelCase : List[Any] = model.cfg UpperCAmelCase : Tuple = fs_config.final_dropout UpperCAmelCase : Tuple = fs_config.layerdrop UpperCAmelCase : int = fs_config.activation_dropout UpperCAmelCase : Union[str, Any] = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 UpperCAmelCase : str = fs_config.attention_dropout UpperCAmelCase : Optional[Any] = fs_config.dropout_input UpperCAmelCase : Optional[int] = fs_config.dropout UpperCAmelCase : str = fs_config.mask_channel_length UpperCAmelCase : Optional[Any] = fs_config.mask_channel_prob UpperCAmelCase : Any = fs_config.mask_length UpperCAmelCase : int = fs_config.mask_prob UpperCAmelCase : Optional[Any] = 'Wav2Vec2FeatureExtractor' UpperCAmelCase : Tuple = 'Wav2Vec2CTCTokenizer' return config @torch.no_grad() def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_=None , UpperCAmelCase_=None , UpperCAmelCase_=True ): if is_finetuned: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Optional[Any] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Optional[int] = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: UpperCAmelCase : List[str] = SEWConfig.from_pretrained(UpperCAmelCase_ ) else: UpperCAmelCase : List[Any] = convert_config(model[0] , UpperCAmelCase_ ) UpperCAmelCase : int = model[0].eval() UpperCAmelCase : Tuple = True if config.feat_extract_norm == 'layer' else False UpperCAmelCase : Optional[int] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0 , do_normalize=UpperCAmelCase_ , return_attention_mask=UpperCAmelCase_ , ) if is_finetuned: if dict_path: UpperCAmelCase : Optional[Any] = Dictionary.load(UpperCAmelCase_ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq UpperCAmelCase : List[Any] = target_dict.pad_index UpperCAmelCase : Optional[Any] = target_dict.bos_index UpperCAmelCase : int = target_dict.pad_index UpperCAmelCase : Tuple = target_dict.bos_index UpperCAmelCase : int = target_dict.eos_index UpperCAmelCase : Optional[int] = len(target_dict.symbols ) UpperCAmelCase : Union[str, Any] = os.path.join(UpperCAmelCase_ , 'vocab.json' ) if not os.path.isdir(UpperCAmelCase_ ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(UpperCAmelCase_ ) ) return os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) with open(UpperCAmelCase_ , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(target_dict.indices , UpperCAmelCase_ ) UpperCAmelCase : Optional[Any] = WavaVecaCTCTokenizer( UpperCAmelCase_ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=UpperCAmelCase_ , ) UpperCAmelCase : Union[str, Any] = WavaVecaProcessor(feature_extractor=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ ) processor.save_pretrained(UpperCAmelCase_ ) UpperCAmelCase : List[str] = SEWForCTC(UpperCAmelCase_ ) else: UpperCAmelCase : Tuple = SEWModel(UpperCAmelCase_ ) feature_extractor.save_pretrained(UpperCAmelCase_ ) recursively_load_weights(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) hf_model.save_pretrained(UpperCAmelCase_ ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--is_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) lowercase__ = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
151
1
"""simple docstring""" import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets __A = datasets.logging.get_logger(__name__) __A = '\\n@InProceedings{moosavi2019minimum,\n author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube},\n title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection},\n year = {2019},\n booktitle = {Proceedings of the 57th Annual Meeting of\n the Association for Computational Linguistics (Volume 1: Long Papers)},\n publisher = {Association for Computational Linguistics},\n address = {Florence, Italy},\n}\n\n@inproceedings{10.3115/1072399.1072405,\nauthor = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette},\ntitle = {A Model-Theoretic Coreference Scoring Scheme},\nyear = {1995},\nisbn = {1558604022},\npublisher = {Association for Computational Linguistics},\naddress = {USA},\nurl = {https://doi.org/10.3115/1072399.1072405},\ndoi = {10.3115/1072399.1072405},\nbooktitle = {Proceedings of the 6th Conference on Message Understanding},\npages = {45–52},\nnumpages = {8},\nlocation = {Columbia, Maryland},\nseries = {MUC6 ’95}\n}\n\n@INPROCEEDINGS{Bagga98algorithmsfor,\n author = {Amit Bagga and Breck Baldwin},\n title = {Algorithms for Scoring Coreference Chains},\n booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference},\n year = {1998},\n pages = {563--566}\n}\n\n@INPROCEEDINGS{Luo05oncoreference,\n author = {Xiaoqiang Luo},\n title = {On coreference resolution performance metrics},\n booktitle = {In Proc. of HLT/EMNLP},\n year = {2005},\n pages = {25--32},\n publisher = {URL}\n}\n\n@inproceedings{moosavi-strube-2016-coreference,\n title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric",\n author = "Moosavi, Nafise Sadat and\n Strube, Michael",\n booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",\n month = aug,\n year = "2016",\n address = "Berlin, Germany",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/P16-1060",\n doi = "10.18653/v1/P16-1060",\n pages = "632--642",\n}\n\n' __A = '\\nCoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which\nimplements of the common evaluation metrics including MUC [Vilain et al, 1995],\nB-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005],\nLEA [Moosavi and Strube, 2016] and the averaged CoNLL score\n(the average of the F1 values of MUC, B-cubed and CEAFe)\n[Denis and Baldridge, 2009a; Pradhan et al., 2011].\n\nThis wrapper of CoVal currently only work with CoNLL line format:\nThe CoNLL format has one word per line with all the annotation for this word in column separated by spaces:\nColumn Type Description\n1 Document ID This is a variation on the document filename\n2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n3 Word number\n4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release.\n5 Part-of-Speech\n6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column.\n7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-"\n8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7.\n9 Word sense This is the word sense of the word in Column 3.\n10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data.\n11 Named Entities These columns identifies the spans representing various named entities.\n12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7.\nN Coreference Coreference chain information encoded in a parenthesis structure.\nMore informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html\n\nDetails on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md\n\nCoVal code was written by @ns-moosavi.\nSome parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py\nThe test suite is taken from https://github.com/conll/reference-coreference-scorers/\nMention evaluation and the test suite are added by @andreasvc.\nParsing CoNLL files is developed by Leo Born.\n' __A = '\nCalculates coreference evaluation metrics.\nArgs:\n predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format.\n Each prediction is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format.\n Each reference is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n keep_singletons: After extracting all mentions of key or system files,\n mentions whose corresponding coreference chain is of size one,\n are considered as singletons. The default evaluation mode will include\n singletons in evaluations if they are included in the key or the system files.\n By setting \'keep_singletons=False\', all singletons in the key and system files\n will be excluded from the evaluation.\n NP_only: Most of the recent coreference resolvers only resolve NP mentions and\n leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs.\n min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans.\n Minimum spans are determined using the MINA algorithm.\n\nReturns:\n \'mentions\': mentions\n \'muc\': MUC metric [Vilain et al, 1995]\n \'bcub\': B-cubed [Bagga and Baldwin, 1998]\n \'ceafe\': CEAFe [Luo et al., 2005]\n \'lea\': LEA [Moosavi and Strube, 2016]\n \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe)\n\nExamples:\n\n >>> coval = datasets.load_metric(\'coval\')\n >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\',\n ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\',\n ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\',\n ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\',\n ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\',\n ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\']\n >>> references = [words]\n >>> predictions = [words]\n >>> results = coval.compute(predictions=predictions, references=references)\n >>> print(results) # doctest:+ELLIPSIS\n {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0}\n' def _lowerCamelCase(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase=False , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=False , __UpperCamelCase="dummy_doc" ) -> Dict: _lowerCAmelCase ={doc: key_lines} _lowerCAmelCase ={doc: sys_lines} _lowerCAmelCase ={} _lowerCAmelCase =0 _lowerCAmelCase =0 _lowerCAmelCase =0 _lowerCAmelCase =0 _lowerCAmelCase =0 _lowerCAmelCase =0 _lowerCAmelCase , _lowerCAmelCase =reader.get_doc_mentions(__UpperCamelCase , key_doc_lines[doc] , __UpperCamelCase ) key_singletons_num += singletons_num if NP_only or min_span: _lowerCAmelCase =reader.set_annotated_parse_trees(__UpperCamelCase , key_doc_lines[doc] , __UpperCamelCase , __UpperCamelCase ) _lowerCAmelCase , _lowerCAmelCase =reader.get_doc_mentions(__UpperCamelCase , sys_doc_lines[doc] , __UpperCamelCase ) sys_singletons_num += singletons_num if NP_only or min_span: _lowerCAmelCase =reader.set_annotated_parse_trees(__UpperCamelCase , key_doc_lines[doc] , __UpperCamelCase , __UpperCamelCase ) if remove_nested: _lowerCAmelCase , _lowerCAmelCase =reader.remove_nested_coref_mentions(__UpperCamelCase , __UpperCamelCase ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters _lowerCAmelCase , _lowerCAmelCase =reader.remove_nested_coref_mentions(__UpperCamelCase , __UpperCamelCase ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters _lowerCAmelCase =reader.get_mention_assignments(__UpperCamelCase , __UpperCamelCase ) _lowerCAmelCase =reader.get_mention_assignments(__UpperCamelCase , __UpperCamelCase ) _lowerCAmelCase =(key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ F'''annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}''' ) logger.info( """Number of resulting singleton clusters in the key """ F'''annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}''' ) if not keep_singletons: logger.info( F'''{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ''' """files, respectively""" ) return doc_coref_infos def _lowerCamelCase(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) -> int: _lowerCAmelCase =get_coref_infos(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) _lowerCAmelCase ={} _lowerCAmelCase =0 _lowerCAmelCase =0 for name, metric in metrics: _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase =evaluator.evaluate_documents(__UpperCamelCase , __UpperCamelCase , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F'''{name}/recall''': recall, F'''{name}/precision''': precision, F'''{name}/f1''': fa} ) logger.info( name.ljust(10 ) , F'''Recall: {recall * 100:.2f}''' , F''' Precision: {precision * 100:.2f}''' , F''' F1: {fa * 100:.2f}''' , ) if conll_subparts_num == 3: _lowerCAmelCase =(conll / 3) * 100 logger.info(F'''CoNLL score: {conll:.2f}''' ) output_scores.update({"""conll_score""": conll} ) return output_scores def _lowerCamelCase(__UpperCamelCase ) -> Tuple: _lowerCAmelCase =False for line in key_lines: if not line.startswith("""#""" ): if len(line.split() ) > 6: _lowerCAmelCase =line.split()[5] if not parse_col == "-": _lowerCAmelCase =True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCamelCase__ ( datasets.Metric ): '''simple docstring''' def _lowerCAmelCase ( self ) -> str: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=False , __UpperCAmelCase=False ) -> Optional[Any]: _lowerCAmelCase =[ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: _lowerCAmelCase =util.check_gold_parse_annotation(__UpperCAmelCase ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use 'min_span'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" _lowerCAmelCase =evaluate( key_lines=__UpperCAmelCase , sys_lines=__UpperCAmelCase , metrics=__UpperCAmelCase , NP_only=__UpperCAmelCase , remove_nested=__UpperCAmelCase , keep_singletons=__UpperCAmelCase , min_span=__UpperCAmelCase , ) return score
368
"""simple docstring""" import warnings from .generation import TFGenerationMixin class lowerCamelCase__ ( __magic_name__ ): '''simple docstring''' # warning at import time warnings.warn( '''Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will ''' '''be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.''' , __magic_name__ , )
341
0
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' lowercase_ = StableDiffusionXLImgaImgPipeline lowercase_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"height", "width"} lowercase_ = PipelineTesterMixin.required_optional_params - {"latents"} lowercase_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowercase_ = IMAGE_TO_IMAGE_IMAGE_PARAMS lowercase_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def SCREAMING_SNAKE_CASE_ (self : Union[str, Any]) ->Tuple: '''simple docstring''' torch.manual_seed(0) lowerCamelCase__: Tuple =UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , attention_head_dim=(2, 4) , use_linear_projection=UpperCAmelCase_ , addition_embed_type="text_time" , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) lowerCamelCase__: Tuple =EulerDiscreteScheduler( beta_start=0.0_0085 , beta_end=0.012 , steps_offset=1 , beta_schedule="scaled_linear" , timestep_spacing="leading" , ) torch.manual_seed(0) lowerCamelCase__: Tuple =AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0) lowerCamelCase__: Union[str, Any] =CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="gelu" , projection_dim=32 , ) lowerCamelCase__: Optional[Any] =CLIPTextModel(UpperCAmelCase_) lowerCamelCase__: List[str] =CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" , local_files_only=UpperCAmelCase_) lowerCamelCase__: Union[str, Any] =CLIPTextModelWithProjection(UpperCAmelCase_) lowerCamelCase__: Optional[int] =CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" , local_files_only=UpperCAmelCase_) lowerCamelCase__: int ={ "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_encoder_2": text_encoder_a, "tokenizer_2": tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def SCREAMING_SNAKE_CASE_ (self : Tuple , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Optional[int]=0) ->List[Any]: '''simple docstring''' lowerCamelCase__: Optional[Any] =floats_tensor((1, 3, 32, 32) , rng=random.Random(UpperCAmelCase_)).to(UpperCAmelCase_) lowerCamelCase__: Any =image / 2 + 0.5 if str(UpperCAmelCase_).startswith("mps"): lowerCamelCase__: str =torch.manual_seed(UpperCAmelCase_) else: lowerCamelCase__: List[str] =torch.Generator(device=UpperCAmelCase_).manual_seed(UpperCAmelCase_) lowerCamelCase__: Any ={ "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 2, "guidance_scale": 5.0, "output_type": "numpy", "strength": 0.75, } return inputs def SCREAMING_SNAKE_CASE_ (self : List[str]) ->Dict: '''simple docstring''' lowerCamelCase__: str ="cpu" # ensure determinism for the device-dependent torch.Generator lowerCamelCase__: List[str] =self.get_dummy_components() lowerCamelCase__: Union[str, Any] =StableDiffusionXLImgaImgPipeline(**UpperCAmelCase_) lowerCamelCase__: Dict =sd_pipe.to(UpperCAmelCase_) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase_) lowerCamelCase__: Dict =self.get_dummy_inputs(UpperCAmelCase_) lowerCamelCase__: Union[str, Any] =sd_pipe(**UpperCAmelCase_).images lowerCamelCase__: int =image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCamelCase__: List[Any] =np.array([0.4656, 0.4840, 0.4439, 0.6698, 0.5574, 0.4524, 0.5799, 0.5943, 0.5165]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def SCREAMING_SNAKE_CASE_ (self : Tuple) ->List[str]: '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=3E-3) def SCREAMING_SNAKE_CASE_ (self : Union[str, Any]) ->Optional[int]: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3) def SCREAMING_SNAKE_CASE_ (self : List[str]) ->List[Any]: '''simple docstring''' pass def SCREAMING_SNAKE_CASE_ (self : List[Any]) ->Dict: '''simple docstring''' lowerCamelCase__: Optional[int] =self.get_dummy_components() lowerCamelCase__: Dict =StableDiffusionXLImgaImgPipeline(**UpperCAmelCase_) lowerCamelCase__: str =sd_pipe.to(UpperCAmelCase_) lowerCamelCase__: List[Any] =sd_pipe.to(UpperCAmelCase_) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase_) # forward without prompt embeds lowerCamelCase__: int =self.get_dummy_inputs(UpperCAmelCase_) lowerCamelCase__: List[Any] =3 * ["this is a negative prompt"] lowerCamelCase__: Tuple =negative_prompt lowerCamelCase__: int =3 * [inputs["prompt"]] lowerCamelCase__: Tuple =sd_pipe(**UpperCAmelCase_) lowerCamelCase__: Tuple =output.images[0, -3:, -3:, -1] # forward with prompt embeds lowerCamelCase__: Union[str, Any] =self.get_dummy_inputs(UpperCAmelCase_) lowerCamelCase__: Dict =3 * ["this is a negative prompt"] lowerCamelCase__: Any =3 * [inputs.pop("prompt")] ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ): Tuple =sd_pipe.encode_prompt(UpperCAmelCase_ , negative_prompt=UpperCAmelCase_) lowerCamelCase__: int =sd_pipe( **UpperCAmelCase_ , prompt_embeds=UpperCAmelCase_ , negative_prompt_embeds=UpperCAmelCase_ , pooled_prompt_embeds=UpperCAmelCase_ , negative_pooled_prompt_embeds=UpperCAmelCase_ , ) lowerCamelCase__: Optional[int] =output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten()).max() < 1E-4 @slow @require_torch_gpu class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE_ (self : List[str]) ->Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE_ (self : Optional[int] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[Any]="cpu" , UpperCAmelCase_ : Optional[int]=torch.floataa , UpperCAmelCase_ : Any=0) ->Tuple: '''simple docstring''' lowerCamelCase__: Optional[int] =torch.Generator(device=UpperCAmelCase_).manual_seed(UpperCAmelCase_) lowerCamelCase__: Union[str, Any] =np.random.RandomState(UpperCAmelCase_).standard_normal((1, 4, 64, 64)) lowerCamelCase__: Union[str, Any] =torch.from_numpy(UpperCAmelCase_).to(device=UpperCAmelCase_ , dtype=UpperCAmelCase_) lowerCamelCase__: Any ={ "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def SCREAMING_SNAKE_CASE_ (self : int) ->List[str]: '''simple docstring''' lowerCamelCase__: List[str] =DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-2-base") pipe.to(UpperCAmelCase_) pipe.set_progress_bar_config(disable=UpperCAmelCase_) lowerCamelCase__: int =self.get_inputs(UpperCAmelCase_) lowerCamelCase__: str =pipe(**UpperCAmelCase_).images lowerCamelCase__: Tuple =image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) lowerCamelCase__: int =np.array([0.4_9493, 0.4_7896, 0.4_0798, 0.5_4214, 0.5_3212, 0.4_8202, 0.4_7656, 0.4_6329, 0.4_8506]) assert np.abs(image_slice - expected_slice).max() < 7E-3
10
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import PoolFormerImageProcessor class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def __init__(self : List[Any] , UpperCAmelCase_ : Any , UpperCAmelCase_ : Dict=7 , UpperCAmelCase_ : List[str]=3 , UpperCAmelCase_ : str=30 , UpperCAmelCase_ : List[str]=400 , UpperCAmelCase_ : Optional[int]=True , UpperCAmelCase_ : Optional[int]=None , UpperCAmelCase_ : Tuple=0.9 , UpperCAmelCase_ : str=None , UpperCAmelCase_ : int=True , UpperCAmelCase_ : Union[str, Any]=[0.5, 0.5, 0.5] , UpperCAmelCase_ : Optional[Any]=[0.5, 0.5, 0.5] , ) ->str: '''simple docstring''' lowerCamelCase__: List[Any] =size if size is not None else {"shortest_edge": 30} lowerCamelCase__: Dict =crop_size if crop_size is not None else {"height": 30, "width": 30} lowerCamelCase__: Any =parent lowerCamelCase__: Any =batch_size lowerCamelCase__: Optional[Any] =num_channels lowerCamelCase__: Tuple =min_resolution lowerCamelCase__: Union[str, Any] =max_resolution lowerCamelCase__: Union[str, Any] =do_resize_and_center_crop lowerCamelCase__: Optional[int] =size lowerCamelCase__: str =crop_pct lowerCamelCase__: Any =crop_size lowerCamelCase__: List[str] =do_normalize lowerCamelCase__: List[str] =image_mean lowerCamelCase__: Tuple =image_std def SCREAMING_SNAKE_CASE_ (self : List[Any]) ->Optional[int]: '''simple docstring''' return { "size": self.size, "do_resize_and_center_crop": self.do_resize_and_center_crop, "crop_pct": self.crop_pct, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' lowercase_ = PoolFormerImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE_ (self : Optional[int]) ->Optional[Any]: '''simple docstring''' lowerCamelCase__: Optional[int] =PoolFormerImageProcessingTester(self) @property def SCREAMING_SNAKE_CASE_ (self : str) ->int: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE_ (self : Optional[Any]) ->Optional[Any]: '''simple docstring''' lowerCamelCase__: Any =self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(UpperCAmelCase_ , "do_resize_and_center_crop")) self.assertTrue(hasattr(UpperCAmelCase_ , "size")) self.assertTrue(hasattr(UpperCAmelCase_ , "crop_pct")) self.assertTrue(hasattr(UpperCAmelCase_ , "do_normalize")) self.assertTrue(hasattr(UpperCAmelCase_ , "image_mean")) self.assertTrue(hasattr(UpperCAmelCase_ , "image_std")) def SCREAMING_SNAKE_CASE_ (self : Any) ->List[str]: '''simple docstring''' lowerCamelCase__: List[str] =self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {"shortest_edge": 30}) self.assertEqual(image_processor.crop_size , {"height": 30, "width": 30}) lowerCamelCase__: Union[str, Any] =self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84) self.assertEqual(image_processor.size , {"shortest_edge": 42}) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84}) def SCREAMING_SNAKE_CASE_ (self : int) ->Optional[Any]: '''simple docstring''' pass def SCREAMING_SNAKE_CASE_ (self : Optional[Any]) ->Any: '''simple docstring''' lowerCamelCase__: Union[str, Any] =self.image_processing_class(**self.image_processor_dict) # create random PIL images lowerCamelCase__: Union[str, Any] =prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase_) for image in image_inputs: self.assertIsInstance(UpperCAmelCase_ , Image.Image) # Test not batched input lowerCamelCase__: Dict =image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched lowerCamelCase__: int =image_processing(UpperCAmelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def SCREAMING_SNAKE_CASE_ (self : Optional[Any]) ->Dict: '''simple docstring''' lowerCamelCase__: Any =self.image_processing_class(**self.image_processor_dict) # create random numpy tensors lowerCamelCase__: Tuple =prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase_ , numpify=UpperCAmelCase_) for image in image_inputs: self.assertIsInstance(UpperCAmelCase_ , np.ndarray) # Test not batched input lowerCamelCase__: Union[str, Any] =image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched lowerCamelCase__: List[str] =image_processing(UpperCAmelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def SCREAMING_SNAKE_CASE_ (self : Union[str, Any]) ->Any: '''simple docstring''' lowerCamelCase__: Optional[int] =self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors lowerCamelCase__: Any =prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase_ , torchify=UpperCAmelCase_) for image in image_inputs: self.assertIsInstance(UpperCAmelCase_ , torch.Tensor) # Test not batched input lowerCamelCase__: Any =image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched lowerCamelCase__: str =image_processing(UpperCAmelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
10
1
"""simple docstring""" import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class __lowerCamelCase : '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase=99 , __UpperCAmelCase=32 , __UpperCAmelCase=5 , __UpperCAmelCase=4 , __UpperCAmelCase=37 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=16 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=3 , __UpperCAmelCase=4 , __UpperCAmelCase=None , ) -> List[str]: _a = parent _a = batch_size _a = seq_length _a = is_training _a = use_input_mask _a = use_token_type_ids _a = use_labels _a = vocab_size _a = hidden_size _a = num_hidden_layers _a = num_attention_heads _a = intermediate_size _a = hidden_act _a = hidden_dropout_prob _a = attention_probs_dropout_prob _a = max_position_embeddings _a = type_vocab_size _a = type_sequence_label_size _a = initializer_range _a = num_labels _a = num_choices _a = scope def _UpperCAmelCase ( self ) -> Optional[Any]: _a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _a = None if self.use_input_mask: _a = random_attention_mask([self.batch_size, self.seq_length] ) _a = None if self.use_token_type_ids: _a = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _a = None _a = None _a = None if self.use_labels: _a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _a = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _a = ids_tensor([self.batch_size] , self.num_choices ) _a = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCAmelCase ( self ) -> Optional[Any]: return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__UpperCAmelCase , initializer_range=self.initializer_range , ) def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Any: _a = LlamaModel(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() _a = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase ) _a = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ) -> List[Any]: _a = True _a = LlamaModel(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() _a = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , ) _a = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , ) _a = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ) -> Optional[Any]: _a = LlamaForCausalLM(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() _a = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ) -> List[str]: _a = True _a = True _a = LlamaForCausalLM(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() # first forward pass _a = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , use_cache=__UpperCAmelCase , ) _a = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids _a = ids_tensor((self.batch_size, 3) , config.vocab_size ) _a = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and _a = torch.cat([input_ids, next_tokens] , dim=-1 ) _a = torch.cat([input_mask, next_mask] , dim=-1 ) _a = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , output_hidden_states=__UpperCAmelCase , )['''hidden_states'''][0] _a = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , past_key_values=__UpperCAmelCase , output_hidden_states=__UpperCAmelCase , )['''hidden_states'''][0] # select random slice _a = ids_tensor((1,) , output_from_past.shape[-1] ).item() _a = output_from_no_past[:, -3:, random_slice_idx].detach() _a = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1e-3 ) ) def _UpperCAmelCase ( self ) -> int: _a = self.prepare_config_and_inputs() ( ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ) = config_and_inputs _a = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __lowerCamelCase ( a__ , a__ , a__ , unittest.TestCase ): '''simple docstring''' A_ : int = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () A_ : List[str] = (LlamaForCausalLM,) if is_torch_available() else () A_ : Union[str, Any] = ( { 'feature-extraction': LlamaModel, 'text-classification': LlamaForSequenceClassification, 'text-generation': LlamaForCausalLM, 'zero-shot': LlamaForSequenceClassification, } if is_torch_available() else {} ) A_ : Tuple = False A_ : Tuple = False def _UpperCAmelCase ( self ) -> Optional[Any]: _a = LlamaModelTester(self ) _a = ConfigTester(self , config_class=__UpperCAmelCase , hidden_size=37 ) def _UpperCAmelCase ( self ) -> Any: self.config_tester.run_common_tests() def _UpperCAmelCase ( self ) -> Any: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def _UpperCAmelCase ( self ) -> List[str]: _a = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _a = type self.model_tester.create_and_check_model(*__UpperCAmelCase ) def _UpperCAmelCase ( self ) -> List[str]: _a , _a = self.model_tester.prepare_config_and_inputs_for_common() _a = 3 _a = input_dict['''input_ids'''] _a = input_ids.ne(1 ).to(__UpperCAmelCase ) _a = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _a = LlamaForSequenceClassification(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() _a = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , labels=__UpperCAmelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _UpperCAmelCase ( self ) -> str: _a , _a = self.model_tester.prepare_config_and_inputs_for_common() _a = 3 _a = '''single_label_classification''' _a = input_dict['''input_ids'''] _a = input_ids.ne(1 ).to(__UpperCAmelCase ) _a = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _a = LlamaForSequenceClassification(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() _a = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , labels=__UpperCAmelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _UpperCAmelCase ( self ) -> Optional[Any]: _a , _a = self.model_tester.prepare_config_and_inputs_for_common() _a = 3 _a = '''multi_label_classification''' _a = input_dict['''input_ids'''] _a = input_ids.ne(1 ).to(__UpperCAmelCase ) _a = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) _a = LlamaForSequenceClassification(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() _a = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , labels=__UpperCAmelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip('''LLaMA buffers include complex numbers, which breaks this test''' ) def _UpperCAmelCase ( self ) -> int: pass @parameterized.expand([('''linear''',), ('''dynamic''',)] ) def _UpperCAmelCase ( self , __UpperCAmelCase ) -> int: _a , _a = self.model_tester.prepare_config_and_inputs_for_common() _a = ids_tensor([1, 10] , config.vocab_size ) _a = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights _a = LlamaModel(__UpperCAmelCase ) original_model.to(__UpperCAmelCase ) original_model.eval() _a = original_model(__UpperCAmelCase ).last_hidden_state _a = original_model(__UpperCAmelCase ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights _a = {'''type''': scaling_type, '''factor''': 10.0} _a = LlamaModel(__UpperCAmelCase ) scaled_model.to(__UpperCAmelCase ) scaled_model.eval() _a = scaled_model(__UpperCAmelCase ).last_hidden_state _a = scaled_model(__UpperCAmelCase ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1e-5 ) ) @require_torch class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @unittest.skip('''Logits are not exactly the same, once we fix the instabalities somehow, will update!''' ) @slow def _UpperCAmelCase ( self ) -> Union[str, Any]: _a = [1, 306, 4658, 278, 6593, 310, 2834, 338] _a = LlamaForCausalLM.from_pretrained('''meta-llama/Llama-2-7b-hf''' , device_map='''auto''' ) _a = model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 _a = torch.tensor([[-6.6550, -4.1227, -4.9859, -3.2406, 0.8262, -3.0033, 1.2964, -3.3699]] ) torch.testing.assert_close(out.mean(-1 ) , __UpperCAmelCase , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off _a = torch.tensor([-12.8281, -7.4453, -0.4639, -8.0625, -7.2500, -8.0000, -6.4883, -7.7695, -7.8438, -7.0312, -6.2188, -7.1328, -1.8496, 1.9961, -8.6250, -6.7227, -12.8281, -6.9492, -7.0742, -7.7852, -7.5820, -7.9062, -6.9375, -7.9805, -8.3438, -8.1562, -8.0469, -7.6250, -7.7422, -7.3398,] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , __UpperCAmelCase , atol=1e-5 , rtol=1e-5 ) @unittest.skip('''Logits are not exactly the same, once we fix the instabalities somehow, will update!''' ) @slow def _UpperCAmelCase ( self ) -> List[Any]: _a = [1, 306, 4658, 278, 6593, 310, 2834, 338] _a = LlamaForCausalLM.from_pretrained('''meta-llama/Llama-2-13b-hf''' , device_map='''auto''' ) _a = model(torch.tensor(__UpperCAmelCase ) ) # Expected mean on dim = -1 _a = torch.tensor([[-2.0622, -1.2794, -1.1638, -0.9788, -1.4603, -1.0238, -1.7893, -1.4411]] ) torch.testing.assert_close(out.mean(-1 ) , __UpperCAmelCase , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off _a = torch.tensor([-8.1406, -8.0547, 2.7461, -1.2344, -0.1448, -1.8262, -1.0020, -1.8154, -1.6895, -1.8516, -2.3574, -0.9277, 3.7598, 6.5742, -1.2998, -0.1177, -8.1406, -2.9688, -2.9199, -3.1699, -3.5254, -2.3555, -2.7988, -3.4141, -2.8262, -4.5195, -3.3379, -3.3164, -2.7832, -3.0273] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , __UpperCAmelCase , atol=1e-5 , rtol=1e-5 ) @unittest.skip('''Logits are not exactly the same, once we fix the instabalities somehow, will update!''' ) @slow def _UpperCAmelCase ( self ) -> Optional[Any]: _a = [1, 306, 4658, 278, 6593, 310, 2834, 338] _a = LlamaForCausalLM.from_pretrained('''meta-llama/Llama-2-13b-chat-hf''' , device_map='''auto''' ) _a = model(torch.tensor(__UpperCAmelCase ) ) # Expected mean on dim = -1 _a = torch.tensor([[-0.8562, -1.8520, -0.7551, -0.4162, -1.5161, -1.2038, -2.4823, -2.3254]] ) torch.testing.assert_close(out.mean(-1 ) , __UpperCAmelCase , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off _a = torch.tensor([-2.2227, 4.8828, 0.9023, -0.4578, -0.7871, -0.1033, -0.6221, -0.5786, -0.7803, -1.0674, -1.2920, -0.1570, 0.8008, 2.0723, -0.9497, 0.2771, -2.2227, -0.7612, -1.4346, -1.2061, -1.6426, -0.3000, -0.7139, -1.1934, -1.8691, -1.6973, -1.5947, -1.2705, -0.3523, -0.5513] ) # fmt: on torch.testing.assert_close(out.mean(-1 ) , __UpperCAmelCase , atol=1e-2 , rtol=1e-2 ) @unittest.skip( '''Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test''' ) @slow def _UpperCAmelCase ( self ) -> List[str]: _a = [1, 306, 4658, 278, 6593, 310, 2834, 338] _a = LlamaForCausalLM.from_pretrained('''meta-llama/Llama-2-70b-hf''' , device_map='''auto''' ) _a = model(torch.tensor(__UpperCAmelCase ) ) _a = torch.tensor( [[-4.2327, -3.3360, -4.6665, -4.7631, -1.8180, -3.4170, -1.4211, -3.1810]] , dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ) , __UpperCAmelCase , atol=1e-2 , rtol=1e-2 ) # fmt: off _a = torch.tensor([-9.4922, -3.9551, 1.7998, -5.6758, -5.1055, -5.8984, -4.8320, -6.8086, -6.5391, -5.6172, -5.5820, -5.5352, 1.7881, 3.6289, -6.5117, -3.4785, -9.5000, -6.0352, -6.8125, -6.0195, -6.6836, -5.4727, -6.2812, -6.0391, -7.3398, -7.4297, -7.4844, -6.5820, -5.8789, -5.5312] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , __UpperCAmelCase , atol=1e-5 , rtol=1e-5 ) @unittest.skip('''Model is curently gated''' ) @slow def _UpperCAmelCase ( self ) -> List[str]: _a = '''Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the "princi''' _a = '''Simply put, the theory of relativity states that ''' _a = LlamaTokenizer.from_pretrained('''meta-llama/Llama-2-13b-chat-hf''' ) _a = tokenizer.encode(__UpperCAmelCase , return_tensors='''pt''' ) _a = LlamaForCausalLM.from_pretrained( '''meta-llama/Llama-2-13b-chat-hf''' , device_map='''sequential''' , use_safetensors=__UpperCAmelCase ) # greedy generation outputs _a = model.generate(__UpperCAmelCase , max_new_tokens=64 , top_p=__UpperCAmelCase , temperature=1 , do_sample=__UpperCAmelCase ) _a = tokenizer.decode(generated_ids[0] , skip_special_tokens=__UpperCAmelCase ) self.assertEqual(__UpperCAmelCase , __UpperCAmelCase )
153
"""simple docstring""" import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class __lowerCamelCase ( a__ ): '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = None , __UpperCAmelCase = None , **__UpperCAmelCase , ) -> List[Any]: super().__init__( __UpperCAmelCase , split=__UpperCAmelCase , features=__UpperCAmelCase , cache_dir=__UpperCAmelCase , keep_in_memory=__UpperCAmelCase , streaming=__UpperCAmelCase , num_proc=__UpperCAmelCase , **__UpperCAmelCase , ) _a = field _a = path_or_paths if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else {self.split: path_or_paths} _a = Json( cache_dir=__UpperCAmelCase , data_files=__UpperCAmelCase , features=__UpperCAmelCase , field=__UpperCAmelCase , **__UpperCAmelCase , ) def _UpperCAmelCase ( self ) -> str: # Build iterable dataset if self.streaming: _a = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: _a = None _a = None _a = None _a = None self.builder.download_and_prepare( download_config=__UpperCAmelCase , download_mode=__UpperCAmelCase , verification_mode=__UpperCAmelCase , base_path=__UpperCAmelCase , num_proc=self.num_proc , ) _a = self.builder.as_dataset( split=self.split , verification_mode=__UpperCAmelCase , in_memory=self.keep_in_memory ) return dataset class __lowerCamelCase : '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , **__UpperCAmelCase , ) -> Tuple: if num_proc is not None and num_proc <= 0: raise ValueError(F'num_proc {num_proc} must be an integer > 0.' ) _a = dataset _a = path_or_buf _a = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE _a = num_proc _a = '''utf-8''' _a = to_json_kwargs def _UpperCAmelCase ( self ) -> int: _a = self.to_json_kwargs.pop('''path_or_buf''' , __UpperCAmelCase ) _a = self.to_json_kwargs.pop('''orient''' , '''records''' ) _a = self.to_json_kwargs.pop('''lines''' , True if orient == '''records''' else False ) _a = self.to_json_kwargs.pop('''index''' , False if orient in ['''split''', '''table'''] else True ) _a = self.to_json_kwargs.pop('''compression''' , __UpperCAmelCase ) if compression not in [None, "infer", "gzip", "bz2", "xz"]: raise NotImplementedError(F'`datasets` currently does not support {compression} compression' ) if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with fsspec.open(self.path_or_buf , '''wb''' , compression=__UpperCAmelCase ) as buffer: _a = self._write(file_obj=__UpperCAmelCase , orient=__UpperCAmelCase , lines=__UpperCAmelCase , index=__UpperCAmelCase , **self.to_json_kwargs ) else: if compression: raise NotImplementedError( F'The compression parameter is not supported when writing to a buffer, but compression={compression}' ''' was passed. Please provide a local path instead.''' ) _a = self._write( file_obj=self.path_or_buf , orient=__UpperCAmelCase , lines=__UpperCAmelCase , index=__UpperCAmelCase , **self.to_json_kwargs ) return written def _UpperCAmelCase ( self , __UpperCAmelCase ) -> Optional[int]: _a , _a , _a , _a , _a = args _a = query_table( table=self.dataset.data , key=slice(__UpperCAmelCase , offset + self.batch_size ) , indices=self.dataset._indices , ) _a = batch.to_pandas().to_json( path_or_buf=__UpperCAmelCase , orient=__UpperCAmelCase , lines=__UpperCAmelCase , index=__UpperCAmelCase , **__UpperCAmelCase ) if not json_str.endswith('''\n''' ): json_str += "\n" return json_str.encode(self.encoding ) def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) -> int: _a = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit='''ba''' , disable=not logging.is_progress_bar_enabled() , desc='''Creating json from Arrow format''' , ): _a = self._batch_json((offset, orient, lines, index, to_json_kwargs) ) written += file_obj.write(__UpperCAmelCase ) else: _a , _a = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for json_str in logging.tqdm( pool.imap( self._batch_json , [(offset, orient, lines, index, to_json_kwargs) for offset in range(0 , __UpperCAmelCase , __UpperCAmelCase )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit='''ba''' , disable=not logging.is_progress_bar_enabled() , desc='''Creating json from Arrow format''' , ): written += file_obj.write(__UpperCAmelCase ) return written
153
1
"""simple docstring""" import json import logging import os import sys from pathlib import Path import finetune_rag from transformers.file_utils import is_apex_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, require_ray, require_torch_gpu, require_torch_multi_gpu, ) logging.basicConfig(level=logging.DEBUG) lowercase__ = logging.getLogger() lowercase__ = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class __lowerCamelCase ( __UpperCAmelCase ): '''simple docstring''' def lowerCamelCase ( self : str , a_ : Dict ): os.makedirs(_lowerCamelCase , exist_ok=_lowerCamelCase ) lowerCAmelCase_ : List[Any] = {'''source''': '''What is love ?''', '''target''': '''life'''} lowerCAmelCase_ : List[Any] = {'''train''': 12, '''val''': 2, '''test''': 2} for split in ["train", "test", "val"]: for field in ["source", "target"]: lowerCAmelCase_ : Union[str, Any] = '''\n'''.join([contents[field]] * n_lines[split] ) with open(os.path.join(_lowerCamelCase , f'''{split}.{field}''' ) , "w" ) as f: f.write(_lowerCamelCase ) def lowerCamelCase ( self : int , a_ : int , a_ : str = "pytorch" ): lowerCAmelCase_ : Dict = self.get_auto_remove_tmp_dir() lowerCAmelCase_ : Tuple = os.path.join(_lowerCamelCase , "output" ) lowerCAmelCase_ : List[Any] = os.path.join(_lowerCamelCase , "data" ) self._create_dummy_data(data_dir=_lowerCamelCase ) lowerCAmelCase_ : List[Any] = f''' --data_dir {data_dir} \ --output_dir {output_dir} \ --model_name_or_path facebook/rag-sequence-base \ --model_type rag_sequence \ --do_train \ --do_predict \ --n_val -1 \ --val_check_interval 1.0 \ --train_batch_size 2 \ --eval_batch_size 1 \ --max_source_length 25 \ --max_target_length 25 \ --val_max_target_length 25 \ --test_max_target_length 25 \ --label_smoothing 0.1 \ --dropout 0.1 \ --attention_dropout 0.1 \ --weight_decay 0.001 \ --adam_epsilon 1e-08 \ --max_grad_norm 0.1 \ --lr_scheduler polynomial \ --learning_rate 3e-04 \ --num_train_epochs 1 \ --warmup_steps 4 \ --gradient_accumulation_steps 1 \ --distributed-port 8787 \ --use_dummy_dataset 1 \ --distributed_retriever {distributed_retriever} \ '''.split() if gpus > 0: testargs.append(f'''--gpus={gpus}''' ) if is_apex_available(): testargs.append("--fp16" ) else: testargs.append("--gpus=0" ) testargs.append("--distributed_backend=ddp_cpu" ) testargs.append("--num_processes=2" ) lowerCAmelCase_ : Optional[Any] = [sys.executable, str(Path(finetune_rag.__file__ ).resolve() )] + testargs execute_subprocess_async(_lowerCamelCase , env=self.get_env() ) lowerCAmelCase_ : str = os.path.join(_lowerCamelCase , "metrics.json" ) with open(_lowerCamelCase ) as f: lowerCAmelCase_ : int = json.load(_lowerCamelCase ) return result @require_torch_gpu def lowerCamelCase ( self : str ): lowerCAmelCase_ : List[Any] = self._run_finetune(gpus=1 ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 ) @require_torch_multi_gpu def lowerCamelCase ( self : Optional[Any] ): lowerCAmelCase_ : Union[str, Any] = self._run_finetune(gpus=2 ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 ) @require_torch_gpu @require_ray def lowerCamelCase ( self : Union[str, Any] ): lowerCAmelCase_ : Optional[Any] = self._run_finetune(gpus=1 , distributed_retriever="ray" ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 ) @require_torch_multi_gpu @require_ray def lowerCamelCase ( self : str ): lowerCAmelCase_ : List[Any] = self._run_finetune(gpus=1 , distributed_retriever="ray" ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 )
241
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from pathlib import Path import torch from ...utils import is_npu_available, is_xpu_available from .config_args import ClusterConfig, default_json_config_file from .config_utils import SubcommandHelpFormatter _lowerCamelCase : Tuple = 'Create a default config file for Accelerate with only a few flags set.' def lowercase_ ( _UpperCAmelCase="no" , _UpperCAmelCase = default_json_config_file , _UpperCAmelCase = False ): """simple docstring""" A_ : str = Path(_UpperCAmelCase ) path.parent.mkdir(parents=_UpperCAmelCase , exist_ok=_UpperCAmelCase ) if path.exists(): print( f"""Configuration already exists at {save_location}, will not override. Run `accelerate config` manually or pass a different `save_location`.""" ) return False A_ : Optional[Any] = mixed_precision.lower() if mixed_precision not in ["no", "fp16", "bf16", "fp8"]: raise ValueError( f"""`mixed_precision` should be one of 'no', 'fp16', 'bf16', or 'fp8'. Received {mixed_precision}""" ) A_ : str = { '''compute_environment''': '''LOCAL_MACHINE''', '''mixed_precision''': mixed_precision, } if torch.cuda.is_available(): A_ : int = torch.cuda.device_count() A_ : int = num_gpus A_ : Tuple = False if num_gpus > 1: A_ : Optional[int] = '''MULTI_GPU''' else: A_ : Union[str, Any] = '''NO''' elif is_xpu_available() and use_xpu: A_ : str = torch.xpu.device_count() A_ : Optional[int] = num_xpus A_ : List[str] = False if num_xpus > 1: A_ : Any = '''MULTI_XPU''' else: A_ : Optional[Any] = '''NO''' elif is_npu_available(): A_ : Union[str, Any] = torch.npu.device_count() A_ : Optional[int] = num_npus A_ : Union[str, Any] = False if num_npus > 1: A_ : List[str] = '''MULTI_NPU''' else: A_ : Tuple = '''NO''' else: A_ : Union[str, Any] = 0 A_ : str = True A_ : str = 1 A_ : List[Any] = '''NO''' A_ : Dict = ClusterConfig(**_UpperCAmelCase ) config.to_json_file(_UpperCAmelCase ) return path def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : List[str] = parser.add_parser('''default''' , parents=_UpperCAmelCase , help=_UpperCAmelCase , formatter_class=_UpperCAmelCase ) parser.add_argument( '''--config_file''' , default=_UpperCAmelCase , help=( '''The path to use to store the config file. Will default to a file named default_config.yaml in the cache ''' '''location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ''' '''such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ''' '''with \'huggingface\'.''' ) , dest='''save_location''' , ) parser.add_argument( '''--mixed_precision''' , choices=['''no''', '''fp16''', '''bf16'''] , type=_UpperCAmelCase , help='''Whether or not to use mixed precision training. ''' '''Choose between FP16 and BF16 (bfloat16) training. ''' '''BF16 training is only supported on Nvidia Ampere GPUs and PyTorch 1.10 or later.''' , default='''no''' , ) parser.set_defaults(func=_UpperCAmelCase ) return parser def lowercase_ ( _UpperCAmelCase ): """simple docstring""" A_ : str = write_basic_config(args.mixed_precision , args.save_location ) if config_file: print(f"""accelerate configuration saved at {config_file}""" )
167
0
'''simple docstring''' # Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING import numpy as np import pyarrow as pa from .. import config from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import torch class UpperCamelCase_ (TensorFormatter[Mapping, 'torch.Tensor', Mapping] ): """simple docstring""" def __init__( self : Optional[int] , _lowerCamelCase : List[str]=None , **_lowerCamelCase : int ): """simple docstring""" super().__init__(features=lowerCamelCase__ ) A_ : Any = torch_tensor_kwargs import torch # noqa import torch at initialization def _a ( self : Optional[Any] , _lowerCamelCase : Optional[int] ): """simple docstring""" import torch if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and column: if all( isinstance(lowerCamelCase__ , torch.Tensor ) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column ): return torch.stack(lowerCamelCase__ ) return column def _a ( self : Optional[Any] , _lowerCamelCase : Union[str, Any] ): """simple docstring""" import torch if isinstance(lowerCamelCase__ , (str, bytes, type(lowerCamelCase__ )) ): return value elif isinstance(lowerCamelCase__ , (np.character, np.ndarray) ) and np.issubdtype(value.dtype , np.character ): return value.tolist() A_ : Tuple = {} if isinstance(lowerCamelCase__ , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.integer ): A_ : int = {'''dtype''': torch.intaa} elif isinstance(lowerCamelCase__ , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.floating ): A_ : Any = {'''dtype''': torch.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(lowerCamelCase__ , PIL.Image.Image ): A_ : Optional[int] = np.asarray(lowerCamelCase__ ) return torch.tensor(lowerCamelCase__ , **{**default_dtype, **self.torch_tensor_kwargs} ) def _a ( self : List[str] , _lowerCamelCase : Any ): """simple docstring""" import torch # support for torch, tf, jax etc. if hasattr(lowerCamelCase__ , '''__array__''' ) and not isinstance(lowerCamelCase__ , torch.Tensor ): A_ : Optional[Any] = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(lowerCamelCase__ , np.ndarray ): if data_struct.dtype == object: # torch tensors cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(lowerCamelCase__ ) for substruct in data_struct] ) elif isinstance(lowerCamelCase__ , (list, tuple) ): return self._consolidate([self.recursive_tensorize(lowerCamelCase__ ) for substruct in data_struct] ) return self._tensorize(lowerCamelCase__ ) def _a ( self : int , _lowerCamelCase : dict ): """simple docstring""" return map_nested(self._recursive_tensorize , lowerCamelCase__ , map_list=lowerCamelCase__ ) def _a ( self : Union[str, Any] , _lowerCamelCase : pa.Table ): """simple docstring""" A_ : int = self.numpy_arrow_extractor().extract_row(lowerCamelCase__ ) A_ : Optional[int] = self.python_features_decoder.decode_row(lowerCamelCase__ ) return self.recursive_tensorize(lowerCamelCase__ ) def _a ( self : str , _lowerCamelCase : pa.Table ): """simple docstring""" A_ : List[Any] = self.numpy_arrow_extractor().extract_column(lowerCamelCase__ ) A_ : str = self.python_features_decoder.decode_column(lowerCamelCase__ , pa_table.column_names[0] ) A_ : int = self.recursive_tensorize(lowerCamelCase__ ) A_ : Tuple = self._consolidate(lowerCamelCase__ ) return column def _a ( self : Tuple , _lowerCamelCase : pa.Table ): """simple docstring""" A_ : Dict = self.numpy_arrow_extractor().extract_batch(lowerCamelCase__ ) A_ : Any = self.python_features_decoder.decode_batch(lowerCamelCase__ ) A_ : Tuple = self.recursive_tensorize(lowerCamelCase__ ) for column_name in batch: A_ : Optional[Any] = self._consolidate(batch[column_name] ) return batch
356
'''simple docstring''' from __future__ import annotations from PIL import Image # Define glider example snake_case__ = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example snake_case__ = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def snake_case__ ( lowerCamelCase__ : list[list[int]] ) -> list[list[int]]: A_ : str = [] for i in range(len(lowerCamelCase__ ) ): A_ : Optional[Any] = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours A_ : Optional[int] = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(lowerCamelCase__ ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(lowerCamelCase__ ) - 1: neighbour_count += cells[i + 1][j] if i < len(lowerCamelCase__ ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. A_ : List[str] = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(lowerCamelCase__ ) return next_generation def snake_case__ ( lowerCamelCase__ : list[list[int]] , lowerCamelCase__ : int ) -> list[Image.Image]: A_ : List[Any] = [] for _ in range(lowerCamelCase__ ): # Create output image A_ : Optional[int] = Image.new('''RGB''' , (len(cells[0] ), len(lowerCamelCase__ )) ) A_ : int = img.load() # Save cells to image for x in range(len(lowerCamelCase__ ) ): for y in range(len(cells[0] ) ): A_ : Optional[Any] = 2_5_5 - cells[y][x] * 2_5_5 A_ : str = (colour, colour, colour) # Save image images.append(lowerCamelCase__ ) A_ : Optional[int] = new_generation(lowerCamelCase__ ) return images if __name__ == "__main__": snake_case__ = generate_images(GLIDER, 16) images[0].save("""out.gif""", save_all=True, append_images=images[1:])
4
0
import math import time from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class SCREAMING_SNAKE_CASE_ ( snake_case_ ): def __init__( self : Optional[Any] , *_A : str , _A : Optional[int]=None , _A : Any=None , **_A : Union[str, Any] ) -> Any: """simple docstring""" super().__init__(*_A , **_A ) snake_case_ : List[str] = eval_examples snake_case_ : List[Any] = post_process_function def UpperCAmelCase_ ( self : Union[str, Any] , _A : Union[str, Any]=None , _A : Optional[int]=None , _A : List[str]=None , _A : str = "eval" ) -> str: """simple docstring""" snake_case_ : List[Any] = self.eval_dataset if eval_dataset is None else eval_dataset snake_case_ : int = self.get_eval_dataloader(_A ) snake_case_ : Optional[int] = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. snake_case_ : Union[str, Any] = self.compute_metrics snake_case_ : Union[str, Any] = None snake_case_ : Optional[int] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop snake_case_ : str = time.time() try: snake_case_ : int = eval_loop( _A , description='Evaluation' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_A , metric_key_prefix=_A , ) finally: snake_case_ : Optional[int] = compute_metrics snake_case_ : Optional[int] = self.args.eval_batch_size * self.args.world_size if F"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[F"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( _A , _A , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default snake_case_ : List[str] = self.post_process_function(_A , _A , output.predictions ) snake_case_ : Any = self.compute_metrics(_A ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): snake_case_ : int = metrics.pop(_A ) metrics.update(output.metrics ) else: snake_case_ : Optional[Any] = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(_A ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) snake_case_ : List[str] = self.callback_handler.on_evaluate(self.args , self.state , self.control , _A ) return metrics def UpperCAmelCase_ ( self : Optional[Any] , _A : List[str] , _A : int , _A : int=None , _A : str = "test" ) -> List[str]: """simple docstring""" snake_case_ : Optional[int] = self.get_test_dataloader(_A ) # Temporarily disable metric computation, we will do it in the loop here. snake_case_ : Optional[Any] = self.compute_metrics snake_case_ : Dict = None snake_case_ : int = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop snake_case_ : Tuple = time.time() try: snake_case_ : Optional[Any] = eval_loop( _A , description='Prediction' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_A , metric_key_prefix=_A , ) finally: snake_case_ : Optional[int] = compute_metrics snake_case_ : Tuple = self.args.eval_batch_size * self.args.world_size if F"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[F"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( _A , _A , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output snake_case_ : Dict = self.post_process_function(_A , _A , output.predictions , 'predict' ) snake_case_ : Optional[Any] = self.compute_metrics(_A ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): snake_case_ : List[str] = metrics.pop(_A ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=_A )
327
def SCREAMING_SNAKE_CASE__ ( __a ): if not isinstance(__a , __a ): snake_case_ : int = f"""Input value of [number={number}] must be an integer""" raise TypeError(__a ) if number < 0: return False snake_case_ : Dict = number * number while number > 0: if number % 10 != number_square % 10: return False number //= 10 number_square //= 10 return True if __name__ == "__main__": import doctest doctest.testmod()
327
1
from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": lowercase__ : Optional[Any] = input("Enter image url: ").strip() print(F'Downloading image from {url} ...') lowercase__ : List[str] = BeautifulSoup(requests.get(url).content, "html.parser") # The image URL is in the content field of the first meta tag with property og:image lowercase__ : Optional[Any] = soup.find("meta", {"property": "og:image"})["content"] lowercase__ : int = requests.get(image_url).content lowercase__ : Union[str, Any] = F'{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg' with open(file_name, "wb") as fp: fp.write(image_data) print(F'Done. Image saved to disk as {file_name}.')
180
import math import unittest def SCREAMING_SNAKE_CASE ( __UpperCamelCase) -> bool: assert isinstance(__UpperCamelCase , __UpperCamelCase) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(__UpperCamelCase) + 1) , 6): if number % i == 0 or number % (i + 2) == 0: return False return True class a__ ( unittest.TestCase ): def lowerCAmelCase_ ( self ) -> List[str]: '''simple docstring''' self.assertTrue(is_prime(2 ) ) self.assertTrue(is_prime(3 ) ) self.assertTrue(is_prime(5 ) ) self.assertTrue(is_prime(7 ) ) self.assertTrue(is_prime(11 ) ) self.assertTrue(is_prime(13 ) ) self.assertTrue(is_prime(17 ) ) self.assertTrue(is_prime(19 ) ) self.assertTrue(is_prime(23 ) ) self.assertTrue(is_prime(29 ) ) def lowerCAmelCase_ ( self ) -> Tuple: '''simple docstring''' with self.assertRaises(A ): is_prime(-19 ) self.assertFalse( is_prime(0 ) , "Zero doesn't have any positive factors, primes must have exactly two." , ) self.assertFalse( is_prime(1 ) , "One only has 1 positive factor, primes must have exactly two." , ) self.assertFalse(is_prime(2 * 2 ) ) self.assertFalse(is_prime(2 * 3 ) ) self.assertFalse(is_prime(3 * 3 ) ) self.assertFalse(is_prime(3 * 5 ) ) self.assertFalse(is_prime(3 * 5 * 7 ) ) if __name__ == "__main__": unittest.main()
180
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __lowerCAmelCase : Union[str, Any] =logging.get_logger(__name__) __lowerCAmelCase : Dict ={ 'google/bit-50': 'https://huggingface.co/google/bit-50/resolve/main/config.json', } class _lowercase ( A__ , A__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = '''bit''' SCREAMING_SNAKE_CASE__ : Tuple = ['''preactivation''', '''bottleneck'''] SCREAMING_SNAKE_CASE__ : List[str] = ['''SAME''', '''VALID'''] def __init__( self :str , lowerCAmelCase__ :Any=3 , lowerCAmelCase__ :Optional[Any]=64 , lowerCAmelCase__ :str=[256, 512, 1_024, 2_048] , lowerCAmelCase__ :List[str]=[3, 4, 6, 3] , lowerCAmelCase__ :str="preactivation" , lowerCAmelCase__ :str="relu" , lowerCAmelCase__ :Dict=None , lowerCAmelCase__ :int=32 , lowerCAmelCase__ :str=0.0 , lowerCAmelCase__ :List[str]=False , lowerCAmelCase__ :Union[str, Any]=32 , lowerCAmelCase__ :List[Any]=1 , lowerCAmelCase__ :Optional[Any]=None , lowerCAmelCase__ :Optional[int]=None , **lowerCAmelCase__ :Union[str, Any] , ) -> Optional[Any]: super().__init__(**lowerCAmelCase__ ) if layer_type not in self.layer_types: raise ValueError(f'''layer_type={layer_type} is not one of {','.join(self.layer_types )}''' ) if global_padding is not None: if global_padding.upper() in self.supported_padding: __SCREAMING_SNAKE_CASE : Tuple = global_padding.upper() else: raise ValueError(f'''Padding strategy {global_padding} not supported''' ) __SCREAMING_SNAKE_CASE : Optional[int] = num_channels __SCREAMING_SNAKE_CASE : str = embedding_size __SCREAMING_SNAKE_CASE : Union[str, Any] = hidden_sizes __SCREAMING_SNAKE_CASE : Union[str, Any] = depths __SCREAMING_SNAKE_CASE : Tuple = layer_type __SCREAMING_SNAKE_CASE : Optional[Any] = hidden_act __SCREAMING_SNAKE_CASE : Optional[int] = global_padding __SCREAMING_SNAKE_CASE : Optional[int] = num_groups __SCREAMING_SNAKE_CASE : List[Any] = drop_path_rate __SCREAMING_SNAKE_CASE : Tuple = embedding_dynamic_padding __SCREAMING_SNAKE_CASE : Union[str, Any] = output_stride __SCREAMING_SNAKE_CASE : List[str] = width_factor __SCREAMING_SNAKE_CASE : List[Any] = ['''stem'''] + [f'''stage{idx}''' for idx in range(1 , len(lowerCAmelCase__ ) + 1 )] __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Dict = get_aligned_output_features_output_indices( out_features=lowerCAmelCase__ , out_indices=lowerCAmelCase__ , stage_names=self.stage_names )
9
from importlib import import_module from .logging import get_logger __lowerCAmelCase : str =get_logger(__name__) class _lowercase : '''simple docstring''' def __init__( self :List[Any] , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :str=None ) -> int: __SCREAMING_SNAKE_CASE : List[str] = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith('''__''' ): setattr(self , lowerCAmelCase__ , getattr(lowerCAmelCase__ , lowerCAmelCase__ ) ) __SCREAMING_SNAKE_CASE : Optional[Any] = module._original_module if isinstance(lowerCAmelCase__ , _PatchedModuleObj ) else module class _lowercase : '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[int] = [] def __init__( self :Tuple , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :str , lowerCAmelCase__ :Any , lowerCAmelCase__ :Dict=None ) -> List[Any]: __SCREAMING_SNAKE_CASE : Optional[int] = obj __SCREAMING_SNAKE_CASE : str = target __SCREAMING_SNAKE_CASE : Dict = new __SCREAMING_SNAKE_CASE : Union[str, Any] = target.split('''.''' )[0] __SCREAMING_SNAKE_CASE : List[str] = {} __SCREAMING_SNAKE_CASE : Tuple = attrs or [] def __enter__( self :int ) -> Dict: *__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Optional[Any] = self.target.split('''.''' ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(lowerCAmelCase__ ) ): try: __SCREAMING_SNAKE_CASE : Any = import_module('''.'''.join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): __SCREAMING_SNAKE_CASE : Union[str, Any] = getattr(self.obj , lowerCAmelCase__ ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(lowerCAmelCase__ , _PatchedModuleObj ) and obj_attr._original_module is submodule) ): __SCREAMING_SNAKE_CASE : int = obj_attr # patch at top level setattr(self.obj , lowerCAmelCase__ , _PatchedModuleObj(lowerCAmelCase__ , attrs=self.attrs ) ) __SCREAMING_SNAKE_CASE : List[str] = getattr(self.obj , lowerCAmelCase__ ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(lowerCAmelCase__ , lowerCAmelCase__ , _PatchedModuleObj(getattr(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) , attrs=self.attrs ) ) __SCREAMING_SNAKE_CASE : Tuple = getattr(lowerCAmelCase__ , lowerCAmelCase__ ) # finally set the target attribute setattr(lowerCAmelCase__ , lowerCAmelCase__ , self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: __SCREAMING_SNAKE_CASE : Union[str, Any] = getattr(import_module('''.'''.join(lowerCAmelCase__ ) ) , lowerCAmelCase__ ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , lowerCAmelCase__ ) is attr_value: __SCREAMING_SNAKE_CASE : Any = getattr(self.obj , lowerCAmelCase__ ) setattr(self.obj , lowerCAmelCase__ , self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" __SCREAMING_SNAKE_CASE : Union[str, Any] = globals()['''__builtins__'''][target_attr] setattr(self.obj , lowerCAmelCase__ , self.new ) else: raise RuntimeError(f'''Tried to patch attribute {target_attr} instead of a submodule.''' ) def __exit__( self :str , *lowerCAmelCase__ :Union[str, Any] ) -> Optional[int]: for attr in list(self.original ): setattr(self.obj , lowerCAmelCase__ , self.original.pop(lowerCAmelCase__ ) ) def __magic_name__( self :List[Any] ) -> List[Any]: self.__enter__() self._active_patches.append(self ) def __magic_name__( self :Optional[int] ) -> int: try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
9
1
from __future__ import annotations from collections.abc import Iterable, Iterator from dataclasses import dataclass lowerCamelCase_ : List[Any] = (3, 9, -11, 0, 7, 5, 1, -1) lowerCamelCase_ : Optional[Any] = (4, 6, 2, 0, 8, 10, 3, -2) @dataclass class a__ : A__ : int A__ : Node | None class a__ : def __init__( self , UpperCAmelCase ) -> None: __a = None for i in sorted(UpperCAmelCase , reverse=UpperCAmelCase ): __a = Node(UpperCAmelCase , self.head ) def __iter__( self ) -> Iterator[int]: __a = self.head while node: yield node.data __a = node.next_node def __len__( self ) -> int: return sum(1 for _ in self ) def __str__( self ) -> str: return " -> ".join([str(UpperCAmelCase ) for node in self] ) def lowerCAmelCase( __lowerCamelCase , __lowerCamelCase ): return SortedLinkedList(list(__lowerCamelCase ) + list(__lowerCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() lowerCamelCase_ : Tuple = SortedLinkedList print(merge_lists(SSL(test_data_odd), SSL(test_data_even)))
357
from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class a__ ( __snake_case ): A__ : torch.FloatTensor A__ : torch.FloatTensor A__ : Optional[torch.FloatTensor] = None class a__ ( __snake_case , __snake_case ): A__ : Optional[Any] = 2 @register_to_config def __init__( self , UpperCAmelCase = 0.02 , UpperCAmelCase = 1_0_0 , UpperCAmelCase = 1.007 , UpperCAmelCase = 8_0 , UpperCAmelCase = 0.05 , UpperCAmelCase = 5_0 , ) -> Optional[Any]: # standard deviation of the initial noise distribution __a = sigma_max # setable values __a = None __a = None __a = None # sigma(t_i) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase = None ) -> torch.FloatTensor: return sample def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase = None ) -> int: __a = num_inference_steps __a = np.arange(0 , self.num_inference_steps )[::-1].copy() __a = torch.from_numpy(UpperCAmelCase ).to(UpperCAmelCase ) __a = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] __a = torch.tensor(UpperCAmelCase , dtype=torch.floataa , device=UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None ) -> Tuple[torch.FloatTensor, float]: if self.config.s_min <= sigma <= self.config.s_max: __a = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: __a = 0 # sample eps ~ N(0, S_noise^2 * I) __a = self.config.s_noise * randn_tensor(sample.shape , generator=UpperCAmelCase ).to(sample.device ) __a = sigma + gamma * sigma __a = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = True , ) -> Union[KarrasVeOutput, Tuple]: __a = sample_hat + sigma_hat * model_output __a = (sample_hat - pred_original_sample) / sigma_hat __a = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=UpperCAmelCase , derivative=UpperCAmelCase , pred_original_sample=UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = True , ) -> Union[KarrasVeOutput, Tuple]: __a = sample_prev + sigma_prev * model_output __a = (sample_prev - pred_original_sample) / sigma_prev __a = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=UpperCAmelCase , derivative=UpperCAmelCase , pred_original_sample=UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> List[Any]: raise NotImplementedError()
197
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __A = { "configuration_falcon": ["FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP", "FalconConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "FALCON_PRETRAINED_MODEL_ARCHIVE_LIST", "FalconForCausalLM", "FalconModel", "FalconPreTrainedModel", "FalconForSequenceClassification", "FalconForTokenClassification", "FalconForQuestionAnswering", ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
164
'''simple docstring''' import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeline, AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_AUDIO_BATCH_PARAMS, TEXT_TO_AUDIO_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A ( __UpperCAmelCase , unittest.TestCase ): lowerCamelCase : List[str] = AudioLDMPipeline lowerCamelCase : Union[str, Any] = TEXT_TO_AUDIO_PARAMS lowerCamelCase : Tuple = TEXT_TO_AUDIO_BATCH_PARAMS lowerCamelCase : Optional[int] = frozenset( [ """num_inference_steps""", """num_waveforms_per_prompt""", """generator""", """latents""", """output_type""", """return_dict""", """callback""", """callback_steps""", ] ) def A__ ( self ) -> List[str]: '''simple docstring''' torch.manual_seed(0 ) lowercase__ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=(32, 64) , class_embed_type="""simple_projection""" , projection_class_embeddings_input_dim=32 , class_embeddings_concat=lowerCamelCase__ , ) lowercase__ = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=lowerCamelCase__ , set_alpha_to_one=lowerCamelCase__ , ) torch.manual_seed(0 ) lowercase__ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=1 , out_channels=1 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) lowercase__ = ClapTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , projection_dim=32 , ) lowercase__ = ClapTextModelWithProjection(lowerCamelCase__ ) lowercase__ = RobertaTokenizer.from_pretrained("""hf-internal-testing/tiny-random-roberta""" , model_max_length=77 ) lowercase__ = SpeechTaHifiGanConfig( model_in_dim=8 , sampling_rate=16_000 , upsample_initial_channel=16 , upsample_rates=[2, 2] , upsample_kernel_sizes=[4, 4] , resblock_kernel_sizes=[3, 7] , resblock_dilation_sizes=[[1, 3, 5], [1, 3, 5]] , normalize_before=lowerCamelCase__ , ) lowercase__ = SpeechTaHifiGan(lowerCamelCase__ ) lowercase__ = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """vocoder""": vocoder, } return components def A__ ( self , lowerCamelCase__ , lowerCamelCase__=0 ) -> Tuple: '''simple docstring''' if str(lowerCamelCase__ ).startswith("""mps""" ): lowercase__ = torch.manual_seed(lowerCamelCase__ ) else: lowercase__ = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) lowercase__ = { """prompt""": """A hammer hitting a wooden surface""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, } return inputs def A__ ( self ) -> Any: '''simple docstring''' lowercase__ = """cpu""" # ensure determinism for the device-dependent torch.Generator lowercase__ = self.get_dummy_components() lowercase__ = AudioLDMPipeline(**lowerCamelCase__ ) lowercase__ = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) lowercase__ = self.get_dummy_inputs(lowerCamelCase__ ) lowercase__ = audioldm_pipe(**lowerCamelCase__ ) lowercase__ = output.audios[0] assert audio.ndim == 1 assert len(lowerCamelCase__ ) == 256 lowercase__ = audio[:10] lowercase__ = np.array( [-0.00_50, 0.00_50, -0.00_60, 0.00_33, -0.00_26, 0.00_33, -0.00_27, 0.00_33, -0.00_28, 0.00_33] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def A__ ( self ) -> Optional[int]: '''simple docstring''' lowercase__ = self.get_dummy_components() lowercase__ = AudioLDMPipeline(**lowerCamelCase__ ) lowercase__ = audioldm_pipe.to(lowerCamelCase__ ) lowercase__ = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) lowercase__ = self.get_dummy_inputs(lowerCamelCase__ ) lowercase__ = 3 * [inputs["""prompt"""]] # forward lowercase__ = audioldm_pipe(**lowerCamelCase__ ) lowercase__ = output.audios[0] lowercase__ = self.get_dummy_inputs(lowerCamelCase__ ) lowercase__ = 3 * [inputs.pop("""prompt""" )] lowercase__ = audioldm_pipe.tokenizer( lowerCamelCase__ , padding="""max_length""" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=lowerCamelCase__ , return_tensors="""pt""" , ) lowercase__ = text_inputs["""input_ids"""].to(lowerCamelCase__ ) lowercase__ = audioldm_pipe.text_encoder( lowerCamelCase__ , ) lowercase__ = prompt_embeds.text_embeds # additional L_2 normalization over each hidden-state lowercase__ = F.normalize(lowerCamelCase__ , dim=-1 ) lowercase__ = prompt_embeds # forward lowercase__ = audioldm_pipe(**lowerCamelCase__ ) lowercase__ = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def A__ ( self ) -> List[Any]: '''simple docstring''' lowercase__ = self.get_dummy_components() lowercase__ = AudioLDMPipeline(**lowerCamelCase__ ) lowercase__ = audioldm_pipe.to(lowerCamelCase__ ) lowercase__ = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) lowercase__ = self.get_dummy_inputs(lowerCamelCase__ ) lowercase__ = 3 * ["""this is a negative prompt"""] lowercase__ = negative_prompt lowercase__ = 3 * [inputs["""prompt"""]] # forward lowercase__ = audioldm_pipe(**lowerCamelCase__ ) lowercase__ = output.audios[0] lowercase__ = self.get_dummy_inputs(lowerCamelCase__ ) lowercase__ = 3 * [inputs.pop("""prompt""" )] lowercase__ = [] for p in [prompt, negative_prompt]: lowercase__ = audioldm_pipe.tokenizer( lowerCamelCase__ , padding="""max_length""" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=lowerCamelCase__ , return_tensors="""pt""" , ) lowercase__ = text_inputs["""input_ids"""].to(lowerCamelCase__ ) lowercase__ = audioldm_pipe.text_encoder( lowerCamelCase__ , ) lowercase__ = text_embeds.text_embeds # additional L_2 normalization over each hidden-state lowercase__ = F.normalize(lowerCamelCase__ , dim=-1 ) embeds.append(lowerCamelCase__ ) lowercase__ , lowercase__ = embeds # forward lowercase__ = audioldm_pipe(**lowerCamelCase__ ) lowercase__ = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def A__ ( self ) -> Optional[Any]: '''simple docstring''' lowercase__ = """cpu""" # ensure determinism for the device-dependent torch.Generator lowercase__ = self.get_dummy_components() lowercase__ = PNDMScheduler(skip_prk_steps=lowerCamelCase__ ) lowercase__ = AudioLDMPipeline(**lowerCamelCase__ ) lowercase__ = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) lowercase__ = self.get_dummy_inputs(lowerCamelCase__ ) lowercase__ = """egg cracking""" lowercase__ = audioldm_pipe(**lowerCamelCase__ , negative_prompt=lowerCamelCase__ ) lowercase__ = output.audios[0] assert audio.ndim == 1 assert len(lowerCamelCase__ ) == 256 lowercase__ = audio[:10] lowercase__ = np.array( [-0.00_51, 0.00_50, -0.00_60, 0.00_34, -0.00_26, 0.00_33, -0.00_27, 0.00_33, -0.00_28, 0.00_32] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def A__ ( self ) -> int: '''simple docstring''' lowercase__ = """cpu""" # ensure determinism for the device-dependent torch.Generator lowercase__ = self.get_dummy_components() lowercase__ = PNDMScheduler(skip_prk_steps=lowerCamelCase__ ) lowercase__ = AudioLDMPipeline(**lowerCamelCase__ ) lowercase__ = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) lowercase__ = """A hammer hitting a wooden surface""" # test num_waveforms_per_prompt=1 (default) lowercase__ = audioldm_pipe(lowerCamelCase__ , num_inference_steps=2 ).audios assert audios.shape == (1, 256) # test num_waveforms_per_prompt=1 (default) for batch of prompts lowercase__ = 2 lowercase__ = audioldm_pipe([prompt] * batch_size , num_inference_steps=2 ).audios assert audios.shape == (batch_size, 256) # test num_waveforms_per_prompt for single prompt lowercase__ = 2 lowercase__ = audioldm_pipe(lowerCamelCase__ , num_inference_steps=2 , num_waveforms_per_prompt=lowerCamelCase__ ).audios assert audios.shape == (num_waveforms_per_prompt, 256) # test num_waveforms_per_prompt for batch of prompts lowercase__ = 2 lowercase__ = audioldm_pipe( [prompt] * batch_size , num_inference_steps=2 , num_waveforms_per_prompt=lowerCamelCase__ ).audios assert audios.shape == (batch_size * num_waveforms_per_prompt, 256) def A__ ( self ) -> Optional[int]: '''simple docstring''' lowercase__ = """cpu""" # ensure determinism for the device-dependent torch.Generator lowercase__ = self.get_dummy_components() lowercase__ = AudioLDMPipeline(**lowerCamelCase__ ) lowercase__ = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) lowercase__ = audioldm_pipe.vocoder.config.sampling_rate lowercase__ = self.get_dummy_inputs(lowerCamelCase__ ) lowercase__ = audioldm_pipe(audio_length_in_s=0.0_16 , **lowerCamelCase__ ) lowercase__ = output.audios[0] assert audio.ndim == 1 assert len(lowerCamelCase__ ) / vocoder_sampling_rate == 0.0_16 lowercase__ = audioldm_pipe(audio_length_in_s=0.0_32 , **lowerCamelCase__ ) lowercase__ = output.audios[0] assert audio.ndim == 1 assert len(lowerCamelCase__ ) / vocoder_sampling_rate == 0.0_32 def A__ ( self ) -> str: '''simple docstring''' lowercase__ = self.get_dummy_components() lowercase__ = AudioLDMPipeline(**lowerCamelCase__ ) lowercase__ = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) lowercase__ = ["""hey"""] lowercase__ = audioldm_pipe(lowerCamelCase__ , num_inference_steps=1 ) lowercase__ = output.audios.shape assert audio_shape == (1, 256) lowercase__ = audioldm_pipe.vocoder.config config.model_in_dim *= 2 lowercase__ = SpeechTaHifiGan(lowerCamelCase__ ).to(lowerCamelCase__ ) lowercase__ = audioldm_pipe(lowerCamelCase__ , num_inference_steps=1 ) lowercase__ = output.audios.shape # waveform shape is unchanged, we just have 2x the number of mel channels in the spectrogram assert audio_shape == (1, 256) def A__ ( self ) -> Optional[int]: '''simple docstring''' self._test_attention_slicing_forward_pass(test_mean_pixel_difference=lowerCamelCase__ ) def A__ ( self ) -> int: '''simple docstring''' self._test_inference_batch_single_identical(test_mean_pixel_difference=lowerCamelCase__ ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def A__ ( self ) -> Any: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCamelCase__ ) @slow class A ( unittest.TestCase ): def A__ ( self ) -> List[str]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def A__ ( self , lowerCamelCase__ , lowerCamelCase__="cpu" , lowerCamelCase__=torch.floataa , lowerCamelCase__=0 ) -> int: '''simple docstring''' lowercase__ = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) lowercase__ = np.random.RandomState(lowerCamelCase__ ).standard_normal((1, 8, 128, 16) ) lowercase__ = torch.from_numpy(lowerCamelCase__ ).to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) lowercase__ = { """prompt""": """A hammer hitting a wooden surface""", """latents""": latents, """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 2.5, } return inputs def A__ ( self ) -> Union[str, Any]: '''simple docstring''' lowercase__ = AudioLDMPipeline.from_pretrained("""cvssp/audioldm""" ) lowercase__ = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) lowercase__ = self.get_inputs(lowerCamelCase__ ) lowercase__ = 25 lowercase__ = audioldm_pipe(**lowerCamelCase__ ).audios[0] assert audio.ndim == 1 assert len(lowerCamelCase__ ) == 81_920 lowercase__ = audio[77_230:77_240] lowercase__ = np.array( [-0.48_84, -0.46_07, 0.00_23, 0.50_07, 0.58_96, 0.51_51, 0.38_13, -0.02_08, -0.36_87, -0.43_15] ) lowercase__ = np.abs(expected_slice - audio_slice ).max() assert max_diff < 1e-2 def A__ ( self ) -> Optional[int]: '''simple docstring''' lowercase__ = AudioLDMPipeline.from_pretrained("""cvssp/audioldm""" ) lowercase__ = LMSDiscreteScheduler.from_config(audioldm_pipe.scheduler.config ) lowercase__ = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) lowercase__ = self.get_inputs(lowerCamelCase__ ) lowercase__ = audioldm_pipe(**lowerCamelCase__ ).audios[0] assert audio.ndim == 1 assert len(lowerCamelCase__ ) == 81_920 lowercase__ = audio[27_780:27_790] lowercase__ = np.array([-0.21_31, -0.08_73, -0.01_24, -0.01_89, 0.05_69, 0.13_73, 0.18_83, 0.28_86, 0.32_97, 0.22_12] ) lowercase__ = np.abs(expected_slice - audio_slice ).max() assert max_diff < 3e-2
164
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) snake_case : Any = { '''configuration_blenderbot''': [ '''BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlenderbotConfig''', '''BlenderbotOnnxConfig''', ], '''tokenization_blenderbot''': ['''BlenderbotTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case : Union[str, Any] = ['''BlenderbotTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case : str = [ '''BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlenderbotForCausalLM''', '''BlenderbotForConditionalGeneration''', '''BlenderbotModel''', '''BlenderbotPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case : int = [ '''TFBlenderbotForConditionalGeneration''', '''TFBlenderbotModel''', '''TFBlenderbotPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case : List[Any] = [ '''FlaxBlenderbotForConditionalGeneration''', '''FlaxBlenderbotModel''', '''FlaxBlenderbotPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys snake_case : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
109
import inspect import jax import jax.lax as lax import jax.numpy as jnp from ..utils import add_start_docstrings from ..utils.logging import get_logger snake_case : Dict = get_logger(__name__) snake_case : str = r''' Args: input_ids (`jnp.ndarray` of shape `(batch_size, sequence_length)`): Indices of input sequence tokens in the vocabulary. Indices can be obtained using [`PreTrainedTokenizer`]. See [`PreTrainedTokenizer.encode`] and [`PreTrainedTokenizer.__call__`] for details. [What are input IDs?](../glossary#input-ids) scores (`jnp.ndarray` of shape `(batch_size, config.vocab_size)`): Prediction scores of a language modeling head. These can be logits for each vocabulary when not using beam search or log softmax for each vocabulary token when using beam search kwargs (`Dict[str, Any]`, *optional*): Additional logits processor specific kwargs. Return: `jnp.ndarray` of shape `(batch_size, config.vocab_size)`: The processed prediction scores. ''' class snake_case_ : @add_start_docstrings(__snake_case ) def __call__( self :Dict ,__snake_case :jnp.ndarray ,__snake_case :jnp.ndarray ) -> jnp.ndarray: raise NotImplementedError( F'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) class snake_case_ : @add_start_docstrings(__snake_case ) def __call__( self :List[str] ,__snake_case :jnp.ndarray ,__snake_case :jnp.ndarray ) -> jnp.ndarray: raise NotImplementedError( F'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) class snake_case_ (lowerCamelCase_ ): @add_start_docstrings(__snake_case ) def __call__( self :Dict ,__snake_case :jnp.ndarray ,__snake_case :jnp.ndarray ,__snake_case :int ,**__snake_case :Any ) -> jnp.ndarray: for processor in self: a__ = inspect.signature(processor.__call__ ).parameters if len(__snake_case ) > 3: if not all(arg in kwargs for arg in list(function_args.keys() )[2:] ): raise ValueError( F'Make sure that all the required parameters: {list(function_args.keys() )} for ' F'{processor.__class__} are passed to the logits processor.' ) a__ = processor(__snake_case ,__snake_case ,__snake_case ,**__snake_case ) else: a__ = processor(__snake_case ,__snake_case ,__snake_case ) return scores class snake_case_ (lowerCamelCase_ ): def __init__( self :str ,__snake_case :float ) -> Tuple: if not isinstance(__snake_case ,__snake_case ) or not (temperature > 0): raise ValueError(F'`temperature` has to be a strictly positive float, but is {temperature}' ) a__ = temperature def __call__( self :Optional[int] ,__snake_case :jnp.ndarray ,__snake_case :jnp.ndarray ,__snake_case :int ) -> jnp.ndarray: a__ = scores / self.temperature return scores class snake_case_ (lowerCamelCase_ ): def __init__( self :Any ,__snake_case :float ,__snake_case :float = -float('Inf' ) ,__snake_case :int = 1 ) -> Dict: if not isinstance(__snake_case ,__snake_case ) or (top_p < 0 or top_p > 1.0): raise ValueError(F'`top_p` has to be a float > 0 and < 1, but is {top_p}' ) if not isinstance(__snake_case ,__snake_case ) or (min_tokens_to_keep < 1): raise ValueError(F'`min_tokens_to_keep` has to be a positive integer, but is {min_tokens_to_keep}' ) a__ = top_p a__ = filter_value a__ = min_tokens_to_keep def __call__( self :Optional[int] ,__snake_case :jnp.ndarray ,__snake_case :jnp.ndarray ,__snake_case :int ) -> jnp.ndarray: a__ , a__ = lax.top_k(__snake_case ,scores.shape[-1] ) a__ = jnp.full_like(__snake_case ,self.filter_value ) a__ = jax.nn.softmax(__snake_case ,axis=-1 ).cumsum(axis=-1 ) a__ = cumulative_probs < self.top_p # include the token that is higher than top_p as well a__ = jnp.roll(__snake_case ,1 ) score_mask |= score_mask.at[:, 0].set(__snake_case ) # min tokens to keep a__ = score_mask.at[:, : self.min_tokens_to_keep].set(__snake_case ) a__ = jnp.where(__snake_case ,__snake_case ,__snake_case ) a__ = jax.lax.sort_key_val(__snake_case ,__snake_case )[-1] return next_scores class snake_case_ (lowerCamelCase_ ): def __init__( self :List[str] ,__snake_case :int ,__snake_case :float = -float('Inf' ) ,__snake_case :int = 1 ) -> Any: if not isinstance(__snake_case ,__snake_case ) or top_k <= 0: raise ValueError(F'`top_k` has to be a strictly positive integer, but is {top_k}' ) a__ = max(__snake_case ,__snake_case ) a__ = filter_value def __call__( self :int ,__snake_case :jnp.ndarray ,__snake_case :jnp.ndarray ,__snake_case :int ) -> jnp.ndarray: a__ , a__ = scores.shape a__ = jnp.full(batch_size * vocab_size ,self.filter_value ) a__ = min(self.top_k ,scores.shape[-1] ) # Safety check a__ , a__ = lax.top_k(__snake_case ,__snake_case ) a__ = jnp.broadcast_to((jnp.arange(__snake_case ) * vocab_size)[:, None] ,(batch_size, topk) ).flatten() a__ = topk_scores.flatten() a__ = topk_indices.flatten() + shift a__ = next_scores_flat.at[topk_indices_flat].set(__snake_case ) a__ = next_scores_flat.reshape(__snake_case ,__snake_case ) return next_scores class snake_case_ (lowerCamelCase_ ): def __init__( self :int ,__snake_case :int ) -> str: a__ = bos_token_id def __call__( self :List[Any] ,__snake_case :jnp.ndarray ,__snake_case :jnp.ndarray ,__snake_case :int ) -> jnp.ndarray: a__ = jnp.full(scores.shape ,-float('inf' ) ) a__ = 1 - jnp.bool_(cur_len - 1 ) a__ = jnp.where(__snake_case ,new_scores.at[:, self.bos_token_id].set(0 ) ,__snake_case ) return scores class snake_case_ (lowerCamelCase_ ): def __init__( self :Union[str, Any] ,__snake_case :int ,__snake_case :int ) -> List[Any]: a__ = max_length a__ = eos_token_id def __call__( self :int ,__snake_case :jnp.ndarray ,__snake_case :jnp.ndarray ,__snake_case :int ) -> jnp.ndarray: a__ = jnp.full(scores.shape ,-float('inf' ) ) a__ = 1 - jnp.bool_(cur_len - self.max_length + 1 ) a__ = jnp.where(__snake_case ,new_scores.at[:, self.eos_token_id].set(0 ) ,__snake_case ) return scores class snake_case_ (lowerCamelCase_ ): def __init__( self :str ,__snake_case :int ,__snake_case :int ) -> List[str]: if not isinstance(__snake_case ,__snake_case ) or min_length < 0: raise ValueError(F'`min_length` has to be a positive integer, but is {min_length}' ) if not isinstance(__snake_case ,__snake_case ) or eos_token_id < 0: raise ValueError(F'`eos_token_id` has to be a positive integer, but is {eos_token_id}' ) a__ = min_length a__ = eos_token_id def __call__( self :Any ,__snake_case :jnp.ndarray ,__snake_case :jnp.ndarray ,__snake_case :int ) -> jnp.ndarray: # create boolean flag to decide if min length penalty should be applied a__ = 1 - jnp.clip(cur_len - self.min_length ,0 ,1 ) a__ = jnp.where(__snake_case ,scores.at[:, self.eos_token_id].set(-float('inf' ) ) ,__snake_case ) return scores class snake_case_ (lowerCamelCase_ ): def __init__( self :Optional[int] ,__snake_case :List[str] ,__snake_case :Optional[int] ) -> Tuple: a__ = list(__snake_case ) a__ = begin_index def __call__( self :str ,__snake_case :List[str] ,__snake_case :str ,__snake_case :int ) -> str: a__ = 1 - jnp.bool_(cur_len - self.begin_index ) a__ = jnp.where(__snake_case ,scores.at[:, self.begin_suppress_tokens].set(-float('inf' ) ) ,__snake_case ) return scores class snake_case_ (lowerCamelCase_ ): def __init__( self :List[str] ,__snake_case :list ) -> List[Any]: a__ = list(__snake_case ) def __call__( self :Dict ,__snake_case :jnp.ndarray ,__snake_case :jnp.ndarray ,__snake_case :int ) -> jnp.ndarray: a__ = scores.at[..., self.suppress_tokens].set(-float('inf' ) ) return scores class snake_case_ (lowerCamelCase_ ): def __init__( self :Dict ,__snake_case :Optional[int] ) -> Union[str, Any]: a__ = dict(__snake_case ) # Converts the dictionary of format {index: token} containing the tokens to be forced to an array, where the # index of the array corresponds to the index of the token to be forced, for XLA compatibility. # Indexes without forced tokens will have a negative value. a__ = jnp.ones((max(force_token_map.keys() ) + 1) ,dtype=jnp.intaa ) * -1 for index, token in force_token_map.items(): if token is not None: a__ = force_token_array.at[index].set(__snake_case ) a__ = jnp.intaa(__snake_case ) def __call__( self :Optional[int] ,__snake_case :jnp.ndarray ,__snake_case :jnp.ndarray ,__snake_case :int ) -> jnp.ndarray: def _force_token(__snake_case :Optional[Any] ): a__ = scores.shape[0] a__ = self.force_token_array[generation_idx] a__ = jnp.ones_like(__snake_case ,dtype=scores.dtype ) * -float('inf' ) a__ = jnp.zeros((batch_size, 1) ,dtype=scores.dtype ) a__ = lax.dynamic_update_slice(__snake_case ,__snake_case ,(0, current_token) ) return new_scores a__ = lax.cond( cur_len >= self.force_token_array.shape[0] ,lambda: scores ,lambda: lax.cond( self.force_token_array[cur_len] >= 0 ,lambda: _force_token(__snake_case ) ,lambda: scores ,) ,) return scores class snake_case_ (lowerCamelCase_ ): def __init__( self :Any ,__snake_case :List[str] ,__snake_case :str ,__snake_case :List[Any] ) -> Optional[int]: a__ = generate_config.eos_token_id a__ = generate_config.no_timestamps_token_id a__ = generate_config.no_timestamps_token_id + 1 a__ = decoder_input_length + 1 if generate_config.is_multilingual: # room for language token and task token self.begin_index += 2 if hasattr(__snake_case ,'max_initial_timestamp_index' ): a__ = generate_config.max_initial_timestamp_index else: a__ = model_config.vocab_size if self.max_initial_timestamp_index is None: a__ = model_config.vocab_size def __call__( self :Any ,__snake_case :List[Any] ,__snake_case :Optional[int] ,__snake_case :Optional[Any] ) -> Tuple: # suppress <|notimestamps|> which is handled by without_timestamps a__ = scores.at[:, self.no_timestamps_token_id].set(-float('inf' ) ) def handle_pairs(__snake_case :List[str] ,__snake_case :Union[str, Any] ): a__ = jnp.where((cur_len - self.begin_index) >= 1 ,__snake_case ,__snake_case ) a__ = jnp.where( input_ids_k[cur_len - 1] >= self.timestamp_begin ,True and last_was_timestamp ,__snake_case ,) a__ = jnp.where((cur_len - self.begin_index) < 2 ,__snake_case ,__snake_case ) a__ = jnp.where( input_ids_k[cur_len - 2] >= self.timestamp_begin ,__snake_case ,__snake_case ,) return jnp.where( __snake_case ,jnp.where( penultimate_was_timestamp > 0 ,scores_k.at[self.timestamp_begin :].set(-float('inf' ) ) ,scores_k.at[: self.eos_token_id].set(-float('inf' ) ) ,) ,__snake_case ,) a__ = jax.vmap(__snake_case )(__snake_case ,__snake_case ) a__ = jnp.where(cur_len == self.begin_index ,__snake_case ,__snake_case ) a__ = jnp.where( self.max_initial_timestamp_index is not None ,True and apply_max_initial_timestamp ,__snake_case ,) a__ = self.timestamp_begin + self.max_initial_timestamp_index a__ = jnp.where( __snake_case ,scores.at[:, last_allowed + 1 :].set(-float('inf' ) ) ,__snake_case ,) # if sum of probability over timestamps is above any other token, sample timestamp a__ = jax.nn.log_softmax(__snake_case ,axis=-1 ) def handle_cumulative_probs(__snake_case :Dict ,__snake_case :List[Any] ): a__ = jax.nn.logsumexp(logprobs_k[self.timestamp_begin :] ,axis=-1 ) a__ = jnp.max(logprobs_k[: self.timestamp_begin] ) return jnp.where( timestamp_logprob > max_text_token_logprob ,scores_k.at[: self.timestamp_begin].set(-float('inf' ) ) ,__snake_case ,) a__ = jax.vmap(__snake_case )(__snake_case ,__snake_case ) return scores
109
1
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets a__: Optional[Any] = '\\n@inproceedings{lin-2004-rouge,\n title = \"{ROUGE}: A Package for Automatic Evaluation of Summaries\",\n author = \"Lin, Chin-Yew\",\n booktitle = \"Text Summarization Branches Out\",\n month = jul,\n year = \"2004\",\n address = \"Barcelona, Spain\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W04-1013\",\n pages = \"74--81\",\n}\n' a__: Any = '\\nROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for\nevaluating automatic summarization and machine translation software in natural language processing.\nThe metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation.\n\nNote that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters.\n\nThis metrics is a wrapper around Google Research reimplementation of ROUGE:\nhttps://github.com/google-research/google-research/tree/master/rouge\n' a__: Any = '\nCalculates average rouge scores for a list of hypotheses and references\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n rouge_types: A list of rouge types to calculate.\n Valid names:\n `\"rouge{n}\"` (e.g. `\"rouge1\"`, `\"rouge2\"`) where: {n} is the n-gram based scoring,\n `\"rougeL\"`: Longest common subsequence based scoring.\n `\"rougeLSum\"`: rougeLsum splits text using `\"\n\"`.\n See details in https://github.com/huggingface/datasets/issues/617\n use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes.\n use_aggregator: Return aggregates if this is set to True\nReturns:\n rouge1: rouge_1 (precision, recall, f1),\n rouge2: rouge_2 (precision, recall, f1),\n rougeL: rouge_l (precision, recall, f1),\n rougeLsum: rouge_lsum (precision, recall, f1)\nExamples:\n\n >>> rouge = datasets.load_metric(\'rouge\')\n >>> predictions = [\"hello there\", \"general kenobi\"]\n >>> references = [\"hello there\", \"general kenobi\"]\n >>> results = rouge.compute(predictions=predictions, references=references)\n >>> print(list(results.keys()))\n [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\']\n >>> print(results[\"rouge1\"])\n AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0))\n >>> print(results[\"rouge1\"].mid.fmeasure)\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE__ ( datasets.Metric ): def UpperCamelCase ( self ): return datasets.MetricInfo( description=_DESCRIPTION,citation=_CITATION,inputs_description=_KWARGS_DESCRIPTION,features=datasets.Features( { '''predictions''': datasets.Value('''string''',id='''sequence''' ), '''references''': datasets.Value('''string''',id='''sequence''' ), } ),codebase_urls=['''https://github.com/google-research/google-research/tree/master/rouge'''],reference_urls=[ '''https://en.wikipedia.org/wiki/ROUGE_(metric)''', '''https://github.com/google-research/google-research/tree/master/rouge''', ],) def UpperCamelCase ( self,__lowerCamelCase,__lowerCamelCase,__lowerCamelCase=None,__lowerCamelCase=True,__lowerCamelCase=False ): if rouge_types is None: A__ = ['''rouge1''', '''rouge2''', '''rougeL''', '''rougeLsum'''] A__ = rouge_scorer.RougeScorer(rouge_types=__lowercase,use_stemmer=__lowercase ) if use_aggregator: A__ = scoring.BootstrapAggregator() else: A__ = [] for ref, pred in zip(__lowercase,__lowercase ): A__ = scorer.score(__lowercase,__lowercase ) if use_aggregator: aggregator.add_scores(__lowercase ) else: scores.append(__lowercase ) if use_aggregator: A__ = aggregator.aggregate() else: A__ = {} for key in scores[0]: A__ = [score[key] for score in scores] return result
193
from __future__ import annotations lowerCamelCase__ = """#""" class SCREAMING_SNAKE_CASE : def __init__( self : Optional[Any] ): '''simple docstring''' __a = {} def UpperCamelCase_ ( self : Optional[Any] , __lowercase : str ): '''simple docstring''' __a = self._trie for char in text: if char not in trie: __a = {} __a = trie[char] __a = True def UpperCamelCase_ ( self : Tuple , __lowercase : str ): '''simple docstring''' __a = self._trie for char in prefix: if char in trie: __a = trie[char] else: return [] return self._elements(__lowercase ) def UpperCamelCase_ ( self : Optional[int] , __lowercase : dict ): '''simple docstring''' __a = [] for c, v in d.items(): __a = [""" """] if c == END else [(c + s) for s in self._elements(__lowercase )] result.extend(__lowercase ) return tuple(__lowercase ) lowerCamelCase__ = Trie() lowerCamelCase__ = ("""depart""", """detergent""", """daring""", """dog""", """deer""", """deal""") for word in words: trie.insert_word(word) def lowerCAmelCase__ ( _SCREAMING_SNAKE_CASE : str ): """simple docstring""" __a = trie.find_word(_SCREAMING_SNAKE_CASE ) return tuple(string + word for word in suffixes ) def lowerCAmelCase__ ( ): """simple docstring""" print(autocomplete_using_trie("""de""" ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
302
0
'''simple docstring''' import operator def a_ ( lowerCamelCase : list , lowerCamelCase : bool = False , lowerCamelCase : list | None = None ): lowerCAmelCase = operator.lt if reverse else operator.gt lowerCAmelCase = solution or [] if not arr: return solution lowerCAmelCase = [arr.pop(0 )] for i, item in enumerate(lowerCamelCase ): if _operator(lowerCamelCase , sublist[-1] ): sublist.append(lowerCamelCase ) arr.pop(lowerCamelCase ) # merging sublist into solution list if not solution: solution.extend(lowerCamelCase ) else: while sublist: lowerCAmelCase = sublist.pop(0 ) for i, xx in enumerate(lowerCamelCase ): if not _operator(lowerCamelCase , lowerCamelCase ): solution.insert(lowerCamelCase , lowerCamelCase ) break else: solution.append(lowerCamelCase ) strand_sort(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return solution if __name__ == "__main__": assert strand_sort([4, 3, 5, 1, 2]) == [1, 2, 3, 4, 5] assert strand_sort([4, 3, 5, 1, 2], reverse=True) == [5, 4, 3, 2, 1]
55
'''simple docstring''' import math def a_ ( lowerCamelCase : int ): lowerCAmelCase = math.loga(math.sqrt(4 * positive_integer + 1 ) / 2 + 1 / 2 ) return exponent == int(lowerCamelCase ) def a_ ( lowerCamelCase : float = 1 / 12345 ): lowerCAmelCase = 0 lowerCAmelCase = 0 lowerCAmelCase = 3 while True: lowerCAmelCase = (integer**2 - 1) / 4 # if candidate is an integer, then there is a partition for k if partition_candidate == int(lowerCamelCase ): lowerCAmelCase = int(lowerCamelCase ) total_partitions += 1 if check_partition_perfect(lowerCamelCase ): perfect_partitions += 1 if perfect_partitions > 0: if perfect_partitions / total_partitions < max_proportion: return int(lowerCamelCase ) integer += 1 if __name__ == "__main__": print(F'''{solution() = }''')
55
1
'''simple docstring''' from __future__ import annotations def lowerCAmelCase_ ( _lowerCamelCase: list[int] , _lowerCamelCase: list[int] , _lowerCamelCase: int ): __SCREAMING_SNAKE_CASE : List[Any] = list(range(len(_lowerCamelCase ) ) ) __SCREAMING_SNAKE_CASE : Optional[Any] = [v / w for v, w in zip(_lowerCamelCase , _lowerCamelCase )] index.sort(key=lambda _lowerCamelCase : ratio[i] , reverse=_lowerCamelCase ) __SCREAMING_SNAKE_CASE : float = 0 __SCREAMING_SNAKE_CASE : list[float] = [0] * len(_lowerCamelCase ) for i in index: if weight[i] <= capacity: __SCREAMING_SNAKE_CASE : str = 1 max_value += value[i] capacity -= weight[i] else: __SCREAMING_SNAKE_CASE : int = capacity / weight[i] max_value += value[i] * capacity / weight[i] break return max_value, fractions if __name__ == "__main__": import doctest doctest.testmod()
112
'''simple docstring''' from __future__ import annotations def lowerCAmelCase_ ( _lowerCamelCase: list[int] ): if not nums: return 0 __SCREAMING_SNAKE_CASE : Optional[int] = nums[0] __SCREAMING_SNAKE_CASE : Union[str, Any] = 0 for num in nums[1:]: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Any = ( max_excluding + num, max(_lowerCamelCase , _lowerCamelCase ), ) return max(_lowerCamelCase , _lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
112
1
'''simple docstring''' import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class UpperCAmelCase_ ( unittest.TestCase ): lowerCamelCase : Optional[Any] = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def __UpperCAmelCase ( self : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Any ) -> List[Any]: lowerCAmelCase = hf_hub_download( repo_id='nateraw/video-demo' , filename='archery.mp4' , repo_type='dataset' ) lowerCAmelCase = VideoClassificationPipeline(model=_lowerCAmelCase , image_processor=_lowerCAmelCase , top_k=2 ) lowerCAmelCase = [ example_video_filepath, """https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4""", ] return video_classifier, examples def __UpperCAmelCase ( self : int , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[Any] ) -> Dict: for example in examples: lowerCAmelCase = video_classifier(_lowerCAmelCase ) self.assertEqual( _lowerCAmelCase , [ {'score': ANY(_lowerCAmelCase ), 'label': ANY(_lowerCAmelCase )}, {'score': ANY(_lowerCAmelCase ), 'label': ANY(_lowerCAmelCase )}, ] , ) @require_torch def __UpperCAmelCase ( self : Optional[Any] ) -> List[Any]: lowerCAmelCase = """hf-internal-testing/tiny-random-VideoMAEForVideoClassification""" lowerCAmelCase = VideoMAEFeatureExtractor( size={'shortest_edge': 1_0} , crop_size={'height': 1_0, 'width': 1_0} ) lowerCAmelCase = pipeline( 'video-classification' , model=_lowerCAmelCase , feature_extractor=_lowerCAmelCase , frame_sampling_rate=4 ) lowerCAmelCase = hf_hub_download(repo_id='nateraw/video-demo' , filename='archery.mp4' , repo_type='dataset' ) lowerCAmelCase = video_classifier(_lowerCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(_lowerCAmelCase , decimals=4 ) , [{'score': 0.5_199, 'label': 'LABEL_0'}, {'score': 0.4_801, 'label': 'LABEL_1'}] , ) lowerCAmelCase = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(_lowerCAmelCase , decimals=4 ) , [ [{'score': 0.5_199, 'label': 'LABEL_0'}, {'score': 0.4_801, 'label': 'LABEL_1'}], [{'score': 0.5_199, 'label': 'LABEL_0'}, {'score': 0.4_801, 'label': 'LABEL_1'}], ] , ) @require_tf def __UpperCAmelCase ( self : int ) -> Optional[Any]: pass
370
'''simple docstring''' import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEmbeddings, BertLayer, BertPooler, BertPreTrainedModel, ) def a_ ( lowerCamelCase : int ): lowerCAmelCase = torch.exp(lowerCamelCase ) lowerCAmelCase = torch.sum(lowerCamelCase , dim=1 ) # sum of exp(x_i) lowerCAmelCase = torch.sum(x * exp_x , dim=1 ) # sum of x_i * exp(x_i) return torch.log(lowerCamelCase ) - B / A class UpperCAmelCase_ ( nn.Module ): def __init__( self : int , UpperCAmelCase__ : int ) -> str: super().__init__() lowerCAmelCase = config.output_attentions lowerCAmelCase = config.output_hidden_states lowerCAmelCase = nn.ModuleList([BertLayer(UpperCAmelCase__ ) for _ in range(config.num_hidden_layers )] ) lowerCAmelCase = nn.ModuleList([BertHighway(UpperCAmelCase__ ) for _ in range(config.num_hidden_layers )] ) lowerCAmelCase = [-1 for _ in range(config.num_hidden_layers )] def __UpperCAmelCase ( self : Union[str, Any] , UpperCAmelCase__ : str ) -> int: if (type(UpperCAmelCase__ ) is float) or (type(UpperCAmelCase__ ) is int): for i in range(len(self.early_exit_entropy ) ): lowerCAmelCase = x else: lowerCAmelCase = x def __UpperCAmelCase ( self : Dict , UpperCAmelCase__ : List[str] ) -> Optional[Any]: lowerCAmelCase = pooler.state_dict() for highway in self.highway: for name, param in highway.pooler.state_dict().items(): param.copy_(loaded_model[name] ) def __UpperCAmelCase ( self : Any , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : int=None , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : Optional[int]=None , ) -> str: lowerCAmelCase = () lowerCAmelCase = () lowerCAmelCase = () for i, layer_module in enumerate(self.layer ): if self.output_hidden_states: lowerCAmelCase = all_hidden_states + (hidden_states,) lowerCAmelCase = layer_module( UpperCAmelCase__ , UpperCAmelCase__ , head_mask[i] , UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = layer_outputs[0] if self.output_attentions: lowerCAmelCase = all_attentions + (layer_outputs[1],) lowerCAmelCase = (hidden_states,) if self.output_hidden_states: lowerCAmelCase = current_outputs + (all_hidden_states,) if self.output_attentions: lowerCAmelCase = current_outputs + (all_attentions,) lowerCAmelCase = self.highway[i](UpperCAmelCase__ ) # logits, pooled_output if not self.training: lowerCAmelCase = highway_exit[0] lowerCAmelCase = entropy(UpperCAmelCase__ ) lowerCAmelCase = highway_exit + (highway_entropy,) # logits, hidden_states(?), entropy lowerCAmelCase = all_highway_exits + (highway_exit,) if highway_entropy < self.early_exit_entropy[i]: lowerCAmelCase = (highway_logits,) + current_outputs[1:] + (all_highway_exits,) raise HighwayException(UpperCAmelCase__ , i + 1 ) else: lowerCAmelCase = all_highway_exits + (highway_exit,) # Add last layer if self.output_hidden_states: lowerCAmelCase = all_hidden_states + (hidden_states,) lowerCAmelCase = (hidden_states,) if self.output_hidden_states: lowerCAmelCase = outputs + (all_hidden_states,) if self.output_attentions: lowerCAmelCase = outputs + (all_attentions,) lowerCAmelCase = outputs + (all_highway_exits,) return outputs # last-layer hidden state, (all hidden states), (all attentions), all highway exits @add_start_docstrings( '''The Bert Model transformer with early exiting (DeeBERT). ''' , __lowercase , ) class UpperCAmelCase_ ( __lowercase ): def __init__( self : Union[str, Any] , UpperCAmelCase__ : Optional[int] ) -> str: super().__init__(UpperCAmelCase__ ) lowerCAmelCase = config lowerCAmelCase = BertEmbeddings(UpperCAmelCase__ ) lowerCAmelCase = DeeBertEncoder(UpperCAmelCase__ ) lowerCAmelCase = BertPooler(UpperCAmelCase__ ) self.init_weights() def __UpperCAmelCase ( self : Any ) -> int: self.encoder.init_highway_pooler(self.pooler ) def __UpperCAmelCase ( self : Optional[Any] ) -> Optional[Any]: return self.embeddings.word_embeddings def __UpperCAmelCase ( self : Any , UpperCAmelCase__ : Dict ) -> List[Any]: lowerCAmelCase = value def __UpperCAmelCase ( self : List[str] , UpperCAmelCase__ : int ) -> Dict: for layer, heads in heads_to_prune.items(): self.encoder.layer[layer].attention.prune_heads(UpperCAmelCase__ ) @add_start_docstrings_to_model_forward(UpperCAmelCase__ ) def __UpperCAmelCase ( self : Union[str, Any] , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : str=None , UpperCAmelCase__ : List[Any]=None , ) -> Optional[int]: if input_ids is not None and inputs_embeds is not None: raise ValueError('You cannot specify both input_ids and inputs_embeds at the same time' ) elif input_ids is not None: lowerCAmelCase = input_ids.size() elif inputs_embeds is not None: lowerCAmelCase = inputs_embeds.size()[:-1] else: raise ValueError('You have to specify either input_ids or inputs_embeds' ) lowerCAmelCase = input_ids.device if input_ids is not None else inputs_embeds.device if attention_mask is None: lowerCAmelCase = torch.ones(UpperCAmelCase__ , device=UpperCAmelCase__ ) if encoder_attention_mask is None: lowerCAmelCase = torch.ones(UpperCAmelCase__ , device=UpperCAmelCase__ ) if token_type_ids is None: lowerCAmelCase = torch.zeros(UpperCAmelCase__ , dtype=torch.long , device=UpperCAmelCase__ ) # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] # ourselves in which case we just need to make it broadcastable to all heads. lowerCAmelCase = self.get_extended_attention_mask(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) # If a 2D ou 3D attention mask is provided for the cross-attention # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] if encoder_attention_mask.dim() == 3: lowerCAmelCase = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.dim() == 2: lowerCAmelCase = encoder_attention_mask[:, None, None, :] lowerCAmelCase = encoder_extended_attention_mask.to( dtype=next(self.parameters() ).dtype ) # fp16 compatibility lowerCAmelCase = (1.0 - encoder_extended_attention_mask) * -10_000.0 # Prepare head mask if needed # 1.0 in head_mask indicate we keep the head # attention_probs has shape bsz x n_heads x N x N # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] lowerCAmelCase = self.get_head_mask(UpperCAmelCase__ , self.config.num_hidden_layers ) lowerCAmelCase = self.embeddings( input_ids=UpperCAmelCase__ , position_ids=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , inputs_embeds=UpperCAmelCase__ ) lowerCAmelCase = self.encoder( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , head_mask=UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , encoder_attention_mask=UpperCAmelCase__ , ) lowerCAmelCase = encoder_outputs[0] lowerCAmelCase = self.pooler(UpperCAmelCase__ ) lowerCAmelCase = ( sequence_output, pooled_output, ) + encoder_outputs[ 1: ] # add hidden_states and attentions if they are here return outputs # sequence_output, pooled_output, (hidden_states), (attentions), highway exits class UpperCAmelCase_ ( __lowercase ): def __init__( self : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : int ) -> Dict: lowerCAmelCase = message lowerCAmelCase = exit_layer # start from 1! class UpperCAmelCase_ ( nn.Module ): def __init__( self : List[Any] , UpperCAmelCase__ : Optional[Any] ) -> List[str]: super().__init__() lowerCAmelCase = BertPooler(UpperCAmelCase__ ) lowerCAmelCase = nn.Dropout(config.hidden_dropout_prob ) lowerCAmelCase = nn.Linear(config.hidden_size , config.num_labels ) def __UpperCAmelCase ( self : str , UpperCAmelCase__ : Dict ) -> Optional[int]: # Pooler lowerCAmelCase = encoder_outputs[0] lowerCAmelCase = self.pooler(UpperCAmelCase__ ) # "return" pooler_output # BertModel lowerCAmelCase = (pooler_input, pooler_output) + encoder_outputs[1:] # "return" bmodel_output # Dropout and classification lowerCAmelCase = bmodel_output[1] lowerCAmelCase = self.dropout(UpperCAmelCase__ ) lowerCAmelCase = self.classifier(UpperCAmelCase__ ) return logits, pooled_output @add_start_docstrings( '''Bert Model (with early exiting - DeeBERT) with a classifier on top, also takes care of multi-layer training. ''' , __lowercase , ) class UpperCAmelCase_ ( __lowercase ): def __init__( self : Dict , UpperCAmelCase__ : Dict ) -> Any: super().__init__(UpperCAmelCase__ ) lowerCAmelCase = config.num_labels lowerCAmelCase = config.num_hidden_layers lowerCAmelCase = DeeBertModel(UpperCAmelCase__ ) lowerCAmelCase = nn.Dropout(config.hidden_dropout_prob ) lowerCAmelCase = nn.Linear(config.hidden_size , self.config.num_labels ) self.init_weights() @add_start_docstrings_to_model_forward(UpperCAmelCase__ ) def __UpperCAmelCase ( self : Dict , UpperCAmelCase__ : List[Any]=None , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Any=None , UpperCAmelCase__ : str=None , UpperCAmelCase__ : List[Any]=None , UpperCAmelCase__ : List[Any]=-1 , UpperCAmelCase__ : Optional[Any]=False , ) -> Dict: lowerCAmelCase = self.num_layers try: lowerCAmelCase = self.bert( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , position_ids=UpperCAmelCase__ , head_mask=UpperCAmelCase__ , inputs_embeds=UpperCAmelCase__ , ) # sequence_output, pooled_output, (hidden_states), (attentions), highway exits lowerCAmelCase = outputs[1] lowerCAmelCase = self.dropout(UpperCAmelCase__ ) lowerCAmelCase = self.classifier(UpperCAmelCase__ ) lowerCAmelCase = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: lowerCAmelCase = e.message lowerCAmelCase = e.exit_layer lowerCAmelCase = outputs[0] if not self.training: lowerCAmelCase = entropy(UpperCAmelCase__ ) lowerCAmelCase = [] lowerCAmelCase = [] if labels is not None: if self.num_labels == 1: # We are doing regression lowerCAmelCase = MSELoss() lowerCAmelCase = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: lowerCAmelCase = CrossEntropyLoss() lowerCAmelCase = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits lowerCAmelCase = [] for highway_exit in outputs[-1]: lowerCAmelCase = highway_exit[0] if not self.training: highway_logits_all.append(UpperCAmelCase__ ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression lowerCAmelCase = MSELoss() lowerCAmelCase = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: lowerCAmelCase = CrossEntropyLoss() lowerCAmelCase = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(UpperCAmelCase__ ) if train_highway: lowerCAmelCase = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: lowerCAmelCase = (loss,) + outputs if not self.training: lowerCAmelCase = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: lowerCAmelCase = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), (highway_exits)
55
0
import unittest import numpy as np from transformers.file_utils import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class _snake_case ( unittest.TestCase ): def __init__( self , _a , _a=7 , _a=3 , _a=18 , _a=30 , _a=400 , _a=True , _a=None , _a=True , _a=[0.5, 0.5, 0.5] , _a=[0.5, 0.5, 0.5] , ): __magic_name__ : int = size if size is not None else {"height": 18, "width": 18} __magic_name__ : Union[str, Any] = parent __magic_name__ : int = batch_size __magic_name__ : List[Any] = num_channels __magic_name__ : str = image_size __magic_name__ : Optional[int] = min_resolution __magic_name__ : List[str] = max_resolution __magic_name__ : str = do_resize __magic_name__ : Optional[int] = size __magic_name__ : Tuple = do_normalize __magic_name__ : Optional[Any] = image_mean __magic_name__ : int = image_std def SCREAMING_SNAKE_CASE ( self ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _snake_case ( snake_case , unittest.TestCase ): UpperCamelCase__ = DPTImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[Any] = DPTImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE ( self ): return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[str] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_a , "image_mean" ) ) self.assertTrue(hasattr(_a , "image_std" ) ) self.assertTrue(hasattr(_a , "do_normalize" ) ) self.assertTrue(hasattr(_a , "do_resize" ) ) self.assertTrue(hasattr(_a , "size" ) ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : int = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"height": 18, "width": 18} ) __magic_name__ : List[str] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"height": 42, "width": 42} ) def SCREAMING_SNAKE_CASE ( self ): # Initialize image_processing __magic_name__ : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __magic_name__ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_a ) for image in image_inputs: self.assertIsInstance(_a , Image.Image ) # Test not batched input __magic_name__ : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __magic_name__ : str = image_processing(_a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def SCREAMING_SNAKE_CASE ( self ): # Initialize image_processing __magic_name__ : str = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __magic_name__ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_a , numpify=_a ) for image in image_inputs: self.assertIsInstance(_a , np.ndarray ) # Test not batched input __magic_name__ : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __magic_name__ : List[Any] = image_processing(_a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def SCREAMING_SNAKE_CASE ( self ): # Initialize image_processing __magic_name__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __magic_name__ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_a , torchify=_a ) for image in image_inputs: self.assertIsInstance(_a , torch.Tensor ) # Test not batched input __magic_name__ : str = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __magic_name__ : Any = image_processing(_a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , )
281
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
281
1
'''simple docstring''' def A_( A : list[list]): UpperCamelCase = current_set.copy() for row_index, row in enumerate(A): UpperCamelCase = row[0] for column_index, column in enumerate(A): if magnitude == 0: UpperCamelCase = column continue UpperCamelCase = column / magnitude # Subtract to cancel term UpperCamelCase = current_set[0] UpperCamelCase = [first_row] UpperCamelCase = current_set[1::] for row in current_set: UpperCamelCase = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(A) continue for column_index in range(len(A)): temp_row.append(first_row[column_index] - row[column_index]) final_set.append(A) # Create next recursion iteration set if len(final_set[0]) != 3: UpperCamelCase = final_set[0] UpperCamelCase = [] UpperCamelCase = [] for row in final_set[1::]: current_first_column.append(row[0]) next_iteration.append(row[1::]) UpperCamelCase = simplify(A) for i in range(len(A)): resultant[i].insert(0 , current_first_column[i]) resultant.insert(0 , A) UpperCamelCase = resultant return final_set def A_( A : list[list]): if len(A) == 0: raise IndexError('solve_simultaneous() requires n lists of length n+1') UpperCamelCase = len(A) + 1 if any(len(A) != _length for item in equations): raise IndexError('solve_simultaneous() requires n lists of length n+1') for row in equations: if any(not isinstance(A , (int, float)) for column in row): raise ValueError('solve_simultaneous() requires lists of integers') if len(A) == 1: return [equations[0][-1] / equations[0][0]] UpperCamelCase = equations.copy() if any(0 in row for row in data_set): UpperCamelCase = data_set.copy() UpperCamelCase = [] for row_index, row in enumerate(A): if 0 not in row: UpperCamelCase = data_set.pop(A) break if not full_row: raise ValueError('solve_simultaneous() requires at least 1 full equation') data_set.insert(0 , A) UpperCamelCase = data_set.copy() UpperCamelCase = simplify(A) UpperCamelCase = simplified[::-1] UpperCamelCase = [] for row in simplified: UpperCamelCase = row[-1] if not solutions: if row[-2] == 0: solutions.append(0) continue solutions.append(current_solution / row[-2]) continue UpperCamelCase = row.copy()[: len(A) - 1 :] while temp_row[0] == 0: temp_row.pop(0) if len(A) == 0: solutions.append(0) continue UpperCamelCase = temp_row[1::] UpperCamelCase = temp_row[::-1] for column_index, column in enumerate(A): current_solution -= column * solutions[column_index] solutions.append(A) UpperCamelCase = [] for item in solutions: final.append(float(round(A , 5))) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase : List[Any] = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
251
'''simple docstring''' import json import os from collections import Counter import torch import torchvision import torchvision.transforms as transforms from PIL import Image from torch import nn from torch.utils.data import Dataset lowerCAmelCase : str = {1: (1, 1), 2: (2, 1), 3: (3, 1), 4: (2, 2), 5: (5, 1), 6: (3, 2), 7: (7, 1), 8: (4, 2), 9: (3, 3)} class SCREAMING_SNAKE_CASE__ ( nn.Module): def __init__( self , A_ )-> int: '''simple docstring''' super().__init__() UpperCamelCase = torchvision.models.resnetaaa(pretrained=A_ ) UpperCamelCase = list(model.children() )[:-2] UpperCamelCase = nn.Sequential(*A_ ) UpperCamelCase = nn.AdaptiveAvgPoolad(POOLING_BREAKDOWN[args.num_image_embeds] ) def UpperCAmelCase_ ( self , A_ )-> List[Any]: '''simple docstring''' UpperCamelCase = self.pool(self.model(A_ ) ) UpperCamelCase = torch.flatten(A_ , start_dim=2 ) UpperCamelCase = out.transpose(1 , 2 ).contiguous() return out # BxNx2048 class SCREAMING_SNAKE_CASE__ ( snake_case_): def __init__( self , A_ , A_ , A_ , A_ , A_ )-> Dict: '''simple docstring''' UpperCamelCase = [json.loads(A_ ) for l in open(A_ )] UpperCamelCase = os.path.dirname(A_ ) UpperCamelCase = tokenizer UpperCamelCase = labels UpperCamelCase = len(A_ ) UpperCamelCase = max_seq_length UpperCamelCase = transforms def __len__( self )-> Union[str, Any]: '''simple docstring''' return len(self.data ) def __getitem__( self , A_ )-> Any: '''simple docstring''' UpperCamelCase = torch.LongTensor(self.tokenizer.encode(self.data[index]['text'] , add_special_tokens=A_ ) ) UpperCamelCase , UpperCamelCase , UpperCamelCase = sentence[0], sentence[1:-1], sentence[-1] UpperCamelCase = sentence[: self.max_seq_length] UpperCamelCase = torch.zeros(self.n_classes ) UpperCamelCase = 1 UpperCamelCase = Image.open(os.path.join(self.data_dir , self.data[index]['img'] ) ).convert('RGB' ) UpperCamelCase = self.transforms(A_ ) return { "image_start_token": start_token, "image_end_token": end_token, "sentence": sentence, "image": image, "label": label, } def UpperCAmelCase_ ( self )-> List[Any]: '''simple docstring''' UpperCamelCase = Counter() for row in self.data: label_freqs.update(row['label'] ) return label_freqs def A_( A : Union[str, Any]): UpperCamelCase = [len(row['sentence']) for row in batch] UpperCamelCase , UpperCamelCase = len(A), max(A) UpperCamelCase = torch.zeros(A , A , dtype=torch.long) UpperCamelCase = torch.zeros(A , A , dtype=torch.long) for i_batch, (input_row, length) in enumerate(zip(A , A)): UpperCamelCase = input_row['sentence'] UpperCamelCase = 1 UpperCamelCase = torch.stack([row['image'] for row in batch]) UpperCamelCase = torch.stack([row['label'] for row in batch]) UpperCamelCase = torch.stack([row['image_start_token'] for row in batch]) UpperCamelCase = torch.stack([row['image_end_token'] for row in batch]) return text_tensor, mask_tensor, img_tensor, img_start_token, img_end_token, tgt_tensor def A_( ): return [ "Crime", "Drama", "Thriller", "Action", "Comedy", "Romance", "Documentary", "Short", "Mystery", "History", "Family", "Adventure", "Fantasy", "Sci-Fi", "Western", "Horror", "Sport", "War", "Music", "Musical", "Animation", "Biography", "Film-Noir", ] def A_( ): return transforms.Compose( [ transforms.Resize(256), transforms.CenterCrop(224), transforms.ToTensor(), transforms.Normalize( mean=[0.46_777_044, 0.44_531_429, 0.40_661_017] , std=[0.12_221_994, 0.12_145_835, 0.14_380_469] , ), ])
251
1
'''simple docstring''' import pandas as pd from matplotlib import pyplot as plt from sklearn.linear_model import LinearRegression # Splitting the dataset into the Training set and Test set from sklearn.model_selection import train_test_split # Fitting Polynomial Regression to the dataset from sklearn.preprocessing import PolynomialFeatures # Importing the dataset lowerCamelCase__ = pd.read_csv( 'https://s3.us-west-2.amazonaws.com/public.gamelab.fun/dataset/' 'position_salaries.csv' ) lowerCamelCase__ = dataset.iloc[:, 1:2].values lowerCamelCase__ = dataset.iloc[:, 2].values lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ = train_test_split(X, y, test_size=0.2, random_state=0) lowerCamelCase__ = PolynomialFeatures(degree=4) lowerCamelCase__ = poly_reg.fit_transform(X) lowerCamelCase__ = LinearRegression() pol_reg.fit(X_poly, y) def __lowerCAmelCase (): plt.scatter(__lowerCAmelCase , __lowerCAmelCase , color="red" ) plt.plot(__lowerCAmelCase , pol_reg.predict(poly_reg.fit_transform(__lowerCAmelCase ) ) , color="blue" ) plt.title("Truth or Bluff (Linear Regression)" ) plt.xlabel("Position level" ) plt.ylabel("Salary" ) plt.show() if __name__ == "__main__": viz_polymonial() # Predicting a new result with Polymonial Regression pol_reg.predict(poly_reg.fit_transform([[5.5]])) # output should be 132148.43750003
234
'''simple docstring''' import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(UpperCAmelCase__ ) , "Tatoeba directory does not exist." ) class lowerCAmelCase__ ( unittest.TestCase ): @cached_property def lowerCAmelCase__ ( self : Optional[Any] ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : List[str] = tempfile.mkdtemp() return TatoebaConverter(save_dir=lowerCamelCase__ ) @slow def lowerCAmelCase__ ( self : Optional[int] ) ->List[str]: '''simple docstring''' self.resolver.convert_models(["heb-eng"] ) @slow def lowerCAmelCase__ ( self : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : Tuple = self.resolver.write_model_card("opus-mt-he-en" , dry_run=lowerCamelCase__ ) assert mmeta["long_pair"] == "heb-eng"
234
1
'''simple docstring''' def lowerCAmelCase__ ( lowerCamelCase : int = 3 ,lowerCamelCase : int = 7 ,lowerCamelCase : int = 1000000 ): _A : List[str] = 0 _A : Any = 1 for current_denominator in range(1 ,limit + 1 ): _A : str = current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: _A : int = current_numerator _A : int = current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=1000000))
370
'''simple docstring''' def lowerCAmelCase__ ( lowerCamelCase : int = 10 ): if not isinstance(lowerCamelCase ,lowerCamelCase ) or n < 0: raise ValueError('Invalid input' ) _A : Optional[Any] = 10**n _A : List[str] = 28433 * (pow(2 ,7830457 ,lowerCamelCase )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(f"""{solution(10) = }""")
227
0