code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCamelCase (_snake_case , _snake_case , unittest.TestCase ): '''simple docstring''' _snake_case : Tuple = IFInpaintingPipeline _snake_case : List[str] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} _snake_case : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _snake_case : Union[str, Any] = PipelineTesterMixin.required_optional_params - {'''latents'''} def __UpperCAmelCase ( self ) -> Dict: return self._get_dummy_components() def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase=0 ) -> int: if str(_UpperCamelCase ).startswith('mps' ): UpperCAmelCase_ : List[Any] = torch.manual_seed(_UpperCamelCase ) else: UpperCAmelCase_ : Union[str, Any] = torch.Generator(device=_UpperCamelCase ).manual_seed(_UpperCamelCase ) UpperCAmelCase_ : Any = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_UpperCamelCase ) ).to(_UpperCamelCase ) UpperCAmelCase_ : Any = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_UpperCamelCase ) ).to(_UpperCamelCase ) UpperCAmelCase_ : List[str] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __UpperCAmelCase ( self ) -> Any: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __UpperCAmelCase ( self ) -> List[Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def __UpperCAmelCase ( self ) -> Optional[int]: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __UpperCAmelCase ( self ) -> Optional[int]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __UpperCAmelCase ( self ) -> Dict: self._test_save_load_local() def __UpperCAmelCase ( self ) -> List[str]: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
29
from string import ascii_lowercase, ascii_uppercase def a__ ( _UpperCamelCase : str ): if not sentence: return "" __lowerCamelCase = dict(zip(_UpperCamelCase ,_UpperCamelCase ) ) return lower_to_upper.get(sentence[0] ,sentence[0] ) + sentence[1:] if __name__ == "__main__": from doctest import testmod testmod()
330
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() snake_case_ = logging.get_logger(__name__) def _lowerCAmelCase ( lowercase_ ): UpperCAmelCase = DPTConfig(embedding_type='hybrid' ) if "large" in checkpoint_url: UpperCAmelCase = 1024 UpperCAmelCase = 4096 UpperCAmelCase = 24 UpperCAmelCase = 16 UpperCAmelCase = [5, 11, 17, 23] UpperCAmelCase = [256, 512, 1024, 1024] UpperCAmelCase = (1, 384, 384) if "nyu" or "midas" in checkpoint_url: UpperCAmelCase = 768 UpperCAmelCase = [1, 1, 1, 0.5] UpperCAmelCase = [256, 512, 768, 768] UpperCAmelCase = 150 UpperCAmelCase = 16 UpperCAmelCase = (1, 384, 384) UpperCAmelCase = False UpperCAmelCase = 'project' if "ade" in checkpoint_url: UpperCAmelCase = True UpperCAmelCase = 768 UpperCAmelCase = [1, 1, 1, 0.5] UpperCAmelCase = 150 UpperCAmelCase = 16 UpperCAmelCase = 'huggingface/label-files' UpperCAmelCase = 'ade20k-id2label.json' UpperCAmelCase = json.load(open(cached_download(hf_hub_url(lowercase_ , lowercase_ , repo_type='dataset' ) ) , 'r' ) ) UpperCAmelCase = {int(lowercase_ ): v for k, v in idalabel.items()} UpperCAmelCase = idalabel UpperCAmelCase = {v: k for k, v in idalabel.items()} UpperCAmelCase = [1, 150, 480, 480] return config, expected_shape def _lowerCAmelCase ( lowercase_ ): UpperCAmelCase = ['pretrained.model.head.weight', 'pretrained.model.head.bias'] for k in ignore_keys: state_dict.pop(lowercase_ , lowercase_ ) def _lowerCAmelCase ( lowercase_ ): if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): UpperCAmelCase = name.replace('pretrained.model' , 'dpt.encoder' ) if "pretrained.model" in name: UpperCAmelCase = name.replace('pretrained.model' , 'dpt.embeddings' ) if "patch_embed" in name: UpperCAmelCase = name.replace('patch_embed' , '' ) if "pos_embed" in name: UpperCAmelCase = name.replace('pos_embed' , 'position_embeddings' ) if "attn.proj" in name: UpperCAmelCase = name.replace('attn.proj' , 'attention.output.dense' ) if "proj" in name and "project" not in name: UpperCAmelCase = name.replace('proj' , 'projection' ) if "blocks" in name: UpperCAmelCase = name.replace('blocks' , 'layer' ) if "mlp.fc1" in name: UpperCAmelCase = name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: UpperCAmelCase = name.replace('mlp.fc2' , 'output.dense' ) if "norm1" in name and "backbone" not in name: UpperCAmelCase = name.replace('norm1' , 'layernorm_before' ) if "norm2" in name and "backbone" not in name: UpperCAmelCase = name.replace('norm2' , 'layernorm_after' ) if "scratch.output_conv" in name: UpperCAmelCase = name.replace('scratch.output_conv' , 'head' ) if "scratch" in name: UpperCAmelCase = name.replace('scratch' , 'neck' ) if "layer1_rn" in name: UpperCAmelCase = name.replace('layer1_rn' , 'convs.0' ) if "layer2_rn" in name: UpperCAmelCase = name.replace('layer2_rn' , 'convs.1' ) if "layer3_rn" in name: UpperCAmelCase = name.replace('layer3_rn' , 'convs.2' ) if "layer4_rn" in name: UpperCAmelCase = name.replace('layer4_rn' , 'convs.3' ) if "refinenet" in name: UpperCAmelCase = int(name[len('neck.refinenet' ) : len('neck.refinenet' ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 UpperCAmelCase = name.replace(F"""refinenet{layer_idx}""" , F"""fusion_stage.layers.{abs(layer_idx-4 )}""" ) if "out_conv" in name: UpperCAmelCase = name.replace('out_conv' , 'projection' ) if "resConfUnit1" in name: UpperCAmelCase = name.replace('resConfUnit1' , 'residual_layer1' ) if "resConfUnit2" in name: UpperCAmelCase = name.replace('resConfUnit2' , 'residual_layer2' ) if "conv1" in name: UpperCAmelCase = name.replace('conv1' , 'convolution1' ) if "conv2" in name: UpperCAmelCase = name.replace('conv2' , 'convolution2' ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: UpperCAmelCase = name.replace('pretrained.act_postprocess1.0.project.0' , 'neck.reassemble_stage.readout_projects.0.0' ) if "pretrained.act_postprocess2.0.project.0" in name: UpperCAmelCase = name.replace('pretrained.act_postprocess2.0.project.0' , 'neck.reassemble_stage.readout_projects.1.0' ) if "pretrained.act_postprocess3.0.project.0" in name: UpperCAmelCase = name.replace('pretrained.act_postprocess3.0.project.0' , 'neck.reassemble_stage.readout_projects.2.0' ) if "pretrained.act_postprocess4.0.project.0" in name: UpperCAmelCase = name.replace('pretrained.act_postprocess4.0.project.0' , 'neck.reassemble_stage.readout_projects.3.0' ) # resize blocks if "pretrained.act_postprocess1.3" in name: UpperCAmelCase = name.replace('pretrained.act_postprocess1.3' , 'neck.reassemble_stage.layers.0.projection' ) if "pretrained.act_postprocess1.4" in name: UpperCAmelCase = name.replace('pretrained.act_postprocess1.4' , 'neck.reassemble_stage.layers.0.resize' ) if "pretrained.act_postprocess2.3" in name: UpperCAmelCase = name.replace('pretrained.act_postprocess2.3' , 'neck.reassemble_stage.layers.1.projection' ) if "pretrained.act_postprocess2.4" in name: UpperCAmelCase = name.replace('pretrained.act_postprocess2.4' , 'neck.reassemble_stage.layers.1.resize' ) if "pretrained.act_postprocess3.3" in name: UpperCAmelCase = name.replace('pretrained.act_postprocess3.3' , 'neck.reassemble_stage.layers.2.projection' ) if "pretrained.act_postprocess4.3" in name: UpperCAmelCase = name.replace('pretrained.act_postprocess4.3' , 'neck.reassemble_stage.layers.3.projection' ) if "pretrained.act_postprocess4.4" in name: UpperCAmelCase = name.replace('pretrained.act_postprocess4.4' , 'neck.reassemble_stage.layers.3.resize' ) if "pretrained" in name: UpperCAmelCase = name.replace('pretrained' , 'dpt' ) if "bn" in name: UpperCAmelCase = name.replace('bn' , 'batch_norm' ) if "head" in name: UpperCAmelCase = name.replace('head' , 'head.head' ) if "encoder.norm" in name: UpperCAmelCase = name.replace('encoder.norm' , 'layernorm' ) if "auxlayer" in name: UpperCAmelCase = name.replace('auxlayer' , 'auxiliary_head.head' ) if "backbone" in name: UpperCAmelCase = name.replace('backbone' , 'backbone.bit.encoder' ) if ".." in name: UpperCAmelCase = name.replace('..' , '.' ) if "stem.conv" in name: UpperCAmelCase = name.replace('stem.conv' , 'bit.embedder.convolution' ) if "blocks" in name: UpperCAmelCase = name.replace('blocks' , 'layers' ) if "convolution" in name and "backbone" in name: UpperCAmelCase = name.replace('convolution' , 'conv' ) if "layer" in name and "backbone" in name: UpperCAmelCase = name.replace('layer' , 'layers' ) if "backbone.bit.encoder.bit" in name: UpperCAmelCase = name.replace('backbone.bit.encoder.bit' , 'backbone.bit' ) if "embedder.conv" in name: UpperCAmelCase = name.replace('embedder.conv' , 'embedder.convolution' ) if "backbone.bit.encoder.stem.norm" in name: UpperCAmelCase = name.replace('backbone.bit.encoder.stem.norm' , 'backbone.bit.embedder.norm' ) return name def _lowerCAmelCase ( lowercase_ , lowercase_ ): for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase = state_dict.pop(F"""dpt.encoder.layer.{i}.attn.qkv.weight""" ) UpperCAmelCase = state_dict.pop(F"""dpt.encoder.layer.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase = in_proj_weight[: config.hidden_size, :] UpperCAmelCase = in_proj_bias[: config.hidden_size] UpperCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( ): UpperCAmelCase = 'http://images.cocodataset.org/val2017/000000039769.jpg' UpperCAmelCase = Image.open(requests.get(lowercase_ , stream=lowercase_ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): UpperCAmelCase , UpperCAmelCase = get_dpt_config(lowercase_ ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") UpperCAmelCase = torch.load(lowercase_ , map_location='cpu' ) # remove certain keys remove_ignore_keys_(lowercase_ ) # rename keys for key in state_dict.copy().keys(): UpperCAmelCase = state_dict.pop(lowercase_ ) UpperCAmelCase = val # read in qkv matrices read_in_q_k_v(lowercase_ , lowercase_ ) # load HuggingFace model UpperCAmelCase = DPTForSemanticSegmentation(lowercase_ ) if 'ade' in checkpoint_url else DPTForDepthEstimation(lowercase_ ) model.load_state_dict(lowercase_ ) model.eval() # Check outputs on an image UpperCAmelCase = 480 if 'ade' in checkpoint_url else 384 UpperCAmelCase = DPTImageProcessor(size=lowercase_ ) UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(lowercase_ , return_tensors='pt' ) # forward pass UpperCAmelCase = model(**lowercase_ ).logits if 'ade' in checkpoint_url else model(**lowercase_ ).predicted_depth if show_prediction: UpperCAmelCase = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode='bicubic' , align_corners=lowercase_ , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 255 ).show() if pytorch_dump_folder_path is not None: Path(lowercase_ ).mkdir(exist_ok=lowercase_ ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowercase_ ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(lowercase_ ) if push_to_hub: model.push_to_hub('ybelkada/dpt-hybrid-midas' ) image_processor.push_to_hub('ybelkada/dpt-hybrid-midas' ) if __name__ == "__main__": snake_case_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint_url""", default="""https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt""", type=str, help="""URL of the original DPT checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=False, help="""Path to the output PyTorch model directory.""", ) parser.add_argument( """--push_to_hub""", action="""store_true""", ) parser.add_argument( """--model_name""", default="""dpt-large""", type=str, help="""Name of the model, in case you're pushing to the hub.""", ) parser.add_argument( """--show_prediction""", action="""store_true""", ) snake_case_ = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
365
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool class A_ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __UpperCamelCase = """philschmid/bart-large-cnn-samsum""" __UpperCamelCase = ( """This is a tool that summarizes an English text. It takes an input `text` containing the text to summarize, """ """and returns a summary of the text.""" ) __UpperCamelCase = """summarizer""" __UpperCamelCase = AutoTokenizer __UpperCamelCase = AutoModelForSeqaSeqLM __UpperCamelCase = ["""text"""] __UpperCamelCase = ["""text"""] def UpperCAmelCase__ ( self :int , lowercase_ :Tuple ) -> int: return self.pre_processor(lowercase_ , return_tensors='pt' , truncation=lowercase_ ) def UpperCAmelCase__ ( self :Dict , lowercase_ :Dict ) -> Optional[int]: return self.model.generate(**lowercase_ )[0] def UpperCAmelCase__ ( self :Dict , lowercase_ :Tuple ) -> List[str]: return self.pre_processor.decode(lowercase_ , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ )
181
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_funnel import FunnelTokenizer __A : int = logging.get_logger(__name__) __A : Optional[Any] = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} __A : List[Any] = [ """small""", """small-base""", """medium""", """medium-base""", """intermediate""", """intermediate-base""", """large""", """large-base""", """xlarge""", """xlarge-base""", ] __A : Optional[int] = { """vocab_file""": { """funnel-transformer/small""": """https://huggingface.co/funnel-transformer/small/resolve/main/vocab.txt""", """funnel-transformer/small-base""": """https://huggingface.co/funnel-transformer/small-base/resolve/main/vocab.txt""", """funnel-transformer/medium""": """https://huggingface.co/funnel-transformer/medium/resolve/main/vocab.txt""", """funnel-transformer/medium-base""": ( """https://huggingface.co/funnel-transformer/medium-base/resolve/main/vocab.txt""" ), """funnel-transformer/intermediate""": ( """https://huggingface.co/funnel-transformer/intermediate/resolve/main/vocab.txt""" ), """funnel-transformer/intermediate-base""": ( """https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/vocab.txt""" ), """funnel-transformer/large""": """https://huggingface.co/funnel-transformer/large/resolve/main/vocab.txt""", """funnel-transformer/large-base""": """https://huggingface.co/funnel-transformer/large-base/resolve/main/vocab.txt""", """funnel-transformer/xlarge""": """https://huggingface.co/funnel-transformer/xlarge/resolve/main/vocab.txt""", """funnel-transformer/xlarge-base""": ( """https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """funnel-transformer/small""": """https://huggingface.co/funnel-transformer/small/resolve/main/tokenizer.json""", """funnel-transformer/small-base""": ( """https://huggingface.co/funnel-transformer/small-base/resolve/main/tokenizer.json""" ), """funnel-transformer/medium""": """https://huggingface.co/funnel-transformer/medium/resolve/main/tokenizer.json""", """funnel-transformer/medium-base""": ( """https://huggingface.co/funnel-transformer/medium-base/resolve/main/tokenizer.json""" ), """funnel-transformer/intermediate""": ( """https://huggingface.co/funnel-transformer/intermediate/resolve/main/tokenizer.json""" ), """funnel-transformer/intermediate-base""": ( """https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/tokenizer.json""" ), """funnel-transformer/large""": """https://huggingface.co/funnel-transformer/large/resolve/main/tokenizer.json""", """funnel-transformer/large-base""": ( """https://huggingface.co/funnel-transformer/large-base/resolve/main/tokenizer.json""" ), """funnel-transformer/xlarge""": """https://huggingface.co/funnel-transformer/xlarge/resolve/main/tokenizer.json""", """funnel-transformer/xlarge-base""": ( """https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/tokenizer.json""" ), }, } __A : Optional[int] = {F'funnel-transformer/{name}': 512 for name in _model_names} __A : Optional[int] = {F'funnel-transformer/{name}': {"""do_lower_case""": True} for name in _model_names} class __A ( UpperCamelCase__ ): lowerCAmelCase_ : str = VOCAB_FILES_NAMES lowerCAmelCase_ : List[Any] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : Dict = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase_ : Union[str, Any] = FunnelTokenizer lowerCAmelCase_ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : int = 2 def __init__( self : str , UpperCAmelCase_ : Union[str, Any]=None , UpperCAmelCase_ : Any=None , UpperCAmelCase_ : int=True , UpperCAmelCase_ : List[Any]="<unk>" , UpperCAmelCase_ : Union[str, Any]="<sep>" , UpperCAmelCase_ : Tuple="<pad>" , UpperCAmelCase_ : str="<cls>" , UpperCAmelCase_ : Optional[Any]="<mask>" , UpperCAmelCase_ : Optional[Any]="<s>" , UpperCAmelCase_ : Optional[int]="</s>" , UpperCAmelCase_ : Any=True , UpperCAmelCase_ : Any=True , UpperCAmelCase_ : str=None , UpperCAmelCase_ : Optional[Any]="##" , **UpperCAmelCase_ : List[str] , ): super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , clean_text=lowerCAmelCase__ , tokenize_chinese_chars=lowerCAmelCase__ , strip_accents=lowerCAmelCase__ , wordpieces_prefix=lowerCAmelCase__ , **lowerCAmelCase__ , ) lowerCAmelCase : Optional[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , lowerCAmelCase__ ) != do_lower_case or normalizer_state.get('strip_accents' , lowerCAmelCase__ ) != strip_accents or normalizer_state.get('handle_chinese_chars' , lowerCAmelCase__ ) != tokenize_chinese_chars ): lowerCAmelCase : List[str] = getattr(lowerCAmelCase__ , normalizer_state.pop('type' ) ) lowerCAmelCase : Union[str, Any] = do_lower_case lowerCAmelCase : Any = strip_accents lowerCAmelCase : Optional[Any] = tokenize_chinese_chars lowerCAmelCase : Dict = normalizer_class(**lowerCAmelCase__ ) lowerCAmelCase : Any = do_lower_case def lowercase__ ( self : List[str] , UpperCAmelCase_ : int , UpperCAmelCase_ : int=None ): lowerCAmelCase : Dict = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowercase__ ( self : int , UpperCAmelCase_ : str , UpperCAmelCase_ : str = None ): lowerCAmelCase : Optional[int] = [self.sep_token_id] lowerCAmelCase : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowercase__ ( self : Tuple , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : int = None ): lowerCAmelCase : Tuple = self._tokenizer.model.save(lowerCAmelCase__ , name=lowerCAmelCase__ ) return tuple(lowerCAmelCase__ )
138
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device UpperCAmelCase : int = False class __lowerCAmelCase ( unittest.TestCase): pass @nightly @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase): def _lowercase ( self ) -> Tuple: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self ) -> Any: '''simple docstring''' a__ : str =VersatileDiffusionPipeline.from_pretrained("shi-labs/versatile-diffusion" , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) a__ : int =load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) a__ : Optional[Any] =torch.manual_seed(0 ) a__ : Optional[Any] =pipe.dual_guided( prompt="first prompt" , image=lowerCAmelCase__ , text_to_image_strength=0.75 , generator=lowerCAmelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCAmelCase__ ) a__ : str =VersatileDiffusionPipeline.from_pretrained(lowerCAmelCase__ , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) a__ : Optional[Any] =generator.manual_seed(0 ) a__ : Tuple =pipe.dual_guided( prompt="first prompt" , image=lowerCAmelCase__ , text_to_image_strength=0.75 , generator=lowerCAmelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def _lowercase ( self ) -> Any: '''simple docstring''' a__ : str =VersatileDiffusionPipeline.from_pretrained("shi-labs/versatile-diffusion" , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) a__ : Optional[Any] ="cyberpunk 2077" a__ : int =load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) a__ : Union[str, Any] =torch.manual_seed(0 ) a__ : Tuple =pipe.dual_guided( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , text_to_image_strength=0.75 , generator=lowerCAmelCase__ , guidance_scale=7.5 , num_inference_steps=5_0 , output_type="numpy" , ).images a__ : int =image[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) a__ : Any =np.array([0.14_48, 0.16_19, 0.17_41, 0.10_86, 0.11_47, 0.11_28, 0.11_99, 0.11_65, 0.10_01] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 a__ : str ="A painting of a squirrel eating a burger " a__ : Optional[int] =torch.manual_seed(0 ) a__ : str =pipe.text_to_image( prompt=lowerCAmelCase__ , generator=lowerCAmelCase__ , guidance_scale=7.5 , num_inference_steps=5_0 , output_type="numpy" ).images a__ : Any =image[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) a__ : Optional[int] =np.array([0.33_67, 0.31_69, 0.26_56, 0.38_70, 0.47_90, 0.37_96, 0.40_09, 0.48_78, 0.47_78] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 a__ : Optional[Any] =pipe.image_variation(lowerCAmelCase__ , generator=lowerCAmelCase__ , output_type="numpy" ).images a__ : Union[str, Any] =image[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) a__ : Any =np.array([0.30_76, 0.31_23, 0.32_84, 0.37_82, 0.37_70, 0.38_94, 0.42_97, 0.43_31, 0.44_56] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
95
0
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowercase__ = version.parse(importlib_metadata.version("nltk")) if NLTK_VERSION >= version.Version("3.6.4"): from nltk import word_tokenize lowercase__ = "\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n" lowercase__ = "\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n" lowercase__ = "\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n 'meteor': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric('meteor')\n >>> predictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"]\n >>> references = [\"It is a guide to action that ensures that the military will forever heed Party commands\"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results[\"meteor\"], 4))\n 0.6944\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ ( datasets.Metric ): '''simple docstring''' def UpperCAmelCase_ ( self : Dict ) -> str: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'] , reference_urls=[ 'https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score', 'https://en.wikipedia.org/wiki/METEOR', ] , ) def UpperCAmelCase_ ( self : Optional[int] , lowercase_ : Any ) -> Optional[Any]: import nltk nltk.download('wordnet' ) if NLTK_VERSION >= version.Version('3.6.5' ): nltk.download('punkt' ) if NLTK_VERSION >= version.Version('3.6.6' ): nltk.download('omw-1.4' ) def UpperCAmelCase_ ( self : List[Any] , lowercase_ : Tuple , lowercase_ : List[Any] , lowercase_ : List[Any]=0.9 , lowercase_ : Tuple=3 , lowercase_ : Optional[Any]=0.5 ) -> Dict: if NLTK_VERSION >= version.Version('3.6.5' ): UpperCAmelCase : Union[str, Any] = [ meteor_score.single_meteor_score( word_tokenize(lowercase_ ) , word_tokenize(lowercase_ ) , alpha=lowercase_ , beta=lowercase_ , gamma=lowercase_ ) for ref, pred in zip(lowercase_ , lowercase_ ) ] else: UpperCAmelCase : Dict = [ meteor_score.single_meteor_score(lowercase_ , lowercase_ , alpha=lowercase_ , beta=lowercase_ , gamma=lowercase_ ) for ref, pred in zip(lowercase_ , lowercase_ ) ] return {"meteor": np.mean(lowercase_ )}
280
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowercase__ = { "configuration_xlm": ["XLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLMConfig", "XLMOnnxConfig"], "tokenization_xlm": ["XLMTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XLMForMultipleChoice", "XLMForQuestionAnswering", "XLMForQuestionAnsweringSimple", "XLMForSequenceClassification", "XLMForTokenClassification", "XLMModel", "XLMPreTrainedModel", "XLMWithLMHeadModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLMForMultipleChoice", "TFXLMForQuestionAnsweringSimple", "TFXLMForSequenceClassification", "TFXLMForTokenClassification", "TFXLMMainLayer", "TFXLMModel", "TFXLMPreTrainedModel", "TFXLMWithLMHeadModel", ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
280
1
from __future__ import annotations from collections.abc import Callable from typing import Any, Generic, TypeVar __lowerCamelCase : Union[str, Any] = TypeVar('''T''') class __snake_case ( Generic[T] ): def __init__( self : Dict , _lowercase : list[T] , _lowercase : Callable[[T, T], T] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = len(_lowercase ) SCREAMING_SNAKE_CASE__ = [any_type for _ in range(self.N )] + arr SCREAMING_SNAKE_CASE__ = fnc self.build() def __a ( self : List[Any] ): """simple docstring""" for p in range(self.N - 1 , 0 , -1 ): SCREAMING_SNAKE_CASE__ = self.fn(self.st[p * 2] , self.st[p * 2 + 1] ) def __a ( self : Optional[int] , _lowercase : int , _lowercase : T ): """simple docstring""" p += self.N SCREAMING_SNAKE_CASE__ = v while p > 1: SCREAMING_SNAKE_CASE__ = p // 2 SCREAMING_SNAKE_CASE__ = self.fn(self.st[p * 2] , self.st[p * 2 + 1] ) def __a ( self : List[str] , _lowercase : int , _lowercase : int ): # noqa: E741 """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = l + self.N, r + self.N SCREAMING_SNAKE_CASE__ = None while l <= r: if l % 2 == 1: SCREAMING_SNAKE_CASE__ = self.st[l] if res is None else self.fn(_lowercase , self.st[l] ) if r % 2 == 0: SCREAMING_SNAKE_CASE__ = self.st[r] if res is None else self.fn(_lowercase , self.st[r] ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = (l + 1) // 2, (r - 1) // 2 return res if __name__ == "__main__": from functools import reduce __lowerCamelCase : List[str] = [1, 10, -2, 9, -3, 8, 4, -7, 5, 6, 11, -12] __lowerCamelCase : Optional[int] = { 0: 7, 1: 2, 2: 6, 3: -14, 4: 5, 5: 4, 6: 7, 7: -10, 8: 9, 9: 10, 10: 12, 11: 1, } __lowerCamelCase : str = SegmentTree(test_array, min) __lowerCamelCase : Any = SegmentTree(test_array, max) __lowerCamelCase : Any = SegmentTree(test_array, lambda a, b: a + b) def __SCREAMING_SNAKE_CASE ( ) -> None: """simple docstring""" for i in range(len(__UpperCamelCase ) ): for j in range(__UpperCamelCase , len(__UpperCamelCase ) ): SCREAMING_SNAKE_CASE__ = reduce(__UpperCamelCase , test_array[i : j + 1] ) SCREAMING_SNAKE_CASE__ = reduce(__UpperCamelCase , test_array[i : j + 1] ) SCREAMING_SNAKE_CASE__ = reduce(lambda __UpperCamelCase , __UpperCamelCase : a + b , test_array[i : j + 1] ) assert min_range == min_segment_tree.query(__UpperCamelCase , __UpperCamelCase ) assert max_range == max_segment_tree.query(__UpperCamelCase , __UpperCamelCase ) assert sum_range == sum_segment_tree.query(__UpperCamelCase , __UpperCamelCase ) test_all_segments() for index, value in test_updates.items(): __lowerCamelCase : int = value min_segment_tree.update(index, value) max_segment_tree.update(index, value) sum_segment_tree.update(index, value) test_all_segments()
219
import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml __lowerCamelCase : List[str] = NewType('''DataClass''', Any) __lowerCamelCase : Dict = NewType('''DataClassType''', Any) def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Any ) -> int: """simple docstring""" if isinstance(__UpperCamelCase , __UpperCamelCase ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( f"""Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).""" ) def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : list ) -> Callable[[str], Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = {str(__UpperCamelCase ): choice for choice in choices} return lambda __UpperCamelCase : str_to_choice.get(__UpperCamelCase , __UpperCamelCase ) def __SCREAMING_SNAKE_CASE ( *, __UpperCamelCase : Union[str, List[str]] = None , __UpperCamelCase : str = None , __UpperCamelCase : Any = dataclasses.MISSING , __UpperCamelCase : Callable[[], Any] = dataclasses.MISSING , __UpperCamelCase : dict = None , **__UpperCamelCase : Dict , ) -> dataclasses.Field: """simple docstring""" if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls SCREAMING_SNAKE_CASE__ = {} if aliases is not None: SCREAMING_SNAKE_CASE__ = aliases if help is not None: SCREAMING_SNAKE_CASE__ = help return dataclasses.field(metadata=__UpperCamelCase , default=__UpperCamelCase , default_factory=__UpperCamelCase , **__UpperCamelCase ) class __snake_case ( lowerCamelCase_ ): lowerCAmelCase_ = 42 def __init__( self : int , _lowercase : Union[DataClassType, Iterable[DataClassType]] , **_lowercase : List[str] ): """simple docstring""" if "formatter_class" not in kwargs: SCREAMING_SNAKE_CASE__ = ArgumentDefaultsHelpFormatter super().__init__(**_lowercase ) if dataclasses.is_dataclass(_lowercase ): SCREAMING_SNAKE_CASE__ = [dataclass_types] SCREAMING_SNAKE_CASE__ = list(_lowercase ) for dtype in self.dataclass_types: self._add_dataclass_arguments(_lowercase ) @staticmethod def __a ( _lowercase : ArgumentParser , _lowercase : dataclasses.Field ): """simple docstring""" SCREAMING_SNAKE_CASE__ = f"""--{field.name}""" SCREAMING_SNAKE_CASE__ = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , _lowercase ): raise RuntimeError( """Unresolved type detected, which should have been done with the help of """ """`typing.get_type_hints` method by default""" ) SCREAMING_SNAKE_CASE__ = kwargs.pop("""aliases""" , [] ) if isinstance(_lowercase , _lowercase ): SCREAMING_SNAKE_CASE__ = [aliases] SCREAMING_SNAKE_CASE__ = getattr(field.type , """__origin__""" , field.type ) if origin_type is Union or (hasattr(_lowercase , """UnionType""" ) and isinstance(_lowercase , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(_lowercase ) not in field.type.__args__ ): raise ValueError( """Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because""" """ the argument parser only supports one type per argument.""" f""" Problem encountered in field '{field.name}'.""" ) if type(_lowercase ) not in field.type.__args__: # filter `str` in Union SCREAMING_SNAKE_CASE__ = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] SCREAMING_SNAKE_CASE__ = getattr(field.type , """__origin__""" , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) SCREAMING_SNAKE_CASE__ = ( field.type.__args__[0] if isinstance(_lowercase , field.type.__args__[1] ) else field.type.__args__[1] ) SCREAMING_SNAKE_CASE__ = getattr(field.type , """__origin__""" , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) SCREAMING_SNAKE_CASE__ = {} if origin_type is Literal or (isinstance(field.type , _lowercase ) and issubclass(field.type , _lowercase )): if origin_type is Literal: SCREAMING_SNAKE_CASE__ = field.type.__args__ else: SCREAMING_SNAKE_CASE__ = [x.value for x in field.type] SCREAMING_SNAKE_CASE__ = make_choice_type_function(kwargs["""choices"""] ) if field.default is not dataclasses.MISSING: SCREAMING_SNAKE_CASE__ = field.default else: SCREAMING_SNAKE_CASE__ = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument SCREAMING_SNAKE_CASE__ = copy(_lowercase ) # Hack because type=bool in argparse does not behave as we want. SCREAMING_SNAKE_CASE__ = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. SCREAMING_SNAKE_CASE__ = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way SCREAMING_SNAKE_CASE__ = default # This tells argparse we accept 0 or 1 value after --field_name SCREAMING_SNAKE_CASE__ = """?""" # This is the value that will get picked if we do --field_name (without value) SCREAMING_SNAKE_CASE__ = True elif isclass(_lowercase ) and issubclass(_lowercase , _lowercase ): SCREAMING_SNAKE_CASE__ = field.type.__args__[0] SCREAMING_SNAKE_CASE__ = """+""" if field.default_factory is not dataclasses.MISSING: SCREAMING_SNAKE_CASE__ = field.default_factory() elif field.default is dataclasses.MISSING: SCREAMING_SNAKE_CASE__ = True else: SCREAMING_SNAKE_CASE__ = field.type if field.default is not dataclasses.MISSING: SCREAMING_SNAKE_CASE__ = field.default elif field.default_factory is not dataclasses.MISSING: SCREAMING_SNAKE_CASE__ = field.default_factory() else: SCREAMING_SNAKE_CASE__ = True parser.add_argument(_lowercase , *_lowercase , **_lowercase ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): SCREAMING_SNAKE_CASE__ = False parser.add_argument(f"""--no_{field.name}""" , action="""store_false""" , dest=field.name , **_lowercase ) def __a ( self : List[str] , _lowercase : DataClassType ): """simple docstring""" if hasattr(_lowercase , """_argument_group_name""" ): SCREAMING_SNAKE_CASE__ = self.add_argument_group(dtype._argument_group_name ) else: SCREAMING_SNAKE_CASE__ = self try: SCREAMING_SNAKE_CASE__ = get_type_hints(_lowercase ) except NameError: raise RuntimeError( f"""Type resolution failed for {dtype}. Try declaring the class in global scope or """ """removing line of `from __future__ import annotations` which opts in Postponed """ """Evaluation of Annotations (PEP 563)""" ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(_lowercase ): SCREAMING_SNAKE_CASE__ = """.""".join(map(_lowercase , sys.version_info[:3] ) ) raise RuntimeError( f"""Type resolution failed for {dtype} on Python {python_version}. Try removing """ """line of `from __future__ import annotations` which opts in union types as """ """`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To """ """support Python versions that lower than 3.10, you need to use """ """`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of """ """`X | None`.""" ) from ex raise for field in dataclasses.fields(_lowercase ): if not field.init: continue SCREAMING_SNAKE_CASE__ = type_hints[field.name] self._parse_dataclass_field(_lowercase , _lowercase ) def __a ( self : str , _lowercase : int=None , _lowercase : Optional[Any]=False , _lowercase : Union[str, Any]=True , _lowercase : Any=None , _lowercase : int=None , ): """simple docstring""" if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): SCREAMING_SNAKE_CASE__ = [] if args_filename: args_files.append(Path(_lowercase ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix(""".args""" ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values SCREAMING_SNAKE_CASE__ = ArgumentParser() args_file_parser.add_argument(_lowercase , type=_lowercase , action="""append""" ) # Use only remaining args for further parsing (remove the args_file_flag) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = args_file_parser.parse_known_args(args=_lowercase ) SCREAMING_SNAKE_CASE__ = vars(_lowercase ).get(args_file_flag.lstrip("""-""" ) , _lowercase ) if cmd_args_file_paths: args_files.extend([Path(_lowercase ) for p in cmd_args_file_paths] ) SCREAMING_SNAKE_CASE__ = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last SCREAMING_SNAKE_CASE__ = file_args + args if args is not None else file_args + sys.argv[1:] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.parse_known_args(args=_lowercase ) SCREAMING_SNAKE_CASE__ = [] for dtype in self.dataclass_types: SCREAMING_SNAKE_CASE__ = {f.name for f in dataclasses.fields(_lowercase ) if f.init} SCREAMING_SNAKE_CASE__ = {k: v for k, v in vars(_lowercase ).items() if k in keys} for k in keys: delattr(_lowercase , _lowercase ) SCREAMING_SNAKE_CASE__ = dtype(**_lowercase ) outputs.append(_lowercase ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(_lowercase ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f"""Some specified arguments are not used by the HfArgumentParser: {remaining_args}""" ) return (*outputs,) def __a ( self : List[str] , _lowercase : Dict[str, Any] , _lowercase : bool = False ): """simple docstring""" SCREAMING_SNAKE_CASE__ = set(args.keys() ) SCREAMING_SNAKE_CASE__ = [] for dtype in self.dataclass_types: SCREAMING_SNAKE_CASE__ = {f.name for f in dataclasses.fields(_lowercase ) if f.init} SCREAMING_SNAKE_CASE__ = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) SCREAMING_SNAKE_CASE__ = dtype(**_lowercase ) outputs.append(_lowercase ) if not allow_extra_keys and unused_keys: raise ValueError(f"""Some keys are not used by the HfArgumentParser: {sorted(_lowercase )}""" ) return tuple(_lowercase ) def __a ( self : List[str] , _lowercase : str , _lowercase : bool = False ): """simple docstring""" with open(Path(_lowercase ) , encoding="""utf-8""" ) as open_json_file: SCREAMING_SNAKE_CASE__ = json.loads(open_json_file.read() ) SCREAMING_SNAKE_CASE__ = self.parse_dict(_lowercase , allow_extra_keys=_lowercase ) return tuple(_lowercase ) def __a ( self : Tuple , _lowercase : str , _lowercase : bool = False ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.parse_dict(yaml.safe_load(Path(_lowercase ).read_text() ) , allow_extra_keys=_lowercase ) return tuple(_lowercase )
219
1
import argparse import os import re import numpy as np import PIL import torch from timm import create_model from torch.optim.lr_scheduler import OneCycleLR from torch.utils.data import DataLoader, Dataset from torchvision.transforms import Compose, RandomResizedCrop, Resize, ToTensor from accelerate import Accelerator def A ( lowercase ) -> List[str]: '''simple docstring''' UpperCamelCase = fname.split(os.path.sep )[-1] return re.search(R'^(.*)_\d+\.jpg$' , _UpperCAmelCase ).groups()[0] class lowercase ( __UpperCamelCase ): def __init__( self , A_ , A_=None , A_=None ) -> List[str]: """simple docstring""" UpperCamelCase = file_names UpperCamelCase = image_transform UpperCamelCase = label_to_id def __len__( self ) -> Dict: """simple docstring""" return len(self.file_names ) def __getitem__( self , A_ ) -> List[str]: """simple docstring""" UpperCamelCase = self.file_names[idx] UpperCamelCase = PIL.Image.open(A_ ) UpperCamelCase = raw_image.convert('RGB' ) if self.image_transform is not None: UpperCamelCase = self.image_transform(A_ ) UpperCamelCase = extract_label(A_ ) if self.label_to_id is not None: UpperCamelCase = self.label_to_id[label] return {"image": image, "label": label} def A ( lowercase , lowercase ) -> Tuple: '''simple docstring''' if args.with_tracking: UpperCamelCase = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , log_with='all' , project_dir=args.project_dir ) else: UpperCamelCase = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs UpperCamelCase = config['lr'] UpperCamelCase = int(config['num_epochs'] ) UpperCamelCase = int(config['seed'] ) UpperCamelCase = int(config['batch_size'] ) UpperCamelCase = config['image_size'] if not isinstance(_UpperCAmelCase , (list, tuple) ): UpperCamelCase = (image_size, image_size) # Parse out whether we are saving every epoch or after a certain number of batches if hasattr(args.checkpointing_steps , 'isdigit' ): if args.checkpointing_steps == "epoch": UpperCamelCase = args.checkpointing_steps elif args.checkpointing_steps.isdigit(): UpperCamelCase = int(args.checkpointing_steps ) else: raise ValueError( f'''Argument `checkpointing_steps` must be either a number or `epoch`. `{args.checkpointing_steps}` passed.''' ) else: UpperCamelCase = None # We need to initialize the trackers we use, and also store our configuration if args.with_tracking: UpperCamelCase = os.path.split(_UpperCAmelCase )[-1].split('.' )[0] accelerator.init_trackers(_UpperCAmelCase , _UpperCAmelCase ) # Grab all the image filenames UpperCamelCase = [os.path.join(args.data_dir , _UpperCAmelCase ) for fname in os.listdir(args.data_dir ) if fname.endswith('.jpg' )] # Build the label correspondences UpperCamelCase = [extract_label(_UpperCAmelCase ) for fname in file_names] UpperCamelCase = list(set(_UpperCAmelCase ) ) id_to_label.sort() UpperCamelCase = {lbl: i for i, lbl in enumerate(_UpperCAmelCase )} # Set the seed before splitting the data. np.random.seed(_UpperCAmelCase ) torch.manual_seed(_UpperCAmelCase ) torch.cuda.manual_seed_all(_UpperCAmelCase ) # Split our filenames between train and validation UpperCamelCase = np.random.permutation(len(_UpperCAmelCase ) ) UpperCamelCase = int(0.8 * len(_UpperCAmelCase ) ) UpperCamelCase = random_perm[:cut] UpperCamelCase = random_perm[cut:] # For training we use a simple RandomResizedCrop UpperCamelCase = Compose([RandomResizedCrop(_UpperCAmelCase , scale=(0.5, 1.0) ), ToTensor()] ) UpperCamelCase = PetsDataset( [file_names[i] for i in train_split] , image_transform=_UpperCAmelCase , label_to_id=_UpperCAmelCase ) # For evaluation, we use a deterministic Resize UpperCamelCase = Compose([Resize(_UpperCAmelCase ), ToTensor()] ) UpperCamelCase = PetsDataset([file_names[i] for i in eval_split] , image_transform=_UpperCAmelCase , label_to_id=_UpperCAmelCase ) # Instantiate dataloaders. UpperCamelCase = DataLoader(_UpperCAmelCase , shuffle=_UpperCAmelCase , batch_size=_UpperCAmelCase , num_workers=4 ) UpperCamelCase = DataLoader(_UpperCAmelCase , shuffle=_UpperCAmelCase , batch_size=_UpperCAmelCase , num_workers=4 ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) UpperCamelCase = create_model('resnet50d' , pretrained=_UpperCAmelCase , num_classes=len(_UpperCAmelCase ) ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). UpperCamelCase = model.to(accelerator.device ) # Freezing the base model for param in model.parameters(): UpperCamelCase = False for param in model.get_classifier().parameters(): UpperCamelCase = True # We normalize the batches of images to be a bit faster. UpperCamelCase = torch.tensor(model.default_cfg['mean'] )[None, :, None, None].to(accelerator.device ) UpperCamelCase = torch.tensor(model.default_cfg['std'] )[None, :, None, None].to(accelerator.device ) # Instantiate optimizer UpperCamelCase = torch.optim.Adam(params=model.parameters() , lr=lr / 25 ) # Instantiate learning rate scheduler UpperCamelCase = OneCycleLR(optimizer=_UpperCAmelCase , max_lr=_UpperCAmelCase , epochs=_UpperCAmelCase , steps_per_epoch=len(_UpperCAmelCase ) ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. UpperCamelCase = accelerator.prepare( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # We need to keep track of how many total steps we have iterated over UpperCamelCase = 0 # We also need to keep track of the starting epoch so files are named properly UpperCamelCase = 0 # Potentially load in the weights and states from a previous save if args.resume_from_checkpoint: if args.resume_from_checkpoint is not None or args.resume_from_checkpoint != "": accelerator.print(f'''Resumed from checkpoint: {args.resume_from_checkpoint}''' ) accelerator.load_state(args.resume_from_checkpoint ) UpperCamelCase = os.path.basename(args.resume_from_checkpoint ) else: # Get the most recent checkpoint UpperCamelCase = [f.name for f in os.scandir(os.getcwd() ) if f.is_dir()] dirs.sort(key=os.path.getctime ) UpperCamelCase = dirs[-1] # Sorts folders by date modified, most recent checkpoint is the last # Extract `epoch_{i}` or `step_{i}` UpperCamelCase = os.path.splitext(_UpperCAmelCase )[0] if "epoch" in training_difference: UpperCamelCase = int(training_difference.replace('epoch_' , '' ) ) + 1 UpperCamelCase = None else: UpperCamelCase = int(training_difference.replace('step_' , '' ) ) UpperCamelCase = resume_step // len(_UpperCAmelCase ) resume_step -= starting_epoch * len(_UpperCAmelCase ) # Now we train the model for epoch in range(_UpperCAmelCase , _UpperCAmelCase ): model.train() if args.with_tracking: UpperCamelCase = 0 if args.resume_from_checkpoint and epoch == starting_epoch and resume_step is not None: # We need to skip steps until we reach the resumed step UpperCamelCase = accelerator.skip_first_batches(_UpperCAmelCase , _UpperCAmelCase ) overall_step += resume_step else: # After the first iteration though, we need to go back to the original dataloader UpperCamelCase = train_dataloader for batch in active_dataloader: # We could avoid this line since we set the accelerator with `device_placement=True`. UpperCamelCase = {k: v.to(accelerator.device ) for k, v in batch.items()} UpperCamelCase = (batch['image'] - mean) / std UpperCamelCase = model(_UpperCAmelCase ) UpperCamelCase = torch.nn.functional.cross_entropy(_UpperCAmelCase , batch['label'] ) # We keep track of the loss at each epoch if args.with_tracking: total_loss += loss.detach().float() accelerator.backward(_UpperCAmelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 if isinstance(_UpperCAmelCase , _UpperCAmelCase ): UpperCamelCase = f'''step_{overall_step}''' if overall_step % checkpointing_steps == 0: if args.output_dir is not None: UpperCamelCase = os.path.join(args.output_dir , _UpperCAmelCase ) accelerator.save_state(_UpperCAmelCase ) model.eval() UpperCamelCase = 0 UpperCamelCase = 0 for step, batch in enumerate(_UpperCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. UpperCamelCase = {k: v.to(accelerator.device ) for k, v in batch.items()} UpperCamelCase = (batch['image'] - mean) / std with torch.no_grad(): UpperCamelCase = model(_UpperCAmelCase ) UpperCamelCase = outputs.argmax(dim=-1 ) UpperCamelCase = accelerator.gather_for_metrics((predictions, batch['label']) ) UpperCamelCase = predictions == references num_elems += accurate_preds.shape[0] accurate += accurate_preds.long().sum() UpperCamelCase = accurate.item() / num_elems # Use accelerator.print to print only on the main process. accelerator.print(f'''epoch {epoch}: {100 * eval_metric:.2f}''' ) if args.with_tracking: accelerator.log( { 'accuracy': 100 * eval_metric, 'train_loss': total_loss.item() / len(_UpperCAmelCase ), 'epoch': epoch, } , step=_UpperCAmelCase , ) if checkpointing_steps == "epoch": UpperCamelCase = f'''epoch_{epoch}''' if args.output_dir is not None: UpperCamelCase = os.path.join(args.output_dir , _UpperCAmelCase ) accelerator.save_state(_UpperCAmelCase ) if args.with_tracking: accelerator.end_training() def A ( ) -> int: '''simple docstring''' UpperCamelCase = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument('--data_dir' , required=_UpperCAmelCase , help='The data folder on disk.' ) parser.add_argument('--fp16' , action='store_true' , help='If passed, will use FP16 training.' ) parser.add_argument( '--mixed_precision' , type=_UpperCAmelCase , default=_UpperCAmelCase , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) parser.add_argument( '--checkpointing_steps' , type=_UpperCAmelCase , default=_UpperCAmelCase , help='Whether the various states should be saved at the end of every n steps, or \'epoch\' for each epoch.' , ) parser.add_argument( '--output_dir' , type=_UpperCAmelCase , default='.' , help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' , ) parser.add_argument( '--resume_from_checkpoint' , type=_UpperCAmelCase , default=_UpperCAmelCase , help='If the training should continue from a checkpoint folder.' , ) parser.add_argument( '--with_tracking' , action='store_true' , help='Whether to load in all available experiment trackers from the environment and use them for logging.' , ) parser.add_argument( '--project_dir' , type=_UpperCAmelCase , default='logs' , help='Location on where to store experiment tracking logs` and relevent project information' , ) UpperCamelCase = parser.parse_args() UpperCamelCase = {'lr': 3e-2, 'num_epochs': 3, 'seed': 42, 'batch_size': 64, 'image_size': 224} training_function(_UpperCAmelCase , _UpperCAmelCase ) if __name__ == "__main__": main()
351
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class lowercase ( unittest.TestCase ): def __UpperCamelCase ( self ) -> List[str]: """simple docstring""" UpperCamelCase = tempfile.mkdtemp() # fmt: off UpperCamelCase = ['', 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'lo', 'l</w>', 'w</w>', 'r</w>', 't</w>', 'low</w>', 'er</w>', 'lowest</w>', 'newer</w>', 'wider', '<unk>', '<|startoftext|>', '<|endoftext|>'] # fmt: on UpperCamelCase = dict(zip(A_ , range(len(A_ ) ) ) ) UpperCamelCase = ['#version: 0.2', 'l o', 'lo w</w>', 'e r</w>', ''] UpperCamelCase = {'unk_token': '<unk>'} UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(A_ ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(A_ ) ) UpperCamelCase = { 'do_resize': True, 'size': 20, 'do_center_crop': True, 'crop_size': 18, 'do_normalize': True, 'image_mean': [0.4814_5466, 0.457_8275, 0.4082_1073], 'image_std': [0.2686_2954, 0.2613_0258, 0.2757_7711], } UpperCamelCase = os.path.join(self.tmpdirname , A_ ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(A_ , A_ ) def __UpperCamelCase ( self , **A_ ) -> Union[str, Any]: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token='!' , **A_ ) def __UpperCamelCase ( self , **A_ ) -> Tuple: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token='!' , **A_ ) def __UpperCamelCase ( self , **A_ ) -> Union[str, Any]: """simple docstring""" return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase ( self ) -> List[str]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def __UpperCamelCase ( self ) -> List[Any]: """simple docstring""" UpperCamelCase = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] UpperCamelCase = [Image.fromarray(np.moveaxis(A_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def __UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase = self.get_tokenizer() UpperCamelCase = self.get_rust_tokenizer() UpperCamelCase = self.get_image_processor() UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) processor_slow.save_pretrained(self.tmpdirname ) UpperCamelCase = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=A_ ) UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) processor_fast.save_pretrained(self.tmpdirname ) UpperCamelCase = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , A_ ) self.assertIsInstance(processor_fast.tokenizer , A_ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , A_ ) self.assertIsInstance(processor_fast.image_processor , A_ ) def __UpperCamelCase ( self ) -> List[Any]: """simple docstring""" UpperCamelCase = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCamelCase = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) UpperCamelCase = self.get_image_processor(do_normalize=A_ ) UpperCamelCase = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=A_ ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , A_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A_ ) def __UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase = self.get_image_processor() UpperCamelCase = self.get_tokenizer() UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase = self.prepare_image_inputs() UpperCamelCase = image_processor(A_ , return_tensors='np' ) UpperCamelCase = processor(images=A_ , return_tensors='np' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def __UpperCamelCase ( self ) -> str: """simple docstring""" UpperCamelCase = self.get_image_processor() UpperCamelCase = self.get_tokenizer() UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase = 'lower newer' UpperCamelCase = processor(text=A_ , return_tensors='np' ) UpperCamelCase = tokenizer(A_ , return_tensors='np' ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def __UpperCamelCase ( self ) -> int: """simple docstring""" UpperCamelCase = self.get_image_processor() UpperCamelCase = self.get_tokenizer() UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase = 'lower newer' UpperCamelCase = self.prepare_image_inputs() UpperCamelCase = processor(text=A_ , images=A_ ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def __UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = 'google/owlvit-base-patch32' UpperCamelCase = OwlViTProcessor.from_pretrained(A_ ) UpperCamelCase = ['cat', 'nasa badge'] UpperCamelCase = processor(text=A_ ) UpperCamelCase = 16 self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def __UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = 'google/owlvit-base-patch32' UpperCamelCase = OwlViTProcessor.from_pretrained(A_ ) UpperCamelCase = [['cat', 'nasa badge'], ['person']] UpperCamelCase = processor(text=A_ ) UpperCamelCase = 16 UpperCamelCase = len(A_ ) UpperCamelCase = max([len(A_ ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def __UpperCamelCase ( self ) -> Dict: """simple docstring""" UpperCamelCase = 'google/owlvit-base-patch32' UpperCamelCase = OwlViTProcessor.from_pretrained(A_ ) UpperCamelCase = ['cat', 'nasa badge'] UpperCamelCase = processor(text=A_ ) UpperCamelCase = 16 UpperCamelCase = inputs['input_ids'] UpperCamelCase = [ [49_406, 2_368, 49_407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [49_406, 6_841, 11_301, 49_407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask'] ) self.assertEqual(inputs['input_ids'].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def __UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase = self.get_image_processor() UpperCamelCase = self.get_tokenizer() UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase = self.prepare_image_inputs() UpperCamelCase = self.prepare_image_inputs() UpperCamelCase = processor(images=A_ , query_images=A_ ) self.assertListEqual(list(inputs.keys() ) , ['query_pixel_values', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(A_ ): processor() def __UpperCamelCase ( self ) -> int: """simple docstring""" UpperCamelCase = self.get_image_processor() UpperCamelCase = self.get_tokenizer() UpperCamelCase = OwlViTProcessor(tokenizer=A_ , image_processor=A_ ) UpperCamelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCamelCase = processor.batch_decode(A_ ) UpperCamelCase = tokenizer.batch_decode(A_ ) self.assertListEqual(A_ , A_ )
110
0
'''simple docstring''' from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if is_torch_available(): import torch a_ : Optional[int] = logging.get_logger(__name__) @add_end_docstrings( lowerCamelCase__ , r'\n top_k (`int`, defaults to 5):\n The number of predictions to return.\n targets (`str` or `List[str]`, *optional*):\n When passed, the model will limit the scores to the passed targets instead of looking up in the whole\n vocab. If the provided targets are not in the model vocab, they will be tokenized and the first resulting\n token will be used (with a warning, and that might be slower).\n\n ' , ) class __UpperCamelCase ( lowerCamelCase__ ): def lowercase__ ( self, lowerCAmelCase ): """simple docstring""" if self.framework == "tf": lowerCamelCase_ =tf.where(input_ids == self.tokenizer.mask_token_id ).numpy() elif self.framework == "pt": lowerCamelCase_ =torch.nonzero(input_ids == self.tokenizer.mask_token_id, as_tuple=lowerCAmelCase ) else: raise ValueError('''Unsupported framework''' ) return masked_index def lowercase__ ( self, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =self.get_masked_index(lowerCAmelCase ) lowerCamelCase_ =np.prod(masked_index.shape ) if numel < 1: raise PipelineException( '''fill-mask''', self.model.base_model_prefix, f'''No mask_token ({self.tokenizer.mask_token}) found on the input''', ) def lowercase__ ( self, lowerCAmelCase ): """simple docstring""" if isinstance(lowerCAmelCase, lowerCAmelCase ): for model_input in model_inputs: self._ensure_exactly_one_mask_token(model_input['''input_ids'''][0] ) else: for input_ids in model_inputs["input_ids"]: self._ensure_exactly_one_mask_token(lowerCAmelCase ) def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase=None, **lowerCAmelCase ): """simple docstring""" if return_tensors is None: lowerCamelCase_ =self.framework lowerCamelCase_ =self.tokenizer(lowerCAmelCase, return_tensors=lowerCAmelCase ) self.ensure_exactly_one_mask_token(lowerCAmelCase ) return model_inputs def lowercase__ ( self, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =self.model(**lowerCAmelCase ) lowerCamelCase_ =model_inputs['''input_ids'''] return model_outputs def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase=5, lowerCAmelCase=None ): """simple docstring""" if target_ids is not None and target_ids.shape[0] < top_k: lowerCamelCase_ =target_ids.shape[0] lowerCamelCase_ =model_outputs['''input_ids'''][0] lowerCamelCase_ =model_outputs['''logits'''] if self.framework == "tf": lowerCamelCase_ =tf.where(input_ids == self.tokenizer.mask_token_id ).numpy()[:, 0] lowerCamelCase_ =outputs.numpy() lowerCamelCase_ =outputs[0, masked_index, :] lowerCamelCase_ =stable_softmax(lowerCAmelCase, axis=-1 ) if target_ids is not None: lowerCamelCase_ =tf.gather_nd(tf.squeeze(lowerCAmelCase, 0 ), target_ids.reshape(-1, 1 ) ) lowerCamelCase_ =tf.expand_dims(lowerCAmelCase, 0 ) lowerCamelCase_ =tf.math.top_k(lowerCAmelCase, k=lowerCAmelCase ) lowerCamelCase_, lowerCamelCase_ =topk.values.numpy(), topk.indices.numpy() else: lowerCamelCase_ =torch.nonzero(input_ids == self.tokenizer.mask_token_id, as_tuple=lowerCAmelCase ).squeeze(-1 ) # Fill mask pipeline supports only one ${mask_token} per sample lowerCamelCase_ =outputs[0, masked_index, :] lowerCamelCase_ =logits.softmax(dim=-1 ) if target_ids is not None: lowerCamelCase_ =probs[..., target_ids] lowerCamelCase_, lowerCamelCase_ =probs.topk(lowerCAmelCase ) lowerCamelCase_ =[] lowerCamelCase_ =values.shape[0] == 1 for i, (_values, _predictions) in enumerate(zip(values.tolist(), predictions.tolist() ) ): lowerCamelCase_ =[] for v, p in zip(_values, _predictions ): # Copy is important since we're going to modify this array in place lowerCamelCase_ =input_ids.numpy().copy() if target_ids is not None: lowerCamelCase_ =target_ids[p].tolist() lowerCamelCase_ =p # Filter padding out: lowerCamelCase_ =tokens[np.where(tokens != self.tokenizer.pad_token_id )] # Originally we skip special tokens to give readable output. # For multi masks though, the other [MASK] would be removed otherwise # making the output look odd, so we add them back lowerCamelCase_ =self.tokenizer.decode(lowerCAmelCase, skip_special_tokens=lowerCAmelCase ) lowerCamelCase_ ={'''score''': v, '''token''': p, '''token_str''': self.tokenizer.decode([p] ), '''sequence''': sequence} row.append(lowerCAmelCase ) result.append(lowerCAmelCase ) if single_mask: return result[0] return result def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase=None ): """simple docstring""" if isinstance(lowerCAmelCase, lowerCAmelCase ): lowerCamelCase_ =[targets] try: lowerCamelCase_ =self.tokenizer.get_vocab() except Exception: lowerCamelCase_ ={} lowerCamelCase_ =[] for target in targets: lowerCamelCase_ =vocab.get(lowerCAmelCase, lowerCAmelCase ) if id_ is None: lowerCamelCase_ =self.tokenizer( lowerCAmelCase, add_special_tokens=lowerCAmelCase, return_attention_mask=lowerCAmelCase, return_token_type_ids=lowerCAmelCase, max_length=1, truncation=lowerCAmelCase, )['''input_ids'''] if len(lowerCAmelCase ) == 0: logger.warning( f'''The specified target token `{target}` does not exist in the model vocabulary. ''' '''We cannot replace it with anything meaningful, ignoring it''' ) continue lowerCamelCase_ =input_ids[0] # XXX: If users encounter this pass # it becomes pretty slow, so let's make sure # The warning enables them to fix the input to # get faster performance. logger.warning( f'''The specified target token `{target}` does not exist in the model vocabulary. ''' f'''Replacing with `{self.tokenizer.convert_ids_to_tokens(id_ )}`.''' ) target_ids.append(id_ ) lowerCamelCase_ =list(set(lowerCAmelCase ) ) if len(lowerCAmelCase ) == 0: raise ValueError('''At least one target must be provided when passed.''' ) lowerCamelCase_ =np.array(lowerCAmelCase ) return target_ids def lowercase__ ( self, lowerCAmelCase=None, lowerCAmelCase=None ): """simple docstring""" lowerCamelCase_ ={} if targets is not None: lowerCamelCase_ =self.get_target_ids(lowerCAmelCase, lowerCAmelCase ) lowerCamelCase_ =target_ids if top_k is not None: lowerCamelCase_ =top_k if self.tokenizer.mask_token_id is None: raise PipelineException( '''fill-mask''', self.model.base_model_prefix, '''The tokenizer does not define a `mask_token`.''' ) return {}, {}, postprocess_params def __call__( self, lowerCAmelCase, *lowerCAmelCase, **lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =super().__call__(lowerCAmelCase, **lowerCAmelCase ) if isinstance(lowerCAmelCase, lowerCAmelCase ) and len(lowerCAmelCase ) == 1: return outputs[0] return outputs
75
'''simple docstring''' from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class __UpperCamelCase : lowercase : Union[str, Any] =XGLMConfig lowercase : Optional[Any] ={} lowercase : Optional[int] ='gelu' def __init__( self, lowerCAmelCase, lowerCAmelCase=14, lowerCAmelCase=7, lowerCAmelCase=True, lowerCAmelCase=True, lowerCAmelCase=True, lowerCAmelCase=99, lowerCAmelCase=32, lowerCAmelCase=2, lowerCAmelCase=4, lowerCAmelCase=37, lowerCAmelCase="gelu", lowerCAmelCase=0.1, lowerCAmelCase=0.1, lowerCAmelCase=512, lowerCAmelCase=0.0_2, ): """simple docstring""" lowerCamelCase_ =parent lowerCamelCase_ =batch_size lowerCamelCase_ =seq_length lowerCamelCase_ =is_training lowerCamelCase_ =use_input_mask lowerCamelCase_ =use_labels lowerCamelCase_ =vocab_size lowerCamelCase_ =d_model lowerCamelCase_ =num_hidden_layers lowerCamelCase_ =num_attention_heads lowerCamelCase_ =ffn_dim lowerCamelCase_ =activation_function lowerCamelCase_ =activation_dropout lowerCamelCase_ =attention_dropout lowerCamelCase_ =max_position_embeddings lowerCamelCase_ =initializer_range lowerCamelCase_ =None lowerCamelCase_ =0 lowerCamelCase_ =2 lowerCamelCase_ =1 def lowercase__ ( self ): """simple docstring""" return XGLMConfig.from_pretrained('''facebook/xglm-564M''' ) def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length], self.vocab_size ), clip_value_min=0, clip_value_max=3 ) lowerCamelCase_ =None if self.use_input_mask: lowerCamelCase_ =random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase_ =self.get_config() lowerCamelCase_ =floats_tensor([self.num_hidden_layers, self.num_attention_heads], 2 ) return ( config, input_ids, input_mask, head_mask, ) def lowercase__ ( self ): """simple docstring""" return XGLMConfig( vocab_size=self.vocab_size, d_model=self.hidden_size, num_layers=self.num_hidden_layers, attention_heads=self.num_attention_heads, ffn_dim=self.ffn_dim, activation_function=self.activation_function, activation_dropout=self.activation_dropout, attention_dropout=self.attention_dropout, max_position_embeddings=self.max_position_embeddings, initializer_range=self.initializer_range, use_cache=lowerCAmelCase, bos_token_id=self.bos_token_id, eos_token_id=self.eos_token_id, pad_token_id=self.pad_token_id, return_dict=lowerCAmelCase, ) def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =self.prepare_config_and_inputs() ( ( lowerCamelCase_ ), ( lowerCamelCase_ ), ( lowerCamelCase_ ), ( lowerCamelCase_ ), ) =config_and_inputs lowerCamelCase_ ={ '''input_ids''': input_ids, '''head_mask''': head_mask, } return config, inputs_dict @require_tf class __UpperCamelCase ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): lowercase : int =(TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () lowercase : Optional[Any] =(TFXGLMForCausalLM,) if is_tf_available() else () lowercase : Tuple =( {'feature-extraction': TFXGLMModel, 'text-generation': TFXGLMForCausalLM} if is_tf_available() else {} ) lowercase : Optional[Any] =False lowercase : Optional[Any] =False lowercase : Optional[int] =False def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =TFXGLMModelTester(self ) lowerCamelCase_ =ConfigTester(self, config_class=lowerCAmelCase, n_embd=37 ) def lowercase__ ( self ): """simple docstring""" self.config_tester.run_common_tests() @slow def lowercase__ ( self ): """simple docstring""" for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase_ =TFXGLMModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) @unittest.skip(reason='''Currently, model embeddings are going to undergo a major refactor.''' ) def lowercase__ ( self ): """simple docstring""" super().test_resize_token_embeddings() @require_tf class __UpperCamelCase ( unittest.TestCase ): @slow def lowercase__ ( self, lowerCAmelCase=True ): """simple docstring""" lowerCamelCase_ =TFXGLMForCausalLM.from_pretrained('''facebook/xglm-564M''' ) lowerCamelCase_ =tf.convert_to_tensor([[2, 268, 9_865]], dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off lowerCamelCase_ =[2, 268, 9_865, 67, 11, 1_988, 57_252, 9_865, 5, 984, 67, 1_988, 213_838, 1_658, 53, 70_446, 33, 6_657, 278, 1_581] # fmt: on lowerCamelCase_ =model.generate(lowerCAmelCase, do_sample=lowerCAmelCase, num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist(), lowerCAmelCase ) @slow def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =XGLMTokenizer.from_pretrained('''facebook/xglm-564M''' ) lowerCamelCase_ =TFXGLMForCausalLM.from_pretrained('''facebook/xglm-564M''' ) tf.random.set_seed(0 ) lowerCamelCase_ =tokenizer('''Today is a nice day and''', return_tensors='''tf''' ) lowerCamelCase_ =tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(''':/CPU:0''' ): lowerCamelCase_ =model.generate(lowerCAmelCase, do_sample=lowerCAmelCase, seed=[7, 0] ) lowerCamelCase_ =tokenizer.decode(output_ids[0], skip_special_tokens=lowerCAmelCase ) lowerCamelCase_ =( '''Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due''' ) self.assertEqual(lowerCAmelCase, lowerCAmelCase ) @slow def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =TFXGLMForCausalLM.from_pretrained('''facebook/xglm-564M''' ) lowerCamelCase_ =XGLMTokenizer.from_pretrained('''facebook/xglm-564M''' ) lowerCamelCase_ ='''left''' # use different length sentences to test batching lowerCamelCase_ =[ '''This is an extremelly long sentence that only exists to test the ability of the model to cope with ''' '''left-padding, such as in batched generation. The output for the sequence below should be the same ''' '''regardless of whether left padding is applied or not. When''', '''Hello, my dog is a little''', ] lowerCamelCase_ =tokenizer(lowerCAmelCase, return_tensors='''tf''', padding=lowerCAmelCase ) lowerCamelCase_ =inputs['''input_ids'''] lowerCamelCase_ =model.generate(input_ids=lowerCAmelCase, attention_mask=inputs['''attention_mask'''], max_new_tokens=12 ) lowerCamelCase_ =tokenizer(sentences[0], return_tensors='''tf''' ).input_ids lowerCamelCase_ =model.generate(input_ids=lowerCAmelCase, max_new_tokens=12 ) lowerCamelCase_ =tokenizer(sentences[1], return_tensors='''tf''' ).input_ids lowerCamelCase_ =model.generate(input_ids=lowerCAmelCase, max_new_tokens=12 ) lowerCamelCase_ =tokenizer.batch_decode(lowerCAmelCase, skip_special_tokens=lowerCAmelCase ) lowerCamelCase_ =tokenizer.decode(output_non_padded[0], skip_special_tokens=lowerCAmelCase ) lowerCamelCase_ =tokenizer.decode(output_padded[0], skip_special_tokens=lowerCAmelCase ) lowerCamelCase_ =[ '''This is an extremelly long sentence that only exists to test the ability of the model to cope with ''' '''left-padding, such as in batched generation. The output for the sequence below should be the same ''' '''regardless of whether left padding is applied or not. When left padding is applied, the sequence will be ''' '''a single''', '''Hello, my dog is a little bit of a shy one, but he is very friendly''', ] self.assertListEqual(lowerCAmelCase, lowerCAmelCase ) self.assertListEqual(lowerCAmelCase, [non_padded_sentence, padded_sentence] )
75
1
def __lowerCamelCase ( __snake_case : float, __snake_case : list[float] ) -> float: """simple docstring""" if discount_rate < 0: raise ValueError("""Discount rate cannot be negative""" ) if not cash_flows: raise ValueError("""Cash flows list cannot be empty""" ) A__ : Optional[Any] =sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(__lowerCAmelCase ) ) return round(__lowerCAmelCase, ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
353
'''simple docstring''' from typing import Dict from .base import GenericTensor, Pipeline class lowerCamelCase ( lowercase_ ): '''simple docstring''' def lowercase__ ( self : List[str] , lowerCAmelCase_ : Tuple=None , lowerCAmelCase_ : Dict=None , lowerCAmelCase_ : Any=None , **lowerCAmelCase_ : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' if tokenize_kwargs is None: A__ : List[Any] ={} if truncation is not None: if "truncation" in tokenize_kwargs: raise ValueError( """truncation parameter defined twice (given as keyword argument as well as in tokenize_kwargs)""" ) A__ : str =truncation A__ : Optional[int] =tokenize_kwargs A__ : List[Any] ={} if return_tensors is not None: A__ : Any =return_tensors return preprocess_params, {}, postprocess_params def lowercase__ ( self : int , lowerCAmelCase_ : Union[str, Any] , **lowerCAmelCase_ : Dict ) -> Dict[str, GenericTensor]: '''simple docstring''' A__ : List[str] =self.framework A__ : Union[str, Any] =self.tokenizer(lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , **lowerCAmelCase_ ) return model_inputs def lowercase__ ( self : Optional[int] , lowerCAmelCase_ : Dict ) -> Optional[Any]: '''simple docstring''' A__ : Union[str, Any] =self.model(**lowerCAmelCase_ ) return model_outputs def lowercase__ ( self : Tuple , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Any=False ) -> List[Any]: '''simple docstring''' # [0] is the first available tensor, logits or last_hidden_state. if return_tensors: return model_outputs[0] if self.framework == "pt": return model_outputs[0].tolist() elif self.framework == "tf": return model_outputs[0].numpy().tolist() def __call__( self : int , *lowerCAmelCase_ : Any , **lowerCAmelCase_ : Union[str, Any] ) -> List[str]: '''simple docstring''' return super().__call__(*lowerCAmelCase_ , **lowerCAmelCase_ )
136
0
"""simple docstring""" import math import time from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class lowerCamelCase__ ( A ): """simple docstring""" def __init__( self : List[Any] , *UpperCamelCase : Any , UpperCamelCase : List[str]=None , UpperCamelCase : List[Any]=None , **UpperCamelCase : Optional[Any] ): '''simple docstring''' super().__init__(*UpperCamelCase , **UpperCamelCase ) __UpperCAmelCase : List[Any] = eval_examples __UpperCAmelCase : Optional[Any] = post_process_function def lowerCamelCase__ ( self : Tuple , UpperCamelCase : Optional[int]=None , UpperCamelCase : List[str]=None , UpperCamelCase : Tuple=None , UpperCamelCase : str = "eval" ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] = self.eval_dataset if eval_dataset is None else eval_dataset __UpperCAmelCase : List[Any] = self.get_eval_dataloader(UpperCamelCase ) __UpperCAmelCase : str = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. __UpperCAmelCase : Optional[Any] = self.compute_metrics __UpperCAmelCase : Optional[Any] = None __UpperCAmelCase : Tuple = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop __UpperCAmelCase : List[str] = time.time() try: __UpperCAmelCase : Optional[Any] = eval_loop( UpperCamelCase , description="""Evaluation""" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=UpperCamelCase , metric_key_prefix=UpperCamelCase , ) finally: __UpperCAmelCase : List[str] = compute_metrics __UpperCAmelCase : Tuple = self.args.eval_batch_size * self.args.world_size if f'''{metric_key_prefix}_jit_compilation_time''' in output.metrics: start_time += output.metrics[f'''{metric_key_prefix}_jit_compilation_time'''] output.metrics.update( speed_metrics( UpperCamelCase , UpperCamelCase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default __UpperCAmelCase : Tuple = self.post_process_function(UpperCamelCase , UpperCamelCase , output.predictions ) __UpperCAmelCase : List[str] = self.compute_metrics(UpperCamelCase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f'''{metric_key_prefix}_''' ): __UpperCAmelCase : Optional[int] = metrics.pop(UpperCamelCase ) metrics.update(output.metrics ) else: __UpperCAmelCase : Dict = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(UpperCamelCase ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) __UpperCAmelCase : Optional[Any] = self.callback_handler.on_evaluate(self.args , self.state , self.control , UpperCamelCase ) return metrics def lowerCamelCase__ ( self : Union[str, Any] , UpperCamelCase : Optional[int] , UpperCamelCase : Dict , UpperCamelCase : Union[str, Any]=None , UpperCamelCase : str = "test" ): '''simple docstring''' __UpperCAmelCase : Any = self.get_test_dataloader(UpperCamelCase ) # Temporarily disable metric computation, we will do it in the loop here. __UpperCAmelCase : List[Any] = self.compute_metrics __UpperCAmelCase : str = None __UpperCAmelCase : int = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop __UpperCAmelCase : Dict = time.time() try: __UpperCAmelCase : Union[str, Any] = eval_loop( UpperCamelCase , description="""Prediction""" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=UpperCamelCase , metric_key_prefix=UpperCamelCase , ) finally: __UpperCAmelCase : Union[str, Any] = compute_metrics __UpperCAmelCase : str = self.args.eval_batch_size * self.args.world_size if f'''{metric_key_prefix}_jit_compilation_time''' in output.metrics: start_time += output.metrics[f'''{metric_key_prefix}_jit_compilation_time'''] output.metrics.update( speed_metrics( UpperCamelCase , UpperCamelCase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output __UpperCAmelCase : List[str] = self.post_process_function(UpperCamelCase , UpperCamelCase , output.predictions , """predict""" ) __UpperCAmelCase : Tuple = self.compute_metrics(UpperCamelCase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f'''{metric_key_prefix}_''' ): __UpperCAmelCase : int = metrics.pop(UpperCamelCase ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=UpperCamelCase )
115
"""simple docstring""" import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() UpperCAmelCase : Optional[Any] = logging.get_logger(__name__) UpperCAmelCase : Optional[int] = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'lm_head', 'mask_emb': 'masked_spec_embed', } UpperCAmelCase : Dict = [ 'lm_head', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def lowerCamelCase ( _UpperCamelCase : Optional[int] , _UpperCamelCase : Optional[int] , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : List[str] , _UpperCamelCase : int ) -> List[str]: '''simple docstring''' for attribute in key.split(""".""" ): __UpperCAmelCase : Tuple = getattr(_UpperCamelCase , _UpperCamelCase ) if weight_type is not None: __UpperCAmelCase : int = getattr(_UpperCamelCase , _UpperCamelCase ).shape else: __UpperCAmelCase : List[Any] = hf_pointer.shape assert hf_shape == value.shape, ( f'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' f''' {value.shape} for {full_name}''' ) if weight_type == "weight": __UpperCAmelCase : Optional[Any] = value elif weight_type == "weight_g": __UpperCAmelCase : List[Any] = value elif weight_type == "weight_v": __UpperCAmelCase : Dict = value elif weight_type == "bias": __UpperCAmelCase : List[str] = value else: __UpperCAmelCase : List[str] = value logger.info(f'''{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.''' ) def lowerCamelCase ( _UpperCamelCase : List[Any] , _UpperCamelCase : Optional[Any] ) -> Any: '''simple docstring''' __UpperCAmelCase : Union[str, Any] = [] __UpperCAmelCase : int = fairseq_model.state_dict() __UpperCAmelCase : Union[str, Any] = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight __UpperCAmelCase : Tuple = None for name, value in fairseq_dict.items(): __UpperCAmelCase : Union[str, Any] = False if "conv_layers" in name: load_conv_layer( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , hf_model.config.feat_extract_norm == """group""" , ) __UpperCAmelCase : int = True elif name.split(""".""" )[0] == "proj": __UpperCAmelCase : Optional[Any] = fairseq_model.proj __UpperCAmelCase : Union[str, Any] = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: __UpperCAmelCase : Dict = True if "*" in mapped_key: __UpperCAmelCase : Any = name.split(_UpperCamelCase )[0].split(""".""" )[-2] __UpperCAmelCase : Tuple = mapped_key.replace("""*""" , _UpperCamelCase ) if "weight_g" in name: __UpperCAmelCase : Any = """weight_g""" elif "weight_v" in name: __UpperCAmelCase : Union[str, Any] = """weight_v""" elif "bias" in name: __UpperCAmelCase : List[Any] = """bias""" elif "weight" in name: __UpperCAmelCase : List[Any] = """weight""" else: __UpperCAmelCase : Dict = None set_recursively(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) continue if not is_used: unused_weights.append(_UpperCamelCase ) logger.warning(f'''Unused weights: {unused_weights}''' ) return proj_weight def lowerCamelCase ( _UpperCamelCase : str , _UpperCamelCase : str , _UpperCamelCase : List[Any] , _UpperCamelCase : List[str] , _UpperCamelCase : Dict ) -> Tuple: '''simple docstring''' __UpperCAmelCase : Any = full_name.split("""conv_layers.""" )[-1] __UpperCAmelCase : int = name.split(""".""" ) __UpperCAmelCase : List[str] = int(items[0] ) __UpperCAmelCase : Union[str, Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) __UpperCAmelCase : Tuple = value logger.info(f'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) __UpperCAmelCase : List[Any] = value logger.info(f'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) __UpperCAmelCase : Optional[Any] = value logger.info(f'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) __UpperCAmelCase : Optional[int] = value logger.info(f'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(_UpperCamelCase ) def lowerCamelCase ( _UpperCamelCase : Tuple ) -> Union[str, Any]: '''simple docstring''' __UpperCAmelCase ,__UpperCAmelCase : Tuple = emb.weight.shape __UpperCAmelCase : Optional[int] = nn.Linear(_UpperCamelCase , _UpperCamelCase , bias=_UpperCamelCase ) __UpperCAmelCase : List[str] = emb.weight.data return lin_layer def lowerCamelCase ( _UpperCamelCase : Any ) -> Any: '''simple docstring''' with open(_UpperCamelCase , """r""" , encoding="""utf-8""" ) as f: __UpperCAmelCase : Optional[int] = f.readlines() __UpperCAmelCase : Optional[Any] = [line.split(""" """ )[0] for line in lines] __UpperCAmelCase : List[Any] = len(_UpperCamelCase ) __UpperCAmelCase : Any = { """<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3, } vocab_dict.update(dict(zip(_UpperCamelCase , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def lowerCamelCase ( _UpperCamelCase : List[Any] , _UpperCamelCase : Optional[Any] , _UpperCamelCase : Optional[Any] , _UpperCamelCase : List[Any] , _UpperCamelCase : Dict , _UpperCamelCase : Tuple , _UpperCamelCase : List[str] , ) -> Any: '''simple docstring''' __UpperCAmelCase : Optional[Any] = WavaVecaConfig.from_pretrained(_UpperCamelCase ) __UpperCAmelCase : Optional[int] = SpeechaTextaConfig.from_pretrained( _UpperCamelCase , vocab_size=_UpperCamelCase , decoder_layers=_UpperCamelCase , do_stable_layer_norm=_UpperCamelCase ) __UpperCAmelCase : Any = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=_UpperCamelCase , return_attention_mask=_UpperCamelCase , ) __UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase : int = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) __UpperCAmelCase : List[Any] = model[0].eval() # set weights for wav2vec2 encoder __UpperCAmelCase : int = WavaVecaModel(_UpperCamelCase ) __UpperCAmelCase : List[str] = recursively_load_weights_wavaveca(model.encoder , _UpperCamelCase ) __UpperCAmelCase : List[str] = SpeechaTextaForCausalLM(_UpperCamelCase ) __UpperCAmelCase ,__UpperCAmelCase : Optional[Any] = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=_UpperCamelCase ) # set output linear layer unexpected_keys.remove("""embed_out""" ) __UpperCAmelCase : List[str] = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(f'''The following keys are missing when loading the decoder weights: {missing_keys}''' ) logger.warning(f'''The following keys are unexpected when loading the decoder weights: {unexpected_keys}''' ) __UpperCAmelCase : str = SpeechEncoderDecoderModel(encoder=_UpperCamelCase , decoder=_UpperCamelCase ) __UpperCAmelCase : Any = False # add projection layer __UpperCAmelCase : str = nn.Parameter(projection_layer.weight ) __UpperCAmelCase : str = nn.Parameter(projection_layer.bias ) __UpperCAmelCase : List[Any] = create_vocab_dict(_UpperCamelCase ) with open(os.path.join(_UpperCamelCase , """vocab.json""" ) , """w""" ) as fp: json.dump(_UpperCamelCase , _UpperCamelCase ) __UpperCAmelCase : List[str] = SpeechaTextaTokenizer(os.path.join(_UpperCamelCase , """vocab.json""" ) ) tokenizer.save_pretrained(_UpperCamelCase ) __UpperCAmelCase : Dict = hf_wavavec.config.to_dict() __UpperCAmelCase : int = tokenizer.pad_token_id __UpperCAmelCase : List[str] = tokenizer.bos_token_id __UpperCAmelCase : Tuple = tokenizer.eos_token_id __UpperCAmelCase : int = """speech_to_text_2""" __UpperCAmelCase : Optional[int] = """wav2vec2""" __UpperCAmelCase : Union[str, Any] = SpeechEncoderDecoderConfig.from_dict(_UpperCamelCase ) hf_wavavec.save_pretrained(_UpperCamelCase ) feature_extractor.save_pretrained(_UpperCamelCase ) if __name__ == "__main__": UpperCAmelCase : Tuple = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument( '--encoder_config_path', default='facebook/wav2vec2-large-lv60', type=str, help='Path to hf encoder wav2vec2 checkpoint config', ) parser.add_argument( '--decoder_config_path', default='facebook/s2t-small-mustc-en-fr-st', type=str, help='Path to hf decoder s2t checkpoint config', ) parser.add_argument('--vocab_size', default=1_0224, type=int, help='Vocab size of decoder') parser.add_argument('--num_decoder_layers', default=7, type=int, help='Number of decoder layers') UpperCAmelCase : int = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
115
1
import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def lowercase_ ( A__ ) -> int: """simple docstring""" snake_case = [] embed.append( ( F'cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight', F'stage{idx}.patch_embed.proj.weight', ) ) embed.append( ( F'cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias', F'stage{idx}.patch_embed.proj.bias', ) ) embed.append( ( F'cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight', F'stage{idx}.patch_embed.norm.weight', ) ) embed.append( ( F'cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias', F'stage{idx}.patch_embed.norm.bias', ) ) return embed def lowercase_ ( A__ , A__ ) -> Any: """simple docstring""" snake_case = [] attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight', F'stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight', F'stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias', F'stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean', F'stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var', F'stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked', F'stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight', F'stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight', F'stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias', F'stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean', F'stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var', F'stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked', F'stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight', F'stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight', F'stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias', F'stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean', F'stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var', F'stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked', F'stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight', F'stage{idx}.blocks.{cnt}.attn.proj_q.weight', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias', F'stage{idx}.blocks.{cnt}.attn.proj_q.bias', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight', F'stage{idx}.blocks.{cnt}.attn.proj_k.weight', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias', F'stage{idx}.blocks.{cnt}.attn.proj_k.bias', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight', F'stage{idx}.blocks.{cnt}.attn.proj_v.weight', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias', F'stage{idx}.blocks.{cnt}.attn.proj_v.bias', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight', F'stage{idx}.blocks.{cnt}.attn.proj.weight', ) ) attention_weights.append( ( F'cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias', F'stage{idx}.blocks.{cnt}.attn.proj.bias', ) ) attention_weights.append( (F'cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight', F'stage{idx}.blocks.{cnt}.mlp.fc1.weight') ) attention_weights.append( (F'cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias', F'stage{idx}.blocks.{cnt}.mlp.fc1.bias') ) attention_weights.append( (F'cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight', F'stage{idx}.blocks.{cnt}.mlp.fc2.weight') ) attention_weights.append( (F'cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias', F'stage{idx}.blocks.{cnt}.mlp.fc2.bias') ) attention_weights.append( (F'cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight', F'stage{idx}.blocks.{cnt}.norm1.weight') ) attention_weights.append( (F'cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias', F'stage{idx}.blocks.{cnt}.norm1.bias') ) attention_weights.append( (F'cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight', F'stage{idx}.blocks.{cnt}.norm2.weight') ) attention_weights.append( (F'cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias', F'stage{idx}.blocks.{cnt}.norm2.bias') ) return attention_weights def lowercase_ ( A__ ) -> Dict: """simple docstring""" snake_case = [] token.append((F'cvt.encoder.stages.{idx}.cls_token', "stage2.cls_token") ) return token def lowercase_ ( ) -> int: """simple docstring""" snake_case = [] head.append(("layernorm.weight", "norm.weight") ) head.append(("layernorm.bias", "norm.bias") ) head.append(("classifier.weight", "head.weight") ) head.append(("classifier.bias", "head.bias") ) return head def lowercase_ ( A__ , A__ , A__ , A__ ) -> List[str]: """simple docstring""" snake_case = "imagenet-1k-id2label.json" snake_case = 1000 snake_case = "huggingface/label-files" snake_case = num_labels snake_case = json.load(open(cached_download(hf_hub_url(A__ , A__ , repo_type="dataset" ) ) , "r" ) ) snake_case = {int(A__ ): v for k, v in idalabel.items()} snake_case = idalabel snake_case = {v: k for k, v in idalabel.items()} snake_case = snake_case = CvtConfig(num_labels=A__ , idalabel=A__ , labelaid=A__ ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit("/" , 1 )[-1][4:6] == "13": snake_case = [1, 2, 10] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit("/" , 1 )[-1][4:6] == "21": snake_case = [1, 4, 16] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: snake_case = [2, 2, 20] snake_case = [3, 12, 16] snake_case = [192, 768, 1024] snake_case = CvtForImageClassification(A__ ) snake_case = AutoImageProcessor.from_pretrained("facebook/convnext-base-224-22k-1k" ) snake_case = image_size snake_case = torch.load(A__ , map_location=torch.device("cpu" ) ) snake_case = OrderedDict() snake_case = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: snake_case = list_of_state_dict + cls_token(A__ ) snake_case = list_of_state_dict + embeddings(A__ ) for cnt in range(config.depth[idx] ): snake_case = list_of_state_dict + attention(A__ , A__ ) snake_case = list_of_state_dict + final() for gg in list_of_state_dict: print(A__ ) for i in range(len(A__ ) ): snake_case = original_weights[list_of_state_dict[i][1]] model.load_state_dict(A__ ) model.save_pretrained(A__ ) image_processor.save_pretrained(A__ ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": _A = argparse.ArgumentParser() parser.add_argument( "--cvt_model", default="cvt-w24", type=str, help="Name of the cvt model you'd like to convert.", ) parser.add_argument( "--image_size", default=3_84, type=int, help="Input Image Size", ) parser.add_argument( "--cvt_file_name", default=R"cvtmodels\CvT-w24-384x384-IN-22k.pth", type=str, help="Input Image Size", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) _A = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
137
import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class lowerCamelCase ( unittest.TestCase ): def UpperCAmelCase(self : Tuple ) -> List[str]: snake_case = ["a", "b", "c"] # Defaults to last layer if both are None snake_case , snake_case = get_aligned_output_features_output_indices(_A , _A , _A ) self.assertEqual(_A , ["c"] ) self.assertEqual(_A , [2] ) # Out indices set to match out features snake_case , snake_case = get_aligned_output_features_output_indices(["a", "c"] , _A , _A ) self.assertEqual(_A , ["a", "c"] ) self.assertEqual(_A , [0, 2] ) # Out features set to match out indices snake_case , snake_case = get_aligned_output_features_output_indices(_A , [0, 2] , _A ) self.assertEqual(_A , ["a", "c"] ) self.assertEqual(_A , [0, 2] ) # Out features selected from negative indices snake_case , snake_case = get_aligned_output_features_output_indices(_A , [-3, -1] , _A ) self.assertEqual(_A , ["a", "c"] ) self.assertEqual(_A , [-3, -1] ) def UpperCAmelCase(self : Optional[int] ) -> str: # Stage names must be set with self.assertRaises(_A ): verify_out_features_out_indices(["a", "b"] , (0, 1) , _A ) # Out features must be a list with self.assertRaises(_A ): verify_out_features_out_indices(("a", "b") , (0, 1) , ["a", "b"] ) # Out features must be a subset of stage names with self.assertRaises(_A ): verify_out_features_out_indices(["a", "b"] , (0, 1) , ["a"] ) # Out indices must be a list or tuple with self.assertRaises(_A ): verify_out_features_out_indices(_A , 0 , ["a", "b"] ) # Out indices must be a subset of stage names with self.assertRaises(_A ): verify_out_features_out_indices(_A , (0, 1) , ["a"] ) # Out features and out indices must be the same length with self.assertRaises(_A ): verify_out_features_out_indices(["a", "b"] , (0,) , ["a", "b", "c"] ) # Out features should match out indices with self.assertRaises(_A ): verify_out_features_out_indices(["a", "b"] , (0, 2) , ["a", "b", "c"] ) # Out features and out indices should be in order with self.assertRaises(_A ): verify_out_features_out_indices(["b", "a"] , (0, 1) , ["a", "b"] ) # Check passes with valid inputs verify_out_features_out_indices(["a", "b", "d"] , (0, 1, -1) , ["a", "b", "c", "d"] ) def UpperCAmelCase(self : List[str] ) -> str: snake_case = BackboneMixin() snake_case = ["a", "b", "c"] snake_case = ["a", "c"] snake_case = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features , ["a", "c"] ) self.assertEqual(backbone.out_indices , [0, 2] ) # Check out features and indices are updated correctly snake_case = ["a", "b"] self.assertEqual(backbone.out_features , ["a", "b"] ) self.assertEqual(backbone.out_indices , [0, 1] ) snake_case = [-3, -1] self.assertEqual(backbone.out_features , ["a", "c"] ) self.assertEqual(backbone.out_indices , [-3, -1] )
137
1
"""simple docstring""" SCREAMING_SNAKE_CASE = { "meter": "m", "kilometer": "km", "megametre": "Mm", "gigametre": "Gm", "terametre": "Tm", "petametre": "Pm", "exametre": "Em", "zettametre": "Zm", "yottametre": "Ym", } # Exponent of the factor(meter) SCREAMING_SNAKE_CASE = { "m": 0, "km": 3, "Mm": 6, "Gm": 9, "Tm": 12, "Pm": 15, "Em": 18, "Zm": 21, "Ym": 24, } def _SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ , lowercase_ ) -> float: A__ = from_type.lower().strip("s" ) A__ = to_type.lower().strip("s" ) A__ = UNIT_SYMBOL.get(UpperCAmelCase__ , UpperCAmelCase__ ) A__ = UNIT_SYMBOL.get(UpperCAmelCase__ , UpperCAmelCase__ ) if from_sanitized not in METRIC_CONVERSION: A__ = ( f"""Invalid \'from_type\' value: {from_type!r}.\n""" f"""Conversion abbreviations are: {", ".join(UpperCAmelCase__ )}""" ) raise ValueError(UpperCAmelCase__ ) if to_sanitized not in METRIC_CONVERSION: A__ = ( f"""Invalid \'to_type\' value: {to_type!r}.\n""" f"""Conversion abbreviations are: {", ".join(UpperCAmelCase__ )}""" ) raise ValueError(UpperCAmelCase__ ) A__ = METRIC_CONVERSION[from_sanitized] A__ = METRIC_CONVERSION[to_sanitized] A__ = 1 if from_exponent > to_exponent: A__ = from_exponent - to_exponent else: A__ = -(to_exponent - from_exponent) return value * pow(10 , UpperCAmelCase__ ) if __name__ == "__main__": from doctest import testmod testmod()
247
'''simple docstring''' def lowerCamelCase ( UpperCAmelCase__ : list ) -> list: if any(not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) or x < 0 for x in sequence ): raise TypeError("""Sequence must be list of non-negative integers""" ) for _ in range(len(UpperCAmelCase__ ) ): for i, (rod_upper, rod_lower) in enumerate(zip(UpperCAmelCase__ , sequence[1:] ) ): if rod_upper > rod_lower: sequence[i] -= rod_upper - rod_lower sequence[i + 1] += rod_upper - rod_lower return sequence if __name__ == "__main__": assert bead_sort([5, 4, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bead_sort([7, 9, 4, 3, 5]) == [3, 4, 5, 7, 9]
239
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_segformer import SegformerImageProcessor lowercase =logging.get_logger(__name__) class __magic_name__ ( lowerCAmelCase ): def __init__( self , *snake_case , **snake_case) -> None: '''simple docstring''' warnings.warn( 'The class SegformerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use SegformerImageProcessor instead.' , _A , ) super().__init__(*_A , **_A)
357
'''simple docstring''' import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever lowercase =logging.getLogger(__name__) class __magic_name__ ( lowerCAmelCase ): def __init__( self , snake_case , snake_case , snake_case , snake_case=None) -> List[Any]: '''simple docstring''' super().__init__( snake_case , question_encoder_tokenizer=snake_case , generator_tokenizer=snake_case , index=snake_case , init_retrieval=snake_case , ) _UpperCAmelCase : Union[str, Any] =None def lowerCAmelCase ( self , snake_case) -> Optional[Any]: '''simple docstring''' logger.info('initializing retrieval') # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info('dist initialized') # needs to be set manually _UpperCAmelCase : Optional[Any] =self._infer_socket_ifname() # avoid clash with the NCCL port _UpperCAmelCase : Optional[int] =str(distributed_port + 1) _UpperCAmelCase : Any =dist.new_group(ranks=snake_case , backend='gloo') # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info('dist not initialized / main') self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group) def lowerCAmelCase ( self) -> Tuple: '''simple docstring''' return dist.get_rank(group=self.process_group) == 0 def lowerCAmelCase ( self , snake_case , snake_case , snake_case=torch.floataa) -> Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Dict =torch.empty(snake_case , dtype=snake_case) dist.scatter(snake_case , src=0 , scatter_list=snake_case , group=self.process_group) return target_tensor def lowerCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCAmelCase : Dict =psutil.net_if_addrs() # a hacky way to deal with varying network interface names _UpperCAmelCase : str =next((addr for addr in addrs if addr.startswith('e')) , snake_case) return ifname def lowerCAmelCase ( self , snake_case , snake_case) -> Tuple[np.ndarray, List[dict]]: '''simple docstring''' # single GPU training if not dist.is_initialized(): _UpperCAmelCase , _UpperCAmelCase : Optional[Any] =self._main_retrieve(snake_case , snake_case) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(snake_case) # distributed training _UpperCAmelCase : Optional[int] =dist.get_world_size(group=self.process_group) # gather logic _UpperCAmelCase : str =None if self._is_main(): _UpperCAmelCase : Union[str, Any] =[torch.empty(question_hidden_states.shape , dtype=torch.floataa) for _ in range(snake_case)] dist.gather(torch.tensor(snake_case) , dst=0 , gather_list=snake_case , group=self.process_group) # scatter logic _UpperCAmelCase : Optional[Any] =question_hidden_states.shape[0] _UpperCAmelCase : List[Any] =[] _UpperCAmelCase : Any =[] if self._is_main(): assert len(snake_case) == world_size _UpperCAmelCase , _UpperCAmelCase : Tuple =self._main_retrieve(torch.cat(snake_case).numpy() , snake_case) _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] =torch.tensor(snake_case), torch.tensor(snake_case) _UpperCAmelCase : List[str] =self._chunk_tensor(snake_case , snake_case) _UpperCAmelCase : Union[str, Any] =self._chunk_tensor(snake_case , snake_case) _UpperCAmelCase : int =self._scattered(snake_case , [n_queries, n_docs] , target_type=torch.intaa) _UpperCAmelCase : Dict =self._scattered(snake_case , [n_queries, n_docs, question_hidden_states.shape[1]]) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(snake_case)
242
0
import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _A = logging.get_logger(__name__) _A = { 'facebook/detr-resnet-50': 'https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json', # See all DETR models at https://huggingface.co/models?filter=detr } class UpperCAmelCase__ ( A_ ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = "detr" UpperCAmelCase__ : Optional[int] = ["past_key_values"] UpperCAmelCase__ : Optional[int] = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self , A_=True , A_=None , A_=3 , A_=100 , A_=6 , A_=2048 , A_=8 , A_=6 , A_=2048 , A_=8 , A_=0.0 , A_=0.0 , A_=True , A_="relu" , A_=256 , A_=0.1 , A_=0.0 , A_=0.0 , A_=0.02 , A_=1.0 , A_=False , A_="sine" , A_="resnet50" , A_=True , A_=False , A_=1 , A_=5 , A_=2 , A_=1 , A_=1 , A_=5 , A_=2 , A_=0.1 , **A_ , ) -> List[Any]: if backbone_config is not None and use_timm_backbone: raise ValueError('You can\'t specify both `backbone_config` and `use_timm_backbone`.' ) if not use_timm_backbone: if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) __UpperCamelCase =CONFIG_MAPPING['resnet'](out_features=['stage4'] ) elif isinstance(A_ , A_ ): __UpperCamelCase =backbone_config.get('model_type' ) __UpperCamelCase =CONFIG_MAPPING[backbone_model_type] __UpperCamelCase =config_class.from_dict(A_ ) # set timm attributes to None __UpperCamelCase , __UpperCamelCase , __UpperCamelCase =None, None, None __UpperCamelCase =use_timm_backbone __UpperCamelCase =backbone_config __UpperCamelCase =num_channels __UpperCamelCase =num_queries __UpperCamelCase =d_model __UpperCamelCase =encoder_ffn_dim __UpperCamelCase =encoder_layers __UpperCamelCase =encoder_attention_heads __UpperCamelCase =decoder_ffn_dim __UpperCamelCase =decoder_layers __UpperCamelCase =decoder_attention_heads __UpperCamelCase =dropout __UpperCamelCase =attention_dropout __UpperCamelCase =activation_dropout __UpperCamelCase =activation_function __UpperCamelCase =init_std __UpperCamelCase =init_xavier_std __UpperCamelCase =encoder_layerdrop __UpperCamelCase =decoder_layerdrop __UpperCamelCase =encoder_layers __UpperCamelCase =auxiliary_loss __UpperCamelCase =position_embedding_type __UpperCamelCase =backbone __UpperCamelCase =use_pretrained_backbone __UpperCamelCase =dilation # Hungarian matcher __UpperCamelCase =class_cost __UpperCamelCase =bbox_cost __UpperCamelCase =giou_cost # Loss coefficients __UpperCamelCase =mask_loss_coefficient __UpperCamelCase =dice_loss_coefficient __UpperCamelCase =bbox_loss_coefficient __UpperCamelCase =giou_loss_coefficient __UpperCamelCase =eos_coefficient super().__init__(is_encoder_decoder=A_ , **A_ ) @property def _a ( self ) -> int: return self.encoder_attention_heads @property def _a ( self ) -> int: return self.d_model @classmethod def _a ( cls , A_ , **A_ ) -> Tuple: return cls(backbone_config=A_ , **A_ ) def _a ( self ) -> Dict[str, any]: __UpperCamelCase =copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: __UpperCamelCase =self.backbone_config.to_dict() __UpperCamelCase =self.__class__.model_type return output class UpperCAmelCase__ ( A_ ): """simple docstring""" UpperCAmelCase__ : Dict = version.parse("1.11" ) @property def _a ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ('pixel_mask', {0: 'batch'}), ] ) @property def _a ( self ) -> float: return 1E-5 @property def _a ( self ) -> int: return 12
62
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def __magic_name__ ( A : Union[str, Any], A : str, A : Optional[int]=None, A : List[str]=None ): '''simple docstring''' if attention_mask is None: a = tf.cast(tf.math.not_equal(A, config.pad_token_id ), tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class snake_case__ : """simple docstring""" SCREAMING_SNAKE_CASE_ : int = OPTConfig SCREAMING_SNAKE_CASE_ : Dict = {} SCREAMING_SNAKE_CASE_ : List[str] = """gelu""" def __init__( self : List[str] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : List[Any]=13 , __lowerCamelCase : int=7 , __lowerCamelCase : Tuple=True , __lowerCamelCase : Union[str, Any]=False , __lowerCamelCase : Optional[Any]=99 , __lowerCamelCase : List[Any]=16 , __lowerCamelCase : List[str]=2 , __lowerCamelCase : int=4 , __lowerCamelCase : Any=4 , __lowerCamelCase : Union[str, Any]="gelu" , __lowerCamelCase : Any=0.1 , __lowerCamelCase : List[Any]=0.1 , __lowerCamelCase : Dict=20 , __lowerCamelCase : List[str]=2 , __lowerCamelCase : Any=1 , __lowerCamelCase : Any=0 , __lowerCamelCase : Optional[Any]=16 , __lowerCamelCase : Optional[Any]=16 , ) -> Any: a = parent a = batch_size a = seq_length a = is_training a = use_labels a = vocab_size a = hidden_size a = num_hidden_layers a = num_attention_heads a = intermediate_size a = hidden_act a = hidden_dropout_prob a = attention_probs_dropout_prob a = max_position_embeddings a = eos_token_id a = pad_token_id a = bos_token_id a = embed_dim a = word_embed_proj_dim a = False def __UpperCAmelCase ( self : str ) -> int: a = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) a = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) a = tf.concat([input_ids, eos_tensor] , axis=1 ) a = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=__lowerCamelCase , **self.config_updates , ) a = prepare_opt_inputs_dict(__lowerCamelCase , __lowerCamelCase ) return config, inputs_dict def __UpperCAmelCase ( self : int , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Dict ) -> List[str]: a = TFOPTModel(config=__lowerCamelCase ) a = inputs_dict["input_ids"] a = input_ids[:1, :] a = inputs_dict["attention_mask"][:1, :] a = 1 # first forward pass a = model(__lowerCamelCase , attention_mask=__lowerCamelCase , use_cache=__lowerCamelCase ) a , a = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids a = ids_tensor((self.batch_size, 3) , config.vocab_size ) a = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and a = tf.concat([input_ids, next_tokens] , axis=-1 ) a = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) a = model(__lowerCamelCase , attention_mask=__lowerCamelCase )[0] a = model(__lowerCamelCase , attention_mask=__lowerCamelCase , past_key_values=__lowerCamelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice a = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) a = output_from_no_past[:, -3:, random_slice_idx] a = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__lowerCamelCase , __lowerCamelCase , rtol=1e-3 ) @require_tf class snake_case__ (_UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () SCREAMING_SNAKE_CASE_ : Optional[Any] = (TFOPTForCausalLM,) if is_tf_available() else () SCREAMING_SNAKE_CASE_ : Tuple = ( {"""feature-extraction""": TFOPTModel, """text-generation""": TFOPTForCausalLM} if is_tf_available() else {} ) SCREAMING_SNAKE_CASE_ : int = False SCREAMING_SNAKE_CASE_ : List[str] = False SCREAMING_SNAKE_CASE_ : int = False SCREAMING_SNAKE_CASE_ : List[str] = 10 def __UpperCAmelCase ( self : Tuple ) -> List[str]: a = TFOPTModelTester(self ) a = ConfigTester(self , config_class=__lowerCamelCase ) def __UpperCAmelCase ( self : Dict ) -> List[str]: self.config_tester.run_common_tests() def __UpperCAmelCase ( self : Optional[int] ) -> Tuple: a = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__lowerCamelCase ) def __UpperCAmelCase ( self : Union[str, Any] ) -> Union[str, Any]: a , a = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(__lowerCamelCase : Tuple , __lowerCamelCase : int ): if hasattr(__lowerCamelCase , "weight" ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(__lowerCamelCase , "weight" ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 10, config.vocab_size + 10]: # build the embeddings a = model_class(config=__lowerCamelCase ) a = _get_word_embedding_weight(__lowerCamelCase , model.get_input_embeddings() ) a = _get_word_embedding_weight(__lowerCamelCase , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(__lowerCamelCase ) a = _get_word_embedding_weight(__lowerCamelCase , model.get_input_embeddings() ) a = _get_word_embedding_weight(__lowerCamelCase , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. a = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , __lowerCamelCase ) # check that weights remain the same after resizing a = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: a = False self.assertTrue(__lowerCamelCase ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , __lowerCamelCase ) a = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: a = False self.assertTrue(__lowerCamelCase ) def __magic_name__ ( A : List[Any] ): '''simple docstring''' return tf.constant(A, dtype=tf.intaa ) @require_tf class snake_case__ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = 99 def __UpperCAmelCase ( self : Dict ) -> Optional[Any]: a = tf.ones((4, 1) , dtype=tf.intaa ) * 2 a = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) a = input_ids.shape[0] a = OPTConfig( vocab_size=self.vocab_size , hidden_size=24 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class snake_case__ (unittest.TestCase ): """simple docstring""" @slow def __UpperCAmelCase ( self : Dict ) -> Optional[Any]: a = TFOPTModel.from_pretrained("facebook/opt-350m" ) a = _long_tensor([[0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2]] ) a = tf.not_equal(__lowerCamelCase , model.config.pad_token_id ) with tf.GradientTape(): a = model(input_ids=__lowerCamelCase , attention_mask=__lowerCamelCase ).last_hidden_state a = (1, 11, 5_12) self.assertEqual(output.shape , __lowerCamelCase ) a = tf.constant( [[-0.2_873, -1.9_218, -0.3_033], [-1.2_710, -0.1_338, -0.1_902], [0.4_095, 0.1_214, -1.3_121]] ) self.assertTrue(np.allclose(output[:, :3, :3] , __lowerCamelCase , atol=4e-3 ) ) a = tf.function(__lowerCamelCase , jit_compile=__lowerCamelCase ) a = xla_generate(__lowerCamelCase , __lowerCamelCase )[0] self.assertTrue(np.allclose(output[:, :3, :3] , __lowerCamelCase , atol=4e-2 ) ) @require_tf @slow class snake_case__ (unittest.TestCase ): """simple docstring""" def __UpperCAmelCase ( self : Optional[Any] ) -> List[str]: super().setUp() a = "facebook/opt-350m" def __UpperCAmelCase ( self : Any ) -> Tuple: a = TFOPTForCausalLM.from_pretrained(self.path_model ) a = GPTaTokenizer.from_pretrained(self.path_model ) a = [ "Today is a beautiful day and I want to", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False a = tokenizer(__lowerCamelCase , return_tensors="tf" , padding=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) a = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) a = tf.constant( [ [1.3_851, -13.8_923, -10.5_229, -10.7_533, -0.2_309, -10.2_384, -0.5_365, -9.0_947, -5.1_670], [-4.7_073, -10.6_276, -3.9_415, -21.5_242, -0.2_822, -0.2_822, -0.2_822, -0.2_822, -0.2_822], [0.6_247, -3.4_229, -8.9_179, -1.4_297, -14.1_650, 1.4_146, -9.0_218, -0.2_703, -0.2_703], [6.4_783, -1.9_913, -10.7_926, -2.3_336, 1.5_092, -0.9_974, -6.8_213, 1.3_477, 1.3_477], ] ) self.assertTrue(np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-4 ) ) a = tf.function(__lowerCamelCase , jit_compile=__lowerCamelCase ) a = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-4 ) ) @require_tf @slow class snake_case__ (unittest.TestCase ): """simple docstring""" @property def __UpperCAmelCase ( self : Any ) -> str: return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def __UpperCAmelCase ( self : Optional[Any] ) -> List[str]: a = "facebook/opt-125m" a = [ "Today is a beautiful day and I want to", "In the city of New York, the city", "Paris is the capital of France and the capital", "Computers and mobile phones have taken over the", ] a = [] a = GPTaTokenizer.from_pretrained(__lowerCamelCase ) a = TFOPTForCausalLM.from_pretrained(__lowerCamelCase ) for prompt in self.prompts: a = tokenizer(__lowerCamelCase , return_tensors="tf" ).input_ids a = model.generate(__lowerCamelCase , max_length=10 ) a = tokenizer.batch_decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) predicted_outputs += generated_string self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def __UpperCAmelCase ( self : str ) -> Dict: a = "facebook/opt-350m" a = GPTaTokenizer.from_pretrained(__lowerCamelCase ) a = TFOPTForCausalLM.from_pretrained(__lowerCamelCase ) a = "left" # use different length sentences to test batching a = [ "Hello, my dog is a little", "Today, I", ] a = tokenizer(__lowerCamelCase , return_tensors="tf" , padding=__lowerCamelCase ) a = inputs["input_ids"] a = model.generate(input_ids=__lowerCamelCase , attention_mask=inputs["attention_mask"] ) a = tokenizer(sentences[0] , return_tensors="tf" ).input_ids a = model.generate(input_ids=__lowerCamelCase ) a = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs["attention_mask"][-1] , tf.intaa ) ) a = tokenizer(sentences[1] , return_tensors="tf" ).input_ids a = model.generate(input_ids=__lowerCamelCase , max_length=model.config.max_length - num_paddings ) a = tokenizer.batch_decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) a = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__lowerCamelCase ) a = tokenizer.decode(output_padded[0] , skip_special_tokens=__lowerCamelCase ) a = [ "Hello, my dog is a little bit of a dork.\nI'm a little bit", "Today, I was in the middle of a conversation with a friend about the", ] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) self.assertListEqual(__lowerCamelCase , [non_padded_sentence, padded_sentence] ) def __UpperCAmelCase ( self : List[Any] ) -> Optional[int]: a = "facebook/opt-350m" a = [ "Today is a beautiful day and I want to", "In the city of San Francisco, the city", "Paris is the capital of France and the capital", "Computers and mobile phones have taken over the", ] a = [] a = GPTaTokenizer.from_pretrained(__lowerCamelCase ) a = TFOPTForCausalLM.from_pretrained(__lowerCamelCase ) for prompt in self.prompts: a = tokenizer(__lowerCamelCase , return_tensors="tf" ).input_ids a = model.generate(__lowerCamelCase , max_length=10 ) a = tokenizer.batch_decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) predicted_outputs += generated_string self.assertListEqual(__lowerCamelCase , __lowerCamelCase )
107
0
"""simple docstring""" import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( 'The converted tokenizer will be the `slow` tokenizer. To use the fast, update your `tokenizers` library and re-run the tokenizer conversion' ) _a = None _a = { '7B': 11_008, '13B': 13_824, '30B': 17_920, '65B': 22_016, '70B': 28_672, } _a = { '7B': 1, '7Bf': 1, '13B': 2, '13Bf': 2, '30B': 4, '65B': 8, '70B': 8, '70Bf': 8, } def __a ( __lowerCamelCase, __lowerCamelCase=1, __lowerCamelCase=256 ): return multiple_of * ((int(ffn_dim_multiplier * int(8 * n / 3 ) ) + multiple_of - 1) // multiple_of) def __a ( __lowerCamelCase ): with open(__lowerCamelCase, "r" ) as f: return json.load(__lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase ): with open(__lowerCamelCase, "w" ) as f: json.dump(__lowerCamelCase, __lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=True ): os.makedirs(__lowerCamelCase, exist_ok=__lowerCamelCase ) UpperCAmelCase_ : Union[str, Any] = os.path.join(__lowerCamelCase, "tmp" ) os.makedirs(__lowerCamelCase, exist_ok=__lowerCamelCase ) UpperCAmelCase_ : List[Any] = read_json(os.path.join(__lowerCamelCase, "params.json" ) ) UpperCAmelCase_ : str = NUM_SHARDS[model_size] UpperCAmelCase_ : Union[str, Any] = params["n_layers"] UpperCAmelCase_ : Union[str, Any] = params["n_heads"] UpperCAmelCase_ : Union[str, Any] = n_heads // num_shards UpperCAmelCase_ : int = params["dim"] UpperCAmelCase_ : Tuple = dim // n_heads UpperCAmelCase_ : Any = 1_0000.0 UpperCAmelCase_ : Optional[int] = 1.0 / (base ** (torch.arange(0, __lowerCamelCase, 2 ).float() / dims_per_head)) if "n_kv_heads" in params: UpperCAmelCase_ : List[str] = params["n_kv_heads"] # for GQA / MQA UpperCAmelCase_ : List[Any] = n_heads_per_shard // num_key_value_heads UpperCAmelCase_ : List[str] = dim // num_key_value_heads else: # compatibility with other checkpoints UpperCAmelCase_ : Dict = n_heads UpperCAmelCase_ : List[Any] = n_heads_per_shard UpperCAmelCase_ : Any = dim # permute for sliced rotary def permute(__lowerCamelCase, __lowerCamelCase=n_heads, __lowerCamelCase=dim, __lowerCamelCase=dim ): return w.view(__lowerCamelCase, dima // n_heads // 2, 2, __lowerCamelCase ).transpose(1, 2 ).reshape(__lowerCamelCase, __lowerCamelCase ) print(f"""Fetching all parameters from the checkpoint at {input_base_path}.""" ) # Load weights if model_size == "7B": # Not sharded # (The sharded implementation would also work, but this is simpler.) UpperCAmelCase_ : Dict = torch.load(os.path.join(__lowerCamelCase, "consolidated.00.pth" ), map_location="cpu" ) else: # Sharded UpperCAmelCase_ : List[Any] = [ torch.load(os.path.join(__lowerCamelCase, f"""consolidated.{i:02d}.pth""" ), map_location="cpu" ) for i in range(__lowerCamelCase ) ] UpperCAmelCase_ : Optional[Any] = 0 UpperCAmelCase_ : Optional[int] = {"weight_map": {}} for layer_i in range(__lowerCamelCase ): UpperCAmelCase_ : List[str] = f"""pytorch_model-{layer_i + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase_ : List[Any] = { f"""model.layers.{layer_i}.self_attn.q_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wq.weight"""] ), f"""model.layers.{layer_i}.self_attn.k_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wk.weight"""] ), f"""model.layers.{layer_i}.self_attn.v_proj.weight""": loaded[f"""layers.{layer_i}.attention.wv.weight"""], f"""model.layers.{layer_i}.self_attn.o_proj.weight""": loaded[f"""layers.{layer_i}.attention.wo.weight"""], f"""model.layers.{layer_i}.mlp.gate_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w1.weight"""], f"""model.layers.{layer_i}.mlp.down_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w2.weight"""], f"""model.layers.{layer_i}.mlp.up_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w3.weight"""], f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[f"""layers.{layer_i}.attention_norm.weight"""], f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[f"""layers.{layer_i}.ffn_norm.weight"""], } else: # Sharded # Note that attention.w{q,k,v,o}, feed_fordward.w[1,2,3], attention_norm.weight and ffn_norm.weight share # the same storage object, saving attention_norm and ffn_norm will save other weights too, which is # redundant as other weights will be stitched from multiple shards. To avoid that, they are cloned. UpperCAmelCase_ : Union[str, Any] = { f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.attention_norm.weight""" ].clone(), f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.ffn_norm.weight""" ].clone(), } UpperCAmelCase_ : Dict = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wq.weight"""].view(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) for i in range(__lowerCamelCase ) ], dim=0, ).reshape(__lowerCamelCase, __lowerCamelCase ) ) UpperCAmelCase_ : List[str] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wk.weight"""].view( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) for i in range(__lowerCamelCase ) ], dim=0, ).reshape(__lowerCamelCase, __lowerCamelCase ), __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, ) UpperCAmelCase_ : int = torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wv.weight"""].view( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) for i in range(__lowerCamelCase ) ], dim=0, ).reshape(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : List[str] = torch.cat( [loaded[i][f"""layers.{layer_i}.attention.wo.weight"""] for i in range(__lowerCamelCase )], dim=1 ) UpperCAmelCase_ : str = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w1.weight"""] for i in range(__lowerCamelCase )], dim=0 ) UpperCAmelCase_ : Dict = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w2.weight"""] for i in range(__lowerCamelCase )], dim=1 ) UpperCAmelCase_ : Optional[Any] = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w3.weight"""] for i in range(__lowerCamelCase )], dim=0 ) UpperCAmelCase_ : List[Any] = inv_freq for k, v in state_dict.items(): UpperCAmelCase_ : List[str] = filename param_count += v.numel() torch.save(__lowerCamelCase, os.path.join(__lowerCamelCase, __lowerCamelCase ) ) UpperCAmelCase_ : str = f"""pytorch_model-{n_layers + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase_ : List[Any] = { "model.embed_tokens.weight": loaded["tok_embeddings.weight"], "model.norm.weight": loaded["norm.weight"], "lm_head.weight": loaded["output.weight"], } else: UpperCAmelCase_ : Tuple = { "model.norm.weight": loaded[0]["norm.weight"], "model.embed_tokens.weight": torch.cat( [loaded[i]["tok_embeddings.weight"] for i in range(__lowerCamelCase )], dim=1 ), "lm_head.weight": torch.cat([loaded[i]["output.weight"] for i in range(__lowerCamelCase )], dim=0 ), } for k, v in state_dict.items(): UpperCAmelCase_ : Dict = filename param_count += v.numel() torch.save(__lowerCamelCase, os.path.join(__lowerCamelCase, __lowerCamelCase ) ) # Write configs UpperCAmelCase_ : Optional[Any] = {"total_size": param_count * 2} write_json(__lowerCamelCase, os.path.join(__lowerCamelCase, "pytorch_model.bin.index.json" ) ) UpperCAmelCase_ : Optional[Any] = params["ffn_dim_multiplier"] if "ffn_dim_multiplier" in params else 1 UpperCAmelCase_ : List[str] = params["multiple_of"] if "multiple_of" in params else 256 UpperCAmelCase_ : str = LlamaConfig( hidden_size=__lowerCamelCase, intermediate_size=compute_intermediate_size(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ), num_attention_heads=params["n_heads"], num_hidden_layers=params["n_layers"], rms_norm_eps=params["norm_eps"], num_key_value_heads=__lowerCamelCase, ) config.save_pretrained(__lowerCamelCase ) # Make space so we can load the model properly now. del state_dict del loaded gc.collect() print("Loading the checkpoint in a Llama model." ) UpperCAmelCase_ : Tuple = LlamaForCausalLM.from_pretrained(__lowerCamelCase, torch_dtype=torch.floataa, low_cpu_mem_usage=__lowerCamelCase ) # Avoid saving this as part of the config. del model.config._name_or_path print("Saving in the Transformers format." ) model.save_pretrained(__lowerCamelCase, safe_serialization=__lowerCamelCase ) shutil.rmtree(__lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase ): # Initialize the tokenizer based on the `spm` model UpperCAmelCase_ : Any = LlamaTokenizer if LlamaTokenizerFast is None else LlamaTokenizerFast print(f"""Saving a {tokenizer_class.__name__} to {tokenizer_path}.""" ) UpperCAmelCase_ : List[Any] = tokenizer_class(__lowerCamelCase ) tokenizer.save_pretrained(__lowerCamelCase ) def __a ( ): UpperCAmelCase_ : Tuple = argparse.ArgumentParser() parser.add_argument( "--input_dir", help="Location of LLaMA weights, which contains tokenizer.model and model folders", ) parser.add_argument( "--model_size", choices=["7B", "7Bf", "13B", "13Bf", "30B", "65B", "70B", "70Bf", "tokenizer_only"], ) parser.add_argument( "--output_dir", help="Location to write HF model and tokenizer", ) parser.add_argument("--safe_serialization", type=__lowerCamelCase, help="Whether or not to save using `safetensors`." ) UpperCAmelCase_ : Dict = parser.parse_args() if args.model_size != "tokenizer_only": write_model( model_path=args.output_dir, input_base_path=os.path.join(args.input_dir, args.model_size ), model_size=args.model_size, safe_serialization=args.safe_serialization, ) UpperCAmelCase_ : str = os.path.join(args.input_dir, "tokenizer.model" ) write_tokenizer(args.output_dir, __lowerCamelCase ) if __name__ == "__main__": main()
365
"""simple docstring""" import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights UpperCAmelCase_ : Optional[Any] = FlaxDiffusionPipeline.from_pretrained( "hf-internal-testing/tiny-stable-diffusion-pipe" , safety_checker=lowercase_ , cache_dir=lowercase_ ) UpperCAmelCase_ : List[Any] = [t[-1] for t in os.walk(os.path.join(lowercase_ , os.listdir(lowercase_ )[0] , "snapshots" ) )] UpperCAmelCase_ : Dict = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith(".bin" ) for f in files ) @slow @require_flax class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( "hf-internal-testing/tiny-stable-diffusion-pipe" , safety_checker=lowercase_ ) UpperCAmelCase_ : Tuple = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : List[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase_ : List[str] = 4 UpperCAmelCase_ : Tuple = jax.device_count() UpperCAmelCase_ : Optional[int] = num_samples * [prompt] UpperCAmelCase_ : List[Any] = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : int = replicate(lowercase_ ) UpperCAmelCase_ : str = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : Dict = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1E-3 assert np.abs(np.abs(lowercase_ , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5E-1 UpperCAmelCase_ : List[Any] = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(lowercase_ ) == num_samples def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="flax" , safety_checker=lowercase_ ) UpperCAmelCase_ : Optional[int] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : str = jax.random.PRNGKey(0 ) UpperCAmelCase_ : Union[str, Any] = 50 UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[str] = num_samples * [prompt] UpperCAmelCase_ : Union[str, Any] = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Any = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : int = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ ) UpperCAmelCase_ : Any = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : str = jax.random.PRNGKey(0 ) UpperCAmelCase_ : str = 50 UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : Any = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Dict = replicate(lowercase_ ) UpperCAmelCase_ : str = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Union[str, Any] = shard(lowercase_ ) UpperCAmelCase_ : List[Any] = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : str = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa ) UpperCAmelCase_ : List[Any] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : Dict = jax.random.PRNGKey(0 ) UpperCAmelCase_ : Optional[int] = 50 UpperCAmelCase_ : Optional[int] = jax.device_count() UpperCAmelCase_ : str = num_samples * [prompt] UpperCAmelCase_ : int = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Union[str, Any] = replicate(lowercase_ ) UpperCAmelCase_ : Union[str, Any] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[int] = shard(lowercase_ ) UpperCAmelCase_ : Any = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="scaled_linear" , set_alpha_to_one=lowercase_ , steps_offset=1 , ) UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , scheduler=lowercase_ , safety_checker=lowercase_ , ) UpperCAmelCase_ : List[Any] = scheduler.create_state() UpperCAmelCase_ : int = scheduler_state UpperCAmelCase_ : Union[str, Any] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : Optional[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase_ : int = 50 UpperCAmelCase_ : str = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : int = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : int = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = shard(lowercase_ ) UpperCAmelCase_ : Any = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : Union[str, Any] = jax.random.split(jax.random.PRNGKey(0 ) , lowercase_ ) UpperCAmelCase_ , UpperCAmelCase_ : str = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ , ) UpperCAmelCase_ : Any = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = pipeline.prepare_inputs(lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : List[Any] = pipeline(lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase_ : int = images[2, 0, 256, 10:17, 1] # With memory efficient attention UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ , use_memory_efficient_attention=lowercase_ , ) UpperCAmelCase_ : str = replicate(lowercase_ ) UpperCAmelCase_ : str = pipeline.prepare_inputs(lowercase_ ) UpperCAmelCase_ : Optional[int] = shard(lowercase_ ) UpperCAmelCase_ : str = pipeline(lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase_ : Optional[int] = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
23
0
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowercase_ = 42 lowercase_ = 42 try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.26.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version(">=", "0.0.12") ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowercase_ = 42 lowercase_ = 42 from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
10
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __A = { "configuration_distilbert": [ "DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DistilBertConfig", "DistilBertOnnxConfig", ], "tokenization_distilbert": ["DistilBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = ["DistilBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "DistilBertForMaskedLM", "DistilBertForMultipleChoice", "DistilBertForQuestionAnswering", "DistilBertForSequenceClassification", "DistilBertForTokenClassification", "DistilBertModel", "DistilBertPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFDistilBertForMaskedLM", "TFDistilBertForMultipleChoice", "TFDistilBertForQuestionAnswering", "TFDistilBertForSequenceClassification", "TFDistilBertForTokenClassification", "TFDistilBertMainLayer", "TFDistilBertModel", "TFDistilBertPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "FlaxDistilBertForMaskedLM", "FlaxDistilBertForMultipleChoice", "FlaxDistilBertForQuestionAnswering", "FlaxDistilBertForSequenceClassification", "FlaxDistilBertForTokenClassification", "FlaxDistilBertModel", "FlaxDistilBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
10
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase__ : Optional[Any] ={'configuration_vit_msn': ['VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ViTMSNConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : Optional[int] =[ 'VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST', 'ViTMSNModel', 'ViTMSNForImageClassification', 'ViTMSNPreTrainedModel', ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys lowerCAmelCase__ : List[Any] =_LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
351
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase__ : Optional[Any] =logging.get_logger(__name__) lowerCAmelCase__ : Optional[int] ={ 'sail/poolformer_s12': 'https://huggingface.co/sail/poolformer_s12/resolve/main/config.json', # See all PoolFormer models at https://huggingface.co/models?filter=poolformer } class __lowercase (__SCREAMING_SNAKE_CASE ): """simple docstring""" _UpperCAmelCase = """poolformer""" def __init__( self , lowerCAmelCase__=3 , lowerCAmelCase__=1_6 , lowerCAmelCase__=1_6 , lowerCAmelCase__=3 , lowerCAmelCase__=4.0 , lowerCAmelCase__=[2, 2, 6, 2] , lowerCAmelCase__=[6_4, 1_2_8, 3_2_0, 5_1_2] , lowerCAmelCase__=[7, 3, 3, 3] , lowerCAmelCase__=[4, 2, 2, 2] , lowerCAmelCase__=[2, 1, 1, 1] , lowerCAmelCase__=4 , lowerCAmelCase__=0.0 , lowerCAmelCase__="gelu" , lowerCAmelCase__=True , lowerCAmelCase__=1E-5 , lowerCAmelCase__=0.02 , **lowerCAmelCase__ , ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = num_channels SCREAMING_SNAKE_CASE_ : List[str] = patch_size SCREAMING_SNAKE_CASE_ : Tuple = stride SCREAMING_SNAKE_CASE_ : List[Any] = padding SCREAMING_SNAKE_CASE_ : Union[str, Any] = pool_size SCREAMING_SNAKE_CASE_ : Tuple = hidden_sizes SCREAMING_SNAKE_CASE_ : Tuple = mlp_ratio SCREAMING_SNAKE_CASE_ : Dict = depths SCREAMING_SNAKE_CASE_ : List[Any] = patch_sizes SCREAMING_SNAKE_CASE_ : List[Any] = strides SCREAMING_SNAKE_CASE_ : int = num_encoder_blocks SCREAMING_SNAKE_CASE_ : List[Any] = drop_path_rate SCREAMING_SNAKE_CASE_ : Tuple = hidden_act SCREAMING_SNAKE_CASE_ : str = use_layer_scale SCREAMING_SNAKE_CASE_ : List[str] = layer_scale_init_value SCREAMING_SNAKE_CASE_ : Tuple = initializer_range super().__init__(**lowerCAmelCase__ ) class __lowercase (__SCREAMING_SNAKE_CASE ): """simple docstring""" _UpperCAmelCase = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self ): """simple docstring""" return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def UpperCamelCase__ ( self ): """simple docstring""" return 2E-3
162
0
from argparse import ArgumentParser from . import BaseTransformersCLICommand def a ( snake_case__: Tuple ): '''simple docstring''' return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase__( UpperCAmelCase ): """simple docstring""" @staticmethod def _lowercase ( SCREAMING_SNAKE_CASE_ : ArgumentParser ) -> int: lowercase_ = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' , type=SCREAMING_SNAKE_CASE_ , default=SCREAMING_SNAKE_CASE_ , help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' , action='''store_true''' , help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' , action='''store_true''' , help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' , ) download_parser.add_argument('''model''' , type=SCREAMING_SNAKE_CASE_ , help='''Name of the model to download''' ) download_parser.set_defaults(func=SCREAMING_SNAKE_CASE_ ) def __init__( self : str , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : bool , SCREAMING_SNAKE_CASE_ : bool ) -> Optional[Any]: lowercase_ = model lowercase_ = cache lowercase_ = force lowercase_ = trust_remote_code def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
30
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. __snake_case :Optional[int] = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. __snake_case :List[str] = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. __snake_case :List[Any] = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): __a = len([g for position, g in enumerate(_UpperCAmelCase ) if g == main_target[position]] ) return (item, float(_UpperCAmelCase )) def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): __a = random.randint(0 , len(_UpperCAmelCase ) - 1 ) __a = parent_a[:random_slice] + parent_a[random_slice:] __a = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): __a = list(_UpperCAmelCase ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: __a = random.choice(_UpperCAmelCase ) return "".join(_UpperCAmelCase ) def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): __a = [] # Generate more children proportionally to the fitness score. __a = int(parent_a[1] * 100 ) + 1 __a = 10 if child_n >= 10 else child_n for _ in range(_UpperCAmelCase ): __a = population_score[random.randint(0 , _UpperCAmelCase )][0] __a , __a = crossover(parent_a[0] , _UpperCAmelCase ) # Append new string to the population list. pop.append(mutate(_UpperCAmelCase , _UpperCAmelCase ) ) pop.append(mutate(_UpperCAmelCase , _UpperCAmelCase ) ) return pop def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = True ): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: __a = f'{N_POPULATION} must be bigger than {N_SELECTED}' raise ValueError(_UpperCAmelCase ) # Verify that the target contains no genes besides the ones inside genes variable. __a = sorted({c for c in target if c not in genes} ) if not_in_genes_list: __a = f'{not_in_genes_list} is not in genes list, evolution cannot converge' raise ValueError(_UpperCAmelCase ) # Generate random starting population. __a = [] for _ in range(_UpperCAmelCase ): population.append(''''''.join([random.choice(_UpperCAmelCase ) for i in range(len(_UpperCAmelCase ) )] ) ) # Just some logs to know what the algorithms is doing. __a , __a = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(_UpperCAmelCase ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. __a = [evaluate(_UpperCAmelCase , _UpperCAmelCase ) for item in population] # Check if there is a matching evolution. __a = sorted(_UpperCAmelCase , key=lambda _UpperCAmelCase : x[1] , reverse=_UpperCAmelCase ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( f'\nGeneration: {generation}' f'\nTotal Population:{total_population}' f'\nBest score: {population_score[0][1]}' f'\nBest string: {population_score[0][0]}' ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. __a = population[: int(N_POPULATION / 3 )] population.clear() population.extend(_UpperCAmelCase ) # Normalize population score to be between 0 and 1. __a = [ (item, score / len(_UpperCAmelCase )) for item, score in population_score ] # This is selection for i in range(_UpperCAmelCase ): population.extend(select(population_score[int(_UpperCAmelCase )] , _UpperCAmelCase , _UpperCAmelCase ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(_UpperCAmelCase ) > N_POPULATION: break if __name__ == "__main__": __snake_case :Optional[int] = ( '''This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!''' ) __snake_case :List[Any] = list( ''' ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm''' '''nopqrstuvwxyz.,;!?+-*#@^\'èéòà€ù=)(&%$£/\\''' ) __snake_case ,__snake_case ,__snake_case :Dict = basic(target_str, genes_list) print( f'\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}' )
49
0
import warnings from typing import Dict import numpy as np from ..utils import ExplicitEnum, add_end_docstrings, is_tf_available, is_torch_available from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING def _snake_case ( lowercase__ : Union[str, Any] ) -> Any: '''simple docstring''' return 1.0 / (1.0 + np.exp(-_outputs )) def _snake_case ( lowercase__ : List[Any] ) -> int: '''simple docstring''' lowerCAmelCase_ :Optional[Any] = np.max(_outputs , axis=-1 , keepdims=lowercase__ ) lowerCAmelCase_ :List[Any] = np.exp(_outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 , keepdims=lowercase__ ) class _SCREAMING_SNAKE_CASE ( A__ ): UpperCAmelCase_ :Tuple = "sigmoid" UpperCAmelCase_ :Dict = "softmax" UpperCAmelCase_ :Tuple = "none" @add_end_docstrings( A__ , r"\n return_all_scores (`bool`, *optional*, defaults to `False`):\n Whether to return all prediction scores or just the one of the predicted class.\n function_to_apply (`str`, *optional*, defaults to `\"default\"`):\n The function to apply to the model outputs in order to retrieve the scores. Accepts four different values:\n\n - `\"default\"`: if the model has a single label, will apply the sigmoid function on the output. If the model\n has several labels, will apply the softmax function on the output.\n - `\"sigmoid\"`: Applies the sigmoid function on the output.\n - `\"softmax\"`: Applies the softmax function on the output.\n - `\"none\"`: Does not apply any function on the output.\n " , ) class _SCREAMING_SNAKE_CASE ( A__ ): UpperCAmelCase_ :Optional[Any] = False UpperCAmelCase_ :List[Any] = ClassificationFunction.NONE def __init__( self , **__A ) -> List[str]: super().__init__(**__A ) self.check_model_type( TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if self.framework == """tf""" else MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING ) def __lowerCAmelCase ( self , __A=None , __A=None , __A="" , **__A ) -> Tuple: # Using "" as default argument because we're going to use `top_k=None` in user code to declare # "No top_k" lowerCAmelCase_ :Tuple = tokenizer_kwargs lowerCAmelCase_ :List[str] = {} if hasattr(self.model.config , """return_all_scores""" ) and return_all_scores is None: lowerCAmelCase_ :int = self.model.config.return_all_scores if isinstance(__A , __A ) or top_k is None: lowerCAmelCase_ :Any = top_k lowerCAmelCase_ :Optional[int] = False elif return_all_scores is not None: warnings.warn( """`return_all_scores` is now deprecated, if want a similar functionality use `top_k=None` instead of""" """ `return_all_scores=True` or `top_k=1` instead of `return_all_scores=False`.""" , __A , ) if return_all_scores: lowerCAmelCase_ :Optional[Any] = None else: lowerCAmelCase_ :Dict = 1 if isinstance(__A , __A ): lowerCAmelCase_ :Any = ClassificationFunction[function_to_apply.upper()] if function_to_apply is not None: lowerCAmelCase_ :Optional[Any] = function_to_apply return preprocess_params, {}, postprocess_params def __call__( self , *__A , **__A ) -> str: lowerCAmelCase_ :Dict = super().__call__(*__A , **__A ) # TODO try and retrieve it in a nicer way from _sanitize_parameters. lowerCAmelCase_ :Optional[int] = """top_k""" not in kwargs if isinstance(args[0] , __A ) and _legacy: # This pipeline is odd, and return a list when single item is run return [result] else: return result def __lowerCAmelCase ( self , __A , **__A ) -> Dict[str, GenericTensor]: lowerCAmelCase_ :List[str] = self.framework if isinstance(__A , __A ): return self.tokenizer(**__A , return_tensors=__A , **__A ) elif isinstance(__A , __A ) and len(__A ) == 1 and isinstance(inputs[0] , __A ) and len(inputs[0] ) == 2: # It used to be valid to use a list of list of list for text pairs, keeping this path for BC return self.tokenizer( text=inputs[0][0] , text_pair=inputs[0][1] , return_tensors=__A , **__A ) elif isinstance(__A , __A ): # This is likely an invalid usage of the pipeline attempting to pass text pairs. raise ValueError( """The pipeline received invalid inputs, if you are trying to send text pairs, you can try to send a""" """ dictionary `{\"text\": \"My text\", \"text_pair\": \"My pair\"}` in order to send a text pair.""" ) return self.tokenizer(__A , return_tensors=__A , **__A ) def __lowerCAmelCase ( self , __A ) -> Optional[int]: return self.model(**__A ) def __lowerCAmelCase ( self , __A , __A=None , __A=1 , __A=True ) -> Union[str, Any]: # `_legacy` is used to determine if we're running the naked pipeline and in backward # compatibility mode, or if running the pipeline with `pipeline(..., top_k=1)` we're running # the more natural result containing the list. # Default value before `set_parameters` if function_to_apply is None: if self.model.config.problem_type == "multi_label_classification" or self.model.config.num_labels == 1: lowerCAmelCase_ :List[str] = ClassificationFunction.SIGMOID elif self.model.config.problem_type == "single_label_classification" or self.model.config.num_labels > 1: lowerCAmelCase_ :Any = ClassificationFunction.SOFTMAX elif hasattr(self.model.config , """function_to_apply""" ) and function_to_apply is None: lowerCAmelCase_ :Optional[int] = self.model.config.function_to_apply else: lowerCAmelCase_ :List[Any] = ClassificationFunction.NONE lowerCAmelCase_ :Any = model_outputs["""logits"""][0] lowerCAmelCase_ :Optional[Any] = outputs.numpy() if function_to_apply == ClassificationFunction.SIGMOID: lowerCAmelCase_ :Optional[int] = sigmoid(__A ) elif function_to_apply == ClassificationFunction.SOFTMAX: lowerCAmelCase_ :Optional[int] = softmax(__A ) elif function_to_apply == ClassificationFunction.NONE: lowerCAmelCase_ :List[Any] = outputs else: raise ValueError(f"""Unrecognized `function_to_apply` argument: {function_to_apply}""" ) if top_k == 1 and _legacy: return {"label": self.model.config.idalabel[scores.argmax().item()], "score": scores.max().item()} lowerCAmelCase_ :int = [ {"""label""": self.model.config.idalabel[i], """score""": score.item()} for i, score in enumerate(__A ) ] if not _legacy: dict_scores.sort(key=lambda __A : x["score"] , reverse=__A ) if top_k is not None: lowerCAmelCase_ :List[str] = dict_scores[:top_k] return dict_scores
351
"""simple docstring""" import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class _SCREAMING_SNAKE_CASE : def __init__( self , __A ) -> Union[str, Any]: if isinstance(__A , __A ): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden lowerCAmelCase_ :Tuple = deepcopy(__A ) elif os.path.exists(__A ): with io.open(__A , """r""" , encoding="""utf-8""" ) as f: lowerCAmelCase_ :str = json.load(__A ) else: try: lowerCAmelCase_ :Dict = baseaa.urlsafe_baadecode(__A ).decode("""utf-8""" ) lowerCAmelCase_ :int = json.loads(__A ) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( f"""Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}""" ) lowerCAmelCase_ :Optional[Any] = config self.set_stage_and_offload() def __lowerCAmelCase ( self ) -> Tuple: # zero stage - this is done as early as possible, before model is created, to allow # ``is_deepspeed_zero3_enabled`` query and getting to the early deepspeed config object # during ``zero.Init()`` which needs to know the dtype, and some other hparams. lowerCAmelCase_ :Tuple = self.get_value("""zero_optimization.stage""" , -1 ) # offload lowerCAmelCase_ :Dict = False if self.is_zeroa() or self.is_zeroa(): lowerCAmelCase_ :Optional[int] = set(["""cpu""", """nvme"""] ) lowerCAmelCase_ :Union[str, Any] = set( [ self.get_value("""zero_optimization.offload_optimizer.device""" ), self.get_value("""zero_optimization.offload_param.device""" ), ] ) if len(offload_devices & offload_devices_valid ) > 0: lowerCAmelCase_ :Optional[int] = True def __lowerCAmelCase ( self , __A ) -> Optional[Any]: lowerCAmelCase_ :str = self.config # find the config node of interest if it exists lowerCAmelCase_ :Tuple = ds_key_long.split(""".""" ) lowerCAmelCase_ :List[str] = nodes.pop() for node in nodes: lowerCAmelCase_ :Tuple = config.get(__A ) if config is None: return None, ds_key return config, ds_key def __lowerCAmelCase ( self , __A , __A=None ) -> Optional[Any]: lowerCAmelCase_ , lowerCAmelCase_ :Optional[Any] = self.find_config_node(__A ) if config is None: return default return config.get(__A , __A ) def __lowerCAmelCase ( self , __A , __A=False ) -> Optional[Any]: lowerCAmelCase_ :Tuple = self.config # find the config node of interest if it exists lowerCAmelCase_ :Union[str, Any] = ds_key_long.split(""".""" ) for node in nodes: lowerCAmelCase_ :int = config lowerCAmelCase_ :Any = config.get(__A ) if config is None: if must_exist: raise ValueError(f"""Can't find {ds_key_long} entry in the config: {self.config}""" ) else: return # if found remove it if parent_config is not None: parent_config.pop(__A ) def __lowerCAmelCase ( self , __A ) -> Union[str, Any]: lowerCAmelCase_ :Optional[int] = self.get_value(__A ) return False if value is None else bool(__A ) def __lowerCAmelCase ( self , __A ) -> Optional[int]: lowerCAmelCase_ :List[str] = self.get_value(__A ) return False if value is None else not bool(__A ) def __lowerCAmelCase ( self ) -> str: return self._stage == 2 def __lowerCAmelCase ( self ) -> Union[str, Any]: return self._stage == 3 def __lowerCAmelCase ( self ) -> Union[str, Any]: return self._offload class _SCREAMING_SNAKE_CASE : def __init__( self , __A ) -> Optional[int]: lowerCAmelCase_ :Dict = engine def __lowerCAmelCase ( self , __A , **__A ) -> str: # runs backpropagation and handles mixed precision self.engine.backward(__A , **__A ) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class _SCREAMING_SNAKE_CASE ( A__ ): def __init__( self , __A ) -> List[str]: super().__init__(__A , device_placement=__A , scaler=__A ) lowerCAmelCase_ :List[str] = hasattr(self.optimizer , """overflow""" ) def __lowerCAmelCase ( self , __A=None ) -> Optional[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def __lowerCAmelCase ( self ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def __lowerCAmelCase ( self ) -> int: if self.__has_overflow__: return self.optimizer.overflow return False class _SCREAMING_SNAKE_CASE ( A__ ): def __init__( self , __A , __A ) -> Optional[int]: super().__init__(__A , __A ) def __lowerCAmelCase ( self ) -> Any: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class _SCREAMING_SNAKE_CASE : def __init__( self , __A , __A=0.0_0_1 , __A=0 , **__A ) -> List[Any]: lowerCAmelCase_ :str = params lowerCAmelCase_ :Any = lr lowerCAmelCase_ :List[Any] = weight_decay lowerCAmelCase_ :Any = kwargs class _SCREAMING_SNAKE_CASE : def __init__( self , __A , __A=None , __A=0 , **__A ) -> List[str]: lowerCAmelCase_ :Optional[int] = optimizer lowerCAmelCase_ :int = total_num_steps lowerCAmelCase_ :List[Any] = warmup_num_steps lowerCAmelCase_ :int = kwargs
1
0
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { '''salesforce/blip2-opt-2.7b''': '''https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json''', } class lowerCAmelCase ( __a ): lowerCAmelCase_ = '''blip_2_vision_model''' def __init__( self : str , __lowercase : List[Any]=1408 , __lowercase : Any=6144 , __lowercase : Optional[Any]=39 , __lowercase : str=16 , __lowercase : Any=224 , __lowercase : Dict=14 , __lowercase : Optional[int]="gelu" , __lowercase : int=0.0_0_0_0_1 , __lowercase : Any=0.0 , __lowercase : List[Any]=1E-10 , __lowercase : Any=True , **__lowercase : Optional[Any] , ): """simple docstring""" super().__init__(**lowerCAmelCase__ ) __lowercase =hidden_size __lowercase =intermediate_size __lowercase =num_hidden_layers __lowercase =num_attention_heads __lowercase =patch_size __lowercase =image_size __lowercase =initializer_range __lowercase =attention_dropout __lowercase =layer_norm_eps __lowercase =hidden_act __lowercase =qkv_bias @classmethod def snake_case ( cls : List[Any] , __lowercase : Optional[Any] , **__lowercase : Dict ): """simple docstring""" cls._set_token_in_kwargs(lowerCAmelCase__ ) __lowercase , __lowercase =cls.get_config_dict(lowerCAmelCase__ , **lowerCAmelCase__ ) # get the vision config dict if we are loading from Blip2Config if config_dict.get('model_type' ) == "blip-2": __lowercase =config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCAmelCase__ , **lowerCAmelCase__ ) class lowerCAmelCase ( __a ): lowerCAmelCase_ = '''blip_2_qformer''' def __init__( self : List[str] , __lowercase : Union[str, Any]=30522 , __lowercase : Tuple=768 , __lowercase : List[Any]=12 , __lowercase : Union[str, Any]=12 , __lowercase : Tuple=3072 , __lowercase : List[Any]="gelu" , __lowercase : Optional[int]=0.1 , __lowercase : Dict=0.1 , __lowercase : Tuple=512 , __lowercase : Optional[int]=0.0_2 , __lowercase : Union[str, Any]=1E-12 , __lowercase : List[str]=0 , __lowercase : List[Any]="absolute" , __lowercase : int=2 , __lowercase : Tuple=1408 , **__lowercase : Optional[int] , ): """simple docstring""" super().__init__(pad_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) __lowercase =vocab_size __lowercase =hidden_size __lowercase =num_hidden_layers __lowercase =num_attention_heads __lowercase =hidden_act __lowercase =intermediate_size __lowercase =hidden_dropout_prob __lowercase =attention_probs_dropout_prob __lowercase =max_position_embeddings __lowercase =initializer_range __lowercase =layer_norm_eps __lowercase =position_embedding_type __lowercase =cross_attention_frequency __lowercase =encoder_hidden_size @classmethod def snake_case ( cls : Any , __lowercase : List[str] , **__lowercase : List[str] ): """simple docstring""" cls._set_token_in_kwargs(lowerCAmelCase__ ) __lowercase , __lowercase =cls.get_config_dict(lowerCAmelCase__ , **lowerCAmelCase__ ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get('model_type' ) == "blip-2": __lowercase =config_dict['qformer_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCAmelCase__ , **lowerCAmelCase__ ) class lowerCAmelCase ( __a ): lowerCAmelCase_ = '''blip-2''' lowerCAmelCase_ = True def __init__( self : Union[str, Any] , __lowercase : Optional[int]=None , __lowercase : List[Any]=None , __lowercase : List[str]=None , __lowercase : str=32 , **__lowercase : List[Any] ): """simple docstring""" super().__init__(**lowerCAmelCase__ ) if vision_config is None: __lowercase ={} logger.info('vision_config is None. initializing the Blip2VisionConfig with default values.' ) if qformer_config is None: __lowercase ={} logger.info('qformer_config is None. Initializing the Blip2QFormerConfig with default values.' ) if text_config is None: __lowercase ={} logger.info('text_config is None. Initializing the text config with default values (`OPTConfig`).' ) __lowercase =BlipaVisionConfig(**lowerCAmelCase__ ) __lowercase =BlipaQFormerConfig(**lowerCAmelCase__ ) __lowercase =text_config['model_type'] if 'model_type' in text_config else 'opt' __lowercase =CONFIG_MAPPING[text_model_type](**lowerCAmelCase__ ) __lowercase =self.text_config.tie_word_embeddings __lowercase =self.text_config.is_encoder_decoder __lowercase =num_query_tokens __lowercase =self.vision_config.hidden_size __lowercase =self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES __lowercase =1.0 __lowercase =0.0_2 @classmethod def snake_case ( cls : Optional[Any] , __lowercase : str , __lowercase : Dict , __lowercase : Optional[Any] , **__lowercase : Any , ): """simple docstring""" return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **lowerCAmelCase__ , ) def snake_case ( self : Dict ): """simple docstring""" __lowercase =copy.deepcopy(self.__dict__ ) __lowercase =self.vision_config.to_dict() __lowercase =self.qformer_config.to_dict() __lowercase =self.text_config.to_dict() __lowercase =self.__class__.model_type return output
141
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING import torch from ..models.auto import AutoModelForVisualQuestionAnswering, AutoProcessor from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class SCREAMING_SNAKE_CASE_ ( __a ): """simple docstring""" __lowercase : Tuple = '''dandelin/vilt-b32-finetuned-vqa''' __lowercase : str = ( '''This is a tool that answers a question about an image. It takes an input named `image` which should be the ''' '''image containing the information, as well as a `question` which should be the question in English. It ''' '''returns a text that is the answer to the question.''' ) __lowercase : Optional[int] = '''image_qa''' __lowercase : Union[str, Any] = AutoProcessor __lowercase : Optional[Any] = AutoModelForVisualQuestionAnswering __lowercase : List[str] = ['''image''', '''text'''] __lowercase : Optional[Any] = ['''text'''] def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__): requires_backends(self , ["""vision"""]) super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__): return self.pre_processor(lowerCAmelCase__ , lowerCAmelCase__ , return_tensors="""pt""") def snake_case_ ( self , lowerCAmelCase__): with torch.no_grad(): return self.model(**lowerCAmelCase__).logits def snake_case_ ( self , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = outputs.argmax(-1).item() return self.model.config.idalabel[idx]
100
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowercase : List[str] = { 'configuration_distilbert': [ 'DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'DistilBertConfig', 'DistilBertOnnxConfig', ], 'tokenization_distilbert': ['DistilBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : int = ['DistilBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : List[str] = [ 'DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'DistilBertForMaskedLM', 'DistilBertForMultipleChoice', 'DistilBertForQuestionAnswering', 'DistilBertForSequenceClassification', 'DistilBertForTokenClassification', 'DistilBertModel', 'DistilBertPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[Any] = [ 'TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFDistilBertForMaskedLM', 'TFDistilBertForMultipleChoice', 'TFDistilBertForQuestionAnswering', 'TFDistilBertForSequenceClassification', 'TFDistilBertForTokenClassification', 'TFDistilBertMainLayer', 'TFDistilBertModel', 'TFDistilBertPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Union[str, Any] = [ 'FlaxDistilBertForMaskedLM', 'FlaxDistilBertForMultipleChoice', 'FlaxDistilBertForQuestionAnswering', 'FlaxDistilBertForSequenceClassification', 'FlaxDistilBertForTokenClassification', 'FlaxDistilBertModel', 'FlaxDistilBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys lowercase : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
151
import qiskit def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : int = 2) -> qiskit.result.counts.Counts: '''simple docstring''' __UpperCamelCase : List[str] = qubits # Using Aer's simulator __UpperCamelCase : int = qiskit.Aer.get_backend("aer_simulator") # Creating a Quantum Circuit acting on the q register __UpperCamelCase : List[str] = qiskit.QuantumCircuit(_lowerCamelCase , _lowerCamelCase) # Adding a H gate on qubit 0 (now q0 in superposition) circuit.h(0) for i in range(1 , _lowerCamelCase): # Adding CX (CNOT) gate circuit.cx(i - 1 , _lowerCamelCase) # Mapping the quantum measurement to the classical bits circuit.measure(list(range(_lowerCamelCase)) , list(range(_lowerCamelCase))) # Now measuring any one qubit would affect other qubits to collapse # their super position and have same state as the measured one. # Executing the circuit on the simulator __UpperCamelCase : Any = qiskit.execute(_lowerCamelCase , _lowerCamelCase , shots=1_000) return job.result().get_counts(_lowerCamelCase) if __name__ == "__main__": print(f"Total count for various states are: {quantum_entanglement(3)}")
151
1
"""simple docstring""" import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def __SCREAMING_SNAKE_CASE ( A_ ): lowerCAmelCase__ : Any = os.path.join(args.tf_model_dir , '''parameters.json''' ) lowerCAmelCase__ : List[Any] = json.loads(open(SCREAMING_SNAKE_CASE_ ).read() ) if not params: raise ValueError( f'It seems that the json file at {parameter_file} is empty. Make sure you have a correct json file.' ) if not args.output.endswith('''.pt''' ): lowerCAmelCase__ : List[Any] = args.output + '''.pt''' lowerCAmelCase__ : Optional[Any] = OrderedDict() with tf.device('''/CPU:0''' ): lowerCAmelCase__ : Dict = tf.train.load_checkpoint(args.tf_model_dir ) lowerCAmelCase__ : Optional[Any] = reader.get_variable_to_shape_map() for key_name in shapes.keys(): lowerCAmelCase__ : List[Any] = reader.get_tensor(SCREAMING_SNAKE_CASE_ ).astype(np.floataa ) if key_name.endswith('''/adam_m''' ) or key_name.endswith('''/adam_v''' ): continue if key_name.startswith('''pasts/''' ): if key_name.startswith('''pasts/mlp''' ): lowerCAmelCase__ : List[Any] = int(key_name[9] ) elif key_name.startswith('''pasts/out''' ): lowerCAmelCase__ : Dict = 8 lowerCAmelCase__ : Union[str, Any] = '''model.sqout.%d.weight''' % (player * 2) # enter to nn.Sequencial with Tanh, so 2 at a time lowerCAmelCase__ : List[Any] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowerCAmelCase__ : Any = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.startswith('''model/moe''' ): lowerCAmelCase__ : List[Any] = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/switch_gating/kernel''' ): lowerCAmelCase__ : Optional[int] = '''model.blocks.%d.feed_forward.mlp.router.classifier.weight''' % player lowerCAmelCase__ : Optional[Any] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowerCAmelCase__ : Dict = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.endswith('''/softmlp/kernel''' ): lowerCAmelCase__ : Any = '''model.blocks.%d.feed_forward.soft_bypass_mlp.weight''' % player lowerCAmelCase__ : int = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowerCAmelCase__ : Optional[Any] = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.endswith('''/wo/kernel''' ) or key_name.endswith('''/wi/kernel''' ): lowerCAmelCase__ : Any = key_name[-9:-7] for i in range(16 ): lowerCAmelCase__ : Any = '''model.blocks.%d.feed_forward.mlp.experts.expert_%d.%s.weight''' % (player, i, nlayer) lowerCAmelCase__ : Union[str, Any] = ( vnp[i].transpose([1, 0] ).copy() ) # In Mesh-Tensorflow, it is one array, so it is divided lowerCAmelCase__ : List[str] = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.startswith('''model/mlp''' ): lowerCAmelCase__ : Tuple = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/p1/kernel''' ): lowerCAmelCase__ : str = '''model.blocks.%d.feed_forward.mlp.wi.weight''' % player lowerCAmelCase__ : Tuple = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowerCAmelCase__ : Optional[Any] = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.endswith('''/p1/bias''' ): lowerCAmelCase__ : Dict = '''model.blocks.%d.feed_forward.mlp.wi.bias''' % player lowerCAmelCase__ : Tuple = vnp.copy() # same because it is one dimensional lowerCAmelCase__ : str = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.endswith('''/p2/kernel''' ): lowerCAmelCase__ : Union[str, Any] = '''model.blocks.%d.feed_forward.mlp.wo.weight''' % player lowerCAmelCase__ : Optional[int] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowerCAmelCase__ : List[Any] = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.endswith('''/p2/bias''' ): lowerCAmelCase__ : Any = '''model.blocks.%d.feed_forward.mlp.wo.bias''' % player lowerCAmelCase__ : Optional[Any] = vnp.copy() # same because it is one dimensional lowerCAmelCase__ : Any = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.startswith('''model/ln''' ): lowerCAmelCase__ : Optional[int] = int(key_name[8:].split('''/''' )[0] ) if key_name.endswith('''/b''' ): lowerCAmelCase__ : int = '''model.blocks.%d.feed_forward.norm.bias''' % player lowerCAmelCase__ : List[Any] = vnp.copy() # same because it is one dimensional lowerCAmelCase__ : int = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.endswith('''/g''' ): lowerCAmelCase__ : Union[str, Any] = '''model.blocks.%d.feed_forward.norm.weight''' % player lowerCAmelCase__ : List[str] = vnp.copy() # same because it is one dimensional lowerCAmelCase__ : Dict = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.startswith('''model/att''' ): lowerCAmelCase__ : Tuple = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/qkv/kernel''' ): lowerCAmelCase__ : Tuple = vnp.copy() # Compute same dimension as Mesh-tensorflow using einsum lowerCAmelCase__ : List[str] = state[:, 0, :, :] lowerCAmelCase__ : Dict = state[:, 1, :, :] lowerCAmelCase__ : Optional[int] = state[:, 2, :, :] lowerCAmelCase__ : str = ( state_q.reshape([state_q.shape[0], state_q.shape[1] * state_q.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix lowerCAmelCase__ : Tuple = ( state_k.reshape([state_k.shape[0], state_k.shape[1] * state_k.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix lowerCAmelCase__ : int = ( state_v.reshape([state_v.shape[0], state_v.shape[1] * state_v.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix lowerCAmelCase__ : Any = '''model.blocks.%d.self_attn.self_attn.q_proj.weight''' % player lowerCAmelCase__ : Tuple = torch.tensor(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Any = '''model.blocks.%d.self_attn.self_attn.k_proj.weight''' % player lowerCAmelCase__ : Tuple = torch.tensor(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : int = '''model.blocks.%d.self_attn.self_attn.v_proj.weight''' % player lowerCAmelCase__ : Optional[Any] = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.endswith('''/o/kernel''' ): lowerCAmelCase__ : Dict = '''model.blocks.%d.self_attn.self_attn.out_proj.weight''' % player lowerCAmelCase__ : Union[str, Any] = ( vnp.reshape([vnp.shape[0] * vnp.shape[1], vnp.shape[2]] ).transpose([1, 0] ).copy() ) # Mesh-Tensorflow is a diagonal matrix lowerCAmelCase__ : Union[str, Any] = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.startswith('''model/an''' ): lowerCAmelCase__ : Union[str, Any] = int(key_name[8:].split('''/''' )[0] ) if key_name.endswith('''/b''' ): lowerCAmelCase__ : List[Any] = '''model.blocks.%d.self_attn.norm.bias''' % player lowerCAmelCase__ : int = vnp.copy() # same because it is one dimensional lowerCAmelCase__ : Dict = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.endswith('''/g''' ): lowerCAmelCase__ : List[Any] = '''model.blocks.%d.self_attn.norm.weight''' % player lowerCAmelCase__ : Dict = vnp.copy() # same because it is one dimensional lowerCAmelCase__ : int = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif ( key_name.startswith('''model/wte''' ) or key_name.startswith('''model/wpe''' ) or key_name.startswith('''model/ete''' ) ): lowerCAmelCase__ : Optional[Any] = {'''wte''': '''embed_tokens''', '''wpe''': '''position_embeddings''', '''ete''': '''extra_position_embeddings'''}[ key_name[-3:] ] lowerCAmelCase__ : Union[str, Any] = '''model.%s.weight''' % nlayer lowerCAmelCase__ : Tuple = vnp.copy() # same in embedded lowerCAmelCase__ : Optional[Any] = torch.tensor(SCREAMING_SNAKE_CASE_ ) if key_name.startswith('''model/wte''' ): lowerCAmelCase__ : Tuple = '''lm_head.weight''' lowerCAmelCase__ : Optional[Any] = vnp.copy() # same in embedded lowerCAmelCase__ : Tuple = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name.startswith('''model/wob''' ): lowerCAmelCase__ : Any = '''final_logits_bias''' lowerCAmelCase__ : Any = vnp.copy() # same in embedded lowerCAmelCase__ : Optional[int] = state.reshape((1, -1) ) lowerCAmelCase__ : List[Any] = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name == "model/dense/kernel": lowerCAmelCase__ : int = '''model.last_project.weight''' lowerCAmelCase__ : str = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix lowerCAmelCase__ : Optional[Any] = torch.tensor(SCREAMING_SNAKE_CASE_ ) elif key_name == "model/dense_1/bias": lowerCAmelCase__ : int = '''model.last_project.bias''' lowerCAmelCase__ : List[str] = vnp.copy() # same because it is one dimensional lowerCAmelCase__ : Optional[Any] = torch.tensor(SCREAMING_SNAKE_CASE_ ) torch.save(SCREAMING_SNAKE_CASE_ , args.output ) if __name__ == "__main__": __UpperCamelCase : str = argparse.ArgumentParser( description='''model converter.''', formatter_class=argparse.ArgumentDefaultsHelpFormatter ) parser.add_argument('''--tf_model_dir''', metavar='''PATH''', type=str, required=True, help='''import model''') parser.add_argument('''--output''', metavar='''PATH''', type=str, required=True, help='''output model''') __UpperCamelCase : Dict = parser.parse_args() convert_tf_gptsan_to_pt(args)
106
"""simple docstring""" from __future__ import annotations import math import random from collections.abc import Collection from typing import overload class lowerCAmelCase : '''simple docstring''' def __init__( self , lowerCAmelCase__ = None ) -> None: if components is None: SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = list(lowerCAmelCase__ ) def __len__( self ) -> int: return len(self.__components ) def __str__( self ) -> str: return "(" + ",".join(map(lowerCAmelCase__ , self.__components ) ) + ")" def __add__( self , lowerCAmelCase__ ) -> Vector: SCREAMING_SNAKE_CASE = len(self ) if size == len(lowerCAmelCase__ ): SCREAMING_SNAKE_CASE = [self.__components[i] + other.component(lowerCAmelCase__ ) for i in range(lowerCAmelCase__ )] return Vector(lowerCAmelCase__ ) else: raise Exception('must have the same size' ) def __sub__( self , lowerCAmelCase__ ) -> Vector: SCREAMING_SNAKE_CASE = len(self ) if size == len(lowerCAmelCase__ ): SCREAMING_SNAKE_CASE = [self.__components[i] - other.component(lowerCAmelCase__ ) for i in range(lowerCAmelCase__ )] return Vector(lowerCAmelCase__ ) else: # error case raise Exception('must have the same size' ) @overload def __mul__( self , lowerCAmelCase__ ) -> Vector: ... @overload def __mul__( self , lowerCAmelCase__ ) -> float: ... def __mul__( self , lowerCAmelCase__ ) -> float | Vector: if isinstance(lowerCAmelCase__ , (float, int) ): SCREAMING_SNAKE_CASE = [c * other for c in self.__components] return Vector(lowerCAmelCase__ ) elif isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) and len(self ) == len(lowerCAmelCase__ ): SCREAMING_SNAKE_CASE = len(self ) SCREAMING_SNAKE_CASE = [self.__components[i] * other.component(lowerCAmelCase__ ) for i in range(lowerCAmelCase__ )] return sum(lowerCAmelCase__ ) else: # error case raise Exception('invalid operand!' ) def __A ( self ) -> Vector: return Vector(self.__components ) def __A ( self , lowerCAmelCase__ ) -> float: if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) and -len(self.__components ) <= i < len(self.__components ): return self.__components[i] else: raise Exception('index out of range' ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: assert -len(self.__components ) <= pos < len(self.__components ) SCREAMING_SNAKE_CASE = value def __A ( self ) -> float: if len(self.__components ) == 0: raise Exception('Vector is empty' ) SCREAMING_SNAKE_CASE = [c**2 for c in self.__components] return math.sqrt(sum(lowerCAmelCase__ ) ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ = False ) -> float: SCREAMING_SNAKE_CASE = self * other SCREAMING_SNAKE_CASE = self.euclidean_length() * other.euclidean_length() if deg: return math.degrees(math.acos(num / den ) ) else: return math.acos(num / den ) def lowercase (SCREAMING_SNAKE_CASE_ : int ) -> Vector: assert isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return Vector([0] * dimension ) def lowercase (SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int ) -> Vector: assert isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) and (isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )) SCREAMING_SNAKE_CASE = [0] * dimension SCREAMING_SNAKE_CASE = 1 return Vector(SCREAMING_SNAKE_CASE_ ) def lowercase (SCREAMING_SNAKE_CASE_ : float , SCREAMING_SNAKE_CASE_ : Vector , SCREAMING_SNAKE_CASE_ : Vector ) -> Vector: assert ( isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) and isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) and (isinstance(SCREAMING_SNAKE_CASE_ , (int, float) )) ) return x * scalar + y def lowercase (SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int ) -> Vector: random.seed(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = [random.randint(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for _ in range(SCREAMING_SNAKE_CASE_ )] return Vector(SCREAMING_SNAKE_CASE_ ) class lowerCAmelCase : '''simple docstring''' def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: SCREAMING_SNAKE_CASE = matrix SCREAMING_SNAKE_CASE = w SCREAMING_SNAKE_CASE = h def __str__( self ) -> str: SCREAMING_SNAKE_CASE = '' for i in range(self.__height ): ans += "|" for j in range(self.__width ): if j < self.__width - 1: ans += str(self.__matrix[i][j] ) + "," else: ans += str(self.__matrix[i][j] ) + "|\n" return ans def __add__( self , lowerCAmelCase__ ) -> Matrix: if self.__width == other.width() and self.__height == other.height(): SCREAMING_SNAKE_CASE = [] for i in range(self.__height ): SCREAMING_SNAKE_CASE = [ self.__matrix[i][j] + other.component(lowerCAmelCase__ , lowerCAmelCase__ ) for j in range(self.__width ) ] matrix.append(lowerCAmelCase__ ) return Matrix(lowerCAmelCase__ , self.__width , self.__height ) else: raise Exception('matrix must have the same dimension!' ) def __sub__( self , lowerCAmelCase__ ) -> Matrix: if self.__width == other.width() and self.__height == other.height(): SCREAMING_SNAKE_CASE = [] for i in range(self.__height ): SCREAMING_SNAKE_CASE = [ self.__matrix[i][j] - other.component(lowerCAmelCase__ , lowerCAmelCase__ ) for j in range(self.__width ) ] matrix.append(lowerCAmelCase__ ) return Matrix(lowerCAmelCase__ , self.__width , self.__height ) else: raise Exception('matrices must have the same dimension!' ) @overload def __mul__( self , lowerCAmelCase__ ) -> Matrix: ... @overload def __mul__( self , lowerCAmelCase__ ) -> Vector: ... def __mul__( self , lowerCAmelCase__ ) -> Vector | Matrix: if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): # matrix-vector if len(lowerCAmelCase__ ) == self.__width: SCREAMING_SNAKE_CASE = zero_vector(self.__height ) for i in range(self.__height ): SCREAMING_SNAKE_CASE = [ self.__matrix[i][j] * other.component(lowerCAmelCase__ ) for j in range(self.__width ) ] ans.change_component(lowerCAmelCase__ , sum(lowerCAmelCase__ ) ) return ans else: raise Exception( 'vector must have the same size as the ' 'number of columns of the matrix!' ) elif isinstance(lowerCAmelCase__ , (int, float) ): # matrix-scalar SCREAMING_SNAKE_CASE = [ [self.__matrix[i][j] * other for j in range(self.__width )] for i in range(self.__height ) ] return Matrix(lowerCAmelCase__ , self.__width , self.__height ) return None def __A ( self ) -> int: return self.__height def __A ( self ) -> int: return self.__width def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> float: if 0 <= x < self.__height and 0 <= y < self.__width: return self.__matrix[x][y] else: raise Exception('change_component: indices out of bounds' ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: if 0 <= x < self.__height and 0 <= y < self.__width: SCREAMING_SNAKE_CASE = value else: raise Exception('change_component: indices out of bounds' ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> float: if self.__height != self.__width: raise Exception('Matrix is not square' ) SCREAMING_SNAKE_CASE = self.__matrix[:x] + self.__matrix[x + 1 :] for i in range(len(lowerCAmelCase__ ) ): SCREAMING_SNAKE_CASE = minor[i][:y] + minor[i][y + 1 :] return Matrix(lowerCAmelCase__ , self.__width - 1 , self.__height - 1 ).determinant() def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> float: if self.__height != self.__width: raise Exception('Matrix is not square' ) if 0 <= x < self.__height and 0 <= y < self.__width: return (-1) ** (x + y) * self.minor(lowerCAmelCase__ , lowerCAmelCase__ ) else: raise Exception('Indices out of bounds' ) def __A ( self ) -> float: if self.__height != self.__width: raise Exception('Matrix is not square' ) if self.__height < 1: raise Exception('Matrix has no element' ) elif self.__height == 1: return self.__matrix[0][0] elif self.__height == 2: return ( self.__matrix[0][0] * self.__matrix[1][1] - self.__matrix[0][1] * self.__matrix[1][0] ) else: SCREAMING_SNAKE_CASE = [ self.__matrix[0][y] * self.cofactor(0 , lowerCAmelCase__ ) for y in range(self.__width ) ] return sum(lowerCAmelCase__ ) def lowercase (SCREAMING_SNAKE_CASE_ : int ) -> Matrix: SCREAMING_SNAKE_CASE = [[0] * n for _ in range(SCREAMING_SNAKE_CASE_ )] return Matrix(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def lowercase (SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int ) -> Matrix: random.seed(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = [ [random.randint(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for _ in range(SCREAMING_SNAKE_CASE_ )] for _ in range(SCREAMING_SNAKE_CASE_ ) ] return Matrix(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )
113
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCAmelCase = logging.get_logger(__name__) _UpperCAmelCase = { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/config.json', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/config.json', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/config.json', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/config.json', 'roberta-base-openai-detector': 'https://huggingface.co/roberta-base-openai-detector/resolve/main/config.json', 'roberta-large-openai-detector': 'https://huggingface.co/roberta-large-openai-detector/resolve/main/config.json', } class snake_case_ ( __lowercase ): A_ = 'roberta' def __init__( self : Union[str, Any] , _snake_case : List[Any]=50265 , _snake_case : Dict=768 , _snake_case : Dict=12 , _snake_case : Optional[Any]=12 , _snake_case : Optional[int]=3072 , _snake_case : Dict="gelu" , _snake_case : Optional[Any]=0.1 , _snake_case : List[str]=0.1 , _snake_case : int=512 , _snake_case : Tuple=2 , _snake_case : Dict=0.02 , _snake_case : Optional[Any]=1E-12 , _snake_case : Dict=1 , _snake_case : Optional[int]=0 , _snake_case : Dict=2 , _snake_case : Any="absolute" , _snake_case : Any=True , _snake_case : int=None , **_snake_case : Optional[int] , )->Union[str, Any]: '''simple docstring''' super().__init__(pad_token_id=_snake_case , bos_token_id=_snake_case , eos_token_id=_snake_case , **_snake_case ) __lowerCAmelCase : Optional[Any] = vocab_size __lowerCAmelCase : Dict = hidden_size __lowerCAmelCase : str = num_hidden_layers __lowerCAmelCase : int = num_attention_heads __lowerCAmelCase : Optional[Any] = hidden_act __lowerCAmelCase : Dict = intermediate_size __lowerCAmelCase : Optional[int] = hidden_dropout_prob __lowerCAmelCase : Union[str, Any] = attention_probs_dropout_prob __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : List[Any] = type_vocab_size __lowerCAmelCase : List[Any] = initializer_range __lowerCAmelCase : Any = layer_norm_eps __lowerCAmelCase : Any = position_embedding_type __lowerCAmelCase : int = use_cache __lowerCAmelCase : List[str] = classifier_dropout class snake_case_ ( __lowercase ): @property def UpperCAmelCase__ ( self : Optional[Any] )->Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": __lowerCAmelCase : Optional[Any] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: __lowerCAmelCase : str = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
232
import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _UpperCAmelCase = logging.get_logger(__name__) _UpperCAmelCase = {'vocab_file': 'vocab.txt', 'emoji_file': 'emoji.json'} _UpperCAmelCase = { 'vocab_file': { 'abeja/gpt-neox-japanese-2.7b': 'https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt', }, 'emoji_file': { 'abeja/gpt-neox-japanese-2.7b': 'https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json', }, } _UpperCAmelCase = { 'abeja/gpt-neox-japanese-2.7b': 2048, } def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :int , SCREAMING_SNAKE_CASE :Optional[int] ) -> Optional[Any]: with open(SCREAMING_SNAKE_CASE , """r""" , encoding="""utf-8""" ) as f: __lowerCAmelCase : int = json.loads(f.read() ) __lowerCAmelCase : Dict = collections.OrderedDict() __lowerCAmelCase : str = collections.OrderedDict() __lowerCAmelCase : Union[str, Any] = collections.OrderedDict() with open(SCREAMING_SNAKE_CASE , """r""" , encoding="""utf-8""" ) as f: __lowerCAmelCase : Tuple = f.readlines() __lowerCAmelCase : Tuple = [[t.rstrip("""\n""" )] if (t == """,""" or """,""" not in t) else t.rstrip("""\n""" ).split(""",""" ) for t in token] for idx, b in enumerate(SCREAMING_SNAKE_CASE ): __lowerCAmelCase : Dict = b __lowerCAmelCase : Dict = idx for wd in b: __lowerCAmelCase : List[str] = idx return vocab, raw_vocab, ids_to_tokens, emoji class snake_case_ ( __lowercase ): A_ = VOCAB_FILES_NAMES A_ = PRETRAINED_VOCAB_FILES_MAP A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ = ['input_ids', 'attention_mask'] def __init__( self : str , _snake_case : Union[str, Any] , _snake_case : Optional[Any] , _snake_case : Any="<|endoftext|>" , _snake_case : str="<|endoftext|>" , _snake_case : str="<|startoftext|>" , _snake_case : List[Any]="<|endoftext|>" , _snake_case : str=False , **_snake_case : List[Any] , )->Union[str, Any]: '''simple docstring''' super().__init__( unk_token=_snake_case , pad_token=_snake_case , bos_token=_snake_case , eos_token=_snake_case , do_clean_text=_snake_case , **_snake_case , ) if not os.path.isfile(_snake_case ): raise ValueError( F'''Can\'t find a vocabulary file at path \'{vocab_file}\'. To load the vocabulary from a Google pretrained''' """ model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`""" ) if not os.path.isfile(_snake_case ): raise ValueError( F'''Can\'t find a emoji file at path \'{emoji_file}\'. To load the emoji information from a Google''' """ pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`""" ) __lowerCAmelCase : Any = do_clean_text __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = load_vocab_and_emoji(_snake_case , _snake_case ) __lowerCAmelCase : int = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji ) @property def UpperCAmelCase__ ( self : int )->str: '''simple docstring''' return len(self.raw_vocab ) def UpperCAmelCase__ ( self : Tuple )->Any: '''simple docstring''' return dict(self.raw_vocab , **self.added_tokens_encoder ) def UpperCAmelCase__ ( self : Any , _snake_case : str )->Optional[int]: '''simple docstring''' return self.subword_tokenizer.tokenize(_snake_case , clean=self.do_clean_text ) def UpperCAmelCase__ ( self : Optional[Any] , _snake_case : Optional[Any] )->Any: '''simple docstring''' return self.vocab.get(_snake_case , self.vocab.get(self.unk_token ) ) def UpperCAmelCase__ ( self : int , _snake_case : Any )->int: '''simple docstring''' return self.subword_tokenizer.convert_id_to_token(_snake_case ) def UpperCAmelCase__ ( self : Optional[int] , _snake_case : int )->List[Any]: '''simple docstring''' __lowerCAmelCase : str = """""".join(_snake_case ).strip() return out_string def UpperCAmelCase__ ( self : List[str] , _snake_case : "Conversation" )->List[int]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(_snake_case , add_special_tokens=_snake_case ) + [self.eos_token_id] ) if len(_snake_case ) > self.model_max_length: __lowerCAmelCase : List[str] = input_ids[-self.model_max_length :] return input_ids def UpperCAmelCase__ ( self : Optional[Any] , _snake_case : str , _snake_case : Optional[str] = None )->Tuple[str]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = 0 if os.path.isdir(_snake_case ): __lowerCAmelCase : Dict = os.path.join( _snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowerCAmelCase : List[Any] = os.path.join( _snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""emoji_file"""] ) else: __lowerCAmelCase : Union[str, Any] = ( (filename_prefix + """-""" if filename_prefix else """""") + save_directory + VOCAB_FILES_NAMES["""vocab_file"""] ) __lowerCAmelCase : Dict = ( (filename_prefix + """-""" if filename_prefix else """""") + save_directory + VOCAB_FILES_NAMES["""emoji_file"""] ) with open(_snake_case , """w""" , encoding="""utf-8""" ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( F'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' """ Please check that the vocabulary is not corrupted!""" ) __lowerCAmelCase : List[str] = token_index writer.write(""",""".join(_snake_case ) + """\n""" ) index += 1 with open(_snake_case , """w""" , encoding="""utf-8""" ) as writer: json.dump(self.emoji , _snake_case ) return vocab_file, emoji_file class snake_case_ ( __lowercase ): def __init__( self : Optional[Any] , _snake_case : str , _snake_case : Union[str, Any] , _snake_case : Optional[int] )->List[Any]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = vocab # same as swe __lowerCAmelCase : str = ids_to_tokens # same as bpe __lowerCAmelCase : Dict = emoji __lowerCAmelCase : int = np.max([len(_snake_case ) for w in self.vocab.keys()] ) __lowerCAmelCase : str = re.compile(R"""(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)""" ) __lowerCAmelCase : Optional[Any] = re.compile(R"""[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*""" ) __lowerCAmelCase : Tuple = re.compile(R"""[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}""" ) __lowerCAmelCase : Optional[Any] = re.compile( R"""([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*""" ) __lowerCAmelCase : Union[str, Any] = re.compile( R"""(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*""" ) __lowerCAmelCase : str = re.compile( R"""((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*""" ) __lowerCAmelCase : List[Any] = """─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿""" __lowerCAmelCase : Union[str, Any] = """▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟""" __lowerCAmelCase : str = str.maketrans({k: """<BLOCK>""" for k in keisen + blocks} ) def __len__( self : int )->int: '''simple docstring''' return len(self.ids_to_tokens ) def UpperCAmelCase__ ( self : List[str] , _snake_case : Any )->str: '''simple docstring''' __lowerCAmelCase : List[str] = self.content_repattera.sub("""<URL>""" , _snake_case ) __lowerCAmelCase : Tuple = self.content_repattera.sub("""<EMAIL>""" , _snake_case ) __lowerCAmelCase : Optional[Any] = self.content_repattera.sub("""<TEL>""" , _snake_case ) __lowerCAmelCase : str = self.content_repattera.sub("""<DATE>""" , _snake_case ) __lowerCAmelCase : Tuple = self.content_repattera.sub("""<DATE>""" , _snake_case ) __lowerCAmelCase : Tuple = self.content_repattera.sub("""<PRICE>""" , _snake_case ) __lowerCAmelCase : List[Any] = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: __lowerCAmelCase : str = content.replace("""<BLOCK><BLOCK>""" , """<BLOCK>""" ) return content def UpperCAmelCase__ ( self : str , _snake_case : List[Any] , _snake_case : Optional[int]=False )->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = text.replace(""" """ , """<SP>""" ) __lowerCAmelCase : Optional[int] = text.replace(""" """ , """<SP>""" ) __lowerCAmelCase : Union[str, Any] = text.replace("""\r\n""" , """<BR>""" ) __lowerCAmelCase : Tuple = text.replace("""\n""" , """<BR>""" ) __lowerCAmelCase : List[str] = text.replace("""\r""" , """<BR>""" ) __lowerCAmelCase : Dict = text.replace("""\t""" , """<TAB>""" ) __lowerCAmelCase : Dict = text.replace("""—""" , """ー""" ) __lowerCAmelCase : Tuple = text.replace("""−""" , """ー""" ) for k, v in self.emoji["emoji"].items(): if k in text: __lowerCAmelCase : Optional[Any] = text.replace(_snake_case , _snake_case ) if clean: __lowerCAmelCase : List[Any] = self.clean_text(_snake_case ) def check_simbol(_snake_case : List[str] ): __lowerCAmelCase : Optional[int] = x.encode() if len(_snake_case ) == 1 and len(_snake_case ) == 2: __lowerCAmelCase : Optional[Any] = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0xc2a1 and c <= 0xc2bf) or (c >= 0xc780 and c <= 0xc783) or (c >= 0xcab9 and c <= 0xcbbf) or (c >= 0xcc80 and c <= 0xcda2) ): return True return False def checkuae(_snake_case : Union[str, Any] ): __lowerCAmelCase : Dict = x.encode() if len(_snake_case ) == 1 and len(_snake_case ) == 3: __lowerCAmelCase : List[str] = (int(e[0] ) << 16) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0xe2_8080 and c <= 0xe2_b07f: return True return False __lowerCAmelCase : Dict = 0 __lowerCAmelCase : Dict = [] while pos < len(_snake_case ): __lowerCAmelCase : str = min(len(_snake_case ) , pos + self.maxlen + 1 ) if text[pos] == """<""" else pos + 3 __lowerCAmelCase : Tuple = [] # (token_id, token, pos) for e in range(_snake_case , _snake_case , -1 ): __lowerCAmelCase : Optional[int] = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(_snake_case ) > 2: __lowerCAmelCase : Tuple = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(_snake_case ) > 0: # the smallest token_id is adopted __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : int = sorted(_snake_case , key=lambda _snake_case : x[0] )[0] result.append(_snake_case ) __lowerCAmelCase : int = e else: __lowerCAmelCase : Dict = pos + 1 __lowerCAmelCase : Dict = text[pos:end] if check_simbol(_snake_case ): result.append("""<KIGOU>""" ) elif checkuae(_snake_case ): result.append("""<U2000U2BFF>""" ) else: for i in wd.encode("""utf-8""" ): result.append("""<|byte%d|>""" % i ) __lowerCAmelCase : int = end return result def UpperCAmelCase__ ( self : List[str] , _snake_case : Optional[int] , _snake_case : List[Any]="\n" )->List[Any]: '''simple docstring''' __lowerCAmelCase : List[str] = [] __lowerCAmelCase : Union[str, Any] = [] __lowerCAmelCase : Optional[Any] = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(_snake_case ) > 0: words.append(bytearray(_snake_case ).decode("""utf-8""" , errors="""replace""" ) ) __lowerCAmelCase : Optional[Any] = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji["""emoji_inv"""][word] ) elif word == "<SP>": words.append(""" """ ) elif word == "<BR>": words.append(_snake_case ) elif word == "<TAB>": words.append("""\t""" ) elif word == "<BLOCK>": words.append("""▀""" ) elif word == "<KIGOU>": words.append("""ǀ""" ) elif word == "<U2000U2BFF>": words.append("""‖""" ) else: words.append(_snake_case ) if len(_snake_case ) > 0: words.append(bytearray(_snake_case ).decode("""utf-8""" , errors="""replace""" ) ) __lowerCAmelCase : Dict = """""".join(_snake_case ) return text
232
1
lowercase : Optional[int] = [ (1000, """M"""), (900, """CM"""), (500, """D"""), (400, """CD"""), (100, """C"""), (90, """XC"""), (50, """L"""), (40, """XL"""), (10, """X"""), (9, """IX"""), (5, """V"""), (4, """IV"""), (1, """I"""), ] def _snake_case( SCREAMING_SNAKE_CASE__ ) -> int: lowercase : str = {"""I""": 1, """V""": 5, """X""": 10, """L""": 50, """C""": 100, """D""": 500, """M""": 1_000} lowercase : List[Any] = 0 lowercase : str = 0 while place < len(SCREAMING_SNAKE_CASE__ ): if (place + 1 < len(SCREAMING_SNAKE_CASE__ )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def _snake_case( SCREAMING_SNAKE_CASE__ ) -> str: lowercase : str = [] for arabic, roman in ROMAN: ((lowercase) , (lowercase)) : Any = divmod(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) result.append(roman * factor ) if number == 0: break return "".join(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": import doctest doctest.testmod()
20
from __future__ import annotations from collections import namedtuple from dataclasses import dataclass @dataclass class __snake_case : _a : int _a : TreeNode | None= None _a : TreeNode | None= None lowercase : Dict = namedtuple("""CoinsDistribResult""", """moves excess""") def _snake_case( SCREAMING_SNAKE_CASE__ ) -> int: if root is None: return 0 # Validation def count_nodes(SCREAMING_SNAKE_CASE__ ) -> int: if node is None: return 0 return count_nodes(node.left ) + count_nodes(node.right ) + 1 def count_coins(SCREAMING_SNAKE_CASE__ ) -> int: if node is None: return 0 return count_coins(node.left ) + count_coins(node.right ) + node.data if count_nodes(SCREAMING_SNAKE_CASE__ ) != count_coins(SCREAMING_SNAKE_CASE__ ): raise ValueError("""The nodes number should be same as the number of coins""" ) # Main calculation def get_distrib(SCREAMING_SNAKE_CASE__ ) -> CoinsDistribResult: if node is None: return CoinsDistribResult(0 , 1 ) lowercase , lowercase : int = get_distrib(node.left ) lowercase , lowercase : List[Any] = get_distrib(node.right ) lowercase : Optional[Any] = 1 - left_distrib_excess lowercase : Union[str, Any] = 1 - right_distrib_excess lowercase : List[Any] = ( left_distrib_moves + right_distrib_moves + abs(SCREAMING_SNAKE_CASE__ ) + abs(SCREAMING_SNAKE_CASE__ ) ) lowercase : Any = node.data - coins_to_left - coins_to_right return CoinsDistribResult(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return get_distrib(SCREAMING_SNAKE_CASE__ )[0] if __name__ == "__main__": import doctest doctest.testmod()
20
1
a__ = { 0: '''0''', 1: '''1''', 2: '''2''', 3: '''3''', 4: '''4''', 5: '''5''', 6: '''6''', 7: '''7''', 8: '''8''', 9: '''9''', 10: '''a''', 11: '''b''', 12: '''c''', 13: '''d''', 14: '''e''', 15: '''f''', } def __UpperCAmelCase ( __a : float ) -> str: """simple docstring""" assert type(__a ) in (int, float) and decimal == int(__a ) _a : List[Any] = int(__a ) _a : int = '''''' _a : Dict = False if decimal < 0: _a : Any = True decimal *= -1 while decimal > 0: _a , _a : Union[str, Any] = divmod(__a ,16 ) _a : Optional[Any] = values[remainder] + hexadecimal _a : int = '''0x''' + hexadecimal if negative: _a : int = '''-''' + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
15
import argparse import os import torch from transformers.utils import WEIGHTS_NAME a__ = ['''small''', '''medium''', '''large'''] a__ = '''lm_head.decoder.weight''' a__ = '''lm_head.weight''' def __UpperCAmelCase ( __a : str ,__a : str ) -> List[str]: """simple docstring""" _a : Any = torch.load(__a ) _a : List[str] = d.pop(__a ) os.makedirs(__a ,exist_ok=__a ) torch.save(__a ,os.path.join(__a ,__a ) ) if __name__ == "__main__": a__ = argparse.ArgumentParser() parser.add_argument('''--dialogpt_path''', default='''.''', type=str) a__ = parser.parse_args() for MODEL in DIALOGPT_MODELS: a__ = os.path.join(args.dialogpt_path, f'''{MODEL}_ft.pkl''') a__ = f'''./DialoGPT-{MODEL}''' convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
15
1
def UpperCAmelCase_ ( __snake_case ) -> int: """simple docstring""" if not grid or not grid[0]: raise TypeError('''The grid does not contain the appropriate information''' ) for cell_n in range(1 , len(grid[0] ) ): grid[0][cell_n] += grid[0][cell_n - 1] _lowercase =grid[0] for row_n in range(1 , len(__snake_case ) ): _lowercase =grid[row_n] _lowercase =fill_row(__snake_case , __snake_case ) _lowercase =grid[row_n] return grid[-1][-1] def UpperCAmelCase_ ( __snake_case , __snake_case ) -> list: """simple docstring""" current_row[0] += row_above[0] for cell_n in range(1 , len(__snake_case ) ): current_row[cell_n] += min(current_row[cell_n - 1] , row_above[cell_n] ) return current_row if __name__ == "__main__": import doctest doctest.testmod()
5
from __future__ import annotations from collections.abc import Callable UpperCAmelCase__ = list[list[float | int]] def UpperCAmelCase_ ( __snake_case , __snake_case ) -> Matrix: """simple docstring""" _lowercase =len(__snake_case ) _lowercase =[[0 for _ in range(size + 1 )] for _ in range(__snake_case )] _lowercase =42 _lowercase =42 _lowercase =42 _lowercase =42 _lowercase =42 _lowercase =42 for row in range(__snake_case ): for col in range(__snake_case ): _lowercase =matrix[row][col] _lowercase =vector[row][0] _lowercase =0 _lowercase =0 while row < size and col < size: # pivoting _lowercase =max((abs(augmented[rowa][col] ), rowa) for rowa in range(__snake_case , __snake_case ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: _lowercase , _lowercase =augmented[pivot_row], augmented[row] for rowa in range(row + 1 , __snake_case ): _lowercase =augmented[rowa][col] / augmented[row][col] _lowercase =0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , __snake_case ): for row in range(__snake_case ): _lowercase =augmented[row][col] / augmented[col][col] for cola in range(__snake_case , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(__snake_case ) ] def UpperCAmelCase_ ( __snake_case ) -> Callable[[int], int]: """simple docstring""" _lowercase =len(__snake_case ) _lowercase =[[0 for _ in range(__snake_case )] for _ in range(__snake_case )] _lowercase =[[0] for _ in range(__snake_case )] _lowercase =42 _lowercase =42 _lowercase =42 _lowercase =42 for x_val, y_val in enumerate(__snake_case ): for col in range(__snake_case ): _lowercase =(x_val + 1) ** (size - col - 1) _lowercase =y_val _lowercase =solve(__snake_case , __snake_case ) def interpolated_func(__snake_case ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(__snake_case ) ) return interpolated_func def UpperCAmelCase_ ( __snake_case ) -> int: """simple docstring""" return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def UpperCAmelCase_ ( __snake_case = question_function , __snake_case = 10 ) -> int: """simple docstring""" _lowercase =[func(__snake_case ) for x_val in range(1 , order + 1 )] _lowercase =[ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] _lowercase =0 _lowercase =42 _lowercase =42 for poly in polynomials: _lowercase =1 while func(__snake_case ) == poly(__snake_case ): x_val += 1 ret += poly(__snake_case ) return ret if __name__ == "__main__": print(f'''{solution() = }''')
5
1
'''simple docstring''' from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def _A (lowerCAmelCase__ :str = "laptop" ) -> DataFrame: '''simple docstring''' _a = f'https://www.amazon.in/laptop/s?k={product}' _a = { 'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36', 'Accept-Language': 'en-US, en;q=0.5', } _a = BeautifulSoup(requests.get(lowerCAmelCase__ , headers=lowerCAmelCase__ ).text ) # Initialize a Pandas dataframe with the column titles _a = DataFrame( columns=[ 'Product Title', 'Product Link', 'Current Price of the product', 'Product Rating', 'MRP of the product', 'Discount', ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( 'div' , attrs={'class': 's-result-item', 'data-component-type': 's-search-result'} , ) , soup.find_all('div' , attrs={'class': 'a-row a-size-base a-color-base'} ) , ): try: _a = item.ha.text _a = 'https://www.amazon.in/' + item.ha.a['href'] _a = item.find('span' , attrs={'class': 'a-offscreen'} ).text try: _a = item.find('span' , attrs={'class': 'a-icon-alt'} ).text except AttributeError: _a = 'Not available' try: _a = ( '₹' + item.find( 'span' , attrs={'class': 'a-price a-text-price'} ).text.split('₹' )[1] ) except AttributeError: _a = '' try: _a = float( ( ( float(product_mrp.strip('₹' ).replace(',' , '' ) ) - float(product_price.strip('₹' ).replace(',' , '' ) ) ) / float(product_mrp.strip('₹' ).replace(',' , '' ) ) ) * 1_00 ) except ValueError: _a = float('nan' ) except AttributeError: pass _a = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] _a = ' ' _a = ' ' data_frame.index += 1 return data_frame if __name__ == "__main__": a_ : int = "headphones" get_amazon_product_data(product).to_csv(f'''Amazon Product Data for {product}.csv''')
104
'''simple docstring''' import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, ) from transformers.models.deberta.modeling_deberta import DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST class a ( _SCREAMING_SNAKE_CASE ): def __init__( self , __magic_name__ , __magic_name__=13 , __magic_name__=7 , __magic_name__=True , __magic_name__=True , __magic_name__=True , __magic_name__=True , __magic_name__=99 , __magic_name__=32 , __magic_name__=5 , __magic_name__=4 , __magic_name__=37 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=5_12 , __magic_name__=16 , __magic_name__=2 , __magic_name__=0.0_2 , __magic_name__=False , __magic_name__=True , __magic_name__="None" , __magic_name__=3 , __magic_name__=4 , __magic_name__=None , ) -> Any: _a = parent _a = batch_size _a = seq_length _a = is_training _a = use_input_mask _a = use_token_type_ids _a = use_labels _a = vocab_size _a = hidden_size _a = num_hidden_layers _a = num_attention_heads _a = intermediate_size _a = hidden_act _a = hidden_dropout_prob _a = attention_probs_dropout_prob _a = max_position_embeddings _a = type_vocab_size _a = type_sequence_label_size _a = initializer_range _a = num_labels _a = num_choices _a = relative_attention _a = position_biased_input _a = pos_att_type _a = scope def __UpperCAmelCase ( self ) -> List[str]: _a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _a = None if self.use_input_mask: _a = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _a = None if self.use_token_type_ids: _a = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _a = None _a = None _a = None if self.use_labels: _a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _a = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _a = ids_tensor([self.batch_size] , self.num_choices ) _a = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCAmelCase ( self ) -> Union[str, Any]: return DebertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def __UpperCAmelCase ( self ) -> Optional[Any]: _a = self.get_config() _a = 3_00 return config def __UpperCAmelCase ( self , __magic_name__ ) -> Dict: self.parent.assertListEqual(list(result.loss.size() ) , [] ) def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Dict: _a = DebertaModel(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() _a = model(__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ )[0] _a = model(__magic_name__ , token_type_ids=__magic_name__ )[0] _a = model(__magic_name__ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Dict: _a = DebertaForMaskedLM(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() _a = model(__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Optional[int]: _a = self.num_labels _a = DebertaForSequenceClassification(__magic_name__ ) model.to(__magic_name__ ) model.eval() _a = model(__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(__magic_name__ ) def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Dict: _a = self.num_labels _a = DebertaForTokenClassification(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() _a = model(__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Optional[int]: _a = DebertaForQuestionAnswering(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() _a = model( __magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , start_positions=__magic_name__ , end_positions=__magic_name__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCAmelCase ( self ) -> Any: _a = self.prepare_config_and_inputs() ( ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ) = config_and_inputs _a = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , unittest.TestCase ): _lowerCAmelCase = ( ( DebertaModel, DebertaForMaskedLM, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaForQuestionAnswering, ) if is_torch_available() else () ) _lowerCAmelCase = ( { """feature-extraction""": DebertaModel, """fill-mask""": DebertaForMaskedLM, """question-answering""": DebertaForQuestionAnswering, """text-classification""": DebertaForSequenceClassification, """token-classification""": DebertaForTokenClassification, """zero-shot""": DebertaForSequenceClassification, } if is_torch_available() else {} ) _lowerCAmelCase = True _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False _lowerCAmelCase = False def __UpperCAmelCase ( self ) -> List[str]: _a = DebertaModelTester(self ) _a = ConfigTester(self , config_class=__magic_name__ , hidden_size=37 ) def __UpperCAmelCase ( self ) -> Tuple: self.config_tester.run_common_tests() def __UpperCAmelCase ( self ) -> Any: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*__magic_name__ ) def __UpperCAmelCase ( self ) -> str: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*__magic_name__ ) def __UpperCAmelCase ( self ) -> Dict: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*__magic_name__ ) def __UpperCAmelCase ( self ) -> Optional[int]: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*__magic_name__ ) def __UpperCAmelCase ( self ) -> Any: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*__magic_name__ ) @slow def __UpperCAmelCase ( self ) -> Optional[Any]: for model_name in DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a = DebertaModel.from_pretrained(__magic_name__ ) self.assertIsNotNone(__magic_name__ ) @require_torch @require_sentencepiece @require_tokenizers class a ( unittest.TestCase ): @unittest.skip(reason='Model not available yet' ) def __UpperCAmelCase ( self ) -> Dict: pass @slow def __UpperCAmelCase ( self ) -> int: _a = DebertaModel.from_pretrained('microsoft/deberta-base' ) _a = torch.tensor([[0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2]] ) _a = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _a = model(__magic_name__ , attention_mask=__magic_name__ )[0] # compare the actual values for a slice. _a = torch.tensor( [[[-0.5_9_8_6, -0.8_0_5_5, -0.8_4_6_2], [1.4_4_8_4, -0.9_3_4_8, -0.8_0_5_9], [0.3_1_2_3, 0.0_0_3_2, -1.4_1_3_1]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __magic_name__ , atol=1e-4 ) , f'{output[:, 1:4, 1:4]}' )
104
1
'''simple docstring''' import copy import tempfile import unittest from transformers import MaMaaaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.utils import cached_property from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaTokenizer from transformers.models.mam_aaa.modeling_mam_aaa import MaMaaaDecoder, MaMaaaEncoder def _a( UpperCamelCase__ : str, UpperCamelCase__ : Union[str, Any], UpperCamelCase__ : Dict, UpperCamelCase__ : int=None, UpperCamelCase__ : List[Any]=None, UpperCamelCase__ : Optional[int]=None, UpperCamelCase__ : Union[str, Any]=None, UpperCamelCase__ : Any=None, ): '''simple docstring''' if attention_mask is None: SCREAMING_SNAKE_CASE__ : Dict =input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: SCREAMING_SNAKE_CASE__ : str =decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: SCREAMING_SNAKE_CASE__ : Union[str, Any] =torch.ones(config.encoder_layers, config.encoder_attention_heads, device=lowerCamelCase__ ) if decoder_head_mask is None: SCREAMING_SNAKE_CASE__ : Union[str, Any] =torch.ones(config.decoder_layers, config.decoder_attention_heads, device=lowerCamelCase__ ) if cross_attn_head_mask is None: SCREAMING_SNAKE_CASE__ : Optional[int] =torch.ones(config.decoder_layers, config.decoder_attention_heads, device=lowerCamelCase__ ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } class __SCREAMING_SNAKE_CASE : def __init__( self : Tuple , __lowercase : str , __lowercase : List[Any]=13 , __lowercase : Union[str, Any]=7 , __lowercase : Any=True , __lowercase : str=False , __lowercase : Optional[Any]=99 , __lowercase : Optional[int]=16 , __lowercase : Union[str, Any]=2 , __lowercase : Tuple=4 , __lowercase : int=4 , __lowercase : Dict="relu" , __lowercase : Optional[int]=0.1 , __lowercase : List[Any]=0.1 , __lowercase : str=0.0 , __lowercase : Dict=0.0 , __lowercase : Any=20 , __lowercase : List[str]=2 , __lowercase : str=1 , __lowercase : Dict=0 , ) -> int: SCREAMING_SNAKE_CASE__ : Optional[Any] =parent SCREAMING_SNAKE_CASE__ : int =batch_size SCREAMING_SNAKE_CASE__ : Optional[Any] =seq_length SCREAMING_SNAKE_CASE__ : List[str] =is_training SCREAMING_SNAKE_CASE__ : str =use_labels SCREAMING_SNAKE_CASE__ : int =vocab_size SCREAMING_SNAKE_CASE__ : Optional[int] =hidden_size SCREAMING_SNAKE_CASE__ : str =num_hidden_layers SCREAMING_SNAKE_CASE__ : Tuple =num_attention_heads SCREAMING_SNAKE_CASE__ : Dict =intermediate_size SCREAMING_SNAKE_CASE__ : str =hidden_act SCREAMING_SNAKE_CASE__ : Optional[Any] =hidden_dropout_prob SCREAMING_SNAKE_CASE__ : Any =attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ : int =encoder_layerdrop SCREAMING_SNAKE_CASE__ : List[Any] =decoder_layerdrop SCREAMING_SNAKE_CASE__ : int =max_position_embeddings SCREAMING_SNAKE_CASE__ : Tuple =eos_token_id SCREAMING_SNAKE_CASE__ : Tuple =pad_token_id SCREAMING_SNAKE_CASE__ : List[Any] =bos_token_id def __magic_name__ ( self : Union[str, Any] ) -> Optional[int]: SCREAMING_SNAKE_CASE__ : Any =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE__ : str =self.eos_token_id # Eos Token SCREAMING_SNAKE_CASE__ : Optional[int] =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for M2M100 the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input SCREAMING_SNAKE_CASE__ : List[str] =input_ids.clamp(self.pad_token_id + 1 ) SCREAMING_SNAKE_CASE__ : Any =decoder_input_ids.clamp(self.pad_token_id + 1 ) SCREAMING_SNAKE_CASE__ : Optional[Any] =self.get_config() SCREAMING_SNAKE_CASE__ : Dict =prepare_mam_aaa_inputs_dict(__lowercase , __lowercase , __lowercase ) return config, inputs_dict def __magic_name__ ( self : Tuple ) -> Any: return MaMaaaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , encoder_layerdrop=self.encoder_layerdrop , decoder_layerdrop=self.decoder_layerdrop , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , ) def __magic_name__ ( self : Dict ) -> Tuple: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Optional[int] =self.prepare_config_and_inputs() return config, inputs_dict def __magic_name__ ( self : Tuple , __lowercase : int , __lowercase : List[str] ) -> Any: SCREAMING_SNAKE_CASE__ : Tuple =MaMaaaModel(config=__lowercase ).get_decoder().to(__lowercase ).eval() SCREAMING_SNAKE_CASE__ : str =inputs_dict['''input_ids'''] SCREAMING_SNAKE_CASE__ : List[str] =inputs_dict['''attention_mask'''] SCREAMING_SNAKE_CASE__ : str =inputs_dict['''head_mask'''] # first forward pass SCREAMING_SNAKE_CASE__ : str =model(__lowercase , attention_mask=__lowercase , head_mask=__lowercase , use_cache=__lowercase ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Optional[Any] =outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids SCREAMING_SNAKE_CASE__ : int =ids_tensor((self.batch_size, 3) , config.vocab_size ) SCREAMING_SNAKE_CASE__ : Tuple =ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and SCREAMING_SNAKE_CASE__ : Optional[int] =torch.cat([input_ids, next_tokens] , dim=-1 ) SCREAMING_SNAKE_CASE__ : Tuple =torch.cat([attention_mask, next_attn_mask] , dim=-1 ) SCREAMING_SNAKE_CASE__ : Optional[int] =model(__lowercase , attention_mask=__lowercase )['''last_hidden_state'''] SCREAMING_SNAKE_CASE__ : Tuple =model(__lowercase , attention_mask=__lowercase , past_key_values=__lowercase )[ '''last_hidden_state''' ] # select random slice SCREAMING_SNAKE_CASE__ : int =ids_tensor((1,) , output_from_past.shape[-1] ).item() SCREAMING_SNAKE_CASE__ : Dict =output_from_no_past[:, -3:, random_slice_idx].detach() SCREAMING_SNAKE_CASE__ : List[str] =output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__lowercase , __lowercase , atol=1e-2 ) ) def __magic_name__ ( self : List[Any] , __lowercase : Optional[int] , __lowercase : List[str] ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__ : List[str] =MaMaaaModel(config=__lowercase ).to(__lowercase ).eval() SCREAMING_SNAKE_CASE__ : List[str] =model(**__lowercase ) SCREAMING_SNAKE_CASE__ : Optional[int] =outputs.encoder_last_hidden_state SCREAMING_SNAKE_CASE__ : Optional[int] =outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: SCREAMING_SNAKE_CASE__ : Optional[int] =model.get_encoder() encoder.save_pretrained(__lowercase ) SCREAMING_SNAKE_CASE__ : int =MaMaaaEncoder.from_pretrained(__lowercase ).to(__lowercase ) SCREAMING_SNAKE_CASE__ : List[str] =encoder(inputs_dict['''input_ids'''] , attention_mask=inputs_dict['''attention_mask'''] )[ 0 ] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1e-3 ) with tempfile.TemporaryDirectory() as tmpdirname: SCREAMING_SNAKE_CASE__ : int =model.get_decoder() decoder.save_pretrained(__lowercase ) SCREAMING_SNAKE_CASE__ : Any =MaMaaaDecoder.from_pretrained(__lowercase ).to(__lowercase ) SCREAMING_SNAKE_CASE__ : Dict =decoder( input_ids=inputs_dict['''decoder_input_ids'''] , attention_mask=inputs_dict['''decoder_attention_mask'''] , encoder_hidden_states=__lowercase , encoder_attention_mask=inputs_dict['''attention_mask'''] , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1e-3 ) @require_torch class __SCREAMING_SNAKE_CASE ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): snake_case_ = ( ( MaMaaaModel, MaMaaaForConditionalGeneration, ) if is_torch_available() else () ) snake_case_ = (MaMaaaForConditionalGeneration,) if is_torch_available() else () snake_case_ = ( { """conversational""": MaMaaaForConditionalGeneration, """feature-extraction""": MaMaaaModel, """summarization""": MaMaaaForConditionalGeneration, """text2text-generation""": MaMaaaForConditionalGeneration, """translation""": MaMaaaForConditionalGeneration, } if is_torch_available() else {} ) snake_case_ = True snake_case_ = True snake_case_ = False snake_case_ = False def __magic_name__ ( self : Any , __lowercase : int , __lowercase : Union[str, Any] , __lowercase : Any , __lowercase : int , __lowercase : Optional[Any] ) -> int: if pipeline_test_casse_name == "TranslationPipelineTests": # Get `ValueError: Translation requires a `src_lang` and a `tgt_lang` for this model`. # `M2M100Config` was never used in pipeline tests: cannot create a simple tokenizer. return True return False def __magic_name__ ( self : Dict ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ : Optional[Any] =MaMaaaModelTester(self ) SCREAMING_SNAKE_CASE__ : List[Any] =ConfigTester(self , config_class=__lowercase ) def __magic_name__ ( self : int ) -> str: self.config_tester.run_common_tests() def __magic_name__ ( self : Optional[Any] ) -> Optional[int]: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Optional[int] =self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ : int =model_class(__lowercase ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__lowercase ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : int =model_class.from_pretrained(__lowercase , output_loading_info=__lowercase ) self.assertEqual(info['''missing_keys'''] , [] ) def __magic_name__ ( self : int ) -> int: SCREAMING_SNAKE_CASE__ : Any =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(*__lowercase ) def __magic_name__ ( self : str ) -> Any: SCREAMING_SNAKE_CASE__ : List[Any] =self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*__lowercase ) def __magic_name__ ( self : List[Any] ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Tuple =self.model_tester.prepare_config_and_inputs_for_common() for model_class in (MaMaaaModel, MaMaaaForConditionalGeneration): SCREAMING_SNAKE_CASE__ : str =model_class(__lowercase ) model.to(__lowercase ) model.eval() SCREAMING_SNAKE_CASE__ : Optional[int] =copy.deepcopy(self._prepare_for_class(__lowercase , __lowercase ) ) if not self.is_encoder_decoder: SCREAMING_SNAKE_CASE__ : Tuple =inputs['''input_ids'''] del inputs["input_ids"] else: SCREAMING_SNAKE_CASE__ : Tuple =inputs['''input_ids'''] SCREAMING_SNAKE_CASE__ : Optional[Any] =inputs.get('''decoder_input_ids''' , __lowercase ) del inputs["input_ids"] inputs.pop('''decoder_input_ids''' , __lowercase ) SCREAMING_SNAKE_CASE__ : Tuple =model.get_input_embeddings() if not self.is_encoder_decoder: SCREAMING_SNAKE_CASE__ : Dict =wte(__lowercase ) else: SCREAMING_SNAKE_CASE__ : List[Any] =wte(__lowercase ) SCREAMING_SNAKE_CASE__ : str =wte(__lowercase ) with torch.no_grad(): model(**__lowercase )[0] def __magic_name__ ( self : List[str] ) -> int: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : str =self.model_tester.prepare_config_and_inputs() SCREAMING_SNAKE_CASE__ : Optional[Any] =input_dict['''input_ids'''] SCREAMING_SNAKE_CASE__ : Tuple =input_ids.ne(1 ).to(__lowercase ) SCREAMING_SNAKE_CASE__ : Union[str, Any] =MaMaaaForConditionalGeneration(__lowercase ).eval().to(__lowercase ) if torch_device == "cuda": model.half() model.generate(__lowercase , attention_mask=__lowercase ) model.generate(num_beams=4 , do_sample=__lowercase , early_stopping=__lowercase , num_return_sequences=3 ) def _a( UpperCamelCase__ : Dict ): '''simple docstring''' return torch.tensor(lowerCamelCase__, dtype=torch.long, device=lowerCamelCase__ ) a_ = 1E-4 @require_torch @require_sentencepiece @require_tokenizers @slow class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @cached_property def __magic_name__ ( self : Dict ) -> Optional[int]: return MaMaaaTokenizer.from_pretrained('''facebook/m2m100_418M''' ) def __magic_name__ ( self : int ) -> List[Any]: SCREAMING_SNAKE_CASE__ : int =MaMaaaModel.from_pretrained('''facebook/m2m100_418M''' ).to(__lowercase ) SCREAMING_SNAKE_CASE__ : Any =_long_tensor([[12_80_28, 98, 12, 3_05_27, 27_32, 1_59, 77_55, 6_19_04, 3_91_44, 38, 2]] ) SCREAMING_SNAKE_CASE__ : List[Any] =_long_tensor([[2, 12_80_28, 98, 12, 3_05_27, 27_32, 1_59, 77_55, 6_19_04, 3_91_44, 38]] ) SCREAMING_SNAKE_CASE__ : Optional[Any] =prepare_mam_aaa_inputs_dict(model.config , __lowercase , __lowercase ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ : Any =model(**__lowercase )[0] SCREAMING_SNAKE_CASE__ : Any =torch.Size((1, 11, 10_24) ) self.assertEqual(output.shape , __lowercase ) # change to expected output here SCREAMING_SNAKE_CASE__ : List[str] =torch.tensor( [[-0.7780, -0.1676, 0.1038], [-6.7556, -1.3992, 0.0567], [-7.5383, -0.5920, -0.2779]] , device=__lowercase ) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowercase , atol=__lowercase ) ) def __magic_name__ ( self : int ) -> Optional[int]: SCREAMING_SNAKE_CASE__ : str =MaMaaaForConditionalGeneration.from_pretrained('''facebook/m2m100_418M''' ).to(__lowercase ) # change to intended input SCREAMING_SNAKE_CASE__ : Tuple =_long_tensor([[12_80_28, 98, 12, 3_05_27, 27_32, 1_59, 77_55, 6_19_04, 3_91_44, 38, 2]] ) SCREAMING_SNAKE_CASE__ : Any =_long_tensor([[2, 12_80_28, 98, 12, 3_05_27, 27_32, 1_59, 77_55, 6_19_04, 3_91_44, 38]] ) SCREAMING_SNAKE_CASE__ : int =prepare_mam_aaa_inputs_dict(model.config , __lowercase , __lowercase ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ : Any =model(**__lowercase )[0] SCREAMING_SNAKE_CASE__ : Optional[Any] =torch.Size((1, 11, model.config.vocab_size) ) self.assertEqual(output.shape , __lowercase ) # change to expected output here SCREAMING_SNAKE_CASE__ : Union[str, Any] =torch.tensor( [[-1.0448, -1.0411, 3.7992], [-3.2191, -3.2386, -1.3451], [-3.6210, -3.5993, 0.4925]] , device=__lowercase ) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowercase , atol=__lowercase ) ) def __magic_name__ ( self : int ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__ : Dict =MaMaaaForConditionalGeneration.from_pretrained('''facebook/m2m100_418M''' ).to(__lowercase ) SCREAMING_SNAKE_CASE__ : Optional[Any] =MaMaaaTokenizer.from_pretrained('''facebook/m2m100_418M''' , src_lang='''fr''' , tgt_lang='''en''' ) SCREAMING_SNAKE_CASE__ : Optional[int] =[ '''L\'affaire NSA souligne l\'absence totale de débat sur le renseignement''', '''Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.''', '''Lorsque François Hollande téléphone à Barack Obama ou quand le ministre des affaires étrangères Laurent''' ''' Fabius convoque l\'ambassadeur des Etats-Unis, ils réagissent à une vraie découverte, qui est celle de''' ''' l\'ampleur de la surveillance américaine sur l\'ensemble des communications en France.''', ] # The below article tests that we don't add any hypotheses outside of the top n_beams SCREAMING_SNAKE_CASE__ : List[str] =tokenizer(__lowercase , padding=__lowercase , return_tensors='''pt''' ) SCREAMING_SNAKE_CASE__ : List[str] =model.generate( input_ids=dct['''input_ids'''].to(__lowercase ) , attention_mask=dct['''attention_mask'''].to(__lowercase ) , num_beams=5 , forced_bos_token_id=tokenizer.get_lang_id('''en''' ) , ) SCREAMING_SNAKE_CASE__ : Optional[Any] =[ '''The NSA case highlights the total absence of intelligence debate''', '''I think there are two levels of response from the French government.''', '''When François Hollande calls Barack Obama or when Foreign Minister Laurent Fabius calls the U.S.''' ''' Ambassador, they respond to a real discovery, which is that of the scale of U.S. surveillance on all''' ''' communications in France.''', ] SCREAMING_SNAKE_CASE__ : str =tokenizer.batch_decode( hypotheses_batch.tolist() , clean_up_tokenization_spaces=__lowercase , skip_special_tokens=__lowercase ) assert generated == expected_en
152
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __A ={'''configuration_wavlm''': ['''WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''WavLMConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A =[ '''WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''WavLMForAudioFrameClassification''', '''WavLMForCTC''', '''WavLMForSequenceClassification''', '''WavLMForXVector''', '''WavLMModel''', '''WavLMPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavlm import WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP, WavLMConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavlm import ( WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST, WavLMForAudioFrameClassification, WavLMForCTC, WavLMForSequenceClassification, WavLMForXVector, WavLMModel, WavLMPreTrainedModel, ) else: import sys __A =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
19
0
"""simple docstring""" import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __A = 1_6 __A = 3_2 def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase = 1_6 ) -> Union[str, Any]: lowercase__: Union[str, Any] = AutoTokenizer.from_pretrained('''bert-base-cased''' ) lowercase__: int = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(__UpperCAmelCase ): # max_length=None => use the model max length (it's actually the default) lowercase__: List[str] = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=__UpperCAmelCase , max_length=__UpperCAmelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowercase__: int = datasets.map( __UpperCAmelCase , batched=__UpperCAmelCase , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowercase__: Any = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(__UpperCAmelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. lowercase__: Any = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowercase__: List[Any] = 1_6 elif accelerator.mixed_precision != "no": lowercase__: Optional[Any] = 8 else: lowercase__: Optional[Any] = None return tokenizer.pad( __UpperCAmelCase , padding='''longest''' , max_length=__UpperCAmelCase , pad_to_multiple_of=__UpperCAmelCase , return_tensors='''pt''' , ) # Instantiate dataloaders. lowercase__: Tuple = DataLoader( tokenized_datasets['''train'''] , shuffle=__UpperCAmelCase , collate_fn=__UpperCAmelCase , batch_size=__UpperCAmelCase ) lowercase__: Tuple = DataLoader( tokenized_datasets['''validation'''] , shuffle=__UpperCAmelCase , collate_fn=__UpperCAmelCase , batch_size=__UpperCAmelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("TESTING_MOCKED_DATALOADERS", None) == "1": from accelerate.test_utils.training import mocked_dataloaders __A = mocked_dataloaders # noqa: F811 def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase ) -> Tuple: # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , __UpperCAmelCase ) == "1": lowercase__: int = 2 # Initialize accelerator lowercase__: str = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowercase__: List[str] = config['''lr'''] lowercase__: Optional[Any] = int(config['''num_epochs'''] ) lowercase__: Optional[Any] = int(config['''seed'''] ) lowercase__: Optional[int] = int(config['''batch_size'''] ) lowercase__: List[str] = evaluate.load('''glue''' , '''mrpc''' ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=__UpperCAmelCase ) def inner_training_loop(__UpperCAmelCase ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(__UpperCAmelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowercase__: Union[str, Any] = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=__UpperCAmelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowercase__: str = model.to(accelerator.device ) # Instantiate optimizer lowercase__: str = AdamW(params=model.parameters() , lr=__UpperCAmelCase ) lowercase__, lowercase__: Tuple = get_dataloaders(__UpperCAmelCase , __UpperCAmelCase ) # Instantiate scheduler lowercase__: int = get_linear_schedule_with_warmup( optimizer=__UpperCAmelCase , num_warmup_steps=1_0_0 , num_training_steps=(len(__UpperCAmelCase ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowercase__, lowercase__, lowercase__, lowercase__, lowercase__: Any = accelerator.prepare( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) # Now we train the model for epoch in range(__UpperCAmelCase ): model.train() for step, batch in enumerate(__UpperCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) lowercase__: List[Any] = model(**__UpperCAmelCase ) lowercase__: Union[str, Any] = outputs.loss accelerator.backward(__UpperCAmelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__UpperCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowercase__: Tuple = model(**__UpperCAmelCase ) lowercase__: int = outputs.logits.argmax(dim=-1 ) lowercase__, lowercase__: Union[str, Any] = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=__UpperCAmelCase , references=__UpperCAmelCase , ) lowercase__: Union[str, Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"""epoch {epoch}:""" , __UpperCAmelCase ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def SCREAMING_SNAKE_CASE__ ( ) -> int: lowercase__: List[str] = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=__UpperCAmelCase , default=__UpperCAmelCase , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) lowercase__: int = parser.parse_args() lowercase__: str = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 4_2, '''batch_size''': 1_6} training_function(__UpperCAmelCase , __UpperCAmelCase ) if __name__ == "__main__": main()
2
"""simple docstring""" import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __A = get_tests_dir("fixtures/spiece.model") @require_sentencepiece @require_tokenizers class UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): """simple docstring""" _UpperCAmelCase :Dict = DebertaVaTokenizer _UpperCAmelCase :Tuple = DebertaVaTokenizerFast _UpperCAmelCase :int = True _UpperCAmelCase :int = True def _snake_case ( self ): super().setUp() # We have a SentencePiece fixture for testing lowercase__: List[Any] = DebertaVaTokenizer(_UpperCAmelCase , unk_token='''<unk>''' ) tokenizer.save_pretrained(self.tmpdirname ) def _snake_case ( self , _UpperCAmelCase ): lowercase__: List[str] = '''this is a test''' lowercase__: int = '''this is a test''' return input_text, output_text def _snake_case ( self ): lowercase__: Optional[int] = '''<pad>''' lowercase__: Optional[int] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_UpperCAmelCase ) , _UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_UpperCAmelCase ) , _UpperCAmelCase ) def _snake_case ( self ): lowercase__: Union[str, Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<pad>''' ) self.assertEqual(vocab_keys[1] , '''<unk>''' ) self.assertEqual(vocab_keys[-1] , '''[PAD]''' ) self.assertEqual(len(_UpperCAmelCase ) , 30001 ) def _snake_case ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 30000 ) def _snake_case ( self ): # fmt: off lowercase__: int = ''' \tHeLLo!how \n Are yoU? ''' lowercase__: List[str] = ['''▁hello''', '''!''', '''how''', '''▁are''', '''▁you''', '''?'''] # fmt: on lowercase__: Any = DebertaVaTokenizer(_UpperCAmelCase , do_lower_case=_UpperCAmelCase ) lowercase__: Union[str, Any] = tokenizer.convert_ids_to_tokens(tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Optional[Any] = DebertaVaTokenizerFast(_UpperCAmelCase , do_lower_case=_UpperCAmelCase ) lowercase__: Optional[Any] = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) @unittest.skip('''There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.''' ) def _snake_case ( self ): pass @unittest.skip('''There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.''' ) def _snake_case ( self ): pass def _snake_case ( self ): # fmt: off lowercase__: Dict = '''I was born in 92000, and this is falsé.''' lowercase__: str = ['''▁''', '''<unk>''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', '''▁''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''<unk>''', '''▁''', '''.''', ] # fmt: on lowercase__: Union[str, Any] = DebertaVaTokenizer(_UpperCAmelCase , split_by_punct=_UpperCAmelCase ) lowercase__: str = tokenizer.convert_ids_to_tokens(tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Tuple = DebertaVaTokenizerFast(_UpperCAmelCase , split_by_punct=_UpperCAmelCase ) lowercase__: Union[str, Any] = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def _snake_case ( self ): # fmt: off lowercase__: Any = '''I was born in 92000, and this is falsé.''' lowercase__: str = ['''▁i''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', '''▁''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''<unk>''', '''▁''', '''.''', ] # fmt: on lowercase__: Optional[int] = DebertaVaTokenizer(_UpperCAmelCase , do_lower_case=_UpperCAmelCase , split_by_punct=_UpperCAmelCase ) lowercase__: List[Any] = tokenizer.convert_ids_to_tokens(tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Optional[int] = DebertaVaTokenizerFast(_UpperCAmelCase , do_lower_case=_UpperCAmelCase , split_by_punct=_UpperCAmelCase ) lowercase__: Tuple = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def _snake_case ( self ): # fmt: off lowercase__: List[str] = '''I was born in 92000, and this is falsé.''' lowercase__: List[str] = ['''▁i''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''<unk>''', '''.''', ] # fmt: on lowercase__: Union[str, Any] = DebertaVaTokenizer(_UpperCAmelCase , do_lower_case=_UpperCAmelCase , split_by_punct=_UpperCAmelCase ) lowercase__: Union[str, Any] = tokenizer.convert_ids_to_tokens(tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: int = DebertaVaTokenizerFast(_UpperCAmelCase , do_lower_case=_UpperCAmelCase , split_by_punct=_UpperCAmelCase ) lowercase__: Tuple = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def _snake_case ( self ): # fmt: off lowercase__: Union[str, Any] = '''I was born in 92000, and this is falsé.''' lowercase__: int = ['''▁''', '''<unk>''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', '''▁''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''<unk>''', '''▁''', '''.''', ] # fmt: on lowercase__: Optional[int] = DebertaVaTokenizer(_UpperCAmelCase , do_lower_case=_UpperCAmelCase , split_by_punct=_UpperCAmelCase ) lowercase__: Dict = tokenizer.convert_ids_to_tokens(tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Union[str, Any] = DebertaVaTokenizerFast(_UpperCAmelCase , do_lower_case=_UpperCAmelCase , split_by_punct=_UpperCAmelCase ) lowercase__: Dict = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def _snake_case ( self ): # fmt: off lowercase__: Optional[int] = ''' \tHeLLo!how \n Are yoU? ''' lowercase__: str = ['''▁''', '''<unk>''', '''e''', '''<unk>''', '''o''', '''!''', '''how''', '''▁''', '''<unk>''', '''re''', '''▁yo''', '''<unk>''', '''?'''] # fmt: on lowercase__: Dict = DebertaVaTokenizer(_UpperCAmelCase , do_lower_case=_UpperCAmelCase , split_by_punct=_UpperCAmelCase ) lowercase__: List[str] = tokenizer.convert_ids_to_tokens(tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Optional[Any] = DebertaVaTokenizerFast(_UpperCAmelCase , do_lower_case=_UpperCAmelCase , split_by_punct=_UpperCAmelCase ) lowercase__: List[str] = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def _snake_case ( self ): lowercase__: int = self.get_tokenizer() lowercase__: List[Any] = self.get_rust_tokenizer() lowercase__: List[str] = '''I was born in 92000, and this is falsé.''' lowercase__: Any = tokenizer.convert_ids_to_tokens(tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) lowercase__: List[str] = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Dict = tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) lowercase__: Tuple = rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Any = self.get_rust_tokenizer() lowercase__: str = tokenizer.encode(_UpperCAmelCase ) lowercase__: Any = rust_tokenizer.encode(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def _snake_case ( self ): lowercase__: Optional[Any] = '''This is a test''' lowercase__: str = [13, 1, 4398, 25, 21, 1289] lowercase__: List[Any] = ['''▁''', '''T''', '''his''', '''▁is''', '''▁a''', '''▁test'''] lowercase__: Any = ['''▁''', '''<unk>''', '''his''', '''▁is''', '''▁a''', '''▁test'''] lowercase__: int = DebertaVaTokenizer(_UpperCAmelCase , keep_accents=_UpperCAmelCase ) lowercase__: int = DebertaVaTokenizerFast(_UpperCAmelCase , keep_accents=_UpperCAmelCase ) lowercase__: Any = tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: str = tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Any = tokenizer.convert_ids_to_tokens(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Union[str, Any] = rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: List[Any] = rust_tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: str = rust_tokenizer.convert_ids_to_tokens(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) # fmt: off lowercase__: str = '''I was born in 92000, and this is falsé.''' lowercase__: Dict = [13, 1, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] lowercase__: Tuple = ['''▁''', '''I''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''é''', '''.''', ] lowercase__: Dict = ['''▁''', '''<unk>''', '''▁was''', '''▁born''', '''▁in''', '''▁9''', '''2000''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁fal''', '''s''', '''<unk>''', '''.''', ] # fmt: on lowercase__: Optional[Any] = tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Dict = tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Optional[Any] = tokenizer.convert_ids_to_tokens(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: List[Any] = rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Dict = rust_tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Optional[Any] = rust_tokenizer.convert_ids_to_tokens(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def _snake_case ( self ): lowercase__: Optional[int] = DebertaVaTokenizer(_UpperCAmelCase ) lowercase__: Optional[int] = tokenizer.encode('''sequence builders''' ) lowercase__: Optional[Any] = tokenizer.encode('''multi-sequence build''' ) lowercase__: Union[str, Any] = tokenizer.build_inputs_with_special_tokens(_UpperCAmelCase ) lowercase__: Dict = tokenizer.build_inputs_with_special_tokens(_UpperCAmelCase , _UpperCAmelCase ) self.assertEqual([tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] , _UpperCAmelCase ) self.assertEqual( [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [tokenizer.sep_token_id] , _UpperCAmelCase , ) @slow def _snake_case ( self ): # fmt: off lowercase__: List[Any] = {'''input_ids''': [[1, 39867, 36, 19390, 486, 27, 35052, 81436, 18, 60685, 1225, 7, 35052, 81436, 18, 9367, 16899, 18, 15937, 53, 594, 773, 18, 16287, 30465, 36, 15937, 6, 41139, 38, 36979, 60763, 191, 6, 34132, 99, 6, 50538, 390, 43230, 6, 34132, 2779, 20850, 14, 699, 1072, 1194, 36, 382, 10901, 53, 7, 699, 1072, 2084, 36, 20422, 630, 53, 19, 105, 3049, 1896, 1053, 16899, 1506, 11, 37978, 4243, 7, 1237, 31869, 200, 16566, 654, 6, 35052, 81436, 7, 55630, 13593, 4, 2], [1, 26, 15011, 13, 667, 8, 1053, 18, 23611, 1237, 72356, 12820, 34, 104134, 1209, 35, 13313, 6627, 21, 202, 347, 7, 164, 2399, 11, 46, 4485, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 5, 1232, 2864, 15785, 14951, 105, 5, 8581, 1250, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''token_type_ids''': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_UpperCAmelCase , model_name='''microsoft/deberta-v2-xlarge''' , revision='''ad6e42c1532ddf3a15c39246b63f5559d558b670''' , )
2
1
'''simple docstring''' from __future__ import annotations def lowerCamelCase ( lowerCAmelCase : list , lowerCAmelCase : int , lowerCAmelCase : int , lowerCAmelCase : int ): """simple docstring""" __magic_name__ : List[Any] = [] __magic_name__ , __magic_name__ : List[str] = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) __magic_name__ : Optional[int] = result + left + right return input_list def lowerCamelCase ( lowerCAmelCase : list ): """simple docstring""" if len(lowerCAmelCase ) <= 1: return input_list __magic_name__ : List[str] = list(lowerCAmelCase ) # iteration for two-way merging __magic_name__ : Union[str, Any] = 2 while p <= len(lowerCAmelCase ): # getting low, high and middle value for merge-sort of single list for i in range(0 , len(lowerCAmelCase ) , lowerCAmelCase ): __magic_name__ : str = i __magic_name__ : List[str] = i + p - 1 __magic_name__ : Tuple = (low + high + 1) // 2 __magic_name__ : Union[str, Any] = merge(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # final merge of last two parts if p * 2 >= len(lowerCAmelCase ): __magic_name__ : Optional[Any] = i __magic_name__ : Optional[int] = merge(lowerCAmelCase , 0 , lowerCAmelCase , len(lowerCAmelCase ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": lowerCAmelCase :Optional[Any] = input('''Enter numbers separated by a comma:\n''').strip() if user_input == "": lowerCAmelCase :Tuple = [] else: lowerCAmelCase :List[Any] = [int(item.strip()) for item in user_input.split(''',''')] print(iter_merge_sort(unsorted))
331
'''simple docstring''' from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax lowerCAmelCase :Any = logging.get_logger(__name__) @add_end_docstrings(lowercase__ ) class _lowerCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self : Optional[Any] , **_A : Union[str, Any] ) -> Tuple: super().__init__(**_A ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING ) def __call__( self : Optional[int] , _A : Union[str, List[str], "Image", List["Image"]] , **_A : Dict ) -> Dict: return super().__call__(_A , **_A ) def __lowerCAmelCase ( self : Any , **_A : Dict ) -> Optional[int]: __magic_name__ : str = {} if "candidate_labels" in kwargs: __magic_name__ : str = kwargs['candidate_labels'] if "hypothesis_template" in kwargs: __magic_name__ : Tuple = kwargs['hypothesis_template'] return preprocess_params, {}, {} def __lowerCAmelCase ( self : str , _A : Dict , _A : Optional[Any]=None , _A : int="This is a photo of {}." ) -> int: __magic_name__ : Dict = load_image(_A ) __magic_name__ : List[str] = self.image_processor(images=[image] , return_tensors=self.framework ) __magic_name__ : Optional[Any] = candidate_labels __magic_name__ : List[Any] = [hypothesis_template.format(_A ) for x in candidate_labels] __magic_name__ : str = self.tokenizer(_A , return_tensors=self.framework , padding=_A ) __magic_name__ : Optional[Any] = [text_inputs] return inputs def __lowerCAmelCase ( self : Union[str, Any] , _A : Tuple ) -> str: __magic_name__ : str = model_inputs.pop('candidate_labels' ) __magic_name__ : str = model_inputs.pop('text_inputs' ) if isinstance(text_inputs[0] , _A ): __magic_name__ : Dict = text_inputs[0] else: # Batching case. __magic_name__ : Optional[Any] = text_inputs[0][0] __magic_name__ : List[Any] = self.model(**_A , **_A ) __magic_name__ : str = { 'candidate_labels': candidate_labels, 'logits': outputs.logits_per_image, } return model_outputs def __lowerCAmelCase ( self : Optional[int] , _A : Optional[Any] ) -> Optional[int]: __magic_name__ : Tuple = model_outputs.pop('candidate_labels' ) __magic_name__ : Union[str, Any] = model_outputs['logits'][0] if self.framework == "pt": __magic_name__ : Tuple = logits.softmax(dim=-1 ).squeeze(-1 ) __magic_name__ : Tuple = probs.tolist() if not isinstance(_A , _A ): __magic_name__ : Any = [scores] elif self.framework == "tf": __magic_name__ : Any = stable_softmax(_A , axis=-1 ) __magic_name__ : Dict = probs.numpy().tolist() else: raise ValueError(F'Unsupported framework: {self.framework}' ) __magic_name__ : Union[str, Any] = [ {'score': score, 'label': candidate_label} for score, candidate_label in sorted(zip(_A , _A ) , key=lambda _A : -x[0] ) ] return result
331
1
import argparse import copy def __snake_case ( _UpperCAmelCase ): __a = {} with open(lowerCamelCase_ ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: __a = [] _list.append([line.split()[1], line.split()[2]] ) __a = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: __a = [] _list.append([line.split()[0], line.split()[2]] ) __a = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): with open(lowerCamelCase_ ) as f: __a = f.read(1 ) __a = start_node __a = [] __a = start_node __a = 0 while visiting not in first_solution: __a = 10000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(lowerCamelCase_ ) and k[0] not in first_solution: __a = k[1] __a = k[0] first_solution.append(lowerCamelCase_ ) __a = distance_of_first_solution + int(lowerCamelCase_ ) __a = best_node first_solution.append(lowerCamelCase_ ) __a = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 __a = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 10000 ) return first_solution, distance_of_first_solution def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): __a = [] for n in solution[1:-1]: __a = solution.index(lowerCamelCase_ ) for kn in solution[1:-1]: __a = solution.index(lowerCamelCase_ ) if n == kn: continue __a = copy.deepcopy(lowerCamelCase_ ) __a = kn __a = n __a = 0 for k in _tmp[:-1]: __a = _tmp[_tmp.index(lowerCamelCase_ ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: __a = distance + int(i[1] ) _tmp.append(lowerCamelCase_ ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) __a = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda _UpperCAmelCase : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): __a = 1 __a = first_solution __a = [] __a = distance_of_first_solution __a = solution while count <= iters: __a = find_neighborhood(lowerCamelCase_ , lowerCamelCase_ ) __a = 0 __a = neighborhood[index_of_best_solution] __a = len(lowerCamelCase_ ) - 1 __a = False while not found: __a = 0 while i < len(lowerCamelCase_ ): if best_solution[i] != solution[i]: __a = best_solution[i] __a = solution[i] break __a = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) __a = True __a = best_solution[:-1] __a = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: __a = cost __a = solution else: __a = index_of_best_solution + 1 __a = neighborhood[index_of_best_solution] if len(lowerCamelCase_ ) >= size: tabu_list.pop(0 ) __a = count + 1 return best_solution_ever, best_cost def __snake_case ( _UpperCAmelCase=None ): __a = generate_neighbours(args.File ) __a = generate_first_solution( args.File , lowerCamelCase_ ) __a = tabu_search( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , args.Iterations , args.Size , ) print(f'Best solution: {best_sol}, with total distance: {best_cost}.' ) if __name__ == "__main__": __snake_case :List[Any] = argparse.ArgumentParser(description='''Tabu Search''') parser.add_argument( '''-f''', '''--File''', type=str, help='''Path to the file containing the data''', required=True, ) parser.add_argument( '''-i''', '''--Iterations''', type=int, help='''How many iterations the algorithm should perform''', required=True, ) parser.add_argument( '''-s''', '''--Size''', type=int, help='''Size of the tabu list''', required=True ) # Pass the arguments to main method main(parser.parse_args())
356
import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils import floats_tensor, nightly, torch_device from diffusers.utils.testing_utils import require_torch_gpu class _A ( unittest.TestCase ): def _lowerCamelCase ( self : str): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowerCamelCase ( self : Any): '''simple docstring''' __a = 1 __a = 3 __a = (32, 32) __a = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0)).to(__SCREAMING_SNAKE_CASE) return image @property def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' torch.manual_seed(0) __a = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) return model @property def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' torch.manual_seed(0) __a = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) return model @property def _lowerCamelCase ( self : Any): '''simple docstring''' torch.manual_seed(0) __a = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) return CLIPTextModel(__SCREAMING_SNAKE_CASE) @property def _lowerCamelCase ( self : List[str]): '''simple docstring''' def extract(*__SCREAMING_SNAKE_CASE : int , **__SCREAMING_SNAKE_CASE : Dict): class _A : def __init__( self : int): '''simple docstring''' __a = torch.ones([0]) def _lowerCamelCase ( self : int , __SCREAMING_SNAKE_CASE : List[Any]): '''simple docstring''' self.pixel_values.to(__SCREAMING_SNAKE_CASE) return self return Out() return extract def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' __a = '''cpu''' # ensure determinism for the device-dependent torch.Generator __a = self.dummy_cond_unet __a = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=__SCREAMING_SNAKE_CASE , set_alpha_to_one=__SCREAMING_SNAKE_CASE , ) __a = self.dummy_vae __a = self.dummy_text_encoder __a = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') # make sure here that pndm scheduler skips prk __a = StableDiffusionPipeline( unet=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE , vae=__SCREAMING_SNAKE_CASE , text_encoder=__SCREAMING_SNAKE_CASE , tokenizer=__SCREAMING_SNAKE_CASE , safety_checker=__SCREAMING_SNAKE_CASE , feature_extractor=self.dummy_extractor , ) __a = sd_pipe.to(__SCREAMING_SNAKE_CASE) sd_pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE) __a = '''A painting of a squirrel eating a burger''' __a = torch.Generator(device=__SCREAMING_SNAKE_CASE).manual_seed(0) __a = sd_pipe([prompt] , generator=__SCREAMING_SNAKE_CASE , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''') __a = output.images __a = torch.Generator(device=__SCREAMING_SNAKE_CASE).manual_seed(0) __a = sd_pipe( [prompt] , generator=__SCREAMING_SNAKE_CASE , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , return_dict=__SCREAMING_SNAKE_CASE , )[0] __a = image[0, -3:, -3:, -1] __a = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __a = np.array([0.57_56, 0.61_18, 0.50_05, 0.50_41, 0.54_71, 0.47_26, 0.49_76, 0.48_65, 0.48_64]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1E-2 def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' __a = '''cpu''' # ensure determinism for the device-dependent torch.Generator __a = self.dummy_cond_unet __a = PNDMScheduler(skip_prk_steps=__SCREAMING_SNAKE_CASE) __a = self.dummy_vae __a = self.dummy_text_encoder __a = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') # make sure here that pndm scheduler skips prk __a = StableDiffusionPipeline( unet=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE , vae=__SCREAMING_SNAKE_CASE , text_encoder=__SCREAMING_SNAKE_CASE , tokenizer=__SCREAMING_SNAKE_CASE , safety_checker=__SCREAMING_SNAKE_CASE , feature_extractor=self.dummy_extractor , ) __a = sd_pipe.to(__SCREAMING_SNAKE_CASE) sd_pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE) __a = '''A painting of a squirrel eating a burger''' __a = torch.Generator(device=__SCREAMING_SNAKE_CASE).manual_seed(0) __a = sd_pipe([prompt] , generator=__SCREAMING_SNAKE_CASE , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''') __a = output.images __a = torch.Generator(device=__SCREAMING_SNAKE_CASE).manual_seed(0) __a = sd_pipe( [prompt] , generator=__SCREAMING_SNAKE_CASE , guidance_scale=6.0 , num_inference_steps=2 , output_type='''np''' , return_dict=__SCREAMING_SNAKE_CASE , )[0] __a = image[0, -3:, -3:, -1] __a = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __a = np.array([0.51_25, 0.57_16, 0.48_28, 0.50_60, 0.56_50, 0.47_68, 0.51_85, 0.48_95, 0.49_93]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1E-2 def _lowerCamelCase ( self : Optional[int]): '''simple docstring''' __a = StableDiffusionPipeline.from_pretrained( '''hf-internal-testing/tiny-stable-diffusion-lms-pipe''' , safety_checker=__SCREAMING_SNAKE_CASE) assert isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) assert isinstance(pipe.scheduler , __SCREAMING_SNAKE_CASE) assert pipe.safety_checker is None __a = pipe('''example prompt''' , num_inference_steps=2).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(__SCREAMING_SNAKE_CASE) __a = StableDiffusionPipeline.from_pretrained(__SCREAMING_SNAKE_CASE) # sanity check that the pipeline still works assert pipe.safety_checker is None __a = pipe('''example prompt''' , num_inference_steps=2).images[0] assert image is not None @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''') def _lowerCamelCase ( self : Tuple): '''simple docstring''' __a = self.dummy_cond_unet __a = PNDMScheduler(skip_prk_steps=__SCREAMING_SNAKE_CASE) __a = self.dummy_vae __a = self.dummy_text_encoder __a = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') # put models in fp16 __a = unet.half() __a = vae.half() __a = bert.half() # make sure here that pndm scheduler skips prk __a = StableDiffusionPipeline( unet=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE , vae=__SCREAMING_SNAKE_CASE , text_encoder=__SCREAMING_SNAKE_CASE , tokenizer=__SCREAMING_SNAKE_CASE , safety_checker=__SCREAMING_SNAKE_CASE , feature_extractor=self.dummy_extractor , ) __a = sd_pipe.to(__SCREAMING_SNAKE_CASE) sd_pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE) __a = '''A painting of a squirrel eating a burger''' __a = sd_pipe([prompt] , num_inference_steps=2 , output_type='''np''').images assert image.shape == (1, 64, 64, 3) @nightly @require_torch_gpu class _A ( unittest.TestCase ): def _lowerCamelCase ( self : Tuple): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : List[str]): '''simple docstring''' __a = StableDiffusionPipeline.from_pretrained('''runwayml/stable-diffusion-v1-5''' , safety_checker=__SCREAMING_SNAKE_CASE) __a = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config) __a = sd_pipe.to(__SCREAMING_SNAKE_CASE) sd_pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE) __a = ( '''portrait of girl with smokey eyes makeup in abandoned hotel, grange clothes, redshift, wide high angle''' ''' coloured polaroid photograph with flash, kodak film, hyper real, stunning moody cinematography, with''' ''' anamorphic lenses, by maripol, fallen angels by wong kar - wai, style of suspiria and neon demon and''' ''' children from bahnhof zoo, detailed ''' ) __a = 4_003_660_346 __a = 7 # without safety guidance (sld_guidance_scale = 0) __a = torch.manual_seed(__SCREAMING_SNAKE_CASE) __a = sd_pipe( [prompt] , generator=__SCREAMING_SNAKE_CASE , guidance_scale=__SCREAMING_SNAKE_CASE , num_inference_steps=50 , output_type='''np''' , width=512 , height=512 , sld_guidance_scale=0 , ) __a = output.images __a = image[0, -3:, -3:, -1] __a = [0.22_78, 0.22_31, 0.22_49, 0.23_33, 0.23_03, 0.18_85, 0.22_73, 0.21_44, 0.21_76] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 # without safety guidance (strong configuration) __a = torch.manual_seed(__SCREAMING_SNAKE_CASE) __a = sd_pipe( [prompt] , generator=__SCREAMING_SNAKE_CASE , guidance_scale=__SCREAMING_SNAKE_CASE , num_inference_steps=50 , output_type='''np''' , width=512 , height=512 , sld_guidance_scale=2_000 , sld_warmup_steps=7 , sld_threshold=0.0_25 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) __a = output.images __a = image[0, -3:, -3:, -1] __a = [0.23_83, 0.22_76, 0.2_36, 0.21_92, 0.21_86, 0.20_53, 0.19_71, 0.19_01, 0.17_19] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def _lowerCamelCase ( self : List[str]): '''simple docstring''' __a = StableDiffusionPipeline.from_pretrained('''runwayml/stable-diffusion-v1-5''' , safety_checker=__SCREAMING_SNAKE_CASE) __a = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config) __a = sd_pipe.to(__SCREAMING_SNAKE_CASE) sd_pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE) __a = '''padme amidala taking a bath artwork, safe for work, no nudity''' __a = 2_734_971_755 __a = 7 __a = torch.manual_seed(__SCREAMING_SNAKE_CASE) __a = sd_pipe( [prompt] , generator=__SCREAMING_SNAKE_CASE , guidance_scale=__SCREAMING_SNAKE_CASE , num_inference_steps=50 , output_type='''np''' , width=512 , height=512 , sld_guidance_scale=0 , ) __a = output.images __a = image[0, -3:, -3:, -1] __a = [0.35_02, 0.36_22, 0.33_96, 0.36_42, 0.34_78, 0.33_18, 0.35, 0.33_48, 0.32_97] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 __a = torch.manual_seed(__SCREAMING_SNAKE_CASE) __a = sd_pipe( [prompt] , generator=__SCREAMING_SNAKE_CASE , guidance_scale=__SCREAMING_SNAKE_CASE , num_inference_steps=50 , output_type='''np''' , width=512 , height=512 , sld_guidance_scale=2_000 , sld_warmup_steps=7 , sld_threshold=0.0_25 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) __a = output.images __a = image[0, -3:, -3:, -1] __a = [0.55_31, 0.52_06, 0.48_95, 0.51_56, 0.51_82, 0.47_51, 0.48_02, 0.48_03, 0.44_43] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def _lowerCamelCase ( self : str): '''simple docstring''' __a = StableDiffusionPipeline.from_pretrained('''runwayml/stable-diffusion-v1-5''') __a = sd_pipe.to(__SCREAMING_SNAKE_CASE) sd_pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE) __a = ( '''the four horsewomen of the apocalypse, painting by tom of finland, gaston bussiere, craig mullins, j. c.''' ''' leyendecker''' ) __a = 1_044_355_234 __a = 12 __a = torch.manual_seed(__SCREAMING_SNAKE_CASE) __a = sd_pipe( [prompt] , generator=__SCREAMING_SNAKE_CASE , guidance_scale=__SCREAMING_SNAKE_CASE , num_inference_steps=50 , output_type='''np''' , width=512 , height=512 , sld_guidance_scale=0 , ) __a = output.images __a = image[0, -3:, -3:, -1] __a = np.array([0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-7 __a = torch.manual_seed(__SCREAMING_SNAKE_CASE) __a = sd_pipe( [prompt] , generator=__SCREAMING_SNAKE_CASE , guidance_scale=__SCREAMING_SNAKE_CASE , num_inference_steps=50 , output_type='''np''' , width=512 , height=512 , sld_guidance_scale=2_000 , sld_warmup_steps=7 , sld_threshold=0.0_25 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) __a = output.images __a = image[0, -3:, -3:, -1] __a = np.array([0.58_18, 0.62_85, 0.68_35, 0.60_19, 0.6_25, 0.67_54, 0.60_96, 0.63_34, 0.65_61]) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2
131
0
import gc import random import unittest import numpy as np import torch from transformers import CLIPImageProcessor, CLIPVisionConfig, CLIPVisionModel from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEImgaImgPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import floats_tensor, load_image, load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class __A( a , unittest.TestCase ): snake_case_ = ShapEImgaImgPipeline snake_case_ = ['''image'''] snake_case_ = ['''image'''] snake_case_ = [ '''num_images_per_prompt''', '''num_inference_steps''', '''generator''', '''latents''', '''guidance_scale''', '''frame_size''', '''output_type''', '''return_dict''', ] snake_case_ = False @property def SCREAMING_SNAKE_CASE_ ( self ) -> List[Any]: '''simple docstring''' return 32 @property def SCREAMING_SNAKE_CASE_ ( self ) -> Dict: '''simple docstring''' return 32 @property def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[int]: '''simple docstring''' return self.time_input_dim * 4 @property def SCREAMING_SNAKE_CASE_ ( self ) -> List[str]: '''simple docstring''' return 8 @property def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[int]: '''simple docstring''' torch.manual_seed(0 ) __a = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=64 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=1 , ) __a = CLIPVisionModel(_snake_case ) return model @property def SCREAMING_SNAKE_CASE_ ( self ) -> List[Any]: '''simple docstring''' __a = CLIPImageProcessor( crop_size=224 , do_center_crop=_snake_case , do_normalize=_snake_case , do_resize=_snake_case , image_mean=[0.4814_5466, 0.457_8275, 0.4082_1073] , image_std=[0.2686_2954, 0.2613_0258, 0.2757_7711] , resample=3 , size=224 , ) return image_processor @property def SCREAMING_SNAKE_CASE_ ( self ) -> Tuple: '''simple docstring''' torch.manual_seed(0 ) __a = { '''num_attention_heads''': 2, '''attention_head_dim''': 16, '''embedding_dim''': self.time_input_dim, '''num_embeddings''': 32, '''embedding_proj_dim''': self.text_embedder_hidden_size, '''time_embed_dim''': self.time_embed_dim, '''num_layers''': 1, '''clip_embed_dim''': self.time_input_dim * 2, '''additional_embeddings''': 0, '''time_embed_act_fn''': '''gelu''', '''norm_in_type''': '''layer''', '''embedding_proj_norm_type''': '''layer''', '''encoder_hid_proj_type''': None, '''added_emb_type''': None, } __a = PriorTransformer(**_snake_case ) return model @property def SCREAMING_SNAKE_CASE_ ( self ) -> Tuple: '''simple docstring''' torch.manual_seed(0 ) __a = { '''param_shapes''': ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), '''d_latent''': self.time_input_dim, '''d_hidden''': self.renderer_dim, '''n_output''': 12, '''background''': ( 0.1, 0.1, 0.1, ), } __a = ShapERenderer(**_snake_case ) return model def SCREAMING_SNAKE_CASE_ ( self ) -> str: '''simple docstring''' __a = self.dummy_prior __a = self.dummy_image_encoder __a = self.dummy_image_processor __a = self.dummy_renderer __a = HeunDiscreteScheduler( beta_schedule='''exp''' , num_train_timesteps=1_024 , prediction_type='''sample''' , use_karras_sigmas=_snake_case , clip_sample=_snake_case , clip_sample_range=1.0 , ) __a = { '''prior''': prior, '''image_encoder''': image_encoder, '''image_processor''': image_processor, '''renderer''': renderer, '''scheduler''': scheduler, } return components def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case=0 ) -> List[Any]: '''simple docstring''' __a = floats_tensor((1, 3, 64, 64) , rng=random.Random(_snake_case ) ).to(_snake_case ) if str(_snake_case ).startswith('''mps''' ): __a = torch.manual_seed(_snake_case ) else: __a = torch.Generator(device=_snake_case ).manual_seed(_snake_case ) __a = { '''image''': input_image, '''generator''': generator, '''num_inference_steps''': 1, '''frame_size''': 32, '''output_type''': '''np''', } return inputs def SCREAMING_SNAKE_CASE_ ( self ) -> Tuple: '''simple docstring''' __a = '''cpu''' __a = self.get_dummy_components() __a = self.pipeline_class(**_snake_case ) __a = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) __a = pipe(**self.get_dummy_inputs(_snake_case ) ) __a = output.images[0] __a = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) __a = np.array( [ 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def SCREAMING_SNAKE_CASE_ ( self ) -> Union[str, Any]: '''simple docstring''' self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def SCREAMING_SNAKE_CASE_ ( self ) -> List[str]: '''simple docstring''' __a = torch_device == '''cpu''' __a = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_snake_case , relax_max_difference=_snake_case , ) def SCREAMING_SNAKE_CASE_ ( self ) -> Union[str, Any]: '''simple docstring''' __a = self.get_dummy_components() __a = self.pipeline_class(**_snake_case ) __a = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) __a = 1 __a = 2 __a = self.get_dummy_inputs(_snake_case ) for key in inputs.keys(): if key in self.batch_params: __a = batch_size * [inputs[key]] __a = pipe(**_snake_case , num_images_per_prompt=_snake_case )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class __A( unittest.TestCase ): def SCREAMING_SNAKE_CASE_ ( self ) -> List[str]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE_ ( self ) -> str: '''simple docstring''' __a = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/shap_e/corgi.png''' ) __a = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/shap_e/test_shap_e_img2img_out.npy''' ) __a = ShapEImgaImgPipeline.from_pretrained('''openai/shap-e-img2img''' ) __a = pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) __a = torch.Generator(device=_snake_case ).manual_seed(0 ) __a = pipe( _snake_case , generator=_snake_case , guidance_scale=3.0 , num_inference_steps=64 , frame_size=64 , output_type='''np''' , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(_snake_case , _snake_case )
6
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __A( a ): snake_case_ = ['''image_processor''', '''tokenizer'''] snake_case_ = '''ChineseCLIPImageProcessor''' snake_case_ = ('''BertTokenizer''', '''BertTokenizerFast''') def __init__( self , _snake_case=None , _snake_case=None , **_snake_case ) -> Tuple: '''simple docstring''' __a = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _snake_case , ) __a = kwargs.pop('''feature_extractor''' ) __a = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_snake_case , _snake_case ) __a = self.image_processor def __call__( self , _snake_case=None , _snake_case=None , _snake_case=None , **_snake_case ) -> Optional[Any]: '''simple docstring''' if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: __a = self.tokenizer(_snake_case , return_tensors=_snake_case , **_snake_case ) if images is not None: __a = self.image_processor(_snake_case , return_tensors=_snake_case , **_snake_case ) if text is not None and images is not None: __a = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_snake_case ) , tensor_type=_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , *_snake_case , **_snake_case ) -> str: '''simple docstring''' return self.tokenizer.batch_decode(*_snake_case , **_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , *_snake_case , **_snake_case ) -> Dict: '''simple docstring''' return self.tokenizer.decode(*_snake_case , **_snake_case ) @property def SCREAMING_SNAKE_CASE_ ( self ) -> Any: '''simple docstring''' __a = self.tokenizer.model_input_names __a = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[Any]: '''simple docstring''' warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , _snake_case , ) return self.image_processor_class
6
1
"""simple docstring""" def UpperCamelCase_ ( lowerCAmelCase__ : str , lowerCAmelCase__ : str ) -> int: """simple docstring""" if len(lowerCAmelCase__ ) != len(lowerCAmelCase__ ): raise ValueError('String lengths must match!' ) lowerCAmelCase_ : List[Any] = 0 for chara, chara in zip(lowerCAmelCase__ , lowerCAmelCase__ ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
289
"""simple docstring""" import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase__ : """simple docstring""" def __init__( self : Optional[Any] , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Dict=1_3 , SCREAMING_SNAKE_CASE_ : List[Any]=7 , SCREAMING_SNAKE_CASE_ : Dict=True , SCREAMING_SNAKE_CASE_ : Optional[Any]=True , SCREAMING_SNAKE_CASE_ : Optional[Any]=True , SCREAMING_SNAKE_CASE_ : str=True , SCREAMING_SNAKE_CASE_ : List[str]=9_9 , SCREAMING_SNAKE_CASE_ : int=1_6 , SCREAMING_SNAKE_CASE_ : List[str]=3_6 , SCREAMING_SNAKE_CASE_ : List[Any]=6 , SCREAMING_SNAKE_CASE_ : Tuple=6 , SCREAMING_SNAKE_CASE_ : List[Any]=6 , SCREAMING_SNAKE_CASE_ : Union[str, Any]=3_7 , SCREAMING_SNAKE_CASE_ : Tuple="gelu" , SCREAMING_SNAKE_CASE_ : Union[str, Any]=0.1 , SCREAMING_SNAKE_CASE_ : int=0.1 , SCREAMING_SNAKE_CASE_ : Optional[Any]=5_1_2 , SCREAMING_SNAKE_CASE_ : List[str]=1_6 , SCREAMING_SNAKE_CASE_ : List[str]=2 , SCREAMING_SNAKE_CASE_ : List[Any]=0.02 , SCREAMING_SNAKE_CASE_ : Dict=3 , SCREAMING_SNAKE_CASE_ : int=4 , SCREAMING_SNAKE_CASE_ : Tuple=None , ): lowerCAmelCase_ : Any = parent lowerCAmelCase_ : Optional[int] = batch_size lowerCAmelCase_ : Dict = seq_length lowerCAmelCase_ : Tuple = is_training lowerCAmelCase_ : str = use_input_mask lowerCAmelCase_ : Union[str, Any] = use_token_type_ids lowerCAmelCase_ : Tuple = use_labels lowerCAmelCase_ : Optional[int] = vocab_size lowerCAmelCase_ : Any = embedding_size lowerCAmelCase_ : Optional[Any] = hidden_size lowerCAmelCase_ : str = num_hidden_layers lowerCAmelCase_ : Optional[Any] = num_hidden_groups lowerCAmelCase_ : Dict = num_attention_heads lowerCAmelCase_ : Optional[Any] = intermediate_size lowerCAmelCase_ : Any = hidden_act lowerCAmelCase_ : Union[str, Any] = hidden_dropout_prob lowerCAmelCase_ : int = attention_probs_dropout_prob lowerCAmelCase_ : int = max_position_embeddings lowerCAmelCase_ : List[Any] = type_vocab_size lowerCAmelCase_ : Any = type_sequence_label_size lowerCAmelCase_ : Optional[int] = initializer_range lowerCAmelCase_ : Tuple = num_labels lowerCAmelCase_ : Dict = num_choices lowerCAmelCase_ : Tuple = scope def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): lowerCAmelCase_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase_ : str = None if self.use_input_mask: lowerCAmelCase_ : int = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase_ : List[Any] = None if self.use_token_type_ids: lowerCAmelCase_ : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase_ : List[str] = None lowerCAmelCase_ : Union[str, Any] = None lowerCAmelCase_ : str = None if self.use_labels: lowerCAmelCase_ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase_ : Dict = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase_ : List[Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE__ ( self : Dict ): return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def SCREAMING_SNAKE_CASE__ ( self : Tuple , SCREAMING_SNAKE_CASE_ : Any , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : str ): lowerCAmelCase_ : Union[str, Any] = AlbertModel(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase_ : List[str] = model(SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ , token_type_ids=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase_ : List[Any] = model(SCREAMING_SNAKE_CASE_ , token_type_ids=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase_ : int = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self : int , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Tuple ): lowerCAmelCase_ : Optional[Any] = AlbertForPreTraining(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase_ : Optional[Any] = model( SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ , token_type_ids=SCREAMING_SNAKE_CASE_ , labels=SCREAMING_SNAKE_CASE_ , sentence_order_label=SCREAMING_SNAKE_CASE_ , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self : Dict , SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : Any , SCREAMING_SNAKE_CASE_ : Dict ): lowerCAmelCase_ : str = AlbertForMaskedLM(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase_ : str = model(SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ , token_type_ids=SCREAMING_SNAKE_CASE_ , labels=SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self : str , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : List[Any] ): lowerCAmelCase_ : List[str] = AlbertForQuestionAnswering(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase_ : Any = model( SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ , token_type_ids=SCREAMING_SNAKE_CASE_ , start_positions=SCREAMING_SNAKE_CASE_ , end_positions=SCREAMING_SNAKE_CASE_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE__ ( self : Dict , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Dict ): lowerCAmelCase_ : Union[str, Any] = self.num_labels lowerCAmelCase_ : Union[str, Any] = AlbertForSequenceClassification(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase_ : Union[str, Any] = model(SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ , token_type_ids=SCREAMING_SNAKE_CASE_ , labels=SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self : Any , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : List[str] ): lowerCAmelCase_ : List[str] = self.num_labels lowerCAmelCase_ : List[Any] = AlbertForTokenClassification(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase_ : str = model(SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ , token_type_ids=SCREAMING_SNAKE_CASE_ , labels=SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Any , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : List[str] ): lowerCAmelCase_ : Optional[Any] = self.num_choices lowerCAmelCase_ : int = AlbertForMultipleChoice(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCAmelCase_ : Optional[int] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase_ : List[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase_ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase_ : List[Any] = model( SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ , token_type_ids=SCREAMING_SNAKE_CASE_ , labels=SCREAMING_SNAKE_CASE_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ): lowerCAmelCase_ : Optional[int] = self.prepare_config_and_inputs() ( ( lowerCAmelCase_ ) ,( lowerCAmelCase_ ) ,( lowerCAmelCase_ ) ,( lowerCAmelCase_ ) ,( lowerCAmelCase_ ) ,( lowerCAmelCase_ ) ,( lowerCAmelCase_ ) , ) : Optional[int] = config_and_inputs lowerCAmelCase_ : Dict = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class UpperCamelCase__ ( lowercase_, lowercase_, unittest.TestCase ): """simple docstring""" _SCREAMING_SNAKE_CASE = ( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) _SCREAMING_SNAKE_CASE = ( { """feature-extraction""": AlbertModel, """fill-mask""": AlbertForMaskedLM, """question-answering""": AlbertForQuestionAnswering, """text-classification""": AlbertForSequenceClassification, """token-classification""": AlbertForTokenClassification, """zero-shot""": AlbertForSequenceClassification, } if is_torch_available() else {} ) _SCREAMING_SNAKE_CASE = True def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : str=False ): lowerCAmelCase_ : List[str] = super()._prepare_for_class(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , return_labels=SCREAMING_SNAKE_CASE_ ) if return_labels: if model_class in get_values(SCREAMING_SNAKE_CASE_ ): lowerCAmelCase_ : List[str] = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase_ : Any = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=SCREAMING_SNAKE_CASE_ ) return inputs_dict def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): lowerCAmelCase_ : str = AlbertModelTester(self ) lowerCAmelCase_ : Optional[int] = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE_ , hidden_size=3_7 ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self : Tuple ): lowerCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_ ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): lowerCAmelCase_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*SCREAMING_SNAKE_CASE_ ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ): lowerCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*SCREAMING_SNAKE_CASE_ ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): lowerCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*SCREAMING_SNAKE_CASE_ ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): lowerCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*SCREAMING_SNAKE_CASE_ ) def SCREAMING_SNAKE_CASE__ ( self : str ): lowerCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*SCREAMING_SNAKE_CASE_ ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): lowerCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowerCAmelCase_ : int = type self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_ ) @slow def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Optional[Any] = AlbertModel.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) @require_torch class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @slow def SCREAMING_SNAKE_CASE__ ( self : List[str] ): lowerCAmelCase_ : Any = AlbertModel.from_pretrained('albert-base-v2' ) lowerCAmelCase_ : Tuple = torch.tensor([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) lowerCAmelCase_ : Tuple = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): lowerCAmelCase_ : List[Any] = model(SCREAMING_SNAKE_CASE_ , attention_mask=SCREAMING_SNAKE_CASE_ )[0] lowerCAmelCase_ : str = torch.Size((1, 1_1, 7_6_8) ) self.assertEqual(output.shape , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase_ : int = torch.tensor( [[[-0.65_13, 1.50_35, -0.27_66], [-0.65_15, 1.50_46, -0.27_80], [-0.65_12, 1.50_49, -0.27_84]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , SCREAMING_SNAKE_CASE_ , atol=1E-4 ) )
289
1
from __future__ import annotations _UpperCAmelCase : Optional[int] = tuple[int, int, int] _UpperCAmelCase : int = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase _UpperCAmelCase : Optional[int] = """ABCDEFGHIJKLMNOPQRSTUVWXYZ""" # -------------------------- default selection -------------------------- # rotors -------------------------- _UpperCAmelCase : List[str] = """EGZWVONAHDCLFQMSIPJBYUKXTR""" _UpperCAmelCase : Optional[int] = """FOBHMDKEXQNRAULPGSJVTYICZW""" _UpperCAmelCase : List[Any] = """ZJXESIUQLHAVRMDOYGTNFWPBKC""" # reflector -------------------------- _UpperCAmelCase : Union[str, Any] = { """A""": """N""", """N""": """A""", """B""": """O""", """O""": """B""", """C""": """P""", """P""": """C""", """D""": """Q""", """Q""": """D""", """E""": """R""", """R""": """E""", """F""": """S""", """S""": """F""", """G""": """T""", """T""": """G""", """H""": """U""", """U""": """H""", """I""": """V""", """V""": """I""", """J""": """W""", """W""": """J""", """K""": """X""", """X""": """K""", """L""": """Y""", """Y""": """L""", """M""": """Z""", """Z""": """M""", } # -------------------------- extra rotors -------------------------- _UpperCAmelCase : List[str] = """RMDJXFUWGISLHVTCQNKYPBEZOA""" _UpperCAmelCase : Any = """SGLCPQWZHKXAREONTFBVIYJUDM""" _UpperCAmelCase : List[str] = """HVSICLTYKQUBXDWAJZOMFGPREN""" _UpperCAmelCase : Tuple = """RZWQHFMVDBKICJLNTUXAGYPSOE""" _UpperCAmelCase : Optional[int] = """LFKIJODBEGAMQPXVUHYSTCZRWN""" _UpperCAmelCase : Tuple = """KOAEGVDHXPQZMLFTYWJNBRCIUS""" def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]: # Checks if there are 3 unique rotors if (unique_rotsel := len(set(_UpperCAmelCase ) )) < 3: lowerCamelCase__ : int = F"""Please use 3 unique rotors (not {unique_rotsel})""" raise Exception(_UpperCAmelCase ) # Checks if rotor positions are valid lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : str = rotpos if not 0 < rotorposa <= len(_UpperCAmelCase ): lowerCamelCase__ : List[Any] = F"""First rotor position is not within range of 1..26 ({rotorposa}""" raise ValueError(_UpperCAmelCase ) if not 0 < rotorposa <= len(_UpperCAmelCase ): lowerCamelCase__ : List[Any] = F"""Second rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(_UpperCAmelCase ) if not 0 < rotorposa <= len(_UpperCAmelCase ): lowerCamelCase__ : int = F"""Third rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(_UpperCAmelCase ) # Validates string and returns dict lowerCamelCase__ : Optional[int] = _plugboard(_UpperCAmelCase ) return rotpos, rotsel, pbdict def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> dict[str, str]: # tests the input string if it # a) is type string # b) has even length (so pairs can be made) if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Union[str, Any] = F"""Plugboard setting isn't type string ({type(_UpperCAmelCase )})""" raise TypeError(_UpperCAmelCase ) elif len(_UpperCAmelCase ) % 2 != 0: lowerCamelCase__ : Dict = F"""Odd number of symbols ({len(_UpperCAmelCase )})""" raise Exception(_UpperCAmelCase ) elif pbstring == "": return {} pbstring.replace(' ' , '' ) # Checks if all characters are unique lowerCamelCase__ : Dict = set() for i in pbstring: if i not in abc: lowerCamelCase__ : Union[str, Any] = F"""'{i}' not in list of symbols""" raise Exception(_UpperCAmelCase ) elif i in tmppbl: lowerCamelCase__ : Optional[Any] = F"""Duplicate symbol ({i})""" raise Exception(_UpperCAmelCase ) else: tmppbl.add(_UpperCAmelCase ) del tmppbl # Created the dictionary lowerCamelCase__ : Dict = {} for j in range(0 , len(_UpperCAmelCase ) - 1 , 2 ): lowerCamelCase__ : int = pbstring[j + 1] lowerCamelCase__ : Union[str, Any] = pbstring[j] return pb def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = (rotora, rotora, rotora) , _UpperCAmelCase = "" , ) -> str: lowerCamelCase__ : List[Any] = text.upper() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = _validator( _UpperCAmelCase , _UpperCAmelCase , plugb.upper() ) lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Tuple = rotor_position lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : List[Any] = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 lowerCamelCase__ : Dict = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: lowerCamelCase__ : Tuple = plugboard[symbol] # rotor ra -------------------------- lowerCamelCase__ : Optional[Any] = abc.index(_UpperCAmelCase ) + rotorposa lowerCamelCase__ : int = rotora[index % len(_UpperCAmelCase )] # rotor rb -------------------------- lowerCamelCase__ : Dict = abc.index(_UpperCAmelCase ) + rotorposa lowerCamelCase__ : Optional[Any] = rotora[index % len(_UpperCAmelCase )] # rotor rc -------------------------- lowerCamelCase__ : str = abc.index(_UpperCAmelCase ) + rotorposa lowerCamelCase__ : Optional[Any] = rotora[index % len(_UpperCAmelCase )] # reflector -------------------------- # this is the reason you don't need another machine to decipher lowerCamelCase__ : List[Any] = reflector[symbol] # 2nd rotors lowerCamelCase__ : Union[str, Any] = abc[rotora.index(_UpperCAmelCase ) - rotorposa] lowerCamelCase__ : Tuple = abc[rotora.index(_UpperCAmelCase ) - rotorposa] lowerCamelCase__ : str = abc[rotora.index(_UpperCAmelCase ) - rotorposa] # 2nd plugboard if symbol in plugboard: lowerCamelCase__ : Dict = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(_UpperCAmelCase ): lowerCamelCase__ : Optional[Any] = 0 rotorposa += 1 if rotorposa >= len(_UpperCAmelCase ): lowerCamelCase__ : Optional[Any] = 0 rotorposa += 1 if rotorposa >= len(_UpperCAmelCase ): lowerCamelCase__ : Union[str, Any] = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(_UpperCAmelCase ) return "".join(_UpperCAmelCase ) if __name__ == "__main__": _UpperCAmelCase : Optional[int] = """This is my Python script that emulates the Enigma machine from WWII.""" _UpperCAmelCase : List[Any] = (1, 1, 1) _UpperCAmelCase : List[Any] = """pictures""" _UpperCAmelCase : int = (rotora, rotora, rotora) _UpperCAmelCase : Any = enigma(message, rotor_pos, rotor_sel, pb) print("""Encrypted message:""", en) print("""Decrypted message:""", enigma(en, rotor_pos, rotor_sel, pb))
50
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> bool: lowerCamelCase__ : List[str] = len(_UpperCAmelCase ) lowerCamelCase__ : str = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): lowerCamelCase__ : Tuple = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): lowerCamelCase__ : Dict = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: lowerCamelCase__ : str = subset[i - 1][j] if arr[i - 1] <= j: lowerCamelCase__ : Dict = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
50
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.generation import DisjunctiveConstraint @require_torch class __lowerCamelCase ( unittest.TestCase ): def lowerCAmelCase_ ( self ) -> Tuple: # For consistency across different places the DisjunctiveConstraint is called, # dc.token_ids is a list of integers. It is also initialized only by integers. snake_case_ = [[1, 2, 4], [1, 2, 3, 4]] snake_case_ = DisjunctiveConstraint(lowerCamelCase ) self.assertTrue(isinstance(dc.token_ids , lowerCamelCase ) ) with self.assertRaises(lowerCamelCase ): DisjunctiveConstraint(torch.LongTensor([[1, 2, 4], [1, 2, 3]] ) ) with self.assertRaises(lowerCamelCase ): DisjunctiveConstraint([torch.LongTensor([1, 2, 4] ), torch.LongTensor([1, 2, 3, 4, 5] )] ) def lowerCAmelCase_ ( self ) -> List[str]: # We can't have constraints that are complete subsets of another. This leads to a preverse # interpretation of "constraint fulfillment": does generating [1,2,3] fulfill the constraint? # It would mean that it generated [1,2] which fulfills it, but it's in the middle of potentially # fulfilling [1,2,3,4]. If we believe that [1,2,3] does fulfill the constraint, then the algorithm # will necessarily never reach [1,2,3,4], giving users a false sense of control (better to just not allow it). snake_case_ = [[1, 2], [1, 2, 3, 4]] with self.assertRaises(lowerCamelCase ): DisjunctiveConstraint(lowerCamelCase ) # fails here def lowerCAmelCase_ ( self ) -> Tuple: snake_case_ = [[1, 2, 3], [1, 2, 4]] snake_case_ = DisjunctiveConstraint(lowerCamelCase ) snake_case_ , snake_case_ , snake_case_ = dc.update(1 ) snake_case_ = stepped is True and completed is False and reset is False self.assertTrue(lowerCamelCase ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) snake_case_ , snake_case_ , snake_case_ = dc.update(2 ) snake_case_ = stepped is True and completed is False and reset is False self.assertTrue(lowerCamelCase ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) snake_case_ , snake_case_ , snake_case_ = dc.update(3 ) snake_case_ = stepped is True and completed is True and reset is False self.assertTrue(lowerCamelCase ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 3] ) def lowerCAmelCase_ ( self ) -> Optional[int]: snake_case_ = [[1, 2, 3], [1, 2, 4, 5], [1, 2, 5]] snake_case_ = DisjunctiveConstraint(lowerCamelCase ) snake_case_ , snake_case_ , snake_case_ = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) snake_case_ , snake_case_ , snake_case_ = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) snake_case_ , snake_case_ , snake_case_ = dc.update(4 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2, 4] ) snake_case_ , snake_case_ , snake_case_ = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 4, 5] ) dc.reset() snake_case_ , snake_case_ , snake_case_ = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 3 ) self.assertTrue(dc.current_seq == [1] ) snake_case_ , snake_case_ , snake_case_ = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 2 ) self.assertTrue(dc.current_seq == [1, 2] ) snake_case_ , snake_case_ , snake_case_ = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.remaining() == 0 ) self.assertTrue(dc.current_seq == [1, 2, 5] )
34
import argparse import json import os import numpy as np import PIL import requests import tensorflow.keras.applications.efficientnet as efficientnet import torch from huggingface_hub import hf_hub_download from PIL import Image from tensorflow.keras.preprocessing import image from transformers import ( EfficientNetConfig, EfficientNetForImageClassification, EfficientNetImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { '''b0''': efficientnet.EfficientNetBa, '''b1''': efficientnet.EfficientNetBa, '''b2''': efficientnet.EfficientNetBa, '''b3''': efficientnet.EfficientNetBa, '''b4''': efficientnet.EfficientNetBa, '''b5''': efficientnet.EfficientNetBa, '''b6''': efficientnet.EfficientNetBa, '''b7''': efficientnet.EfficientNetBa, } lowerCamelCase_ = { '''b0''': { '''hidden_dim''': 1280, '''width_coef''': 1.0, '''depth_coef''': 1.0, '''image_size''': 224, '''dropout_rate''': 0.2, '''dw_padding''': [], }, '''b1''': { '''hidden_dim''': 1280, '''width_coef''': 1.0, '''depth_coef''': 1.1, '''image_size''': 240, '''dropout_rate''': 0.2, '''dw_padding''': [16], }, '''b2''': { '''hidden_dim''': 1408, '''width_coef''': 1.1, '''depth_coef''': 1.2, '''image_size''': 260, '''dropout_rate''': 0.3, '''dw_padding''': [5, 8, 16], }, '''b3''': { '''hidden_dim''': 1536, '''width_coef''': 1.2, '''depth_coef''': 1.4, '''image_size''': 300, '''dropout_rate''': 0.3, '''dw_padding''': [5, 18], }, '''b4''': { '''hidden_dim''': 1792, '''width_coef''': 1.4, '''depth_coef''': 1.8, '''image_size''': 380, '''dropout_rate''': 0.4, '''dw_padding''': [6], }, '''b5''': { '''hidden_dim''': 2048, '''width_coef''': 1.6, '''depth_coef''': 2.2, '''image_size''': 456, '''dropout_rate''': 0.4, '''dw_padding''': [13, 27], }, '''b6''': { '''hidden_dim''': 2304, '''width_coef''': 1.8, '''depth_coef''': 2.6, '''image_size''': 528, '''dropout_rate''': 0.5, '''dw_padding''': [31], }, '''b7''': { '''hidden_dim''': 2560, '''width_coef''': 2.0, '''depth_coef''': 3.1, '''image_size''': 600, '''dropout_rate''': 0.5, '''dw_padding''': [18], }, } def UpperCamelCase( lowercase_ ) -> Tuple: '''simple docstring''' snake_case_ = EfficientNetConfig() snake_case_ = CONFIG_MAP[model_name]["""hidden_dim"""] snake_case_ = CONFIG_MAP[model_name]["""width_coef"""] snake_case_ = CONFIG_MAP[model_name]["""depth_coef"""] snake_case_ = CONFIG_MAP[model_name]["""image_size"""] snake_case_ = CONFIG_MAP[model_name]["""dropout_rate"""] snake_case_ = CONFIG_MAP[model_name]["""dw_padding"""] snake_case_ = """huggingface/label-files""" snake_case_ = """imagenet-1k-id2label.json""" snake_case_ = 1000 snake_case_ = json.load(open(hf_hub_download(lowercase_ , lowercase_ , repo_type="""dataset""" ) , """r""" ) ) snake_case_ = {int(lowercase_ ): v for k, v in idalabel.items()} snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} return config def UpperCamelCase( ) -> Tuple: '''simple docstring''' snake_case_ = """http://images.cocodataset.org/val2017/000000039769.jpg""" snake_case_ = Image.open(requests.get(lowercase_ , stream=lowercase_ ).raw ) return im def UpperCamelCase( lowercase_ ) -> Union[str, Any]: '''simple docstring''' snake_case_ = CONFIG_MAP[model_name]["""image_size"""] snake_case_ = EfficientNetImageProcessor( size={"""height""": size, """width""": size} , image_mean=[0.4_85, 0.4_56, 0.4_06] , image_std=[0.47_85_39_44, 0.4_73_28_64, 0.47_43_41_63] , do_center_crop=lowercase_ , ) return preprocessor def UpperCamelCase( lowercase_ ) -> str: '''simple docstring''' snake_case_ = [v.split("""_""" )[0].split("""block""" )[1] for v in original_param_names if v.startswith("""block""" )] snake_case_ = sorted(set(lowercase_ ) ) snake_case_ = len(lowercase_ ) snake_case_ = {b: str(lowercase_ ) for b, i in zip(lowercase_ , range(lowercase_ ) )} snake_case_ = [] rename_keys.append(("""stem_conv/kernel:0""", """embeddings.convolution.weight""") ) rename_keys.append(("""stem_bn/gamma:0""", """embeddings.batchnorm.weight""") ) rename_keys.append(("""stem_bn/beta:0""", """embeddings.batchnorm.bias""") ) rename_keys.append(("""stem_bn/moving_mean:0""", """embeddings.batchnorm.running_mean""") ) rename_keys.append(("""stem_bn/moving_variance:0""", """embeddings.batchnorm.running_var""") ) for b in block_names: snake_case_ = block_name_mapping[b] rename_keys.append((f'''block{b}_expand_conv/kernel:0''', f'''encoder.blocks.{hf_b}.expansion.expand_conv.weight''') ) rename_keys.append((f'''block{b}_expand_bn/gamma:0''', f'''encoder.blocks.{hf_b}.expansion.expand_bn.weight''') ) rename_keys.append((f'''block{b}_expand_bn/beta:0''', f'''encoder.blocks.{hf_b}.expansion.expand_bn.bias''') ) rename_keys.append( (f'''block{b}_expand_bn/moving_mean:0''', f'''encoder.blocks.{hf_b}.expansion.expand_bn.running_mean''') ) rename_keys.append( (f'''block{b}_expand_bn/moving_variance:0''', f'''encoder.blocks.{hf_b}.expansion.expand_bn.running_var''') ) rename_keys.append( (f'''block{b}_dwconv/depthwise_kernel:0''', f'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_conv.weight''') ) rename_keys.append((f'''block{b}_bn/gamma:0''', f'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.weight''') ) rename_keys.append((f'''block{b}_bn/beta:0''', f'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.bias''') ) rename_keys.append( (f'''block{b}_bn/moving_mean:0''', f'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_mean''') ) rename_keys.append( (f'''block{b}_bn/moving_variance:0''', f'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_var''') ) rename_keys.append((f'''block{b}_se_reduce/kernel:0''', f'''encoder.blocks.{hf_b}.squeeze_excite.reduce.weight''') ) rename_keys.append((f'''block{b}_se_reduce/bias:0''', f'''encoder.blocks.{hf_b}.squeeze_excite.reduce.bias''') ) rename_keys.append((f'''block{b}_se_expand/kernel:0''', f'''encoder.blocks.{hf_b}.squeeze_excite.expand.weight''') ) rename_keys.append((f'''block{b}_se_expand/bias:0''', f'''encoder.blocks.{hf_b}.squeeze_excite.expand.bias''') ) rename_keys.append( (f'''block{b}_project_conv/kernel:0''', f'''encoder.blocks.{hf_b}.projection.project_conv.weight''') ) rename_keys.append((f'''block{b}_project_bn/gamma:0''', f'''encoder.blocks.{hf_b}.projection.project_bn.weight''') ) rename_keys.append((f'''block{b}_project_bn/beta:0''', f'''encoder.blocks.{hf_b}.projection.project_bn.bias''') ) rename_keys.append( (f'''block{b}_project_bn/moving_mean:0''', f'''encoder.blocks.{hf_b}.projection.project_bn.running_mean''') ) rename_keys.append( (f'''block{b}_project_bn/moving_variance:0''', f'''encoder.blocks.{hf_b}.projection.project_bn.running_var''') ) rename_keys.append(("""top_conv/kernel:0""", """encoder.top_conv.weight""") ) rename_keys.append(("""top_bn/gamma:0""", """encoder.top_bn.weight""") ) rename_keys.append(("""top_bn/beta:0""", """encoder.top_bn.bias""") ) rename_keys.append(("""top_bn/moving_mean:0""", """encoder.top_bn.running_mean""") ) rename_keys.append(("""top_bn/moving_variance:0""", """encoder.top_bn.running_var""") ) snake_case_ = {} for item in rename_keys: if item[0] in original_param_names: snake_case_ = """efficientnet.""" + item[1] snake_case_ = """classifier.weight""" snake_case_ = """classifier.bias""" return key_mapping def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ ) -> Tuple: '''simple docstring''' for key, value in tf_params.items(): if "normalization" in key: continue snake_case_ = key_mapping[key] if "_conv" in key and "kernel" in key: snake_case_ = torch.from_numpy(lowercase_ ).permute(3 , 2 , 0 , 1 ) elif "depthwise_kernel" in key: snake_case_ = torch.from_numpy(lowercase_ ).permute(2 , 3 , 0 , 1 ) elif "kernel" in key: snake_case_ = torch.from_numpy(np.transpose(lowercase_ ) ) else: snake_case_ = torch.from_numpy(lowercase_ ) # Replace HF parameters with original TF model parameters assert hf_params[hf_key].shape == new_hf_value.shape hf_params[hf_key].copy_(lowercase_ ) @torch.no_grad() def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> int: '''simple docstring''' snake_case_ = model_classes[model_name]( include_top=lowercase_ , weights="""imagenet""" , input_tensor=lowercase_ , input_shape=lowercase_ , pooling=lowercase_ , classes=1000 , classifier_activation="""softmax""" , ) snake_case_ = original_model.trainable_variables snake_case_ = original_model.non_trainable_variables snake_case_ = {param.name: param.numpy() for param in tf_params} for param in tf_non_train_params: snake_case_ = param.numpy() snake_case_ = list(tf_params.keys() ) # Load HuggingFace model snake_case_ = get_efficientnet_config(lowercase_ ) snake_case_ = EfficientNetForImageClassification(lowercase_ ).eval() snake_case_ = hf_model.state_dict() # Create src-to-dst parameter name mapping dictionary print("""Converting parameters...""" ) snake_case_ = rename_keys(lowercase_ ) replace_params(lowercase_ , lowercase_ , lowercase_ ) # Initialize preprocessor and preprocess input image snake_case_ = convert_image_processor(lowercase_ ) snake_case_ = preprocessor(images=prepare_img() , return_tensors="""pt""" ) # HF model inference hf_model.eval() with torch.no_grad(): snake_case_ = hf_model(**lowercase_ ) snake_case_ = outputs.logits.detach().numpy() # Original model inference snake_case_ = False snake_case_ = CONFIG_MAP[model_name]["""image_size"""] snake_case_ = prepare_img().resize((image_size, image_size) , resample=PIL.Image.NEAREST ) snake_case_ = image.img_to_array(lowercase_ ) snake_case_ = np.expand_dims(lowercase_ , axis=0 ) snake_case_ = original_model.predict(lowercase_ ) # Check whether original and HF model outputs match -> np.allclose assert np.allclose(lowercase_ , lowercase_ , atol=1e-3 ), "The predicted logits are not the same." print("""Model outputs match!""" ) if save_model: # Create folder to save model if not os.path.isdir(lowercase_ ): os.mkdir(lowercase_ ) # Save converted model and image processor hf_model.save_pretrained(lowercase_ ) preprocessor.save_pretrained(lowercase_ ) if push_to_hub: # Push model and image processor to hub print(f'''Pushing converted {model_name} to the hub...''' ) snake_case_ = f'''efficientnet-{model_name}''' preprocessor.push_to_hub(lowercase_ ) hf_model.push_to_hub(lowercase_ ) if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''b0''', type=str, help='''Version name of the EfficientNet model you want to convert, select from [b0, b1, b2, b3, b4, b5, b6, b7].''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''hf_model''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--save_model''', action='''store_true''', help='''Save model to local''') parser.add_argument('''--push_to_hub''', action='''store_true''', help='''Push model and image processor to the hub''') lowerCamelCase_ = parser.parse_args() convert_efficientnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.save_model, args.push_to_hub)
34
1
'''simple docstring''' import argparse from transformers import BigBirdConfig, BigBirdForPreTraining, BigBirdForQuestionAnswering, load_tf_weights_in_big_bird from transformers.utils import logging logging.set_verbosity_info() def UpperCamelCase ( _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Optional[Any] , _lowerCamelCase : Union[str, Any] ): A__ = BigBirdConfig.from_json_file(_lowercase ) print(F"Building PyTorch model from configuration: {config}" ) if is_trivia_qa: A__ = BigBirdForQuestionAnswering(_lowercase ) else: A__ = BigBirdForPreTraining(_lowercase ) # Load weights from tf checkpoint load_tf_weights_in_big_bird(_lowercase , _lowercase , is_trivia_qa=_lowercase ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(_lowercase ) if __name__ == "__main__": __lowerCAmelCase : Union[str, Any] =argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--big_bird_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained BERT model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--is_trivia_qa", action="store_true", help="Whether to convert a model with a trivia_qa head." ) __lowerCAmelCase : Any =parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.big_bird_config_file, args.pytorch_dump_path, args.is_trivia_qa )
237
import math import sys def lowerCAmelCase_ ( _lowercase : str) -> str: """simple docstring""" a__ : str = """""" try: with open(_lowercase , """rb""") as binary_file: a__ : Any = binary_file.read() for dat in data: a__ : Dict = F'''{dat:08b}''' result += curr_byte return result except OSError: print("""File not accessible""") sys.exit() def lowerCAmelCase_ ( _lowercase : str) -> str: """simple docstring""" a__ : Optional[Any] = {"""0""": """0""", """1""": """1"""} a__ , a__ : Optional[int] = """""", """""" a__ : int = len(_lowercase) for i in range(len(_lowercase)): curr_string += data_bits[i] if curr_string not in lexicon: continue a__ : List[str] = lexicon[curr_string] result += last_match_id a__ : Any = last_match_id + """0""" if math.loga(_lowercase).is_integer(): a__ : Union[str, Any] = {} for curr_key in list(_lowercase): a__ : Optional[Any] = lexicon.pop(_lowercase) a__ : Union[str, Any] = new_lex a__ : str = last_match_id + """1""" index += 1 a__ : List[Any] = """""" return result def lowerCAmelCase_ ( _lowercase : str , _lowercase : str) -> None: """simple docstring""" a__ : List[Any] = 8 try: with open(_lowercase , """wb""") as opened_file: a__ : Dict = [ to_write[i : i + byte_length] for i in range(0 , len(_lowercase) , _lowercase) ] if len(result_byte_array[-1]) % byte_length == 0: result_byte_array.append("""10000000""") else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1]) - 1 ) for elem in result_byte_array[:-1]: opened_file.write(int(_lowercase , 2).to_bytes(1 , byteorder="""big""")) except OSError: print("""File not accessible""") sys.exit() def lowerCAmelCase_ ( _lowercase : str) -> str: """simple docstring""" a__ : Any = 0 for letter in data_bits: if letter == "1": break counter += 1 a__ : Optional[Any] = data_bits[counter:] a__ : Tuple = data_bits[counter + 1 :] return data_bits def lowerCAmelCase_ ( _lowercase : str , _lowercase : str) -> None: """simple docstring""" a__ : Dict = read_file_binary(_lowercase) a__ : str = remove_prefix(_lowercase) a__ : List[str] = decompress_data(_lowercase) write_file_binary(_lowercase , _lowercase) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
170
0
import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class _lowercase ( unittest.TestCase ): def __init__( self : Union[str, Any] , snake_case : int , snake_case : Tuple = True , snake_case : Union[str, Any] = None , snake_case : int = 3_2 , snake_case : Optional[Any] = True , snake_case : Any = 1 / 2_5_5 , snake_case : Optional[Any] = True , snake_case : List[Any] = True , snake_case : Union[str, Any] = [0.48145466, 0.4578275, 0.40821073] , snake_case : List[Any] = [0.26862954, 0.26130258, 0.27577711] , snake_case : Union[str, Any] = True , snake_case : Any=7 , snake_case : List[Any]=3_0 , snake_case : Union[str, Any]=4_0_0 , snake_case : str=3 , ) -> Tuple: """simple docstring""" UpperCamelCase_ : int = parent UpperCamelCase_ : str = do_resize UpperCamelCase_ : Tuple = size if size is not None else {'shortest_edge': 2_8_8} UpperCamelCase_ : Any = size_divisor UpperCamelCase_ : str = do_rescale UpperCamelCase_ : Dict = rescale_factor UpperCamelCase_ : Union[str, Any] = do_normalize UpperCamelCase_ : Any = do_center_crop UpperCamelCase_ : Optional[Any] = image_mean UpperCamelCase_ : Tuple = image_std UpperCamelCase_ : Optional[int] = do_pad UpperCamelCase_ : Dict = batch_size UpperCamelCase_ : Tuple = num_channels UpperCamelCase_ : int = min_resolution UpperCamelCase_ : Optional[int] = max_resolution def SCREAMING_SNAKE_CASE__ ( self : int ) -> Union[str, Any]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , snake_case : Any , snake_case : Any=False ) -> Optional[Any]: """simple docstring""" if not batched: UpperCamelCase_ : Optional[int] = self.size['shortest_edge'] UpperCamelCase_ : Any = image_inputs[0] if isinstance(SCREAMING_SNAKE_CASE_ , Image.Image ): UpperCamelCase_ : List[Any] = image.size else: UpperCamelCase_ : Tuple = image.shape[1], image.shape[2] UpperCamelCase_ : Union[str, Any] = size / min(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if h < w: UpperCamelCase_ : int = size, scale * w else: UpperCamelCase_ : str = scale * h, size UpperCamelCase_ : int = int((1_3_3_3 / 8_0_0) * size ) if max(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) > max_size: UpperCamelCase_ : Union[str, Any] = max_size / max(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase_ : int = newh * scale UpperCamelCase_ : Any = neww * scale UpperCamelCase_ : List[str] = int(newh + 0.5 ), int(neww + 0.5 ) UpperCamelCase_ : Optional[int] = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: UpperCamelCase_ : Any = [] for image in image_inputs: UpperCamelCase_ : int = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCamelCase_ : Dict = max(SCREAMING_SNAKE_CASE_ , key=lambda snake_case : item[0] )[0] UpperCamelCase_ : Tuple = max(SCREAMING_SNAKE_CASE_ , key=lambda snake_case : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _lowercase ( a__ , unittest.TestCase ): lowercase = BridgeTowerImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE__ ( self : str ) -> Dict: """simple docstring""" UpperCamelCase_ : int = BridgeTowerImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> Optional[Any]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> str: """simple docstring""" UpperCamelCase_ : Dict = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , 'image_mean' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , 'image_std' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , 'do_normalize' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , 'do_resize' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , 'size' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , 'size_divisor' ) ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> int: """simple docstring""" pass def SCREAMING_SNAKE_CASE__ ( self : List[Any] ) -> Dict: """simple docstring""" UpperCamelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCamelCase_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE_ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE_ , Image.Image ) # Test not batched input UpperCamelCase_ : Optional[Any] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values UpperCamelCase_ : str = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase_ : Union[str, Any] = image_processing(SCREAMING_SNAKE_CASE_ , return_tensors='pt' ).pixel_values UpperCamelCase_ : Tuple = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> List[Any]: """simple docstring""" UpperCamelCase_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase_ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE_ , numpify=SCREAMING_SNAKE_CASE_ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE_ , np.ndarray ) # Test not batched input UpperCamelCase_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values UpperCamelCase_ : Optional[Any] = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase_ : List[str] = image_processing(SCREAMING_SNAKE_CASE_ , return_tensors='pt' ).pixel_values UpperCamelCase_ : Union[str, Any] = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def SCREAMING_SNAKE_CASE__ ( self : str ) -> Tuple: """simple docstring""" UpperCamelCase_ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase_ : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE_ , torchify=SCREAMING_SNAKE_CASE_ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE_ , torch.Tensor ) # Test not batched input UpperCamelCase_ : List[Any] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values UpperCamelCase_ : Tuple = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase_ : Any = image_processing(SCREAMING_SNAKE_CASE_ , return_tensors='pt' ).pixel_values UpperCamelCase_ : Any = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
364
import math import flax.linen as nn import jax.numpy as jnp def __lowercase ( lowerCamelCase : jnp.ndarray , lowerCamelCase : int , lowerCamelCase : float = 1 , lowerCamelCase : float = 1 , lowerCamelCase : float = 1.0e4 , lowerCamelCase : bool = False , lowerCamelCase : float = 1.0 , ): assert timesteps.ndim == 1, "Timesteps should be a 1d-array" assert embedding_dim % 2 == 0, F"Embedding dimension {embedding_dim} should be even" UpperCamelCase_ : Dict = float(embedding_dim // 2 ) UpperCamelCase_ : Dict = math.log(max_timescale / min_timescale ) / (num_timescales - freq_shift) UpperCamelCase_ : Optional[int] = min_timescale * jnp.exp(jnp.arange(lowerCamelCase , dtype=jnp.floataa ) * -log_timescale_increment ) UpperCamelCase_ : int = jnp.expand_dims(lowerCamelCase , 1 ) * jnp.expand_dims(lowerCamelCase , 0 ) # scale embeddings UpperCamelCase_ : Tuple = scale * emb if flip_sin_to_cos: UpperCamelCase_ : Tuple = jnp.concatenate([jnp.cos(lowerCamelCase ), jnp.sin(lowerCamelCase )] , axis=1 ) else: UpperCamelCase_ : Optional[int] = jnp.concatenate([jnp.sin(lowerCamelCase ), jnp.cos(lowerCamelCase )] , axis=1 ) UpperCamelCase_ : Optional[Any] = jnp.reshape(lowerCamelCase , [jnp.shape(lowerCamelCase )[0], embedding_dim] ) return signal class _lowercase ( nn.Module ): lowercase = 3_2 lowercase = jnp.floataa @nn.compact def __call__( self : str , snake_case : Union[str, Any] ) -> Dict: """simple docstring""" UpperCamelCase_ : str = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_1' )(snake_case ) UpperCamelCase_ : int = nn.silu(snake_case ) UpperCamelCase_ : Optional[Any] = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_2' )(snake_case ) return temb class _lowercase ( nn.Module ): lowercase = 3_2 lowercase = False lowercase = 1 @nn.compact def __call__( self : int , snake_case : Any ) -> str: """simple docstring""" return get_sinusoidal_embeddings( snake_case , embedding_dim=self.dim , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.freq_shift )
50
0
from string import ascii_uppercase __snake_case :str = {str(ord(c) - 55): c for c in ascii_uppercase} def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): if isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('''int() can\'t convert non-string with explicit base''' ) if num < 0: raise ValueError('''parameter must be positive int''' ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('''\'str\' object cannot be interpreted as an integer''' ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError('''\'float\' object cannot be interpreted as an integer''' ) if base in (0, 1): raise ValueError('''base must be >= 2''' ) if base > 36: raise ValueError('''base must be <= 36''' ) __a = '''''' __a = 0 __a = 0 while div != 1: __a , __a = divmod(_UpperCAmelCase , _UpperCAmelCase ) if base >= 11 and 9 < mod < 36: __a = ALPHABET_VALUES[str(_UpperCAmelCase )] else: __a = str(_UpperCAmelCase ) new_value += actual_value __a = num // base __a = div if div == 0: return str(new_value[::-1] ) elif div == 1: new_value += str(_UpperCAmelCase ) return str(new_value[::-1] ) return new_value[::-1] if __name__ == "__main__": import doctest doctest.testmod() for base in range(2, 37): for num in range(1000): assert int(decimal_to_any(num, base), base) == num, ( num, base, decimal_to_any(num, base), int(decimal_to_any(num, base), base), )
49
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer __snake_case :str = logging.get_logger(__name__) __snake_case :int = {'''vocab_file''': '''vocab.txt'''} __snake_case :List[Any] = { '''vocab_file''': { '''YituTech/conv-bert-base''': '''https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt''', '''YituTech/conv-bert-medium-small''': ( '''https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt''' ), '''YituTech/conv-bert-small''': '''https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt''', } } __snake_case :List[str] = { '''YituTech/conv-bert-base''': 512, '''YituTech/conv-bert-medium-small''': 512, '''YituTech/conv-bert-small''': 512, } __snake_case :Optional[int] = { '''YituTech/conv-bert-base''': {'''do_lower_case''': True}, '''YituTech/conv-bert-medium-small''': {'''do_lower_case''': True}, '''YituTech/conv-bert-small''': {'''do_lower_case''': True}, } class _A ( __UpperCAmelCase ): UpperCamelCase__ : Optional[int] = VOCAB_FILES_NAMES UpperCamelCase__ : str = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : Optional[Any] = PRETRAINED_INIT_CONFIGURATION UpperCamelCase__ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : int = ConvBertTokenizer def __init__( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=None , __SCREAMING_SNAKE_CASE : Dict=None , __SCREAMING_SNAKE_CASE : Tuple=True , __SCREAMING_SNAKE_CASE : Optional[int]="[UNK]" , __SCREAMING_SNAKE_CASE : int="[SEP]" , __SCREAMING_SNAKE_CASE : List[Any]="[PAD]" , __SCREAMING_SNAKE_CASE : int="[CLS]" , __SCREAMING_SNAKE_CASE : Optional[int]="[MASK]" , __SCREAMING_SNAKE_CASE : Optional[int]=True , __SCREAMING_SNAKE_CASE : Union[str, Any]=None , **__SCREAMING_SNAKE_CASE : Dict , ): '''simple docstring''' super().__init__( __SCREAMING_SNAKE_CASE , tokenizer_file=__SCREAMING_SNAKE_CASE , do_lower_case=__SCREAMING_SNAKE_CASE , unk_token=__SCREAMING_SNAKE_CASE , sep_token=__SCREAMING_SNAKE_CASE , pad_token=__SCREAMING_SNAKE_CASE , cls_token=__SCREAMING_SNAKE_CASE , mask_token=__SCREAMING_SNAKE_CASE , tokenize_chinese_chars=__SCREAMING_SNAKE_CASE , strip_accents=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) __a = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( normalizer_state.get('''lowercase''' , __SCREAMING_SNAKE_CASE) != do_lower_case or normalizer_state.get('''strip_accents''' , __SCREAMING_SNAKE_CASE) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , __SCREAMING_SNAKE_CASE) != tokenize_chinese_chars ): __a = getattr(__SCREAMING_SNAKE_CASE , normalizer_state.pop('''type''')) __a = do_lower_case __a = strip_accents __a = tokenize_chinese_chars __a = normalizer_class(**__SCREAMING_SNAKE_CASE) __a = do_lower_case def _lowerCamelCase ( self : int , __SCREAMING_SNAKE_CASE : Dict , __SCREAMING_SNAKE_CASE : Any=None): '''simple docstring''' __a = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowerCamelCase ( self : str , __SCREAMING_SNAKE_CASE : List[int] , __SCREAMING_SNAKE_CASE : Optional[List[int]] = None): '''simple docstring''' __a = [self.sep_token_id] __a = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def _lowerCamelCase ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[str] = None): '''simple docstring''' __a = self._tokenizer.model.save(__SCREAMING_SNAKE_CASE , name=__SCREAMING_SNAKE_CASE) return tuple(__SCREAMING_SNAKE_CASE)
49
1
from __future__ import annotations class __lowercase : '''simple docstring''' def __init__( self : List[Any] , _a : int ): UpperCamelCase__ = data UpperCamelCase__ = None UpperCamelCase__ = None def lowerCamelCase_ ( UpperCamelCase__ : Node | None ): # In Order traversal of the tree '''simple docstring''' if tree: display(tree.left ) print(tree.data ) display(tree.right ) def lowerCamelCase_ ( UpperCamelCase__ : Node | None ): '''simple docstring''' return 1 + max(depth_of_tree(tree.left ), depth_of_tree(tree.right ) ) if tree else 0 def lowerCamelCase_ ( UpperCamelCase__ : Node ): '''simple docstring''' if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def lowerCamelCase_ ( ): # Main function for testing. '''simple docstring''' UpperCamelCase__ = Node(1 ) UpperCamelCase__ = Node(2 ) UpperCamelCase__ = Node(3 ) UpperCamelCase__ = Node(4 ) UpperCamelCase__ = Node(5 ) UpperCamelCase__ = Node(6 ) UpperCamelCase__ = Node(7 ) UpperCamelCase__ = Node(8 ) UpperCamelCase__ = Node(9 ) print(is_full_binary_tree(UpperCamelCase__ ) ) print(depth_of_tree(UpperCamelCase__ ) ) print('''Tree is: ''' ) display(UpperCamelCase__ ) if __name__ == "__main__": main()
35
import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor lowercase = logging.get_logger(__name__) class __lowercase ( A ): '''simple docstring''' def __init__( self : Any , *_a : Optional[Any] , **_a : Any ): warnings.warn( '''The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use CLIPImageProcessor instead.''' , _a , ) super().__init__(*_a , **_a )
35
1
'''simple docstring''' import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Optional[int] = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece @require_tokenizers class lowerCamelCase_ (A__ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase: str = XLMRobertaTokenizer __UpperCamelCase: Any = XLMRobertaTokenizerFast __UpperCamelCase: Union[str, Any] = True __UpperCamelCase: int = True def _A ( self : Union[str, Any] ): super().setUp() # We have a SentencePiece fixture for testing _UpperCAmelCase : Any = XLMRobertaTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def _A ( self : Optional[int] ): _UpperCAmelCase : Any = "<pad>" _UpperCAmelCase : Optional[int] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCAmelCase ) , __UpperCAmelCase ) def _A ( self : Optional[Any] ): _UpperCAmelCase : Dict = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<s>" ) self.assertEqual(vocab_keys[1] , "<pad>" ) self.assertEqual(vocab_keys[-1] , "<mask>" ) self.assertEqual(len(__UpperCAmelCase ) , 1002 ) def _A ( self : List[Any] ): self.assertEqual(self.get_tokenizer().vocab_size , 1002 ) def _A ( self : int ): _UpperCAmelCase : Any = XLMRobertaTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) _UpperCAmelCase : Union[str, Any] = tokenizer.tokenize("This is a test" ) self.assertListEqual(__UpperCAmelCase , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) _UpperCAmelCase : Union[str, Any] = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) _UpperCAmelCase : Optional[Any] = tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) _UpperCAmelCase : int = tokenizer.convert_ids_to_tokens(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def _A ( self : Tuple ): if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _UpperCAmelCase : Optional[int] = (self.rust_tokenizer_class, "hf-internal-testing/tiny-xlm-roberta", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _UpperCAmelCase : Tuple = self.rust_tokenizer_class.from_pretrained(__UpperCAmelCase , **__UpperCAmelCase ) _UpperCAmelCase : str = self.tokenizer_class.from_pretrained(__UpperCAmelCase , **__UpperCAmelCase ) _UpperCAmelCase : Optional[Any] = tempfile.mkdtemp() _UpperCAmelCase : List[str] = tokenizer_r.save_pretrained(__UpperCAmelCase ) _UpperCAmelCase : Union[str, Any] = tokenizer_p.save_pretrained(__UpperCAmelCase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) _UpperCAmelCase : Any = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f ) self.assertSequenceEqual(__UpperCAmelCase , __UpperCAmelCase ) # Checks everything loads correctly in the same way _UpperCAmelCase : Optional[Any] = tokenizer_r.from_pretrained(__UpperCAmelCase ) _UpperCAmelCase : str = tokenizer_p.from_pretrained(__UpperCAmelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__UpperCAmelCase , __UpperCAmelCase ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(__UpperCAmelCase ) # Save tokenizer rust, legacy_format=True _UpperCAmelCase : Optional[int] = tempfile.mkdtemp() _UpperCAmelCase : Optional[Any] = tokenizer_r.save_pretrained(__UpperCAmelCase , legacy_format=__UpperCAmelCase ) _UpperCAmelCase : Union[str, Any] = tokenizer_p.save_pretrained(__UpperCAmelCase ) # Checks it save with the same files self.assertSequenceEqual(__UpperCAmelCase , __UpperCAmelCase ) # Checks everything loads correctly in the same way _UpperCAmelCase : Tuple = tokenizer_r.from_pretrained(__UpperCAmelCase ) _UpperCAmelCase : Tuple = tokenizer_p.from_pretrained(__UpperCAmelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__UpperCAmelCase , __UpperCAmelCase ) ) shutil.rmtree(__UpperCAmelCase ) # Save tokenizer rust, legacy_format=False _UpperCAmelCase : List[Any] = tempfile.mkdtemp() _UpperCAmelCase : List[Any] = tokenizer_r.save_pretrained(__UpperCAmelCase , legacy_format=__UpperCAmelCase ) _UpperCAmelCase : Any = tokenizer_p.save_pretrained(__UpperCAmelCase ) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _UpperCAmelCase : Optional[int] = tokenizer_r.from_pretrained(__UpperCAmelCase ) _UpperCAmelCase : Any = tokenizer_p.from_pretrained(__UpperCAmelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__UpperCAmelCase , __UpperCAmelCase ) ) shutil.rmtree(__UpperCAmelCase ) @cached_property def _A ( self : Any ): return XLMRobertaTokenizer.from_pretrained("xlm-roberta-base" ) def _A ( self : Union[str, Any] ): with tempfile.NamedTemporaryFile() as f: shutil.copyfile(__UpperCAmelCase , f.name ) _UpperCAmelCase : str = XLMRobertaTokenizer(f.name , keep_accents=__UpperCAmelCase ) _UpperCAmelCase : int = pickle.dumps(__UpperCAmelCase ) pickle.loads(__UpperCAmelCase ) def _A ( self : List[str] ): if not self.test_rust_tokenizer: return _UpperCAmelCase : Dict = self.get_tokenizer() _UpperCAmelCase : Union[str, Any] = self.get_rust_tokenizer() _UpperCAmelCase : Optional[Any] = "I was born in 92000, and this is falsé." _UpperCAmelCase : Optional[int] = tokenizer.tokenize(__UpperCAmelCase ) _UpperCAmelCase : Optional[int] = rust_tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) _UpperCAmelCase : Optional[int] = tokenizer.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) _UpperCAmelCase : Optional[int] = rust_tokenizer.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) _UpperCAmelCase : Union[str, Any] = self.get_rust_tokenizer() _UpperCAmelCase : Optional[int] = tokenizer.encode(__UpperCAmelCase ) _UpperCAmelCase : Dict = rust_tokenizer.encode(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) @slow def _A ( self : Tuple ): _UpperCAmelCase : Tuple = "Hello World!" _UpperCAmelCase : Optional[Any] = [0, 35378, 6661, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__UpperCAmelCase , self.big_tokenizer.encode(__UpperCAmelCase ) ) @slow def _A ( self : List[Any] ): _UpperCAmelCase : Optional[Any] = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) _UpperCAmelCase : Any = [ 0, 3293, 83, 10, 4552, 4989, 7986, 678, 10, 5915, 111, 179459, 124850, 4, 6044, 237, 12, 6, 5, 6, 4, 6780, 705, 15, 1388, 44, 378, 10114, 711, 152, 20, 6, 5, 22376, 642, 1221, 15190, 34153, 450, 5608, 959, 1119, 57702, 136, 186, 47, 1098, 29367, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6044, 237, 6284, 50901, 528, 31, 90, 34, 927, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__UpperCAmelCase , self.big_tokenizer.encode(__UpperCAmelCase ) ) @slow def _A ( self : Any ): # fmt: off _UpperCAmelCase : Tuple = {"input_ids": [[0, 11062, 82772, 7, 15, 82772, 538, 51529, 237, 17198, 1290, 206, 9, 215175, 1314, 136, 17198, 1290, 206, 9, 56359, 42, 122009, 9, 16466, 16, 87344, 4537, 9, 4717, 78381, 6, 159958, 7, 15, 24480, 618, 4, 527, 22693, 5428, 4, 2777, 24480, 9874, 4, 43523, 594, 4, 803, 18392, 33189, 18, 4, 43523, 24447, 12399, 100, 24955, 83658, 9626, 144057, 15, 839, 22335, 16, 136, 24955, 83658, 83479, 15, 39102, 724, 16, 678, 645, 2789, 1328, 4589, 42, 122009, 115774, 23, 805, 1328, 46876, 7, 136, 53894, 1940, 42227, 41159, 17721, 823, 425, 4, 27512, 98722, 206, 136, 5531, 4970, 919, 17336, 5, 2], [0, 20080, 618, 83, 82775, 47, 479, 9, 1517, 73, 53894, 333, 80581, 110117, 18811, 5256, 1295, 51, 152526, 297, 7986, 390, 124416, 538, 35431, 214, 98, 15044, 25737, 136, 7108, 43701, 23, 756, 135355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 581, 63773, 119455, 6, 147797, 88203, 7, 645, 70, 21, 3285, 10269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__UpperCAmelCase , model_name="xlm-roberta-base" , revision="d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3" , )
31
"""simple docstring""" import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class lowerCamelCase (unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE ( self : Dict ) -> Dict: SCREAMING_SNAKE_CASE__ = AutoImageProcessor.from_pretrained("""microsoft/dit-base-finetuned-rvlcdip""" ) SCREAMING_SNAKE_CASE__ = AutoModelForImageClassification.from_pretrained("""microsoft/dit-base-finetuned-rvlcdip""" ) model.to(__UpperCAmelCase ) from datasets import load_dataset SCREAMING_SNAKE_CASE__ = load_dataset("""nielsr/rvlcdip-demo""" ) SCREAMING_SNAKE_CASE__ = dataset["""train"""][0]["""image"""].convert("""RGB""" ) SCREAMING_SNAKE_CASE__ = image_processor(__UpperCAmelCase , return_tensors="""pt""" ).to(__UpperCAmelCase ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**__UpperCAmelCase ) SCREAMING_SNAKE_CASE__ = outputs.logits SCREAMING_SNAKE_CASE__ = torch.Size((1, 1_6) ) self.assertEqual(logits.shape , __UpperCAmelCase ) SCREAMING_SNAKE_CASE__ = torch.tensor( [-0.4_158, -0.4_092, -0.4_347] , device=__UpperCAmelCase , dtype=torch.float , ) self.assertTrue(torch.allclose(logits[0, :3] , __UpperCAmelCase , atol=1e-4 ) )
165
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __a = { 'configuration_deberta': ['DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'DebertaConfig', 'DebertaOnnxConfig'], 'tokenization_deberta': ['DebertaTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ['DebertaTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ 'DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'DebertaForMaskedLM', 'DebertaForQuestionAnswering', 'DebertaForSequenceClassification', 'DebertaForTokenClassification', 'DebertaModel', 'DebertaPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ 'TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFDebertaForMaskedLM', 'TFDebertaForQuestionAnswering', 'TFDebertaForSequenceClassification', 'TFDebertaForTokenClassification', 'TFDebertaModel', 'TFDebertaPreTrainedModel', ] if TYPE_CHECKING: from .configuration_deberta import DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, DebertaConfig, DebertaOnnxConfig from .tokenization_deberta import DebertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_deberta_fast import DebertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deberta import ( DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, DebertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deberta import ( TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFDebertaForMaskedLM, TFDebertaForQuestionAnswering, TFDebertaForSequenceClassification, TFDebertaForTokenClassification, TFDebertaModel, TFDebertaPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
356
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) __a = { 'configuration_layoutlmv2': ['LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv2Config'], 'processing_layoutlmv2': ['LayoutLMv2Processor'], 'tokenization_layoutlmv2': ['LayoutLMv2Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ['LayoutLMv2TokenizerFast'] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ['LayoutLMv2FeatureExtractor'] __a = ['LayoutLMv2ImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ 'LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv2ForQuestionAnswering', 'LayoutLMv2ForSequenceClassification', 'LayoutLMv2ForTokenClassification', 'LayoutLMv2Layer', 'LayoutLMv2Model', 'LayoutLMv2PreTrainedModel', ] if TYPE_CHECKING: from .configuration_layoutlmva import LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor, LayoutLMvaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaLayer, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
17
0
import os import unittest from transformers import MobileBertTokenizer, MobileBertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = MobileBertTokenizer __UpperCAmelCase : int = MobileBertTokenizerFast __UpperCAmelCase : Union[str, Any] = True __UpperCAmelCase : Union[str, Any] = True __UpperCAmelCase : Tuple = filter_non_english __UpperCAmelCase : Tuple = "google/mobilebert-uncased" def __snake_case ( self : Union[str, Any] ) -> List[str]: super().setUp() __snake_case : int = [ "[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] __snake_case : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) __snake_case : Optional[int] = [ (tokenizer_def[0], self.pre_trained_model_path, tokenizer_def[2]) # else the 'google/' prefix is stripped for tokenizer_def in self.tokenizers_list ] def __snake_case ( self : int , lowerCamelCase : Dict ) -> List[str]: __snake_case : List[Any] = "UNwant\u00E9d,running" __snake_case : Optional[int] = "unwanted, running" return input_text, output_text def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : Optional[Any] = self.tokenizer_class(self.vocab_file ) __snake_case : Optional[int] = tokenizer.tokenize("UNwant\u00E9d,running" ) self.assertListEqual(lowerCamelCase , ["un", "##want", "##ed", ",", "runn", "##ing"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase ) , [9, 6, 7, 12, 10, 11] ) def __snake_case ( self : List[str] ) -> Optional[int]: if not self.test_rust_tokenizer: return __snake_case : List[Any] = self.get_tokenizer() __snake_case : List[str] = self.get_rust_tokenizer() __snake_case : Tuple = "UNwant\u00E9d,running" __snake_case : Optional[int] = tokenizer.tokenize(lowerCamelCase ) __snake_case : str = rust_tokenizer.tokenize(lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) __snake_case : Union[str, Any] = tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) __snake_case : Optional[int] = rust_tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) __snake_case : Any = self.get_rust_tokenizer() __snake_case : Dict = tokenizer.encode(lowerCamelCase ) __snake_case : Optional[int] = rust_tokenizer.encode(lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) # With lower casing __snake_case : Any = self.get_tokenizer(do_lower_case=lowerCamelCase ) __snake_case : List[str] = self.get_rust_tokenizer(do_lower_case=lowerCamelCase ) __snake_case : Any = "UNwant\u00E9d,running" __snake_case : int = tokenizer.tokenize(lowerCamelCase ) __snake_case : Tuple = rust_tokenizer.tokenize(lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) __snake_case : str = tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) __snake_case : Tuple = rust_tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) __snake_case : str = self.get_rust_tokenizer() __snake_case : Optional[Any] = tokenizer.encode(lowerCamelCase ) __snake_case : List[str] = rust_tokenizer.encode(lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : List[str] = BasicTokenizer() self.assertListEqual(tokenizer.tokenize("ah\u535A\u63A8zz" ) , ["ah", "\u535A", "\u63A8", "zz"] ) def __snake_case ( self : str ) -> Dict: __snake_case : Tuple = BasicTokenizer(do_lower_case=lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ) , ["hello", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def __snake_case ( self : Dict ) -> int: __snake_case : List[str] = BasicTokenizer(do_lower_case=lowerCamelCase , strip_accents=lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hällo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["h\u00E9llo"] ) def __snake_case ( self : List[str] ) -> str: __snake_case : int = BasicTokenizer(do_lower_case=lowerCamelCase , strip_accents=lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def __snake_case ( self : Tuple ) -> Optional[int]: __snake_case : Dict = BasicTokenizer(do_lower_case=lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def __snake_case ( self : List[str] ) -> Any: __snake_case : Optional[Any] = BasicTokenizer(do_lower_case=lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ) , ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def __snake_case ( self : Optional[Any] ) -> str: __snake_case : str = BasicTokenizer(do_lower_case=lowerCamelCase , strip_accents=lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["HäLLo", "!", "how", "Are", "yoU", "?"] ) def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : Optional[Any] = BasicTokenizer(do_lower_case=lowerCamelCase , strip_accents=lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["HaLLo", "!", "how", "Are", "yoU", "?"] ) def __snake_case ( self : int ) -> Union[str, Any]: __snake_case : int = BasicTokenizer(do_lower_case=lowerCamelCase , never_split=["[UNK]"] ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? [UNK]" ) , ["HeLLo", "!", "how", "Are", "yoU", "?", "[UNK]"] ) def __snake_case ( self : Any ) -> str: __snake_case : Optional[Any] = ["[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing"] __snake_case : List[str] = {} for i, token in enumerate(lowerCamelCase ): __snake_case : str = i __snake_case : Tuple = WordpieceTokenizer(vocab=lowerCamelCase , unk_token="[UNK]" ) self.assertListEqual(tokenizer.tokenize("" ) , [] ) self.assertListEqual(tokenizer.tokenize("unwanted running" ) , ["un", "##want", "##ed", "runn", "##ing"] ) self.assertListEqual(tokenizer.tokenize("unwantedX running" ) , ["[UNK]", "runn", "##ing"] ) def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: self.assertTrue(_is_whitespace(" " ) ) self.assertTrue(_is_whitespace("\t" ) ) self.assertTrue(_is_whitespace("\r" ) ) self.assertTrue(_is_whitespace("\n" ) ) self.assertTrue(_is_whitespace("\u00A0" ) ) self.assertFalse(_is_whitespace("A" ) ) self.assertFalse(_is_whitespace("-" ) ) def __snake_case ( self : Optional[int] ) -> Union[str, Any]: self.assertTrue(_is_control("\u0005" ) ) self.assertFalse(_is_control("A" ) ) self.assertFalse(_is_control(" " ) ) self.assertFalse(_is_control("\t" ) ) self.assertFalse(_is_control("\r" ) ) def __snake_case ( self : Dict ) -> Dict: self.assertTrue(_is_punctuation("-" ) ) self.assertTrue(_is_punctuation("$" ) ) self.assertTrue(_is_punctuation("`" ) ) self.assertTrue(_is_punctuation("." ) ) self.assertFalse(_is_punctuation("A" ) ) self.assertFalse(_is_punctuation(" " ) ) def __snake_case ( self : int ) -> Dict: __snake_case : List[str] = self.get_tokenizer() __snake_case : Dict = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(lowerCamelCase ) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]] ) self.assertListEqual( [rust_tokenizer.tokenize(lowerCamelCase ) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]] ) @slow def __snake_case ( self : Dict ) -> int: __snake_case : Optional[Any] = self.tokenizer_class.from_pretrained("google/mobilebert-uncased" ) __snake_case : Optional[int] = tokenizer.encode("sequence builders" , add_special_tokens=lowerCamelCase ) __snake_case : List[str] = tokenizer.encode("multi-sequence build" , add_special_tokens=lowerCamelCase ) __snake_case : Tuple = tokenizer.build_inputs_with_special_tokens(lowerCamelCase ) __snake_case : str = tokenizer.build_inputs_with_special_tokens(lowerCamelCase , lowerCamelCase ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def __snake_case ( self : Optional[Any] ) -> int: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): __snake_case : str = self.rust_tokenizer_class.from_pretrained(lowerCamelCase , **lowerCamelCase ) __snake_case : Dict = F'A, naïve {tokenizer_r.mask_token} AllenNLP sentence.' __snake_case : Tuple = tokenizer_r.encode_plus( lowerCamelCase , return_attention_mask=lowerCamelCase , return_token_type_ids=lowerCamelCase , return_offsets_mapping=lowerCamelCase , add_special_tokens=lowerCamelCase , ) __snake_case : Tuple = tokenizer_r.do_lower_case if hasattr(lowerCamelCase , "do_lower_case" ) else False __snake_case : Optional[Any] = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "A"), ((1, 2), ","), ((3, 5), "na"), ((5, 6), "##ï"), ((6, 8), "##ve"), ((9, 15), tokenizer_r.mask_token), ((16, 21), "Allen"), ((21, 23), "##NL"), ((23, 24), "##P"), ((25, 33), "sentence"), ((33, 34), "."), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "a"), ((1, 2), ","), ((3, 8), "naive"), ((9, 15), tokenizer_r.mask_token), ((16, 21), "allen"), ((21, 23), "##nl"), ((23, 24), "##p"), ((25, 33), "sentence"), ((33, 34), "."), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["input_ids"] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["offset_mapping"] ) def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : List[Any] = ["的", "人", "有"] __snake_case : Tuple = "".join(lowerCamelCase ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): __snake_case : int = True __snake_case : Tuple = self.tokenizer_class.from_pretrained(lowerCamelCase , **lowerCamelCase ) __snake_case : Any = self.rust_tokenizer_class.from_pretrained(lowerCamelCase , **lowerCamelCase ) __snake_case : int = tokenizer_p.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) __snake_case : Union[str, Any] = tokenizer_r.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) __snake_case : Optional[Any] = tokenizer_r.convert_ids_to_tokens(lowerCamelCase ) __snake_case : Tuple = tokenizer_p.convert_ids_to_tokens(lowerCamelCase ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(lowerCamelCase , lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) __snake_case : int = False __snake_case : Any = self.rust_tokenizer_class.from_pretrained(lowerCamelCase , **lowerCamelCase ) __snake_case : Optional[Any] = self.tokenizer_class.from_pretrained(lowerCamelCase , **lowerCamelCase ) __snake_case : Dict = tokenizer_r.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) __snake_case : Any = tokenizer_p.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) __snake_case : Union[str, Any] = tokenizer_r.convert_ids_to_tokens(lowerCamelCase ) __snake_case : Optional[Any] = tokenizer_p.convert_ids_to_tokens(lowerCamelCase ) # it is expected that only the first Chinese character is not preceded by "##". __snake_case : Optional[int] = [ F'##{token}' if idx != 0 else token for idx, token in enumerate(lowerCamelCase ) ] self.assertListEqual(lowerCamelCase , lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase )
123
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _snake_case : Optional[int] = { "configuration_distilbert": [ "DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DistilBertConfig", "DistilBertOnnxConfig", ], "tokenization_distilbert": ["DistilBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Union[str, Any] = ["DistilBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = [ "DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "DistilBertForMaskedLM", "DistilBertForMultipleChoice", "DistilBertForQuestionAnswering", "DistilBertForSequenceClassification", "DistilBertForTokenClassification", "DistilBertModel", "DistilBertPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[str] = [ "TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFDistilBertForMaskedLM", "TFDistilBertForMultipleChoice", "TFDistilBertForQuestionAnswering", "TFDistilBertForSequenceClassification", "TFDistilBertForTokenClassification", "TFDistilBertMainLayer", "TFDistilBertModel", "TFDistilBertPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : str = [ "FlaxDistilBertForMaskedLM", "FlaxDistilBertForMultipleChoice", "FlaxDistilBertForQuestionAnswering", "FlaxDistilBertForSequenceClassification", "FlaxDistilBertForTokenClassification", "FlaxDistilBertModel", "FlaxDistilBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys _snake_case : Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
123
1
import argparse import json import subprocess def _lowerCAmelCase ( A__: Any , A__: Optional[int] ): '''simple docstring''' UpperCAmelCase = [] UpperCAmelCase = ( F"""curl -H \"Accept: application/vnd.github+json\" -H \"Authorization: Bearer {token}\"""" ''' https://api.github.com/repos/huggingface/transformers/actions/runners''' ) UpperCAmelCase = subprocess.run(A__ , shell=A__ , stdout=subprocess.PIPE ) UpperCAmelCase = output.stdout.decode('''utf-8''' ) UpperCAmelCase = json.loads(A__ ) UpperCAmelCase = status['''runners'''] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(A__ ) # save the result so we can report them on Slack with open('''offline_runners.txt''' , '''w''' ) as fp: fp.write(json.dumps(A__ ) ) if len(A__ ) > 0: UpperCAmelCase = '''\n'''.join([x['''name'''] for x in offline_runners] ) raise ValueError(F"""The following runners are offline:\n{failed}""" ) if __name__ == "__main__": def _lowerCAmelCase ( A__: List[str] ): '''simple docstring''' return values.split(''',''' ) __magic_name__ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--target_runners", default=None, type=list_str, required=True, help="Comma-separated list of runners to check status.", ) parser.add_argument( "--token", default=None, type=str, required=True, help="A token that has actions:read permission." ) __magic_name__ = parser.parse_args() get_runner_status(args.target_runners, args.token)
152
def _lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase = [31, 28, 31, 30, 31, 30, 31, 31, 30, 31, 30, 31] UpperCAmelCase = 6 UpperCAmelCase = 1 UpperCAmelCase = 1901 UpperCAmelCase = 0 while year < 2001: day += 7 if (year % 4 == 0 and year % 100 != 0) or (year % 400 == 0): if day > days_per_month[month - 1] and month != 2: month += 1 UpperCAmelCase = day - days_per_month[month - 2] elif day > 29 and month == 2: month += 1 UpperCAmelCase = day - 29 else: if day > days_per_month[month - 1]: month += 1 UpperCAmelCase = day - days_per_month[month - 2] if month > 12: year += 1 UpperCAmelCase = 1 if year < 2001 and day == 1: sundays += 1 return sundays if __name__ == "__main__": print(solution())
152
1
from __future__ import annotations from collections.abc import Callable from typing import Any, Generic, TypeVar __A : Any = TypeVar('''T''') class __A ( Generic[T] ): def __init__( self : Dict , UpperCAmelCase_ : list[T] , UpperCAmelCase_ : Callable[[T, T], T] ): lowerCAmelCase : Any | T = None lowerCAmelCase : int = len(UpperCAmelCase_ ) lowerCAmelCase : list[T] = [any_type for _ in range(self.N )] + arr lowerCAmelCase : List[Any] = fnc self.build() def lowercase__ ( self : str ): for p in range(self.N - 1 , 0 , -1 ): lowerCAmelCase : Optional[Any] = self.fn(self.st[p * 2] , self.st[p * 2 + 1] ) def lowercase__ ( self : int , UpperCAmelCase_ : int , UpperCAmelCase_ : T ): p += self.N lowerCAmelCase : int = v while p > 1: lowerCAmelCase : List[Any] = p // 2 lowerCAmelCase : List[Any] = self.fn(self.st[p * 2] , self.st[p * 2 + 1] ) def lowercase__ ( self : Optional[Any] , UpperCAmelCase_ : int , UpperCAmelCase_ : int ): # noqa: E741 lowerCAmelCase , lowerCAmelCase : str = l + self.N, r + self.N lowerCAmelCase : T | None = None while l <= r: if l % 2 == 1: lowerCAmelCase : Any = self.st[l] if res is None else self.fn(UpperCAmelCase_ , self.st[l] ) if r % 2 == 0: lowerCAmelCase : Optional[int] = self.st[r] if res is None else self.fn(UpperCAmelCase_ , self.st[r] ) lowerCAmelCase , lowerCAmelCase : Optional[Any] = (l + 1) // 2, (r - 1) // 2 return res if __name__ == "__main__": from functools import reduce __A : str = [1, 10, -2, 9, -3, 8, 4, -7, 5, 6, 11, -12] __A : List[Any] = { 0: 7, 1: 2, 2: 6, 3: -14, 4: 5, 5: 4, 6: 7, 7: -10, 8: 9, 9: 10, 10: 12, 11: 1, } __A : Optional[int] = SegmentTree(test_array, min) __A : Optional[int] = SegmentTree(test_array, max) __A : Dict = SegmentTree(test_array, lambda a, b: a + b) def SCREAMING_SNAKE_CASE__ ( ) -> None: '''simple docstring''' for i in range(len(_UpperCAmelCase ) ): for j in range(_UpperCAmelCase, len(_UpperCAmelCase ) ): lowerCAmelCase : str = reduce(_UpperCAmelCase, test_array[i : j + 1] ) lowerCAmelCase : Dict = reduce(_UpperCAmelCase, test_array[i : j + 1] ) lowerCAmelCase : str = reduce(lambda _UpperCAmelCase, _UpperCAmelCase : a + b, test_array[i : j + 1] ) assert min_range == min_segment_tree.query(_UpperCAmelCase, _UpperCAmelCase ) assert max_range == max_segment_tree.query(_UpperCAmelCase, _UpperCAmelCase ) assert sum_range == sum_segment_tree.query(_UpperCAmelCase, _UpperCAmelCase ) test_all_segments() for index, value in test_updates.items(): __A : int = value min_segment_tree.update(index, value) max_segment_tree.update(index, value) sum_segment_tree.update(index, value) test_all_segments()
138
from pathlib import Path import cva import numpy as np from matplotlib import pyplot as plt def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) -> np.ndarray: '''simple docstring''' lowerCAmelCase : Dict = cva.getAffineTransform(_UpperCAmelCase, _UpperCAmelCase ) return cva.warpAffine(_UpperCAmelCase, _UpperCAmelCase, (rows, cols) ) if __name__ == "__main__": # read original image __A : List[str] = cva.imread( str(Path(__file__).resolve().parent.parent / '''image_data''' / '''lena.jpg''') ) # turn image in gray scale value __A : int = cva.cvtColor(image, cva.COLOR_BGR2GRAY) # get image shape __A , __A : Optional[Any] = gray_img.shape # set different points to rotate image __A : int = np.array([[50, 50], [200, 50], [50, 200]], np.floataa) __A : Any = np.array([[10, 100], [200, 50], [100, 250]], np.floataa) __A : Optional[int] = np.array([[50, 50], [150, 50], [120, 200]], np.floataa) __A : List[Any] = np.array([[10, 100], [80, 50], [180, 250]], np.floataa) # add all rotated images in a list __A : List[str] = [ gray_img, get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), get_rotation(gray_img, ptsa, ptsa, img_rows, img_cols), ] # plot different image rotations __A : Union[str, Any] = plt.figure(1) __A : Optional[Any] = ['''Original''', '''Rotation 1''', '''Rotation 2''', '''Rotation 3'''] for i, image in enumerate(images): plt.subplot(2, 2, i + 1), plt.imshow(image, '''gray''') plt.title(titles[i]) plt.axis('''off''') plt.subplots_adjust(left=0.0, bottom=0.05, right=1.0, top=0.95) plt.show()
138
1
'''simple docstring''' # Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position __lowerCAmelCase = """2.13.1""" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("""3.7"""): raise ImportWarning( """To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition.""" ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( """To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n""" """If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.""" ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip __lowerCAmelCase = concatenate_datasets __lowerCAmelCase = DownloadConfig __lowerCAmelCase = DownloadManager __lowerCAmelCase = DownloadMode __lowerCAmelCase = DownloadConfig __lowerCAmelCase = DownloadMode __lowerCAmelCase = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
5
'''simple docstring''' import qiskit def UpperCAmelCase_ (__a : int , __a : int ): """simple docstring""" _a : Any = qiskit.Aer.get_backend('aer_simulator' ) # Create a Quantum Circuit acting on the q register _a : List[Any] = qiskit.QuantumCircuit(__a , __a ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator _a : Tuple = qiskit.execute(__a , __a , shots=1_0_0_0 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(__a ) if __name__ == "__main__": print(f'''Total count for various states are: {single_qubit_measure(1, 1)}''')
5
1
"""simple docstring""" from abc import ABC, abstractmethod from argparse import ArgumentParser class __A ( _SCREAMING_SNAKE_CASE ): """simple docstring""" @staticmethod @abstractmethod def SCREAMING_SNAKE_CASE ( __A ) -> str: raise NotImplementedError() @abstractmethod def SCREAMING_SNAKE_CASE ( self ) -> str: raise NotImplementedError()
81
"""simple docstring""" import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def _A ( ): """simple docstring""" a =ArgumentParser( description=( '''PyTorch TPU distributed training launch ''' '''helper utility that will spawn up ''' '''multiple distributed processes''' ) ) # Optional arguments for the launch helper parser.add_argument('''--num_cores''' , type=lowercase , default=1 , help='''Number of TPU cores to use (1 or 8).''' ) # positional parser.add_argument( '''training_script''' , type=lowercase , help=( '''The full path to the single TPU training ''' '''program/script to be launched in parallel, ''' '''followed by all the arguments for the ''' '''training script''' ) , ) # rest from the training program parser.add_argument('''training_script_args''' , nargs=lowercase ) return parser.parse_args() def _A ( ): """simple docstring""" a =parse_args() # Import training_script as a module. a =Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) a =script_fpath.stem a =importlib.import_module(lowercase ) # Patch sys.argv a =[args.training_script] + args.training_script_args + ['''--tpu_num_cores''', str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
81
1
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase_ : List[str] = { "configuration_vivit": ["VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "VivitConfig"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : List[Any] = ["VivitImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Optional[int] = [ "VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "VivitModel", "VivitPreTrainedModel", "VivitForVideoClassification", ] if TYPE_CHECKING: from .configuration_vivit import VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, VivitConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_vivit import VivitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vivit import ( VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST, VivitForVideoClassification, VivitModel, VivitPreTrainedModel, ) else: import sys lowerCamelCase_ : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
359
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ : Tuple = logging.get_logger(__name__) lowerCamelCase_ : List[Any] = { """google/fnet-base""": """https://huggingface.co/google/fnet-base/resolve/main/config.json""", """google/fnet-large""": """https://huggingface.co/google/fnet-large/resolve/main/config.json""" # See all FNet models at https://huggingface.co/models?filter=fnet } class a__ ( __snake_case ): A__ : Dict = 'fnet' def __init__( self , UpperCAmelCase=3_2_0_0_0 , UpperCAmelCase=7_6_8 , UpperCAmelCase=1_2 , UpperCAmelCase=3_0_7_2 , UpperCAmelCase="gelu_new" , UpperCAmelCase=0.1 , UpperCAmelCase=5_1_2 , UpperCAmelCase=4 , UpperCAmelCase=0.02 , UpperCAmelCase=1e-12 , UpperCAmelCase=False , UpperCAmelCase=5_1_2 , UpperCAmelCase=3 , UpperCAmelCase=1 , UpperCAmelCase=2 , **UpperCAmelCase , ) -> Dict: super().__init__(pad_token_id=UpperCAmelCase , bos_token_id=UpperCAmelCase , eos_token_id=UpperCAmelCase , **UpperCAmelCase ) __a = vocab_size __a = max_position_embeddings __a = hidden_size __a = num_hidden_layers __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = initializer_range __a = type_vocab_size __a = layer_norm_eps __a = use_tpu_fourier_optimizations __a = tpu_short_seq_length
197
0
import os # Precomputes a list of the 100 first triangular numbers _lowerCamelCase =[int(0.5 * n * (n + 1)) for n in range(1, 1_0_1)] def _a ( ): lowerCamelCase : Any = os.path.dirname(os.path.realpath(lowerCamelCase ) ) lowerCamelCase : str = os.path.join(lowerCamelCase, """words.txt""" ) lowerCamelCase : Any = """""" with open(lowerCamelCase ) as f: lowerCamelCase : Tuple = f.readline() lowerCamelCase : Tuple = [word.strip("""\"""" ) for word in words.strip("""\r\n""" ).split(""",""" )] lowerCamelCase : Tuple = [ word for word in [sum(ord(lowerCamelCase ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(lowerCamelCase ) if __name__ == "__main__": print(solution())
287
import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 _lowerCamelCase ={ """return_dict""": False, """output_hidden_states""": True, """output_attentions""": True, """torchscript""": True, """torch_dtype""": """float16""", """use_bfloat16""": True, """tf_legacy_loss""": True, """pruned_heads""": {"""a""": 1}, """tie_word_embeddings""": False, """is_decoder""": True, """cross_attention_hidden_size""": 1_2_8, """add_cross_attention""": True, """tie_encoder_decoder""": True, """max_length""": 5_0, """min_length""": 3, """do_sample""": True, """early_stopping""": True, """num_beams""": 3, """num_beam_groups""": 3, """diversity_penalty""": 0.5, """temperature""": 2.0, """top_k""": 1_0, """top_p""": 0.7, """typical_p""": 0.2, """repetition_penalty""": 0.8, """length_penalty""": 0.8, """no_repeat_ngram_size""": 5, """encoder_no_repeat_ngram_size""": 5, """bad_words_ids""": [1, 2, 3], """num_return_sequences""": 3, """chunk_size_feed_forward""": 5, """output_scores""": True, """return_dict_in_generate""": True, """forced_bos_token_id""": 2, """forced_eos_token_id""": 3, """remove_invalid_values""": True, """architectures""": ["""BertModel"""], """finetuning_task""": """translation""", """id2label""": {0: """label"""}, """label2id""": {"""label""": """0"""}, """tokenizer_class""": """BertTokenizerFast""", """prefix""": """prefix""", """bos_token_id""": 6, """pad_token_id""": 7, """eos_token_id""": 8, """sep_token_id""": 9, """decoder_start_token_id""": 1_0, """exponential_decay_length_penalty""": (5, 1.01), """suppress_tokens""": [0, 1], """begin_suppress_tokens""": 2, """task_specific_params""": {"""translation""": """some_params"""}, """problem_type""": """regression""", } @is_staging_test class A__ ( unittest.TestCase): @classmethod def UpperCamelCase__ ( cls ): lowerCamelCase : int = TOKEN HfFolder.save_token(__magic_name__ ) @classmethod def UpperCamelCase__ ( cls ): try: delete_repo(token=cls._token , repo_id="""test-config""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-config-org""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""test-dynamic-config""" ) except HTTPError: pass def UpperCamelCase__ ( self ): lowerCamelCase : List[Any] = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""test-config""" , use_auth_token=self._token ) lowerCamelCase : Any = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""test-config""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__magic_name__ , repo_id="""test-config""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : Optional[Any] = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): lowerCamelCase : Dict = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub("""valid_org/test-config-org""" , use_auth_token=self._token ) lowerCamelCase : Optional[int] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-config-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( __magic_name__ , repo_id="""valid_org/test-config-org""" , push_to_hub=__magic_name__ , use_auth_token=self._token ) lowerCamelCase : List[str] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__magic_name__ , getattr(__magic_name__ , __magic_name__ ) ) def UpperCamelCase__ ( self ): CustomConfig.register_for_auto_class() lowerCamelCase : Optional[Any] = CustomConfig(attribute=4_2 ) config.push_to_hub("""test-dynamic-config""" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"""AutoConfig""": """custom_configuration.CustomConfig"""} ) lowerCamelCase : List[str] = AutoConfig.from_pretrained(F'''{USER}/test-dynamic-config''' , trust_remote_code=__magic_name__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , """CustomConfig""" ) self.assertEqual(new_config.attribute , 4_2 ) class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : str = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated lowerCamelCase : Optional[int] = c.n_embd + 1 # int lowerCamelCase : Optional[int] = c.resid_pdrop + 1.0 # float lowerCamelCase : Tuple = not c.scale_attn_weights # bool lowerCamelCase : Any = c.summary_type + """foo""" # str c.update_from_string( F'''n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}''' ) self.assertEqual(__magic_name__ , c.n_embd , """mismatch for key: n_embd""" ) self.assertEqual(__magic_name__ , c.resid_pdrop , """mismatch for key: resid_pdrop""" ) self.assertEqual(__magic_name__ , c.scale_attn_weights , """mismatch for key: scale_attn_weights""" ) self.assertEqual(__magic_name__ , c.summary_type , """mismatch for key: summary_type""" ) def UpperCamelCase__ ( self ): lowerCamelCase : str = PretrainedConfig() lowerCamelCase : int = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( __magic_name__ , ["""is_encoder_decoder""", """_name_or_path""", """_commit_hash""", """transformers_version"""] ) lowerCamelCase : List[str] = [key for key, value in config_common_kwargs.items() if value == getattr(__magic_name__ , __magic_name__ )] if len(__magic_name__ ) > 0: raise ValueError( """The following keys are set with the default values in""" """ `test_configuration_common.config_common_kwargs` pick another value for them:""" F''' {", ".join(__magic_name__ )}.''' ) def UpperCamelCase__ ( self ): with self.assertRaises(__magic_name__ ): # config is in subfolder, the following should not work without specifying the subfolder lowerCamelCase : Dict = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" ) lowerCamelCase : str = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" , subfolder="""bert""" ) self.assertIsNotNone(__magic_name__ ) def UpperCamelCase__ ( self ): # A mock response for an HTTP head request to emulate server down lowerCamelCase : Dict = mock.Mock() lowerCamelCase : Optional[int] = 5_0_0 lowerCamelCase : List[Any] = {} lowerCamelCase : Tuple = HTTPError lowerCamelCase : Union[str, Any] = {} # Download this model to make sure it's in the cache. lowerCamelCase : List[str] = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("""requests.Session.request""" , return_value=__magic_name__ ) as mock_head: lowerCamelCase : Any = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # This check we did call the fake head request mock_head.assert_called() def UpperCamelCase__ ( self ): # This test is for deprecated behavior and can be removed in v5 lowerCamelCase : List[str] = BertConfig.from_pretrained( """https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json""" ) def UpperCamelCase__ ( self ): lowerCamelCase : Optional[int] = AutoConfig.from_pretrained("""bert-base-cased""" ) lowerCamelCase : Optional[Any] = ["""config.4.0.0.json"""] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(__magic_name__ ) lowerCamelCase : str = 2 json.dump(configuration.to_dict() , open(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , """w""" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 lowerCamelCase : Any = ["""config.42.0.0.json"""] lowerCamelCase : Optional[Any] = 7_6_8 configuration.save_pretrained(__magic_name__ ) shutil.move(os.path.join(__magic_name__ , """config.4.0.0.json""" ) , os.path.join(__magic_name__ , """config.42.0.0.json""" ) ) lowerCamelCase : int = AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 7_6_8 ) def UpperCamelCase__ ( self ): # This repo has two configuration files, one for v4.0.0 and above with a different hidden size. lowerCamelCase : str = """hf-internal-testing/test-two-configs""" import transformers as new_transformers lowerCamelCase : Tuple = """v4.0.0""" lowerCamelCase , lowerCamelCase : Optional[int] = new_transformers.models.auto.AutoConfig.from_pretrained( __magic_name__ , return_unused_kwargs=__magic_name__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(__magic_name__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers lowerCamelCase : Tuple = """v3.0.0""" lowerCamelCase : Any = old_transformers.models.auto.AutoConfig.from_pretrained(__magic_name__ ) self.assertEqual(old_configuration.hidden_size , 7_6_8 )
287
1
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class A_ ( _UpperCAmelCase ): lowerCAmelCase__ = ["image_processor", "tokenizer"] lowerCAmelCase__ = "ViTImageProcessor" lowerCAmelCase__ = ("CLIPTokenizer", "CLIPTokenizerFast") def __init__( self: Tuple ,__lowerCAmelCase: List[Any]=None ,__lowerCAmelCase: Any=None ,**__lowerCAmelCase: Union[str, Any] ): '''simple docstring''' _lowerCamelCase : Optional[Any] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." ,_UpperCAmelCase ,) _lowerCamelCase : Optional[int] = kwargs.pop("feature_extractor" ) _lowerCamelCase : List[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(_UpperCAmelCase ,_UpperCAmelCase ) def __call__( self: int ,__lowerCAmelCase: Dict=None ,__lowerCAmelCase: Union[str, Any]=None ,__lowerCAmelCase: int=None ,__lowerCAmelCase: List[Any]=None ,**__lowerCAmelCase: Tuple ): '''simple docstring''' if text is None and visual_prompt is None and images is None: raise ValueError("You have to specify either text, visual prompt or images." ) if text is not None and visual_prompt is not None: raise ValueError("You have to specify exactly one type of prompt. Either text or visual prompt." ) if text is not None: _lowerCamelCase : Optional[Any] = self.tokenizer(_UpperCAmelCase ,return_tensors=_UpperCAmelCase ,**_UpperCAmelCase ) if visual_prompt is not None: _lowerCamelCase : str = self.image_processor(_UpperCAmelCase ,return_tensors=_UpperCAmelCase ,**_UpperCAmelCase ) if images is not None: _lowerCamelCase : Tuple = self.image_processor(_UpperCAmelCase ,return_tensors=_UpperCAmelCase ,**_UpperCAmelCase ) if visual_prompt is not None and images is not None: _lowerCamelCase : Any = { '''pixel_values''': image_features.pixel_values, '''conditional_pixel_values''': prompt_features.pixel_values, } return encoding elif text is not None and images is not None: _lowerCamelCase : str = image_features.pixel_values return encoding elif text is not None: return encoding elif visual_prompt is not None: _lowerCamelCase : Union[str, Any] = { '''conditional_pixel_values''': prompt_features.pixel_values, } return encoding else: return BatchEncoding(data=dict(**_UpperCAmelCase ) ,tensor_type=_UpperCAmelCase ) def _lowercase ( self: int ,*__lowerCAmelCase: int ,**__lowerCAmelCase: List[str] ): '''simple docstring''' return self.tokenizer.batch_decode(*_UpperCAmelCase ,**_UpperCAmelCase ) def _lowercase ( self: Any ,*__lowerCAmelCase: Any ,**__lowerCAmelCase: Optional[int] ): '''simple docstring''' return self.tokenizer.decode(*_UpperCAmelCase ,**_UpperCAmelCase ) @property def _lowercase ( self: List[Any] ): '''simple docstring''' warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." ,_UpperCAmelCase ,) return self.image_processor_class @property def _lowercase ( self: List[str] ): '''simple docstring''' warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." ,_UpperCAmelCase ,) return self.image_processor
352
"""simple docstring""" import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging _lowerCAmelCase : Dict = logging.get_logger(__name__) class A_ ( _a ): def __init__( self: List[Any] ,__lowerCAmelCase: Union[List[ControlNetModel], Tuple[ControlNetModel]] ): '''simple docstring''' super().__init__() _lowerCamelCase : Tuple = nn.ModuleList(__lowerCAmelCase ) def _lowercase ( self: Union[str, Any] ,__lowerCAmelCase: torch.FloatTensor ,__lowerCAmelCase: Union[torch.Tensor, float, int] ,__lowerCAmelCase: torch.Tensor ,__lowerCAmelCase: List[torch.tensor] ,__lowerCAmelCase: List[float] ,__lowerCAmelCase: Optional[torch.Tensor] = None ,__lowerCAmelCase: Optional[torch.Tensor] = None ,__lowerCAmelCase: Optional[torch.Tensor] = None ,__lowerCAmelCase: Optional[Dict[str, Any]] = None ,__lowerCAmelCase: bool = False ,__lowerCAmelCase: bool = True ,): '''simple docstring''' for i, (image, scale, controlnet) in enumerate(zip(__lowerCAmelCase ,__lowerCAmelCase ,self.nets ) ): _lowerCamelCase, _lowerCamelCase : Union[str, Any] = controlnet( __lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ,) # merge samples if i == 0: _lowerCamelCase, _lowerCamelCase : Optional[Any] = down_samples, mid_sample else: _lowerCamelCase : Optional[int] = [ samples_prev + samples_curr for samples_prev, samples_curr in zip(__lowerCAmelCase ,__lowerCAmelCase ) ] mid_block_res_sample += mid_sample return down_block_res_samples, mid_block_res_sample def _lowercase ( self: Union[str, Any] ,__lowerCAmelCase: Union[str, os.PathLike] ,__lowerCAmelCase: bool = True ,__lowerCAmelCase: Callable = None ,__lowerCAmelCase: bool = False ,__lowerCAmelCase: Optional[str] = None ,): '''simple docstring''' _lowerCamelCase : List[Any] = 0 _lowerCamelCase : str = save_directory for controlnet in self.nets: controlnet.save_pretrained( __lowerCAmelCase ,is_main_process=__lowerCAmelCase ,save_function=__lowerCAmelCase ,safe_serialization=__lowerCAmelCase ,variant=__lowerCAmelCase ,) idx += 1 _lowerCamelCase : int = model_path_to_save + F"""_{idx}""" @classmethod def _lowercase ( cls: Any ,__lowerCAmelCase: Optional[Union[str, os.PathLike]] ,**__lowerCAmelCase: int ): '''simple docstring''' _lowerCamelCase : int = 0 _lowerCamelCase : str = [] # load controlnet and append to list until no controlnet directory exists anymore # first controlnet has to be saved under `./mydirectory/controlnet` to be compliant with `DiffusionPipeline.from_prertained` # second, third, ... controlnets have to be saved under `./mydirectory/controlnet_1`, `./mydirectory/controlnet_2`, ... _lowerCamelCase : Dict = pretrained_model_path while os.path.isdir(__lowerCAmelCase ): _lowerCamelCase : List[Any] = ControlNetModel.from_pretrained(__lowerCAmelCase ,**__lowerCAmelCase ) controlnets.append(__lowerCAmelCase ) idx += 1 _lowerCamelCase : Tuple = pretrained_model_path + F"""_{idx}""" logger.info(F"""{len(__lowerCAmelCase )} controlnets loaded from {pretrained_model_path}.""" ) if len(__lowerCAmelCase ) == 0: raise ValueError( F"""No ControlNets found under {os.path.dirname(__lowerCAmelCase )}. Expected at least {pretrained_model_path + '_0'}.""" ) return cls(__lowerCAmelCase )
340
0
"""simple docstring""" import math import flax.linen as nn import jax.numpy as jnp def a_ ( _lowerCAmelCase : jnp.ndarray , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1.0E4 , _lowerCAmelCase : bool = False , _lowerCAmelCase : float = 1.0 , ): '''simple docstring''' assert timesteps.ndim == 1, "Timesteps should be a 1d-array" assert embedding_dim % 2 == 0, f"""Embedding dimension {embedding_dim} should be even""" lowercase__ : Optional[Any] = float(embedding_dim // 2 ) lowercase__ : Optional[Any] = math.log(max_timescale / min_timescale ) / (num_timescales - freq_shift) lowercase__ : Any = min_timescale * jnp.exp(jnp.arange(_lowerCAmelCase , dtype=jnp.floataa ) * -log_timescale_increment ) lowercase__ : Dict = jnp.expand_dims(_lowerCAmelCase , 1 ) * jnp.expand_dims(_lowerCAmelCase , 0 ) # scale embeddings lowercase__ : List[str] = scale * emb if flip_sin_to_cos: lowercase__ : Dict = jnp.concatenate([jnp.cos(_lowerCAmelCase ), jnp.sin(_lowerCAmelCase )] , axis=1 ) else: lowercase__ : Optional[int] = jnp.concatenate([jnp.sin(_lowerCAmelCase ), jnp.cos(_lowerCAmelCase )] , axis=1 ) lowercase__ : List[Any] = jnp.reshape(_lowerCAmelCase , [jnp.shape(_lowerCAmelCase )[0], embedding_dim] ) return signal class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : jnp.dtype = jnp.floataa @nn.compact def __call__( self , a ) -> Any: lowercase__ : str = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_1' )(a ) lowercase__ : Union[str, Any] = nn.silu(a ) lowercase__ : Optional[Any] = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_2' )(a ) return temb class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : bool = False lowerCamelCase__ : float = 1 @nn.compact def __call__( self , a ) -> str: return get_sinusoidal_embeddings( a , embedding_dim=self.dim , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.freq_shift )
77
from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_ ( _snake_case : str = "laptop" ) -> DataFrame: '''simple docstring''' __magic_name__ : Tuple = F'''https://www.amazon.in/laptop/s?k={product}''' __magic_name__ : Dict = { "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36", "Accept-Language": "en-US, en;q=0.5", } __magic_name__ : Tuple = BeautifulSoup(requests.get(_snake_case , headers=_snake_case ).text ) # Initialize a Pandas dataframe with the column titles __magic_name__ : int = DataFrame( columns=[ "Product Title", "Product Link", "Current Price of the product", "Product Rating", "MRP of the product", "Discount", ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( "div" , attrs={"class": "s-result-item", "data-component-type": "s-search-result"} , ) , soup.find_all("div" , attrs={"class": "a-row a-size-base a-color-base"} ) , ): try: __magic_name__ : Dict = item.ha.text __magic_name__ : Optional[int] = "https://www.amazon.in/" + item.ha.a["href"] __magic_name__ : Optional[Any] = item.find("span" , attrs={"class": "a-offscreen"} ).text try: __magic_name__ : Union[str, Any] = item.find("span" , attrs={"class": "a-icon-alt"} ).text except AttributeError: __magic_name__ : Dict = "Not available" try: __magic_name__ : Optional[int] = ( "₹" + item.find( "span" , attrs={"class": "a-price a-text-price"} ).text.split("₹" )[1] ) except AttributeError: __magic_name__ : List[str] = "" try: __magic_name__ : int = float( ( ( float(product_mrp.strip("₹" ).replace("," , "" ) ) - float(product_price.strip("₹" ).replace("," , "" ) ) ) / float(product_mrp.strip("₹" ).replace("," , "" ) ) ) * 100 ) except ValueError: __magic_name__ : str = float("nan" ) except AttributeError: pass __magic_name__ : Optional[int] = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] __magic_name__ : Optional[Any] = " " __magic_name__ : str = " " data_frame.index += 1 return data_frame if __name__ == "__main__": snake_case : Any = "headphones" get_amazon_product_data(product).to_csv(F"Amazon Product Data for {product}.csv")
281
0
from collections import defaultdict def __snake_case ( __UpperCamelCase : str ,__UpperCamelCase : str ): """simple docstring""" A_ = first_str.lower().strip() A_ = second_str.lower().strip() # Remove whitespace A_ = first_str.replace(" " ,"" ) A_ = second_str.replace(" " ,"" ) # Strings of different lengths are not anagrams if len(__UpperCamelCase ) != len(__UpperCamelCase ): return False # Default values for count should be 0 A_ = defaultdict(__UpperCamelCase ) # For each character in input strings, # increment count in the corresponding for i in range(len(__UpperCamelCase ) ): count[first_str[i]] += 1 count[second_str[i]] -= 1 return all(_count == 0 for _count in count.values() ) if __name__ == "__main__": from doctest import testmod testmod() __a :Any = input('Enter the first string ').strip() __a :Any = input('Enter the second string ').strip() __a :Optional[int] = check_anagrams(input_a, input_b) print(F"{input_a} and {input_b} are {'' if status else 'not '}anagrams.")
329
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __a :Union[str, Any] = { 'configuration_biogpt': ['BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BioGptConfig'], 'tokenization_biogpt': ['BioGptTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a :Optional[int] = [ 'BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BioGptForCausalLM', 'BioGptForTokenClassification', 'BioGptForSequenceClassification', 'BioGptModel', 'BioGptPreTrainedModel', ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys __a :str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
329
1
def __UpperCamelCase ( _A = 1000000 ): lowerCAmelCase_ = 1 lowerCAmelCase_ = 1 lowerCAmelCase_ = {1: 1} for inputa in range(2 , _A ): lowerCAmelCase_ = 0 lowerCAmelCase_ = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: lowerCAmelCase_ = (3 * number) + 1 counter += 1 if inputa not in counters: lowerCAmelCase_ = counter if counter > pre_counter: lowerCAmelCase_ = inputa lowerCAmelCase_ = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
278
import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class A ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = AutoImageProcessor.from_pretrained('''microsoft/dit-base-finetuned-rvlcdip''' ) lowerCAmelCase_ = AutoModelForImageClassification.from_pretrained('''microsoft/dit-base-finetuned-rvlcdip''' ) model.to(UpperCamelCase__ ) from datasets import load_dataset lowerCAmelCase_ = load_dataset('''nielsr/rvlcdip-demo''' ) lowerCAmelCase_ = dataset['''train'''][0]['''image'''].convert('''RGB''' ) lowerCAmelCase_ = image_processor(UpperCamelCase__, return_tensors='''pt''' ).to(UpperCamelCase__ ) # forward pass with torch.no_grad(): lowerCAmelCase_ = model(**UpperCamelCase__ ) lowerCAmelCase_ = outputs.logits lowerCAmelCase_ = torch.Size((1, 16) ) self.assertEqual(logits.shape, UpperCamelCase__ ) lowerCAmelCase_ = torch.tensor( [-0.4_158, -0.4_092, -0.4_347], device=UpperCamelCase__, dtype=torch.float, ) self.assertTrue(torch.allclose(logits[0, :3], UpperCamelCase__, atol=1E-4 ) )
278
1
import inspect import unittest from datasets import load_dataset from packaging import version from transformers import BeitConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_MAPPING, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, ) from transformers.models.beit.modeling_beit import BEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): import PIL from PIL import Image from transformers import BeitImageProcessor class a__ : def __init__( self , UpperCAmelCase , UpperCAmelCase=1_0_0 , UpperCAmelCase=1_3 , UpperCAmelCase=3_0 , UpperCAmelCase=2 , UpperCAmelCase=3 , UpperCAmelCase=True , UpperCAmelCase=True , UpperCAmelCase=3_2 , UpperCAmelCase=4 , UpperCAmelCase=4 , UpperCAmelCase=3_7 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=1_0 , UpperCAmelCase=0.02 , UpperCAmelCase=3 , UpperCAmelCase=None , UpperCAmelCase=[0, 1, 2, 3] , ) -> List[str]: __a = parent __a = 1_0_0 __a = batch_size __a = image_size __a = patch_size __a = num_channels __a = is_training __a = use_labels __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = type_sequence_label_size __a = initializer_range __a = scope __a = out_indices __a = num_labels # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __a = (image_size // patch_size) ** 2 __a = num_patches + 1 def __SCREAMING_SNAKE_CASE ( self ) -> List[str]: __a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a = None __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __a = self.get_config() return config, pixel_values, labels, pixel_labels def __SCREAMING_SNAKE_CASE ( self ) -> Dict: return BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCAmelCase , initializer_range=self.initializer_range , out_indices=self.out_indices , ) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> List[str]: __a = BeitModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() __a = model(UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Any: __a = BeitForMaskedImageModeling(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() __a = model(UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Any: __a = self.type_sequence_label_size __a = BeitForImageClassification(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() __a = model(UpperCAmelCase , labels=UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __a = 1 __a = BeitForImageClassification(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() __a = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a = model(UpperCAmelCase , labels=UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> int: __a = self.num_labels __a = BeitForSemanticSegmentation(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() __a = model(UpperCAmelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) __a = model(UpperCAmelCase , labels=UpperCAmelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: __a = self.prepare_config_and_inputs() __a , __a , __a , __a = config_and_inputs __a = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class a__ ( __snake_case , __snake_case , unittest.TestCase ): A__ : Union[str, Any] = ( (BeitModel, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation) if is_torch_available() else () ) A__ : Optional[Any] = ( { 'feature-extraction': BeitModel, 'image-classification': BeitForImageClassification, 'image-segmentation': BeitForSemanticSegmentation, } if is_torch_available() else {} ) A__ : Optional[int] = False A__ : List[str] = False A__ : List[Any] = False def __SCREAMING_SNAKE_CASE ( self ) -> Tuple: __a = BeitModelTester(self ) __a = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase , hidden_size=3_7 ) def __SCREAMING_SNAKE_CASE ( self ) -> Tuple: self.config_tester.run_common_tests() @unittest.skip(reason='BEiT does not use inputs_embeds' ) def __SCREAMING_SNAKE_CASE ( self ) -> str: pass @require_torch_multi_gpu @unittest.skip(reason='BEiT has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def __SCREAMING_SNAKE_CASE ( self ) -> List[Any]: pass def __SCREAMING_SNAKE_CASE ( self ) -> Any: __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(UpperCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __a = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase , nn.Linear ) ) def __SCREAMING_SNAKE_CASE ( self ) -> Any: __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(UpperCAmelCase ) __a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a = [*signature.parameters.keys()] __a = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( self ) -> str: __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( self ) -> Dict: __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( self ) -> str: __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( self ) -> Tuple: if not self.model_tester.is_training: return __a , __a = self.model_tester.prepare_config_and_inputs_for_common() __a = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if model_class in [*get_values(UpperCAmelCase ), BeitForMaskedImageModeling]: continue __a = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.train() __a = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) __a = model(**UpperCAmelCase ).loss loss.backward() def __SCREAMING_SNAKE_CASE ( self ) -> List[str]: __a , __a = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return __a = False __a = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if ( model_class in [*get_values(UpperCAmelCase ), BeitForMaskedImageModeling] or not model_class.supports_gradient_checkpointing ): continue __a = model_class(UpperCAmelCase ) model.gradient_checkpointing_enable() model.to(UpperCAmelCase ) model.train() __a = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase , return_labels=UpperCAmelCase ) __a = model(**UpperCAmelCase ).loss loss.backward() def __SCREAMING_SNAKE_CASE ( self ) -> Dict: __a , __a = self.model_tester.prepare_config_and_inputs_for_common() __a = _config_zero_init(UpperCAmelCase ) for model_class in self.all_model_classes: __a = model_class(config=UpperCAmelCase ) for name, param in model.named_parameters(): # we skip lambda parameters as these require special initial values # determined by config.layer_scale_init_value if "lambda" in name: continue if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @slow def __SCREAMING_SNAKE_CASE ( self ) -> Tuple: for model_name in BEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = BeitModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def lowerCAmelCase( ): __a = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class a__ ( unittest.TestCase ): @cached_property def __SCREAMING_SNAKE_CASE ( self ) -> Dict: return BeitImageProcessor.from_pretrained('microsoft/beit-base-patch16-224' ) if is_vision_available() else None @slow def __SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: __a = BeitForMaskedImageModeling.from_pretrained('microsoft/beit-base-patch16-224-pt22k' ).to(UpperCAmelCase ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=UpperCAmelCase , return_tensors='pt' ).pixel_values.to(UpperCAmelCase ) # prepare bool_masked_pos __a = torch.ones((1, 1_9_6) , dtype=torch.bool ).to(UpperCAmelCase ) # forward pass with torch.no_grad(): __a = model(pixel_values=UpperCAmelCase , bool_masked_pos=UpperCAmelCase ) __a = outputs.logits # verify the logits __a = torch.Size((1, 1_9_6, 8_1_9_2) ) self.assertEqual(logits.shape , UpperCAmelCase ) __a = torch.tensor( [[-3.2_437, 0.5_072, -13.9_174], [-3.2_456, 0.4_948, -13.9_401], [-3.2_033, 0.5_121, -13.8_550]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(logits[bool_masked_pos][:3, :3] , UpperCAmelCase , atol=1e-2 ) ) @slow def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: __a = BeitForImageClassification.from_pretrained('microsoft/beit-base-patch16-224' ).to(UpperCAmelCase ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=UpperCAmelCase , return_tensors='pt' ).to(UpperCAmelCase ) # forward pass with torch.no_grad(): __a = model(**UpperCAmelCase ) __a = outputs.logits # verify the logits __a = torch.Size((1, 1_0_0_0) ) self.assertEqual(logits.shape , UpperCAmelCase ) __a = torch.tensor([-1.2_385, -1.0_987, -1.0_108] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(logits[0, :3] , UpperCAmelCase , atol=1e-4 ) ) __a = 2_8_1 self.assertEqual(logits.argmax(-1 ).item() , UpperCAmelCase ) @slow def __SCREAMING_SNAKE_CASE ( self ) -> str: __a = BeitForImageClassification.from_pretrained('microsoft/beit-large-patch16-224-pt22k-ft22k' ).to( UpperCAmelCase ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=UpperCAmelCase , return_tensors='pt' ).to(UpperCAmelCase ) # forward pass with torch.no_grad(): __a = model(**UpperCAmelCase ) __a = outputs.logits # verify the logits __a = torch.Size((1, 2_1_8_4_1) ) self.assertEqual(logits.shape , UpperCAmelCase ) __a = torch.tensor([1.6_881, -0.2_787, 0.5_901] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(logits[0, :3] , UpperCAmelCase , atol=1e-4 ) ) __a = 2_3_9_6 self.assertEqual(logits.argmax(-1 ).item() , UpperCAmelCase ) @slow def __SCREAMING_SNAKE_CASE ( self ) -> Tuple: __a = BeitForSemanticSegmentation.from_pretrained('microsoft/beit-base-finetuned-ade-640-640' ) __a = model.to(UpperCAmelCase ) __a = BeitImageProcessor(do_resize=UpperCAmelCase , size=6_4_0 , do_center_crop=UpperCAmelCase ) __a = load_dataset('hf-internal-testing/fixtures_ade20k' , split='test' ) __a = Image.open(ds[0]['file'] ) __a = image_processor(images=UpperCAmelCase , return_tensors='pt' ).to(UpperCAmelCase ) # forward pass with torch.no_grad(): __a = model(**UpperCAmelCase ) __a = outputs.logits # verify the logits __a = torch.Size((1, 1_5_0, 1_6_0, 1_6_0) ) self.assertEqual(logits.shape , UpperCAmelCase ) __a = version.parse(PIL.__version__ ) < version.parse('9.0.0' ) if is_pillow_less_than_a: __a = torch.tensor( [ [[-4.9_225, -2.3_954, -3.0_522], [-2.8_822, -1.0_046, -1.7_561], [-2.9_549, -1.3_228, -2.1_347]], [[-5.8_168, -3.4_129, -4.0_778], [-3.8_651, -2.2_214, -3.0_277], [-3.8_356, -2.4_643, -3.3_535]], [[-0.0_078, 3.9_952, 4.0_754], [2.9_856, 4.6_944, 5.0_035], [3.2_413, 4.7_813, 4.9_969]], ] , device=UpperCAmelCase , ) else: __a = torch.tensor( [ [[-4.8_960, -2.3_688, -3.0_355], [-2.8_478, -0.9_836, -1.7_418], [-2.9_449, -1.3_332, -2.1_456]], [[-5.8_081, -3.4_124, -4.1_006], [-3.8_561, -2.2_081, -3.0_323], [-3.8_365, -2.4_601, -3.3_669]], [[-0.0_309, 3.9_868, 4.0_540], [2.9_640, 4.6_877, 4.9_976], [3.2_081, 4.7_690, 4.9_942]], ] , device=UpperCAmelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , UpperCAmelCase , atol=1e-4 ) ) @slow def __SCREAMING_SNAKE_CASE ( self ) -> Dict: __a = BeitForSemanticSegmentation.from_pretrained('microsoft/beit-base-finetuned-ade-640-640' ) __a = model.to(UpperCAmelCase ) __a = BeitImageProcessor(do_resize=UpperCAmelCase , size=6_4_0 , do_center_crop=UpperCAmelCase ) __a = load_dataset('hf-internal-testing/fixtures_ade20k' , split='test' ) __a = Image.open(ds[0]['file'] ) __a = image_processor(images=UpperCAmelCase , return_tensors='pt' ).to(UpperCAmelCase ) # forward pass with torch.no_grad(): __a = model(**UpperCAmelCase ) __a = outputs.logits.detach().cpu() __a = image_processor.post_process_semantic_segmentation(outputs=UpperCAmelCase , target_sizes=[(5_0_0, 3_0_0)] ) __a = torch.Size((5_0_0, 3_0_0) ) self.assertEqual(segmentation[0].shape , UpperCAmelCase ) __a = image_processor.post_process_semantic_segmentation(outputs=UpperCAmelCase ) __a = torch.Size((1_6_0, 1_6_0) ) self.assertEqual(segmentation[0].shape , UpperCAmelCase )
362
from .glue import glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels from .squad import SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features from .utils import DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor from .xnli import xnli_output_modes, xnli_processors, xnli_tasks_num_labels
197
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __UpperCAmelCase = logging.get_logger(__name__) class UpperCamelCase__ ( lowercase_ ): """simple docstring""" def __init__( self : Any , *lowerCamelCase_ : str , **lowerCamelCase_ : Dict ): '''simple docstring''' warnings.warn( """The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use BeitImageProcessor instead.""" , lowerCamelCase_ , ) super().__init__(*lowerCamelCase_ , **lowerCamelCase_ )
323
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class UpperCamelCase__ ( lowercase_ ): """simple docstring""" SCREAMING_SNAKE_CASE__ = '''git_vision_model''' def __init__( self : int , lowerCamelCase_ : Dict=7_68 , lowerCamelCase_ : Tuple=30_72 , lowerCamelCase_ : List[Any]=12 , lowerCamelCase_ : Tuple=12 , lowerCamelCase_ : int=3 , lowerCamelCase_ : List[str]=2_24 , lowerCamelCase_ : Optional[Any]=16 , lowerCamelCase_ : Optional[Any]="quick_gelu" , lowerCamelCase_ : List[Any]=1e-5 , lowerCamelCase_ : List[Any]=0.0 , lowerCamelCase_ : Optional[Any]=0.02 , **lowerCamelCase_ : str , ): '''simple docstring''' super().__init__(**lowerCamelCase_ ) SCREAMING_SNAKE_CASE : Union[str, Any] = hidden_size SCREAMING_SNAKE_CASE : int = intermediate_size SCREAMING_SNAKE_CASE : List[str] = num_hidden_layers SCREAMING_SNAKE_CASE : int = num_attention_heads SCREAMING_SNAKE_CASE : int = num_channels SCREAMING_SNAKE_CASE : Optional[Any] = patch_size SCREAMING_SNAKE_CASE : Optional[int] = image_size SCREAMING_SNAKE_CASE : List[str] = initializer_range SCREAMING_SNAKE_CASE : str = attention_dropout SCREAMING_SNAKE_CASE : Any = layer_norm_eps SCREAMING_SNAKE_CASE : List[str] = hidden_act @classmethod def lowerCamelCase_ ( cls : List[str] , lowerCamelCase_ : Union[str, os.PathLike] , **lowerCamelCase_ : int ): '''simple docstring''' cls._set_token_in_kwargs(lowerCamelCase_ ) SCREAMING_SNAKE_CASE, SCREAMING_SNAKE_CASE : Any = cls.get_config_dict(lowerCamelCase_ , **lowerCamelCase_ ) # get the vision config dict if we are loading from GITConfig if config_dict.get("""model_type""" ) == "git": SCREAMING_SNAKE_CASE : Optional[Any] = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCamelCase_ , **lowerCamelCase_ ) class UpperCamelCase__ ( lowercase_ ): """simple docstring""" SCREAMING_SNAKE_CASE__ = '''git''' def __init__( self : List[str] , lowerCamelCase_ : Dict=None , lowerCamelCase_ : Tuple=3_05_22 , lowerCamelCase_ : Optional[Any]=7_68 , lowerCamelCase_ : Any=6 , lowerCamelCase_ : List[str]=12 , lowerCamelCase_ : List[str]=30_72 , lowerCamelCase_ : Union[str, Any]="gelu" , lowerCamelCase_ : List[str]=0.1 , lowerCamelCase_ : Dict=0.1 , lowerCamelCase_ : Optional[int]=10_24 , lowerCamelCase_ : int=0.02 , lowerCamelCase_ : Optional[int]=1e-12 , lowerCamelCase_ : Union[str, Any]=0 , lowerCamelCase_ : Optional[Any]="absolute" , lowerCamelCase_ : Dict=True , lowerCamelCase_ : Optional[Any]=False , lowerCamelCase_ : Optional[int]=1_01 , lowerCamelCase_ : Optional[Any]=1_02 , lowerCamelCase_ : List[str]=None , **lowerCamelCase_ : str , ): '''simple docstring''' super().__init__(bos_token_id=lowerCamelCase_ , eos_token_id=lowerCamelCase_ , pad_token_id=lowerCamelCase_ , **lowerCamelCase_ ) if vision_config is None: SCREAMING_SNAKE_CASE : Any = {} logger.info("""vision_config is None. initializing the GitVisionConfig with default values.""" ) SCREAMING_SNAKE_CASE : Union[str, Any] = GitVisionConfig(**lowerCamelCase_ ) SCREAMING_SNAKE_CASE : List[Any] = vocab_size SCREAMING_SNAKE_CASE : Tuple = hidden_size SCREAMING_SNAKE_CASE : int = num_hidden_layers SCREAMING_SNAKE_CASE : List[Any] = num_attention_heads SCREAMING_SNAKE_CASE : str = hidden_act SCREAMING_SNAKE_CASE : Dict = intermediate_size SCREAMING_SNAKE_CASE : Optional[Any] = hidden_dropout_prob SCREAMING_SNAKE_CASE : List[str] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE : List[Any] = max_position_embeddings SCREAMING_SNAKE_CASE : Dict = initializer_range SCREAMING_SNAKE_CASE : Optional[int] = layer_norm_eps SCREAMING_SNAKE_CASE : Union[str, Any] = position_embedding_type SCREAMING_SNAKE_CASE : Union[str, Any] = use_cache SCREAMING_SNAKE_CASE : List[str] = tie_word_embeddings SCREAMING_SNAKE_CASE : int = num_image_with_embedding SCREAMING_SNAKE_CASE : Optional[Any] = bos_token_id SCREAMING_SNAKE_CASE : str = eos_token_id def lowerCamelCase_ ( self : Optional[int] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE : Tuple = self.vision_config.to_dict() SCREAMING_SNAKE_CASE : Any = self.__class__.model_type return output
323
1
'''simple docstring''' def __lowerCamelCase ( _lowercase ) -> str: if isinstance(_lowercase , _lowercase ): raise TypeError("""'float' object cannot be interpreted as an integer""" ) if isinstance(_lowercase , _lowercase ): raise TypeError("""'str' object cannot be interpreted as an integer""" ) if num == 0: return "0b0" UpperCAmelCase : int = False if num < 0: UpperCAmelCase : Union[str, Any] = True UpperCAmelCase : Tuple = -num UpperCAmelCase : list[int] = [] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(_lowercase ) for e in binary ) return "0b" + "".join(str(_lowercase ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
338
'''simple docstring''' a : Dict = """ABCDEFGHIJKLMNOPQRSTUVWXYZ""" def __lowerCamelCase ( ) -> None: UpperCAmelCase : Optional[int] = input("""Enter message: """ ) UpperCAmelCase : Dict = input("""Enter key [alphanumeric]: """ ) UpperCAmelCase : Optional[Any] = input("""Encrypt/Decrypt [e/d]: """ ) if mode.lower().startswith("""e""" ): UpperCAmelCase : List[str] = """encrypt""" UpperCAmelCase : List[str] = encrypt_message(_lowercase , _lowercase ) elif mode.lower().startswith("""d""" ): UpperCAmelCase : Tuple = """decrypt""" UpperCAmelCase : str = decrypt_message(_lowercase , _lowercase ) print(F'''\n{mode.title()}ed message:''' ) print(_lowercase ) def __lowerCamelCase ( _lowercase , _lowercase ) -> str: return translate_message(_lowercase , _lowercase , """encrypt""" ) def __lowerCamelCase ( _lowercase , _lowercase ) -> str: return translate_message(_lowercase , _lowercase , """decrypt""" ) def __lowerCamelCase ( _lowercase , _lowercase , _lowercase ) -> str: UpperCAmelCase : Optional[int] = [] UpperCAmelCase : Optional[Any] = 0 UpperCAmelCase : Tuple = key.upper() for symbol in message: UpperCAmelCase : Dict = LETTERS.find(symbol.upper() ) if num != -1: if mode == "encrypt": num += LETTERS.find(key[key_index] ) elif mode == "decrypt": num -= LETTERS.find(key[key_index] ) num %= len(_lowercase ) if symbol.isupper(): translated.append(LETTERS[num] ) elif symbol.islower(): translated.append(LETTERS[num].lower() ) key_index += 1 if key_index == len(_lowercase ): UpperCAmelCase : Optional[int] = 0 else: translated.append(_lowercase ) return "".join(_lowercase ) if __name__ == "__main__": main()
338
1
"""simple docstring""" import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class _UpperCAmelCase ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,unittest.TestCase ): '''simple docstring''' a__ =StableUnCLIPPipeline a__ =TEXT_TO_IMAGE_PARAMS a__ =TEXT_TO_IMAGE_BATCH_PARAMS a__ =TEXT_TO_IMAGE_IMAGE_PARAMS a__ =TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false a__ =False def __lowerCAmelCase ( self ) -> Union[str, Any]: _UpperCAmelCase : Any = 3_2 _UpperCAmelCase : Dict = embedder_hidden_size # prior components torch.manual_seed(0 ) _UpperCAmelCase : int = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) torch.manual_seed(0 ) _UpperCAmelCase : Dict = CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=_lowerCAmelCase , projection_dim=_lowerCAmelCase , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) ) torch.manual_seed(0 ) _UpperCAmelCase : str = PriorTransformer( num_attention_heads=2 , attention_head_dim=1_2 , embedding_dim=_lowerCAmelCase , num_layers=1 , ) torch.manual_seed(0 ) _UpperCAmelCase : Union[str, Any] = DDPMScheduler( variance_type='''fixed_small_log''' , prediction_type='''sample''' , num_train_timesteps=1_0_0_0 , clip_sample=_lowerCAmelCase , clip_sample_range=5.0 , beta_schedule='''squaredcos_cap_v2''' , ) # regular denoising components torch.manual_seed(0 ) _UpperCAmelCase : List[Any] = StableUnCLIPImageNormalizer(embedding_dim=_lowerCAmelCase ) _UpperCAmelCase : Dict = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''' ) torch.manual_seed(0 ) _UpperCAmelCase : Dict = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) torch.manual_seed(0 ) _UpperCAmelCase : Optional[Any] = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=_lowerCAmelCase , projection_dim=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) ) torch.manual_seed(0 ) _UpperCAmelCase : Optional[Any] = UNetaDConditionModel( sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(3_2, 6_4) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=_lowerCAmelCase , layers_per_block=1 , upcast_attention=_lowerCAmelCase , use_linear_projection=_lowerCAmelCase , ) torch.manual_seed(0 ) _UpperCAmelCase : List[Any] = DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.00_085 , beta_end=0.012 , prediction_type='''v_prediction''' , set_alpha_to_one=_lowerCAmelCase , steps_offset=1 , ) torch.manual_seed(0 ) _UpperCAmelCase : int = AutoencoderKL() _UpperCAmelCase : Any = { # prior components '''prior_tokenizer''': prior_tokenizer, '''prior_text_encoder''': prior_text_encoder, '''prior''': prior, '''prior_scheduler''': prior_scheduler, # image noising components '''image_normalizer''': image_normalizer, '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder, '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, } return components def __lowerCAmelCase ( self , A , A=0 ) -> Tuple: if str(_lowerCAmelCase ).startswith('''mps''' ): _UpperCAmelCase : Dict = torch.manual_seed(_lowerCAmelCase ) else: _UpperCAmelCase : int = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _UpperCAmelCase : int = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''prior_num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def __lowerCAmelCase ( self ) -> List[Any]: _UpperCAmelCase : List[Any] = torch_device == '''cpu''' self._test_attention_slicing_forward_pass(test_max_difference=_lowerCAmelCase ) def __lowerCAmelCase ( self ) -> int: _UpperCAmelCase : str = torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=_lowerCAmelCase ) @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __lowerCAmelCase ( self ) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self ) -> Optional[int]: _UpperCAmelCase : Union[str, Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy''' ) _UpperCAmelCase : List[Any] = StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() _UpperCAmelCase : List[str] = torch.Generator(device='''cpu''' ).manual_seed(0 ) _UpperCAmelCase : Optional[int] = pipe('''anime turle''' , generator=_lowerCAmelCase , output_type='''np''' ) _UpperCAmelCase : Optional[Any] = output.images[0] assert image.shape == (7_6_8, 7_6_8, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase ) def __lowerCAmelCase ( self ) -> List[str]: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() _UpperCAmelCase : List[str] = StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa ) _UpperCAmelCase : List[str] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() _UpperCAmelCase : Tuple = pipe( '''anime turtle''' , prior_num_inference_steps=2 , num_inference_steps=2 , output_type='''np''' , ) _UpperCAmelCase : List[Any] = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 1_0**9
263
import unittest import numpy as np from diffusers import OnnxStableDiffusionInpaintPipelineLegacy from diffusers.utils.testing_utils import ( is_onnx_available, load_image, load_numpy, nightly, require_onnxruntime, require_torch_gpu, ) if is_onnx_available(): import onnxruntime as ort @nightly @require_onnxruntime @require_torch_gpu class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' @property def lowerCAmelCase_ ( self : Optional[Any] ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def lowerCAmelCase_ ( self : str ): SCREAMING_SNAKE_CASE_ = ort.SessionOptions() SCREAMING_SNAKE_CASE_ = False return options def lowerCAmelCase_ ( self : Tuple ): SCREAMING_SNAKE_CASE_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo.png' ) SCREAMING_SNAKE_CASE_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo_mask.png' ) SCREAMING_SNAKE_CASE_ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/red_cat_sitting_on_a_park_bench_onnx.npy' ) # using the PNDM scheduler by default SCREAMING_SNAKE_CASE_ = OnnxStableDiffusionInpaintPipelineLegacy.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='onnx' , safety_checker=_lowerCAmelCase , feature_extractor=_lowerCAmelCase , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = 'A red cat sitting on a park bench' SCREAMING_SNAKE_CASE_ = np.random.RandomState(0 ) SCREAMING_SNAKE_CASE_ = pipe( prompt=_lowerCAmelCase , image=_lowerCAmelCase , mask_image=_lowerCAmelCase , strength=0.75 , guidance_scale=7.5 , num_inference_steps=15 , generator=_lowerCAmelCase , output_type='np' , ) SCREAMING_SNAKE_CASE_ = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1E-2
225
0
'''simple docstring''' import numpy as np import torch from imwatermark import WatermarkEncoder # Copied from https://github.com/Stability-AI/generative-models/blob/613af104c6b85184091d42d374fef420eddb356d/scripts/demo/streamlit_helpers.py#L66 __lowerCAmelCase = 0B1011_0011_1110_1100_1001_0000_0111_1011_1011_0001_1001_1110 # bin(x)[2:] gives bits of x as str, use int to convert them to 0/1 __lowerCAmelCase = [int(bit) for bit in bin(WATERMARK_MESSAGE)[2:]] class UpperCAmelCase__ : """simple docstring""" def __init__( self : List[str] ): '''simple docstring''' _a : List[Any] = WATERMARK_BITS _a : int = WatermarkEncoder() self.encoder.set_watermark('bits' ,self.watermark ) def __lowercase ( self : str ,_a : torch.FloatTensor ): '''simple docstring''' if images.shape[-1] < 256: return images _a : Tuple = (255 * (images / 2 + 0.5)).cpu().permute(0 ,2 ,3 ,1 ).float().numpy() _a : Any = [self.encoder.encode(_a ,'dwtDct' ) for image in images] _a : Optional[Any] = torch.from_numpy(np.array(_a ) ).permute(0 ,3 ,1 ,2 ) _a : str = torch.clamp(2 * (images / 255 - 0.5) ,min=-1.0 ,max=1.0 ) return images
5
'''simple docstring''' def UpperCAmelCase_ (__a : str , __a : str ): """simple docstring""" _a : int = len(__a ) + 1 _a : List[str] = len(__a ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length i of input_string matches with prefix string of length j of # given pattern. # "dp" stands for dynamic programming. _a : Optional[int] = [[0 for i in range(__a )] for j in range(__a )] # since string of zero length match pattern of zero length _a : str = 1 # since pattern of zero length will never match with string of non-zero length for i in range(1 , __a ): _a : Optional[Any] = 0 # since string of zero length will match with pattern where there # is at least one * alternatively for j in range(1 , __a ): _a : Dict = dp[0][j - 2] if pattern[j - 1] == '*' else 0 # now using bottom-up approach to find for all remaining lengths for i in range(1 , __a ): for j in range(1 , __a ): if input_string[i - 1] == pattern[j - 1] or pattern[j - 1] == ".": _a : Tuple = dp[i - 1][j - 1] elif pattern[j - 1] == "*": if dp[i][j - 2] == 1: _a : List[str] = 1 elif pattern[j - 2] in (input_string[i - 1], "."): _a : int = dp[i - 1][j] else: _a : Any = 0 else: _a : Optional[Any] = 0 return bool(dp[-1][-1] ) if __name__ == "__main__": import doctest doctest.testmod() # inputing the strings # input_string = input("input a string :") # pattern = input("input a pattern :") __lowerCAmelCase = """aab""" __lowerCAmelCase = """c*a*b""" # using function to check whether given string matches the given pattern if match_pattern(input_string, pattern): print(f'''{input_string} matches the given pattern {pattern}''') else: print(f'''{input_string} does not match with the given pattern {pattern}''')
5
1
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" , "False" ) ) is not True , reason="Skipping test because should only be run when releasing minor transformers version" , ) @pytest.mark.usefixtures("sm_env" ) @parameterized_class( [ { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.g4dn.xlarge", "results": {"train_runtime": 650, "eval_accuracy": 0.6, "eval_loss": 0.9}, }, { "framework": "tensorflow", "script": "run_tf.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.g4dn.xlarge", "results": {"train_runtime": 600, "eval_accuracy": 0.3, "eval_loss": 0.9}, }, ] ) class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self : Optional[int] ): if self.framework == "pytorch": subprocess.run( F'cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'.split() ,encoding='''utf-8''' ,check=lowercase_ ,) assert hasattr(self ,'''env''' ) def __lowerCAmelCase ( self : str ,lowercase_ : Dict=1 ): # creates estimator return HuggingFace( entry_point=self.script ,source_dir=self.env.test_path ,role=self.env.role ,image_uri=self.env.image_uri ,base_job_name=F'{self.env.base_job_name}-single' ,instance_count=lowercase_ ,instance_type=self.instance_type ,debugger_hook_config=lowercase_ ,hyperparameters={**self.env.hyperparameters, '''model_name_or_path''': self.model_name_or_path} ,metric_definitions=self.env.metric_definitions ,py_version='''py36''' ,) def __lowerCAmelCase ( self : Dict ,lowercase_ : Optional[int] ): TrainingJobAnalytics(lowercase_ ).export_csv(F'{self.env.test_path}/{job_name}_metrics.csv' ) def __lowerCAmelCase ( self : Optional[Any] ): # create estimator lowerCAmelCase__ : Any = self.create_estimator() # run training estimator.fit() # result dataframe lowerCAmelCase__ : Any = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis lowerCAmelCase__ : Any = list(result_metrics_df[result_metrics_df.metric_name == '''eval_accuracy''']['''value'''] ) lowerCAmelCase__ : str = list(result_metrics_df[result_metrics_df.metric_name == '''eval_loss''']['''value'''] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping lowerCAmelCase__ : int = ( Session().describe_training_job(estimator.latest_training_job.name ).get('''TrainingTimeInSeconds''' ,9_9_9_9_9_9 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results['''eval_accuracy'''] for t in eval_accuracy ) assert all(t <= self.results['''eval_loss'''] for t in eval_loss ) # dump tests result into json file to share in PR with open(F'{estimator.latest_training_job.name}.json' ,'''w''' ) as outfile: json.dump({'''train_time''': train_runtime, '''eval_accuracy''': eval_accuracy, '''eval_loss''': eval_loss} ,lowercase_ )
106
def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: '''simple docstring''' __UpperCAmelCase = [[] for _ in range(SCREAMING_SNAKE_CASE )] __UpperCAmelCase = key - 1 if key <= 0: raise ValueError('''Height of grid can\'t be 0 or negative''' ) if key == 1 or len(SCREAMING_SNAKE_CASE ) <= key: return input_string for position, character in enumerate(SCREAMING_SNAKE_CASE ): __UpperCAmelCase = position % (lowest * 2) # puts it in bounds __UpperCAmelCase = min(SCREAMING_SNAKE_CASE , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(SCREAMING_SNAKE_CASE ) __UpperCAmelCase = [''''''.join(SCREAMING_SNAKE_CASE ) for row in temp_grid] __UpperCAmelCase = ''''''.join(SCREAMING_SNAKE_CASE ) return output_string def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: '''simple docstring''' __UpperCAmelCase = [] __UpperCAmelCase = key - 1 if key <= 0: raise ValueError('''Height of grid can\'t be 0 or negative''' ) if key == 1: return input_string __UpperCAmelCase = [[] for _ in range(SCREAMING_SNAKE_CASE )] # generates template for position in range(len(SCREAMING_SNAKE_CASE ) ): __UpperCAmelCase = position % (lowest * 2) # puts it in bounds __UpperCAmelCase = min(SCREAMING_SNAKE_CASE , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append('''*''' ) __UpperCAmelCase = 0 for row in temp_grid: # fills in the characters __UpperCAmelCase = input_string[counter : counter + len(SCREAMING_SNAKE_CASE )] grid.append(list(SCREAMING_SNAKE_CASE ) ) counter += len(SCREAMING_SNAKE_CASE ) __UpperCAmelCase = '''''' # reads as zigzag for position in range(len(SCREAMING_SNAKE_CASE ) ): __UpperCAmelCase = position % (lowest * 2) # puts it in bounds __UpperCAmelCase = min(SCREAMING_SNAKE_CASE , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def __a ( SCREAMING_SNAKE_CASE ) -> dict[int, str]: '''simple docstring''' __UpperCAmelCase = {} for key_guess in range(1 , len(SCREAMING_SNAKE_CASE ) ): # tries every key __UpperCAmelCase = decrypt(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return results if __name__ == "__main__": import doctest doctest.testmod()
333
0
"""simple docstring""" import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def snake_case ( A__=None ,A__=None ): return field(default_factory=lambda: default ,metadata=A__ ) @dataclass class UpperCamelCase_ : __magic_name__ = field( metadata={'''help''': '''The csv file to plot.'''} , ) __magic_name__ = field( default=__A , metadata={'''help''': '''Whether to plot along batch size or sequence length. Defaults to sequence length.'''} , ) __magic_name__ = field( default=__A , metadata={'''help''': '''Whether the csv file has time results or memory results. Defaults to memory results.'''} , ) __magic_name__ = field( default=__A , metadata={'''help''': '''Disable logarithmic scale when plotting'''} , ) __magic_name__ = field( default=__A , metadata={ '''help''': '''Whether the csv file has training results or inference results. Defaults to inference results.''' } , ) __magic_name__ = field( default=__A , metadata={'''help''': '''Filename under which the plot will be saved. If unused no plot is saved.'''} , ) __magic_name__ = list_field( default=__A , metadata={'''help''': '''List of model names that are used instead of the ones in the csv file.'''} ) def snake_case ( A__ ): try: int(A__ ) return True except ValueError: return False def snake_case ( A__ ): try: float(A__ ) return True except ValueError: return False class UpperCamelCase_ : def __init__( self : Optional[int] , lowerCAmelCase_ : Tuple ) -> Any: UpperCAmelCase_ : str = args UpperCAmelCase_ : Optional[Any] = defaultdict(lambda: {"bsz": [], "seq_len": [], "result": {}} ) with open(self.args.csv_file , newline="" ) as csv_file: UpperCAmelCase_ : Dict = csv.DictReader(lowerCAmelCase_ ) for row in reader: UpperCAmelCase_ : int = row["model"] self.result_dict[model_name]["bsz"].append(int(row["batch_size"] ) ) self.result_dict[model_name]["seq_len"].append(int(row["sequence_length"] ) ) if can_convert_to_int(row["result"] ): # value is not None UpperCAmelCase_ : Optional[int] = int(row["result"] ) elif can_convert_to_float(row["result"] ): # value is not None UpperCAmelCase_ : int = float(row["result"] ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Any: UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = plt.subplots() UpperCAmelCase_ : Optional[Any] = "Time usage" if self.args.is_time else "Memory usage" UpperCAmelCase_ : Optional[Any] = title_str + " for training" if self.args.is_train else title_str + " for inference" if not self.args.no_log_scale: # set logarithm scales ax.set_xscale("log" ) ax.set_yscale("log" ) for axis in [ax.xaxis, ax.yaxis]: axis.set_major_formatter(ScalarFormatter() ) for model_name_idx, model_name in enumerate(self.result_dict.keys() ): UpperCAmelCase_ : List[Any] = sorted(set(self.result_dict[model_name]["bsz"] ) ) UpperCAmelCase_ : List[Any] = sorted(set(self.result_dict[model_name]["seq_len"] ) ) UpperCAmelCase_ : str = self.result_dict[model_name]["result"] ((UpperCAmelCase_) , (UpperCAmelCase_)) : Optional[Any] = ( (batch_sizes, sequence_lengths) if self.args.plot_along_batch else (sequence_lengths, batch_sizes) ) UpperCAmelCase_ : List[Any] = ( model_name if self.args.short_model_names is None else self.args.short_model_names[model_name_idx] ) for inner_loop_value in inner_loop_array: if self.args.plot_along_batch: UpperCAmelCase_ : List[Any] = np.asarray( [results[(x, inner_loop_value)] for x in x_axis_array if (x, inner_loop_value) in results] , dtype=lowerCAmelCase_ , ) else: UpperCAmelCase_ : List[Any] = np.asarray( [results[(inner_loop_value, x)] for x in x_axis_array if (inner_loop_value, x) in results] , dtype=np.floataa , ) ((UpperCAmelCase_) , (UpperCAmelCase_)) : Optional[Any] = ( ("batch_size", "len") if self.args.plot_along_batch else ("in #tokens", "bsz") ) UpperCAmelCase_ : Any = np.asarray(lowerCAmelCase_ , lowerCAmelCase_ )[: len(lowerCAmelCase_ )] plt.scatter( lowerCAmelCase_ , lowerCAmelCase_ , label=f"""{label_model_name} - {inner_loop_label}: {inner_loop_value}""" ) plt.plot(lowerCAmelCase_ , lowerCAmelCase_ , "--" ) title_str += f""" {label_model_name} vs.""" UpperCAmelCase_ : Union[str, Any] = title_str[:-4] UpperCAmelCase_ : Any = "Time in s" if self.args.is_time else "Memory in MB" # plot plt.title(lowerCAmelCase_ ) plt.xlabel(lowerCAmelCase_ ) plt.ylabel(lowerCAmelCase_ ) plt.legend() if self.args.figure_png_file is not None: plt.savefig(self.args.figure_png_file ) else: plt.show() def snake_case ( ): UpperCAmelCase_ : Any = HfArgumentParser(A__ ) UpperCAmelCase_ : List[Any] = parser.parse_args_into_dataclasses()[0] UpperCAmelCase_ : Optional[Any] = Plot(args=A__ ) plot.plot() if __name__ == "__main__": main()
253
"""simple docstring""" import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class UpperCamelCase_ (__A ): __magic_name__ = '''M-CLIP''' def __init__( self : Any , lowerCAmelCase_ : str=1_024 , lowerCAmelCase_ : str=768 , **lowerCAmelCase_ : Union[str, Any] ) -> List[str]: UpperCAmelCase_ : Tuple = transformerDimSize UpperCAmelCase_ : List[str] = imageDimSize super().__init__(**lowerCAmelCase_ ) class UpperCamelCase_ (__A ): __magic_name__ = MCLIPConfig def __init__( self : str , lowerCAmelCase_ : int , *lowerCAmelCase_ : int , **lowerCAmelCase_ : List[Any] ) -> Any: super().__init__(lowerCAmelCase_ , *lowerCAmelCase_ , **lowerCAmelCase_ ) UpperCAmelCase_ : Union[str, Any] = XLMRobertaModel(lowerCAmelCase_ ) UpperCAmelCase_ : Optional[int] = torch.nn.Linear( in_features=config.transformerDimensions , out_features=config.numDims ) def _SCREAMING_SNAKE_CASE ( self : int , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Union[str, Any] ) -> List[str]: UpperCAmelCase_ : Any = self.transformer(input_ids=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] UpperCAmelCase_ : Tuple = (embs * attention_mask.unsqueeze(2 )).sum(dim=1 ) / attention_mask.sum(dim=1 )[:, None] return self.LinearTransformation(lowerCAmelCase_ ), embs
253
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import numpy as np import torch from ..models.clipseg import CLIPSegForImageSegmentation from ..utils import is_vision_available, requires_backends from .base import PipelineTool if is_vision_available(): from PIL import Image class _lowercase ( A__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = ( '''This is a tool that creates a segmentation mask of an image according to a label. It cannot create an image.''' '''It takes two arguments named `image` which should be the original image, and `label` which should be a text ''' '''describing the elements what should be identified in the segmentation mask. The tool returns the mask.''' ) SCREAMING_SNAKE_CASE__ : Dict = '''CIDAS/clipseg-rd64-refined''' SCREAMING_SNAKE_CASE__ : Optional[int] = '''image_segmenter''' SCREAMING_SNAKE_CASE__ : Tuple = CLIPSegForImageSegmentation SCREAMING_SNAKE_CASE__ : List[Any] = ['''image''', '''text'''] SCREAMING_SNAKE_CASE__ : str = ['''image'''] def __init__( self :Dict , *lowerCAmelCase__ :str , **lowerCAmelCase__ :str ) -> Tuple: requires_backends(self , ['''vision'''] ) super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__ ) def __magic_name__( self :Union[str, Any] , lowerCAmelCase__ :"Image" , lowerCAmelCase__ :str ) -> Dict: return self.pre_processor(text=[label] , images=[image] , padding=lowerCAmelCase__ , return_tensors='''pt''' ) def __magic_name__( self :List[Any] , lowerCAmelCase__ :Tuple ) -> Optional[int]: with torch.no_grad(): __SCREAMING_SNAKE_CASE : Optional[Any] = self.model(**lowerCAmelCase__ ).logits return logits def __magic_name__( self :Optional[Any] , lowerCAmelCase__ :Optional[int] ) -> Any: __SCREAMING_SNAKE_CASE : Tuple = outputs.cpu().detach().numpy() __SCREAMING_SNAKE_CASE : Any = 0 __SCREAMING_SNAKE_CASE : Union[str, Any] = 1 return Image.fromarray((array * 255).astype(np.uinta ) )
9
from collections import Counter import numpy as np from sklearn import datasets from sklearn.model_selection import train_test_split __lowerCAmelCase : List[Any] =datasets.load_iris() __lowerCAmelCase : Tuple =np.array(data['data']) __lowerCAmelCase : Dict =np.array(data['target']) __lowerCAmelCase : List[str] =data['target_names'] __lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase : str =train_test_split(X, y) def _UpperCamelCase ( lowercase__ , lowercase__ ): return np.linalg.norm(np.array(lowercase__ ) - np.array(lowercase__ ) ) def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__=5 ): __SCREAMING_SNAKE_CASE : Optional[int] = zip(lowercase__ , lowercase__ ) # List of distances of all points from the point to be classified __SCREAMING_SNAKE_CASE : Dict = [] for data_point in data: __SCREAMING_SNAKE_CASE : Tuple = euclidean_distance(data_point[0] , lowercase__ ) distances.append((distance, data_point[1]) ) # Choosing 'k' points with the least distances. __SCREAMING_SNAKE_CASE : int = [i[1] for i in sorted(lowercase__ )[:k]] # Most commonly occurring class among them # is the class into which the point is classified __SCREAMING_SNAKE_CASE : Any = Counter(lowercase__ ).most_common(1 )[0][0] return classes[result] if __name__ == "__main__": print(classifier(X_train, y_train, classes, [4.4, 3.1, 1.3, 1.4]))
9
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'weiweishi/roc-bert-base-zh': 'https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json', } class lowerCamelCase (__SCREAMING_SNAKE_CASE ): '''simple docstring''' _snake_case : str = '''roc_bert''' def __init__( self , _UpperCamelCase=3_0_5_2_2 , _UpperCamelCase=7_6_8 , _UpperCamelCase=1_2 , _UpperCamelCase=1_2 , _UpperCamelCase=3_0_7_2 , _UpperCamelCase="gelu" , _UpperCamelCase=0.1 , _UpperCamelCase=0.1 , _UpperCamelCase=5_1_2 , _UpperCamelCase=2 , _UpperCamelCase=0.02 , _UpperCamelCase=1E-12 , _UpperCamelCase=True , _UpperCamelCase=0 , _UpperCamelCase="absolute" , _UpperCamelCase=None , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=7_6_8 , _UpperCamelCase=9_1_0 , _UpperCamelCase=5_1_2 , _UpperCamelCase=2_4_8_5_8 , _UpperCamelCase=True , **_UpperCamelCase , ) -> List[str]: UpperCAmelCase_ : int = vocab_size UpperCAmelCase_ : str = max_position_embeddings UpperCAmelCase_ : str = hidden_size UpperCAmelCase_ : List[str] = num_hidden_layers UpperCAmelCase_ : List[Any] = num_attention_heads UpperCAmelCase_ : str = intermediate_size UpperCAmelCase_ : Dict = hidden_act UpperCAmelCase_ : Any = hidden_dropout_prob UpperCAmelCase_ : str = attention_probs_dropout_prob UpperCAmelCase_ : Dict = initializer_range UpperCAmelCase_ : List[Any] = type_vocab_size UpperCAmelCase_ : List[str] = layer_norm_eps UpperCAmelCase_ : Optional[Any] = use_cache UpperCAmelCase_ : str = enable_pronunciation UpperCAmelCase_ : Dict = enable_shape UpperCAmelCase_ : List[str] = pronunciation_embed_dim UpperCAmelCase_ : List[Any] = pronunciation_vocab_size UpperCAmelCase_ : List[Any] = shape_embed_dim UpperCAmelCase_ : Tuple = shape_vocab_size UpperCAmelCase_ : Optional[int] = concat_input UpperCAmelCase_ : Optional[Any] = position_embedding_type UpperCAmelCase_ : Optional[Any] = classifier_dropout super().__init__(pad_token_id=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE )
364
def lowercase__ ( __snake_case : str , __snake_case : int , __snake_case : List[str] ): '''simple docstring''' if n == 0: return 1 elif n % 2 == 1: return (binary_exponentiation(__snake_case , n - 1 , __snake_case ) * a) % mod else: UpperCAmelCase_ : Optional[int] = binary_exponentiation(__snake_case , n / 2 , __snake_case ) return (b * b) % mod # a prime number __UpperCAmelCase = 701 __UpperCAmelCase = 1000000000 __UpperCAmelCase = 10 # using binary exponentiation function, O(log(p)): print((a / b) % p == (a * binary_exponentiation(b, p - 2, p)) % p) print((a / b) % p == (a * b ** (p - 2)) % p)
145
0
'''simple docstring''' import numpy as np # Importing the Keras libraries and packages import tensorflow as tf from tensorflow.keras import layers, models if __name__ == "__main__": # Initialising the CNN # (Sequential- Building the model layer by layer) a : Dict = models.Sequential() # Step 1 - Convolution # Here 64,64 is the length & breadth of dataset images and 3 is for the RGB channel # (3,3) is the kernel size (filter matrix) classifier.add( layers.ConvaD(32, (3, 3), input_shape=(64, 64, 3), activation="relu") ) # Step 2 - Pooling classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Adding a second convolutional layer classifier.add(layers.ConvaD(32, (3, 3), activation="relu")) classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Step 3 - Flattening classifier.add(layers.Flatten()) # Step 4 - Full connection classifier.add(layers.Dense(units=1_28, activation="relu")) classifier.add(layers.Dense(units=1, activation="sigmoid")) # Compiling the CNN classifier.compile( optimizer="adam", loss="binary_crossentropy", metrics=["accuracy"] ) # Part 2 - Fitting the CNN to the images # Load Trained model weights # from keras.models import load_model # regressor=load_model('cnn.h5') a : int = tf.keras.preprocessing.image.ImageDataGenerator( rescale=1.0 / 2_55, shear_range=0.2, zoom_range=0.2, horizontal_flip=True ) a : List[Any] = tf.keras.preprocessing.image.ImageDataGenerator(rescale=1.0 / 2_55) a : List[str] = train_datagen.flow_from_directory( "dataset/training_set", target_size=(64, 64), batch_size=32, class_mode="binary" ) a : List[Any] = test_datagen.flow_from_directory( "dataset/test_set", target_size=(64, 64), batch_size=32, class_mode="binary" ) classifier.fit_generator( training_set, steps_per_epoch=5, epochs=30, validation_data=test_set ) classifier.save("cnn.h5") # Part 3 - Making new predictions a : Any = tf.keras.preprocessing.image.load_img( "dataset/single_prediction/image.png", target_size=(64, 64) ) a : Any = tf.keras.preprocessing.image.img_to_array(test_image) a : List[str] = np.expand_dims(test_image, axis=0) a : int = classifier.predict(test_image) # training_set.class_indices if result[0][0] == 0: a : Dict = "Normal" if result[0][0] == 1: a : str = "Abnormality detected"
311
'''simple docstring''' import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class UpperCamelCase__ : """simple docstring""" @staticmethod def A_ ( *snake_case , **snake_case ): '''simple docstring''' pass @is_pipeline_test @require_vision @require_torch class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def A_ ( self , snake_case , snake_case , snake_case ): '''simple docstring''' UpperCAmelCase : str = pipeline( "zero-shot-object-detection" , model="hf-internal-testing/tiny-random-owlvit-object-detection" ) UpperCAmelCase : Union[str, Any] = [ { "image": "./tests/fixtures/tests_samples/COCO/000000039769.png", "candidate_labels": ["cat", "remote", "couch"], } ] return object_detector, examples def A_ ( self , snake_case , snake_case ): '''simple docstring''' UpperCAmelCase : List[Any] = object_detector(examples[0] , threshold=0.0 ) UpperCAmelCase : Dict = len(snake_case ) self.assertGreater(snake_case , 0 ) self.assertEqual( snake_case , [ { "score": ANY(snake_case ), "label": ANY(snake_case ), "box": {"xmin": ANY(snake_case ), "ymin": ANY(snake_case ), "xmax": ANY(snake_case ), "ymax": ANY(snake_case )}, } for i in range(snake_case ) ] , ) @require_tf @unittest.skip("Zero Shot Object Detection not implemented in TF" ) def A_ ( self ): '''simple docstring''' pass @require_torch def A_ ( self ): '''simple docstring''' UpperCAmelCase : Optional[Any] = pipeline( "zero-shot-object-detection" , model="hf-internal-testing/tiny-random-owlvit-object-detection" ) UpperCAmelCase : Optional[Any] = object_detector( "./tests/fixtures/tests_samples/COCO/000000039769.png" , candidate_labels=["cat", "remote", "couch"] , threshold=0.64 , ) self.assertEqual( nested_simplify(snake_case , decimals=4 ) , [ {"score": 0.7235, "label": "cat", "box": {"xmin": 2_0_4, "ymin": 1_6_7, "xmax": 2_3_2, "ymax": 1_9_0}}, {"score": 0.7218, "label": "remote", "box": {"xmin": 2_0_4, "ymin": 1_6_7, "xmax": 2_3_2, "ymax": 1_9_0}}, {"score": 0.7184, "label": "couch", "box": {"xmin": 2_0_4, "ymin": 1_6_7, "xmax": 2_3_2, "ymax": 1_9_0}}, {"score": 0.6748, "label": "remote", "box": {"xmin": 5_7_1, "ymin": 8_3, "xmax": 5_9_8, "ymax": 1_0_3}}, {"score": 0.6656, "label": "cat", "box": {"xmin": 5_7_1, "ymin": 8_3, "xmax": 5_9_8, "ymax": 1_0_3}}, {"score": 0.6614, "label": "couch", "box": {"xmin": 5_7_1, "ymin": 8_3, "xmax": 5_9_8, "ymax": 1_0_3}}, {"score": 0.6456, "label": "remote", "box": {"xmin": 4_9_4, "ymin": 1_0_5, "xmax": 5_2_1, "ymax": 1_2_7}}, {"score": 0.642, "label": "remote", "box": {"xmin": 6_7, "ymin": 2_7_4, "xmax": 9_3, "ymax": 2_9_7}}, {"score": 0.6419, "label": "cat", "box": {"xmin": 4_9_4, "ymin": 1_0_5, "xmax": 5_2_1, "ymax": 1_2_7}}, ] , ) UpperCAmelCase : Tuple = object_detector( [ { "image": "./tests/fixtures/tests_samples/COCO/000000039769.png", "candidate_labels": ["cat", "remote", "couch"], } ] , threshold=0.64 , ) self.assertEqual( nested_simplify(snake_case , decimals=4 ) , [ [ {"score": 0.7235, "label": "cat", "box": {"xmin": 2_0_4, "ymin": 1_6_7, "xmax": 2_3_2, "ymax": 1_9_0}}, {"score": 0.7218, "label": "remote", "box": {"xmin": 2_0_4, "ymin": 1_6_7, "xmax": 2_3_2, "ymax": 1_9_0}}, {"score": 0.7184, "label": "couch", "box": {"xmin": 2_0_4, "ymin": 1_6_7, "xmax": 2_3_2, "ymax": 1_9_0}}, {"score": 0.6748, "label": "remote", "box": {"xmin": 5_7_1, "ymin": 8_3, "xmax": 5_9_8, "ymax": 1_0_3}}, {"score": 0.6656, "label": "cat", "box": {"xmin": 5_7_1, "ymin": 8_3, "xmax": 5_9_8, "ymax": 1_0_3}}, {"score": 0.6614, "label": "couch", "box": {"xmin": 5_7_1, "ymin": 8_3, "xmax": 5_9_8, "ymax": 1_0_3}}, {"score": 0.6456, "label": "remote", "box": {"xmin": 4_9_4, "ymin": 1_0_5, "xmax": 5_2_1, "ymax": 1_2_7}}, {"score": 0.642, "label": "remote", "box": {"xmin": 6_7, "ymin": 2_7_4, "xmax": 9_3, "ymax": 2_9_7}}, {"score": 0.6419, "label": "cat", "box": {"xmin": 4_9_4, "ymin": 1_0_5, "xmax": 5_2_1, "ymax": 1_2_7}}, ] ] , ) @require_torch @slow def A_ ( self ): '''simple docstring''' UpperCAmelCase : Tuple = pipeline("zero-shot-object-detection" ) UpperCAmelCase : Optional[int] = object_detector( "http://images.cocodataset.org/val2017/000000039769.jpg" , candidate_labels=["cat", "remote", "couch"] , ) self.assertEqual( nested_simplify(snake_case , decimals=4 ) , [ {"score": 0.2868, "label": "cat", "box": {"xmin": 3_2_4, "ymin": 2_0, "xmax": 6_4_0, "ymax": 3_7_3}}, {"score": 0.277, "label": "remote", "box": {"xmin": 4_0, "ymin": 7_2, "xmax": 1_7_7, "ymax": 1_1_5}}, {"score": 0.2537, "label": "cat", "box": {"xmin": 1, "ymin": 5_5, "xmax": 3_1_5, "ymax": 4_7_2}}, {"score": 0.1474, "label": "remote", "box": {"xmin": 3_3_5, "ymin": 7_4, "xmax": 3_7_1, "ymax": 1_8_7}}, {"score": 0.1208, "label": "couch", "box": {"xmin": 4, "ymin": 0, "xmax": 6_4_2, "ymax": 4_7_6}}, ] , ) UpperCAmelCase : Union[str, Any] = object_detector( [ { "image": "http://images.cocodataset.org/val2017/000000039769.jpg", "candidate_labels": ["cat", "remote", "couch"], }, { "image": "http://images.cocodataset.org/val2017/000000039769.jpg", "candidate_labels": ["cat", "remote", "couch"], }, ] , ) self.assertEqual( nested_simplify(snake_case , decimals=4 ) , [ [ {"score": 0.2868, "label": "cat", "box": {"xmin": 3_2_4, "ymin": 2_0, "xmax": 6_4_0, "ymax": 3_7_3}}, {"score": 0.277, "label": "remote", "box": {"xmin": 4_0, "ymin": 7_2, "xmax": 1_7_7, "ymax": 1_1_5}}, {"score": 0.2537, "label": "cat", "box": {"xmin": 1, "ymin": 5_5, "xmax": 3_1_5, "ymax": 4_7_2}}, {"score": 0.1474, "label": "remote", "box": {"xmin": 3_3_5, "ymin": 7_4, "xmax": 3_7_1, "ymax": 1_8_7}}, {"score": 0.1208, "label": "couch", "box": {"xmin": 4, "ymin": 0, "xmax": 6_4_2, "ymax": 4_7_6}}, ], [ {"score": 0.2868, "label": "cat", "box": {"xmin": 3_2_4, "ymin": 2_0, "xmax": 6_4_0, "ymax": 3_7_3}}, {"score": 0.277, "label": "remote", "box": {"xmin": 4_0, "ymin": 7_2, "xmax": 1_7_7, "ymax": 1_1_5}}, {"score": 0.2537, "label": "cat", "box": {"xmin": 1, "ymin": 5_5, "xmax": 3_1_5, "ymax": 4_7_2}}, {"score": 0.1474, "label": "remote", "box": {"xmin": 3_3_5, "ymin": 7_4, "xmax": 3_7_1, "ymax": 1_8_7}}, {"score": 0.1208, "label": "couch", "box": {"xmin": 4, "ymin": 0, "xmax": 6_4_2, "ymax": 4_7_6}}, ], ] , ) @require_tf @unittest.skip("Zero Shot Object Detection not implemented in TF" ) def A_ ( self ): '''simple docstring''' pass @require_torch @slow def A_ ( self ): '''simple docstring''' UpperCAmelCase : Any = 0.2 UpperCAmelCase : Union[str, Any] = pipeline("zero-shot-object-detection" ) UpperCAmelCase : str = object_detector( "http://images.cocodataset.org/val2017/000000039769.jpg" , candidate_labels=["cat", "remote", "couch"] , threshold=snake_case , ) self.assertEqual( nested_simplify(snake_case , decimals=4 ) , [ {"score": 0.2868, "label": "cat", "box": {"xmin": 3_2_4, "ymin": 2_0, "xmax": 6_4_0, "ymax": 3_7_3}}, {"score": 0.277, "label": "remote", "box": {"xmin": 4_0, "ymin": 7_2, "xmax": 1_7_7, "ymax": 1_1_5}}, {"score": 0.2537, "label": "cat", "box": {"xmin": 1, "ymin": 5_5, "xmax": 3_1_5, "ymax": 4_7_2}}, ] , ) @require_torch @slow def A_ ( self ): '''simple docstring''' UpperCAmelCase : Dict = 2 UpperCAmelCase : Optional[Any] = pipeline("zero-shot-object-detection" ) UpperCAmelCase : List[str] = object_detector( "http://images.cocodataset.org/val2017/000000039769.jpg" , candidate_labels=["cat", "remote", "couch"] , top_k=snake_case , ) self.assertEqual( nested_simplify(snake_case , decimals=4 ) , [ {"score": 0.2868, "label": "cat", "box": {"xmin": 3_2_4, "ymin": 2_0, "xmax": 6_4_0, "ymax": 3_7_3}}, {"score": 0.277, "label": "remote", "box": {"xmin": 4_0, "ymin": 7_2, "xmax": 1_7_7, "ymax": 1_1_5}}, ] , )
311
1
from manim import * class a ( _A ): '''simple docstring''' def lowerCamelCase_ ( self : Optional[int] ): UpperCAmelCase_ = Rectangle(height=0.5 , width=0.5 ) UpperCAmelCase_ = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) UpperCAmelCase_ = [mem.copy() for i in range(6 )] UpperCAmelCase_ = [mem.copy() for i in range(6 )] UpperCAmelCase_ = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase_ = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase_ = VGroup(__snake_case , __snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase_ = Text('''CPU''' , font_size=24 ) UpperCAmelCase_ = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__snake_case ) UpperCAmelCase_ = [mem.copy() for i in range(1 )] UpperCAmelCase_ = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase_ = Text('''GPU''' , font_size=24 ) UpperCAmelCase_ = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) gpu.align_to(__snake_case , __snake_case ) gpu.set_x(gpu.get_x() - 1 ) self.add(__snake_case ) UpperCAmelCase_ = [mem.copy() for i in range(6 )] UpperCAmelCase_ = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase_ = Text('''Model''' , font_size=24 ) UpperCAmelCase_ = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) model.move_to([3, -1.0, 0] ) self.play( Create(__snake_case , run_time=1 ) , Create(__snake_case , run_time=1 ) , Create(__snake_case , run_time=1 ) , ) UpperCAmelCase_ = MarkupText( F'First, an empty model skeleton is loaded\ninto <span fgcolor=\'{YELLOW}\'>memory</span> without using much RAM.' , font_size=24 , ) UpperCAmelCase_ = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCAmelCase_ = MarkupText( F'<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) step_a.move_to([2, 2, 0] ) self.play(Write(__snake_case , run_time=2.5 ) , Write(__snake_case ) , Write(__snake_case ) ) self.add(__snake_case ) UpperCAmelCase_ = [] UpperCAmelCase_ = [] UpperCAmelCase_ = [] for i, rect in enumerate(__snake_case ): UpperCAmelCase_ = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(__snake_case , opacity=0.7 ) cpu_target.move_to(__snake_case ) cpu_target.generate_target() UpperCAmelCase_ = 0.46 / 4 UpperCAmelCase_ = 0.46 / 3 if i == 0: cpu_target.target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=__snake_case ) cpu_target.target.set_x(cpu_target.target.get_x() + 0.1 ) elif i == 3: cpu_target.target.next_to(cpu_targs[0].target , direction=__snake_case , buff=0.0 ) else: cpu_target.target.next_to(cpu_targs[i - 1].target , direction=__snake_case , buff=0.0 ) cpu_targs.append(__snake_case ) first_animations.append(rect.animate(run_time=0.5 ).set_stroke(__snake_case ) ) second_animations.append(MoveToTarget(__snake_case , run_time=1.5 ) ) self.play(*__snake_case ) self.play(*__snake_case ) self.wait()
364
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class a ( unittest.TestCase ): '''simple docstring''' def __init__( self : Tuple , __snake_case : int , __snake_case : List[Any]=7 , __snake_case : Any=3 , __snake_case : Any=18 , __snake_case : str=30 , __snake_case : Any=4_00 , __snake_case : Optional[int]=True , __snake_case : str=None , __snake_case : Any=True , __snake_case : List[Any]=None , ): UpperCAmelCase_ = size if size is not None else {'''shortest_edge''': 20} UpperCAmelCase_ = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = image_size UpperCAmelCase_ = min_resolution UpperCAmelCase_ = max_resolution UpperCAmelCase_ = do_resize UpperCAmelCase_ = size UpperCAmelCase_ = do_center_crop UpperCAmelCase_ = crop_size def lowerCamelCase_ ( self : Any ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, } @require_torch @require_vision class a ( _A , unittest.TestCase ): '''simple docstring''' lowerCAmelCase : Union[str, Any] = MobileNetVaImageProcessor if is_vision_available() else None def lowerCamelCase_ ( self : Any ): UpperCAmelCase_ = MobileNetVaImageProcessingTester(self ) @property def lowerCamelCase_ ( self : List[str] ): return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase_ ( self : Optional[int] ): UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__snake_case , '''do_resize''' ) ) self.assertTrue(hasattr(__snake_case , '''size''' ) ) self.assertTrue(hasattr(__snake_case , '''do_center_crop''' ) ) self.assertTrue(hasattr(__snake_case , '''crop_size''' ) ) def lowerCamelCase_ ( self : Dict ): UpperCAmelCase_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 20} ) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18} ) UpperCAmelCase_ = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42} ) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84} ) def lowerCamelCase_ ( self : Optional[int] ): pass def lowerCamelCase_ ( self : Tuple ): # Initialize image_processing UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=__snake_case ) for image in image_inputs: self.assertIsInstance(__snake_case , Image.Image ) # Test not batched input UpperCAmelCase_ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched UpperCAmelCase_ = image_processing(__snake_case , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowerCamelCase_ ( self : str ): # Initialize image_processing UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=__snake_case , numpify=__snake_case ) for image in image_inputs: self.assertIsInstance(__snake_case , np.ndarray ) # Test not batched input UpperCAmelCase_ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched UpperCAmelCase_ = image_processing(__snake_case , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def lowerCamelCase_ ( self : int ): # Initialize image_processing UpperCAmelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=__snake_case , torchify=__snake_case ) for image in image_inputs: self.assertIsInstance(__snake_case , torch.Tensor ) # Test not batched input UpperCAmelCase_ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched UpperCAmelCase_ = image_processing(__snake_case , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
177
0
'''simple docstring''' import os def lowerCamelCase ( UpperCAmelCase__ : str = "input.txt" ) -> int: with open(os.path.join(os.path.dirname(UpperCAmelCase__ ) , UpperCAmelCase__ ) ) as input_file: lowercase_ : str = [ [int(UpperCAmelCase__ ) for element in line.split(""",""" )] for line in input_file.readlines() ] lowercase_ : Optional[Any] = len(UpperCAmelCase__ ) lowercase_ : Any = len(matrix[0] ) lowercase_ : Union[str, Any] = [[-1 for _ in range(UpperCAmelCase__ )] for _ in range(UpperCAmelCase__ )] for i in range(UpperCAmelCase__ ): lowercase_ : int = matrix[i][0] for j in range(1 , UpperCAmelCase__ ): for i in range(UpperCAmelCase__ ): lowercase_ : Union[str, Any] = minimal_path_sums[i][j - 1] + matrix[i][j] for i in range(1 , UpperCAmelCase__ ): lowercase_ : Tuple = min( minimal_path_sums[i][j] , minimal_path_sums[i - 1][j] + matrix[i][j] ) for i in range(rows - 2 , -1 , -1 ): lowercase_ : Dict = min( minimal_path_sums[i][j] , minimal_path_sums[i + 1][j] + matrix[i][j] ) return min(minimal_path_sums_row[-1] for minimal_path_sums_row in minimal_path_sums ) if __name__ == "__main__": print(f"""{solution() = }""")
239
'''simple docstring''' def lowerCamelCase ( UpperCAmelCase__ : int , UpperCAmelCase__ : int ) -> bool: return numa ^ numa < 0 if __name__ == "__main__": import doctest doctest.testmod()
239
1
'''simple docstring''' import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin A_ : List[Any] = get_tests_dir("""fixtures/test_sentencepiece.model""") A_ : Any = get_tests_dir("""fixtures/test_sentencepiece_bpe.model""") A_ : Dict = """pt""" if is_torch_available() else """tf""" @require_sentencepiece @require_tokenizers class lowercase ( _lowerCamelCase , unittest.TestCase ): """simple docstring""" UpperCAmelCase = CamembertTokenizer UpperCAmelCase = CamembertTokenizerFast UpperCAmelCase = True UpperCAmelCase = True def _snake_case ( self ) -> Dict: super().setUp() # We have a SentencePiece fixture for testing _UpperCAmelCase : str = CamembertTokenizer(a_ ) tokenizer.save_pretrained(self.tmpdirname ) def _snake_case ( self ) -> str: _UpperCAmelCase : Optional[int] = """<pad>""" _UpperCAmelCase : Any = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) ,a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) ,a_ ) def _snake_case ( self ) -> str: _UpperCAmelCase : int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,"""<s>NOTUSED""" ) self.assertEqual(vocab_keys[1] ,"""<pad>""" ) self.assertEqual(vocab_keys[-1] ,"""<mask>""" ) self.assertEqual(len(a_ ) ,1_004 ) def _snake_case ( self ) -> Union[str, Any]: self.assertEqual(self.get_tokenizer().vocab_size ,1_005 ) def _snake_case ( self ) -> int: _UpperCAmelCase : Union[str, Any] = CamembertTokenizer(a_ ) tokenizer.save_pretrained(self.tmpdirname ) _UpperCAmelCase : Tuple = CamembertTokenizerFast.from_pretrained(self.tmpdirname ) _UpperCAmelCase : Union[str, Any] = """I was born in 92000, and this is falsé.""" _UpperCAmelCase : List[Any] = tokenizer.encode(a_ ) _UpperCAmelCase : str = rust_tokenizer.encode(a_ ) self.assertListEqual(a_ ,a_ ) _UpperCAmelCase : Optional[int] = tokenizer.encode(a_ ,add_special_tokens=a_ ) _UpperCAmelCase : Optional[int] = rust_tokenizer.encode(a_ ,add_special_tokens=a_ ) self.assertListEqual(a_ ,a_ ) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) _UpperCAmelCase : int = tokenizer.convert_ids_to_tokens(a_ ) _UpperCAmelCase : List[str] = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_ ,a_ ) def _snake_case ( self ) -> Union[str, Any]: if not self.test_rust_tokenizer: return _UpperCAmelCase : str = self.get_tokenizer() _UpperCAmelCase : Any = self.get_rust_tokenizer() _UpperCAmelCase : Union[str, Any] = """I was born in 92000, and this is falsé.""" _UpperCAmelCase : Optional[int] = tokenizer.tokenize(a_ ) _UpperCAmelCase : Union[str, Any] = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_ ,a_ ) _UpperCAmelCase : Any = tokenizer.encode(a_ ,add_special_tokens=a_ ) _UpperCAmelCase : int = rust_tokenizer.encode(a_ ,add_special_tokens=a_ ) self.assertListEqual(a_ ,a_ ) _UpperCAmelCase : List[Any] = self.get_rust_tokenizer() _UpperCAmelCase : Tuple = tokenizer.encode(a_ ) _UpperCAmelCase : Dict = rust_tokenizer.encode(a_ ) self.assertListEqual(a_ ,a_ ) @slow def _snake_case ( self ) -> Any: # fmt: off _UpperCAmelCase : List[str] = {"""input_ids""": [[5, 54, 7_196, 297, 30, 23, 776, 18, 11, 3_215, 3_705, 8_252, 22, 3_164, 1_181, 2_116, 29, 16, 813, 25, 791, 3_314, 20, 3_446, 38, 27_575, 120, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 468, 17, 11, 9_088, 20, 1_517, 8, 22_804, 18_818, 10, 38, 629, 607, 607, 142, 19, 7_196, 867, 56, 10_326, 24, 2_267, 20, 416, 5_072, 15_612, 233, 734, 7, 2_399, 27, 16, 3_015, 1_649, 7, 24, 20, 4_338, 2_399, 27, 13, 3_400, 14, 13, 6_189, 8, 930, 9, 6]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. _UpperCAmelCase : List[str] = [ """Le transformeur est un modèle d'apprentissage profond introduit en 2017, """ """utilisé principalement dans le domaine du traitement automatique des langues (TAL).""", """À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus """ """pour gérer des données séquentielles, telles que le langage naturel, pour des tâches """ """telles que la traduction et la synthèse de texte.""", ] self.tokenizer_integration_test_util( expected_encoding=a_ ,model_name="""camembert-base""" ,revision="""3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf""" ,sequences=a_ ,)
349
'''simple docstring''' import argparse import copy def snake_case_ ( lowerCAmelCase_ )-> Dict: '''simple docstring''' _UpperCAmelCase : Dict = {} with open(lowerCAmelCase_ ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: _UpperCAmelCase : Optional[int] = [] _list.append([line.split()[1], line.split()[2]] ) _UpperCAmelCase : List[str] = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: _UpperCAmelCase : List[str] = [] _list.append([line.split()[0], line.split()[2]] ) _UpperCAmelCase : Optional[int] = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ )-> List[Any]: '''simple docstring''' with open(lowerCAmelCase_ ) as f: _UpperCAmelCase : List[Any] = f.read(1 ) _UpperCAmelCase : int = start_node _UpperCAmelCase : List[Any] = [] _UpperCAmelCase : Dict = start_node _UpperCAmelCase : Any = 0 while visiting not in first_solution: _UpperCAmelCase : Optional[int] = 10000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(lowerCAmelCase_ ) and k[0] not in first_solution: _UpperCAmelCase : Optional[int] = k[1] _UpperCAmelCase : List[str] = k[0] first_solution.append(lowerCAmelCase_ ) _UpperCAmelCase : Optional[int] = distance_of_first_solution + int(lowerCAmelCase_ ) _UpperCAmelCase : Dict = best_node first_solution.append(lowerCAmelCase_ ) _UpperCAmelCase : List[str] = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 _UpperCAmelCase : int = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 10000 ) return first_solution, distance_of_first_solution def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ )-> int: '''simple docstring''' _UpperCAmelCase : int = [] for n in solution[1:-1]: _UpperCAmelCase : Tuple = solution.index(lowerCAmelCase_ ) for kn in solution[1:-1]: _UpperCAmelCase : int = solution.index(lowerCAmelCase_ ) if n == kn: continue _UpperCAmelCase : Tuple = copy.deepcopy(lowerCAmelCase_ ) _UpperCAmelCase : Union[str, Any] = kn _UpperCAmelCase : List[str] = n _UpperCAmelCase : Optional[int] = 0 for k in _tmp[:-1]: _UpperCAmelCase : List[str] = _tmp[_tmp.index(lowerCAmelCase_ ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: _UpperCAmelCase : Dict = distance + int(i[1] ) _tmp.append(lowerCAmelCase_ ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) _UpperCAmelCase : Dict = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda lowerCAmelCase_ : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> int: '''simple docstring''' _UpperCAmelCase : List[Any] = 1 _UpperCAmelCase : Optional[Any] = first_solution _UpperCAmelCase : List[Any] = [] _UpperCAmelCase : List[Any] = distance_of_first_solution _UpperCAmelCase : Dict = solution while count <= iters: _UpperCAmelCase : Any = find_neighborhood(lowerCAmelCase_ , lowerCAmelCase_ ) _UpperCAmelCase : Dict = 0 _UpperCAmelCase : Optional[Any] = neighborhood[index_of_best_solution] _UpperCAmelCase : Optional[Any] = len(lowerCAmelCase_ ) - 1 _UpperCAmelCase : Optional[Any] = False while not found: _UpperCAmelCase : Tuple = 0 while i < len(lowerCAmelCase_ ): if best_solution[i] != solution[i]: _UpperCAmelCase : Any = best_solution[i] _UpperCAmelCase : str = solution[i] break _UpperCAmelCase : int = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) _UpperCAmelCase : Tuple = True _UpperCAmelCase : List[Any] = best_solution[:-1] _UpperCAmelCase : str = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: _UpperCAmelCase : Tuple = cost _UpperCAmelCase : List[Any] = solution else: _UpperCAmelCase : Any = index_of_best_solution + 1 _UpperCAmelCase : Dict = neighborhood[index_of_best_solution] if len(lowerCAmelCase_ ) >= size: tabu_list.pop(0 ) _UpperCAmelCase : Optional[Any] = count + 1 return best_solution_ever, best_cost def snake_case_ ( lowerCAmelCase_=None )-> Optional[int]: '''simple docstring''' _UpperCAmelCase : Tuple = generate_neighbours(args.File ) _UpperCAmelCase ,_UpperCAmelCase : Tuple = generate_first_solution( args.File , lowerCAmelCase_ ) _UpperCAmelCase ,_UpperCAmelCase : str = tabu_search( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , args.Iterations , args.Size , ) print(F'''Best solution: {best_sol}, with total distance: {best_cost}.''' ) if __name__ == "__main__": A_ : Optional[int] = argparse.ArgumentParser(description="""Tabu Search""") parser.add_argument( """-f""", """--File""", type=str, help="""Path to the file containing the data""", required=True, ) parser.add_argument( """-i""", """--Iterations""", type=int, help="""How many iterations the algorithm should perform""", required=True, ) parser.add_argument( """-s""", """--Size""", type=int, help="""Size of the tabu list""", required=True ) # Pass the arguments to main method main(parser.parse_args())
349
1
from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class _lowerCAmelCase : def __init__( self , _UpperCamelCase , _UpperCamelCase=13 , _UpperCamelCase=7 , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=99 , _UpperCamelCase=[1, 1, 2] , _UpperCamelCase=1 , _UpperCamelCase=32 , _UpperCamelCase=4 , _UpperCamelCase=8 , _UpperCamelCase=37 , _UpperCamelCase="gelu_new" , _UpperCamelCase=0.1 , _UpperCamelCase=0.1 , _UpperCamelCase=0.0 , _UpperCamelCase=512 , _UpperCamelCase=3 , _UpperCamelCase=0.02 , _UpperCamelCase=3 , _UpperCamelCase=4 , _UpperCamelCase=None , _UpperCamelCase=False , ) -> Tuple: lowerCAmelCase_ = parent lowerCAmelCase_ = batch_size lowerCAmelCase_ = seq_length lowerCAmelCase_ = is_training lowerCAmelCase_ = use_input_mask lowerCAmelCase_ = use_token_type_ids lowerCAmelCase_ = use_labels lowerCAmelCase_ = vocab_size lowerCAmelCase_ = block_sizes lowerCAmelCase_ = num_decoder_layers lowerCAmelCase_ = d_model lowerCAmelCase_ = n_head lowerCAmelCase_ = d_head lowerCAmelCase_ = d_inner lowerCAmelCase_ = hidden_act lowerCAmelCase_ = hidden_dropout lowerCAmelCase_ = attention_dropout lowerCAmelCase_ = activation_dropout lowerCAmelCase_ = max_position_embeddings lowerCAmelCase_ = type_vocab_size lowerCAmelCase_ = 2 lowerCAmelCase_ = num_labels lowerCAmelCase_ = num_choices lowerCAmelCase_ = scope lowerCAmelCase_ = initializer_std # Used in the tests to check the size of the first attention layer lowerCAmelCase_ = n_head # Used in the tests to check the size of the first hidden state lowerCAmelCase_ = self.d_model # Used in the tests to check the number of output hidden states/attentions lowerCAmelCase_ = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: lowerCAmelCase_ = self.num_hidden_layers + 2 def __a ( self ) -> List[str]: lowerCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase_ = None if self.use_input_mask: lowerCAmelCase_ = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase_ = None if self.use_token_type_ids: lowerCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase_ = None lowerCAmelCase_ = None lowerCAmelCase_ = None if self.use_labels: lowerCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase_ = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase_ = FunnelConfig( vocab_size=self.vocab_size , block_sizes=self.block_sizes , num_decoder_layers=self.num_decoder_layers , d_model=self.d_model , n_head=self.n_head , d_head=self.d_head , d_inner=self.d_inner , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , activation_dropout=self.activation_dropout , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_std=self.initializer_std , ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ) -> Tuple: lowerCAmelCase_ = TFFunnelModel(config=_UpperCamelCase ) lowerCAmelCase_ = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} lowerCAmelCase_ = model(_UpperCamelCase ) lowerCAmelCase_ = [input_ids, input_mask] lowerCAmelCase_ = model(_UpperCamelCase ) lowerCAmelCase_ = model(_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) lowerCAmelCase_ = False lowerCAmelCase_ = TFFunnelModel(config=_UpperCamelCase ) lowerCAmelCase_ = model(_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) lowerCAmelCase_ = False lowerCAmelCase_ = TFFunnelModel(config=_UpperCamelCase ) lowerCAmelCase_ = model(_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ) -> List[Any]: lowerCAmelCase_ = TFFunnelBaseModel(config=_UpperCamelCase ) lowerCAmelCase_ = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} lowerCAmelCase_ = model(_UpperCamelCase ) lowerCAmelCase_ = [input_ids, input_mask] lowerCAmelCase_ = model(_UpperCamelCase ) lowerCAmelCase_ = model(_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) lowerCAmelCase_ = False lowerCAmelCase_ = TFFunnelBaseModel(config=_UpperCamelCase ) lowerCAmelCase_ = model(_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 3, self.d_model) ) lowerCAmelCase_ = False lowerCAmelCase_ = TFFunnelBaseModel(config=_UpperCamelCase ) lowerCAmelCase_ = model(_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ) -> Dict: lowerCAmelCase_ = TFFunnelForPreTraining(config=_UpperCamelCase ) lowerCAmelCase_ = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} lowerCAmelCase_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length) ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ) -> Tuple: lowerCAmelCase_ = TFFunnelForMaskedLM(config=_UpperCamelCase ) lowerCAmelCase_ = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} lowerCAmelCase_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ) -> int: lowerCAmelCase_ = self.num_labels lowerCAmelCase_ = TFFunnelForSequenceClassification(config=_UpperCamelCase ) lowerCAmelCase_ = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} lowerCAmelCase_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ) -> str: lowerCAmelCase_ = self.num_choices lowerCAmelCase_ = TFFunnelForMultipleChoice(config=_UpperCamelCase ) lowerCAmelCase_ = tf.tile(tf.expand_dims(_UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase_ = tf.tile(tf.expand_dims(_UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase_ = tf.tile(tf.expand_dims(_UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase_ = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } lowerCAmelCase_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ) -> Optional[Any]: lowerCAmelCase_ = self.num_labels lowerCAmelCase_ = TFFunnelForTokenClassification(config=_UpperCamelCase ) lowerCAmelCase_ = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} lowerCAmelCase_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ) -> int: lowerCAmelCase_ = TFFunnelForQuestionAnswering(config=_UpperCamelCase ) lowerCAmelCase_ = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} lowerCAmelCase_ = model(_UpperCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __a ( self ) -> Tuple: lowerCAmelCase_ = self.prepare_config_and_inputs() ( ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ) = config_and_inputs lowerCAmelCase_ = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class _lowerCAmelCase ( __a , __a , unittest.TestCase ): _lowercase =( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) _lowercase =( { '''feature-extraction''': (TFFunnelBaseModel, TFFunnelModel), '''fill-mask''': TFFunnelForMaskedLM, '''question-answering''': TFFunnelForQuestionAnswering, '''text-classification''': TFFunnelForSequenceClassification, '''token-classification''': TFFunnelForTokenClassification, '''zero-shot''': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) _lowercase =False _lowercase =False def __a ( self ) -> Any: lowerCAmelCase_ = TFFunnelModelTester(self ) lowerCAmelCase_ = ConfigTester(self , config_class=_UpperCamelCase ) def __a ( self ) -> List[str]: self.config_tester.run_common_tests() def __a ( self ) -> List[Any]: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCamelCase ) def __a ( self ) -> Tuple: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*_UpperCamelCase ) def __a ( self ) -> Any: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_UpperCamelCase ) def __a ( self ) -> Any: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_UpperCamelCase ) def __a ( self ) -> Any: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_UpperCamelCase ) @require_tf class _lowerCAmelCase ( __a , unittest.TestCase ): _lowercase =( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) _lowercase =False _lowercase =False def __a ( self ) -> str: lowerCAmelCase_ = TFFunnelModelTester(self , base=_UpperCamelCase ) lowerCAmelCase_ = ConfigTester(self , config_class=_UpperCamelCase ) def __a ( self ) -> List[Any]: self.config_tester.run_common_tests() def __a ( self ) -> List[str]: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*_UpperCamelCase ) def __a ( self ) -> Optional[int]: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_UpperCamelCase ) def __a ( self ) -> List[str]: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_UpperCamelCase )
231
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _A = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = ["NllbTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = ["NllbTokenizerFast"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys _A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
231
1
import argparse import json import os import numpy as np import PIL import requests import tensorflow.keras.applications.efficientnet as efficientnet import torch from huggingface_hub import hf_hub_download from PIL import Image from tensorflow.keras.preprocessing import image from transformers import ( EfficientNetConfig, EfficientNetForImageClassification, EfficientNetImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() snake_case = logging.get_logger(__name__) snake_case = { """b0""": efficientnet.EfficientNetBa, """b1""": efficientnet.EfficientNetBa, """b2""": efficientnet.EfficientNetBa, """b3""": efficientnet.EfficientNetBa, """b4""": efficientnet.EfficientNetBa, """b5""": efficientnet.EfficientNetBa, """b6""": efficientnet.EfficientNetBa, """b7""": efficientnet.EfficientNetBa, } snake_case = { """b0""": { """hidden_dim""": 1_280, """width_coef""": 1.0, """depth_coef""": 1.0, """image_size""": 224, """dropout_rate""": 0.2, """dw_padding""": [], }, """b1""": { """hidden_dim""": 1_280, """width_coef""": 1.0, """depth_coef""": 1.1, """image_size""": 240, """dropout_rate""": 0.2, """dw_padding""": [16], }, """b2""": { """hidden_dim""": 1_408, """width_coef""": 1.1, """depth_coef""": 1.2, """image_size""": 260, """dropout_rate""": 0.3, """dw_padding""": [5, 8, 16], }, """b3""": { """hidden_dim""": 1_536, """width_coef""": 1.2, """depth_coef""": 1.4, """image_size""": 300, """dropout_rate""": 0.3, """dw_padding""": [5, 18], }, """b4""": { """hidden_dim""": 1_792, """width_coef""": 1.4, """depth_coef""": 1.8, """image_size""": 380, """dropout_rate""": 0.4, """dw_padding""": [6], }, """b5""": { """hidden_dim""": 2_048, """width_coef""": 1.6, """depth_coef""": 2.2, """image_size""": 456, """dropout_rate""": 0.4, """dw_padding""": [13, 27], }, """b6""": { """hidden_dim""": 2_304, """width_coef""": 1.8, """depth_coef""": 2.6, """image_size""": 528, """dropout_rate""": 0.5, """dw_padding""": [31], }, """b7""": { """hidden_dim""": 2_560, """width_coef""": 2.0, """depth_coef""": 3.1, """image_size""": 600, """dropout_rate""": 0.5, """dw_padding""": [18], }, } def lowerCamelCase__ ( lowercase ): """simple docstring""" SCREAMING_SNAKE_CASE : str = EfficientNetConfig() SCREAMING_SNAKE_CASE : str = CONFIG_MAP[model_name]["hidden_dim"] SCREAMING_SNAKE_CASE : Tuple = CONFIG_MAP[model_name]["width_coef"] SCREAMING_SNAKE_CASE : Optional[int] = CONFIG_MAP[model_name]["depth_coef"] SCREAMING_SNAKE_CASE : Union[str, Any] = CONFIG_MAP[model_name]["image_size"] SCREAMING_SNAKE_CASE : Any = CONFIG_MAP[model_name]["dropout_rate"] SCREAMING_SNAKE_CASE : str = CONFIG_MAP[model_name]["dw_padding"] SCREAMING_SNAKE_CASE : str = "huggingface/label-files" SCREAMING_SNAKE_CASE : str = "imagenet-1k-id2label.json" SCREAMING_SNAKE_CASE : str = 1000 SCREAMING_SNAKE_CASE : List[Any] = json.load(open(hf_hub_download(lowercase , lowercase , repo_type="dataset" ) , "r" ) ) SCREAMING_SNAKE_CASE : Tuple = {int(lowercase ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE : Union[str, Any] = idalabel SCREAMING_SNAKE_CASE : Union[str, Any] = {v: k for k, v in idalabel.items()} return config def lowerCamelCase__ ( ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = "http://images.cocodataset.org/val2017/000000039769.jpg" SCREAMING_SNAKE_CASE : List[Any] = Image.open(requests.get(lowercase , stream=lowercase ).raw ) return im def lowerCamelCase__ ( lowercase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = CONFIG_MAP[model_name]["image_size"] SCREAMING_SNAKE_CASE : int = EfficientNetImageProcessor( size={"height": size, "width": size} , image_mean=[0.485, 0.456, 0.406] , image_std=[0.47853944, 0.4732864, 0.47434163] , do_center_crop=lowercase , ) return preprocessor def lowerCamelCase__ ( lowercase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = [v.split("_" )[0].split("block" )[1] for v in original_param_names if v.startswith("block" )] SCREAMING_SNAKE_CASE : List[str] = sorted(set(lowercase ) ) SCREAMING_SNAKE_CASE : List[str] = len(lowercase ) SCREAMING_SNAKE_CASE : Optional[int] = {b: str(lowercase ) for b, i in zip(lowercase , range(lowercase ) )} SCREAMING_SNAKE_CASE : Dict = [] rename_keys.append(("stem_conv/kernel:0", "embeddings.convolution.weight") ) rename_keys.append(("stem_bn/gamma:0", "embeddings.batchnorm.weight") ) rename_keys.append(("stem_bn/beta:0", "embeddings.batchnorm.bias") ) rename_keys.append(("stem_bn/moving_mean:0", "embeddings.batchnorm.running_mean") ) rename_keys.append(("stem_bn/moving_variance:0", "embeddings.batchnorm.running_var") ) for b in block_names: SCREAMING_SNAKE_CASE : Tuple = block_name_mapping[b] rename_keys.append((F'''block{b}_expand_conv/kernel:0''', F'''encoder.blocks.{hf_b}.expansion.expand_conv.weight''') ) rename_keys.append((F'''block{b}_expand_bn/gamma:0''', F'''encoder.blocks.{hf_b}.expansion.expand_bn.weight''') ) rename_keys.append((F'''block{b}_expand_bn/beta:0''', F'''encoder.blocks.{hf_b}.expansion.expand_bn.bias''') ) rename_keys.append( (F'''block{b}_expand_bn/moving_mean:0''', F'''encoder.blocks.{hf_b}.expansion.expand_bn.running_mean''') ) rename_keys.append( (F'''block{b}_expand_bn/moving_variance:0''', F'''encoder.blocks.{hf_b}.expansion.expand_bn.running_var''') ) rename_keys.append( (F'''block{b}_dwconv/depthwise_kernel:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_conv.weight''') ) rename_keys.append((F'''block{b}_bn/gamma:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.weight''') ) rename_keys.append((F'''block{b}_bn/beta:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.bias''') ) rename_keys.append( (F'''block{b}_bn/moving_mean:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_mean''') ) rename_keys.append( (F'''block{b}_bn/moving_variance:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_var''') ) rename_keys.append((F'''block{b}_se_reduce/kernel:0''', F'''encoder.blocks.{hf_b}.squeeze_excite.reduce.weight''') ) rename_keys.append((F'''block{b}_se_reduce/bias:0''', F'''encoder.blocks.{hf_b}.squeeze_excite.reduce.bias''') ) rename_keys.append((F'''block{b}_se_expand/kernel:0''', F'''encoder.blocks.{hf_b}.squeeze_excite.expand.weight''') ) rename_keys.append((F'''block{b}_se_expand/bias:0''', F'''encoder.blocks.{hf_b}.squeeze_excite.expand.bias''') ) rename_keys.append( (F'''block{b}_project_conv/kernel:0''', F'''encoder.blocks.{hf_b}.projection.project_conv.weight''') ) rename_keys.append((F'''block{b}_project_bn/gamma:0''', F'''encoder.blocks.{hf_b}.projection.project_bn.weight''') ) rename_keys.append((F'''block{b}_project_bn/beta:0''', F'''encoder.blocks.{hf_b}.projection.project_bn.bias''') ) rename_keys.append( (F'''block{b}_project_bn/moving_mean:0''', F'''encoder.blocks.{hf_b}.projection.project_bn.running_mean''') ) rename_keys.append( (F'''block{b}_project_bn/moving_variance:0''', F'''encoder.blocks.{hf_b}.projection.project_bn.running_var''') ) rename_keys.append(("top_conv/kernel:0", "encoder.top_conv.weight") ) rename_keys.append(("top_bn/gamma:0", "encoder.top_bn.weight") ) rename_keys.append(("top_bn/beta:0", "encoder.top_bn.bias") ) rename_keys.append(("top_bn/moving_mean:0", "encoder.top_bn.running_mean") ) rename_keys.append(("top_bn/moving_variance:0", "encoder.top_bn.running_var") ) SCREAMING_SNAKE_CASE : int = {} for item in rename_keys: if item[0] in original_param_names: SCREAMING_SNAKE_CASE : Any = "efficientnet." + item[1] SCREAMING_SNAKE_CASE : Optional[Any] = "classifier.weight" SCREAMING_SNAKE_CASE : List[str] = "classifier.bias" return key_mapping def lowerCamelCase__ ( lowercase , lowercase , lowercase ): """simple docstring""" for key, value in tf_params.items(): if "normalization" in key: continue SCREAMING_SNAKE_CASE : str = key_mapping[key] if "_conv" in key and "kernel" in key: SCREAMING_SNAKE_CASE : Dict = torch.from_numpy(lowercase ).permute(3 , 2 , 0 , 1 ) elif "depthwise_kernel" in key: SCREAMING_SNAKE_CASE : int = torch.from_numpy(lowercase ).permute(2 , 3 , 0 , 1 ) elif "kernel" in key: SCREAMING_SNAKE_CASE : List[str] = torch.from_numpy(np.transpose(lowercase ) ) else: SCREAMING_SNAKE_CASE : Dict = torch.from_numpy(lowercase ) # Replace HF parameters with original TF model parameters assert hf_params[hf_key].shape == new_hf_value.shape hf_params[hf_key].copy_(lowercase ) @torch.no_grad() def lowerCamelCase__ ( lowercase , lowercase , lowercase , lowercase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = model_classes[model_name]( include_top=lowercase , weights="imagenet" , input_tensor=lowercase , input_shape=lowercase , pooling=lowercase , classes=1000 , classifier_activation="softmax" , ) SCREAMING_SNAKE_CASE : List[Any] = original_model.trainable_variables SCREAMING_SNAKE_CASE : Dict = original_model.non_trainable_variables SCREAMING_SNAKE_CASE : Dict = {param.name: param.numpy() for param in tf_params} for param in tf_non_train_params: SCREAMING_SNAKE_CASE : Tuple = param.numpy() SCREAMING_SNAKE_CASE : Tuple = list(tf_params.keys() ) # Load HuggingFace model SCREAMING_SNAKE_CASE : Tuple = get_efficientnet_config(lowercase ) SCREAMING_SNAKE_CASE : str = EfficientNetForImageClassification(lowercase ).eval() SCREAMING_SNAKE_CASE : Dict = hf_model.state_dict() # Create src-to-dst parameter name mapping dictionary print("Converting parameters..." ) SCREAMING_SNAKE_CASE : Dict = rename_keys(lowercase ) replace_params(lowercase , lowercase , lowercase ) # Initialize preprocessor and preprocess input image SCREAMING_SNAKE_CASE : Optional[int] = convert_image_processor(lowercase ) SCREAMING_SNAKE_CASE : int = preprocessor(images=prepare_img() , return_tensors="pt" ) # HF model inference hf_model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE : List[str] = hf_model(**lowercase ) SCREAMING_SNAKE_CASE : Optional[int] = outputs.logits.detach().numpy() # Original model inference SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : List[str] = CONFIG_MAP[model_name]["image_size"] SCREAMING_SNAKE_CASE : Any = prepare_img().resize((image_size, image_size) , resample=PIL.Image.NEAREST ) SCREAMING_SNAKE_CASE : Tuple = image.img_to_array(lowercase ) SCREAMING_SNAKE_CASE : Tuple = np.expand_dims(lowercase , axis=0 ) SCREAMING_SNAKE_CASE : Any = original_model.predict(lowercase ) # Check whether original and HF model outputs match -> np.allclose assert np.allclose(lowercase , lowercase , atol=1E-3 ), "The predicted logits are not the same." print("Model outputs match!" ) if save_model: # Create folder to save model if not os.path.isdir(lowercase ): os.mkdir(lowercase ) # Save converted model and image processor hf_model.save_pretrained(lowercase ) preprocessor.save_pretrained(lowercase ) if push_to_hub: # Push model and image processor to hub print(F'''Pushing converted {model_name} to the hub...''' ) SCREAMING_SNAKE_CASE : Union[str, Any] = F'''efficientnet-{model_name}''' preprocessor.push_to_hub(lowercase ) hf_model.push_to_hub(lowercase ) if __name__ == "__main__": snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""b0""", type=str, help="""Version name of the EfficientNet model you want to convert, select from [b0, b1, b2, b3, b4, b5, b6, b7].""", ) parser.add_argument( """--pytorch_dump_folder_path""", default="""hf_model""", type=str, help="""Path to the output PyTorch model directory.""", ) parser.add_argument("""--save_model""", action="""store_true""", help="""Save model to local""") parser.add_argument("""--push_to_hub""", action="""store_true""", help="""Push model and image processor to the hub""") snake_case = parser.parse_args() convert_efficientnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.save_model, args.push_to_hub)
319
import inspect import jax import jax.lax as lax import jax.numpy as jnp from ..utils import add_start_docstrings from ..utils.logging import get_logger snake_case = get_logger(__name__) snake_case = r""" Args: input_ids (`jnp.ndarray` of shape `(batch_size, sequence_length)`): Indices of input sequence tokens in the vocabulary. Indices can be obtained using [`PreTrainedTokenizer`]. See [`PreTrainedTokenizer.encode`] and [`PreTrainedTokenizer.__call__`] for details. [What are input IDs?](../glossary#input-ids) scores (`jnp.ndarray` of shape `(batch_size, config.vocab_size)`): Prediction scores of a language modeling head. These can be logits for each vocabulary when not using beam search or log softmax for each vocabulary token when using beam search kwargs (`Dict[str, Any]`, *optional*): Additional logits processor specific kwargs. Return: `jnp.ndarray` of shape `(batch_size, config.vocab_size)`: The processed prediction scores. """ class SCREAMING_SNAKE_CASE : '''simple docstring''' @add_start_docstrings(UpperCAmelCase_ ) def __call__( self : str , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : jnp.ndarray ): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) class SCREAMING_SNAKE_CASE : '''simple docstring''' @add_start_docstrings(UpperCAmelCase_ ) def __call__( self : Optional[Any] , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : jnp.ndarray ): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' @add_start_docstrings(UpperCAmelCase_ ) def __call__( self : Optional[int] , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : int , **UpperCAmelCase_ : Tuple ): for processor in self: SCREAMING_SNAKE_CASE : Optional[int] = inspect.signature(processor.__call__ ).parameters if len(UpperCAmelCase_ ) > 3: if not all(arg in kwargs for arg in list(function_args.keys() )[2:] ): raise ValueError( f'''Make sure that all the required parameters: {list(function_args.keys() )} for ''' f'''{processor.__class__} are passed to the logits processor.''' ) SCREAMING_SNAKE_CASE : int = processor(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , **UpperCAmelCase_ ) else: SCREAMING_SNAKE_CASE : Dict = processor(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) return scores class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' def __init__( self : int , UpperCAmelCase_ : float ): if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) or not (temperature > 0): raise ValueError(f'''`temperature` has to be a strictly positive float, but is {temperature}''' ) SCREAMING_SNAKE_CASE : Optional[int] = temperature def __call__( self : List[Any] , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE : Dict = scores / self.temperature return scores class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' def __init__( self : str , UpperCAmelCase_ : float , UpperCAmelCase_ : float = -float("Inf" ) , UpperCAmelCase_ : int = 1 ): if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) or (top_p < 0 or top_p > 1.0): raise ValueError(f'''`top_p` has to be a float > 0 and < 1, but is {top_p}''' ) if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) or (min_tokens_to_keep < 1): raise ValueError(f'''`min_tokens_to_keep` has to be a positive integer, but is {min_tokens_to_keep}''' ) SCREAMING_SNAKE_CASE : Optional[int] = top_p SCREAMING_SNAKE_CASE : str = filter_value SCREAMING_SNAKE_CASE : List[str] = min_tokens_to_keep def __call__( self : Dict , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Tuple = lax.top_k(UpperCAmelCase_ , scores.shape[-1] ) SCREAMING_SNAKE_CASE : str = jnp.full_like(UpperCAmelCase_ , self.filter_value ) SCREAMING_SNAKE_CASE : Optional[int] = jax.nn.softmax(UpperCAmelCase_ , axis=-1 ).cumsum(axis=-1 ) SCREAMING_SNAKE_CASE : Tuple = cumulative_probs < self.top_p # include the token that is higher than top_p as well SCREAMING_SNAKE_CASE : Optional[int] = jnp.roll(UpperCAmelCase_ , 1 ) score_mask |= score_mask.at[:, 0].set(UpperCAmelCase_ ) # min tokens to keep SCREAMING_SNAKE_CASE : Union[str, Any] = score_mask.at[:, : self.min_tokens_to_keep].set(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : str = jnp.where(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : List[str] = jax.lax.sort_key_val(UpperCAmelCase_ , UpperCAmelCase_ )[-1] return next_scores class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Optional[int] , UpperCAmelCase_ : int , UpperCAmelCase_ : float = -float("Inf" ) , UpperCAmelCase_ : int = 1 ): if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) or top_k <= 0: raise ValueError(f'''`top_k` has to be a strictly positive integer, but is {top_k}''' ) SCREAMING_SNAKE_CASE : List[str] = max(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : int = filter_value def __call__( self : Dict , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : int = scores.shape SCREAMING_SNAKE_CASE : List[str] = jnp.full(batch_size * vocab_size , self.filter_value ) SCREAMING_SNAKE_CASE : List[str] = min(self.top_k , scores.shape[-1] ) # Safety check SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[int] = lax.top_k(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Any = jnp.broadcast_to((jnp.arange(UpperCAmelCase_ ) * vocab_size)[:, None] , (batch_size, topk) ).flatten() SCREAMING_SNAKE_CASE : List[str] = topk_scores.flatten() SCREAMING_SNAKE_CASE : List[Any] = topk_indices.flatten() + shift SCREAMING_SNAKE_CASE : Dict = next_scores_flat.at[topk_indices_flat].set(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Any = next_scores_flat.reshape(UpperCAmelCase_ , UpperCAmelCase_ ) return next_scores class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Dict , UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE : List[str] = bos_token_id def __call__( self : Tuple , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE : Dict = jnp.full(scores.shape , -float("inf" ) ) SCREAMING_SNAKE_CASE : Optional[int] = 1 - jnp.bool_(cur_len - 1 ) SCREAMING_SNAKE_CASE : Union[str, Any] = jnp.where(UpperCAmelCase_ , new_scores.at[:, self.bos_token_id].set(0 ) , UpperCAmelCase_ ) return scores class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' def __init__( self : int , UpperCAmelCase_ : int , UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE : Optional[Any] = max_length SCREAMING_SNAKE_CASE : Tuple = eos_token_id def __call__( self : List[str] , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE : List[str] = jnp.full(scores.shape , -float("inf" ) ) SCREAMING_SNAKE_CASE : str = 1 - jnp.bool_(cur_len - self.max_length + 1 ) SCREAMING_SNAKE_CASE : Optional[Any] = jnp.where(UpperCAmelCase_ , new_scores.at[:, self.eos_token_id].set(0 ) , UpperCAmelCase_ ) return scores class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Optional[Any] , UpperCAmelCase_ : int , UpperCAmelCase_ : int ): if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) or min_length < 0: raise ValueError(f'''`min_length` has to be a positive integer, but is {min_length}''' ) if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) or eos_token_id < 0: raise ValueError(f'''`eos_token_id` has to be a positive integer, but is {eos_token_id}''' ) SCREAMING_SNAKE_CASE : List[str] = min_length SCREAMING_SNAKE_CASE : Tuple = eos_token_id def __call__( self : Optional[Any] , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : int ): # create boolean flag to decide if min length penalty should be applied SCREAMING_SNAKE_CASE : Optional[int] = 1 - jnp.clip(cur_len - self.min_length , 0 , 1 ) SCREAMING_SNAKE_CASE : Optional[int] = jnp.where(UpperCAmelCase_ , scores.at[:, self.eos_token_id].set(-float("inf" ) ) , UpperCAmelCase_ ) return scores class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Dict , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE : Optional[Any] = list(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Tuple = begin_index def __call__( self : Dict , UpperCAmelCase_ : str , UpperCAmelCase_ : int , UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE : Union[str, Any] = 1 - jnp.bool_(cur_len - self.begin_index ) SCREAMING_SNAKE_CASE : List[str] = jnp.where(UpperCAmelCase_ , scores.at[:, self.begin_suppress_tokens].set(-float("inf" ) ) , UpperCAmelCase_ ) return scores class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' def __init__( self : List[str] , UpperCAmelCase_ : list ): SCREAMING_SNAKE_CASE : List[Any] = list(UpperCAmelCase_ ) def __call__( self : Any , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE : Tuple = scores.at[..., self.suppress_tokens].set(-float("inf" ) ) return scores class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' def __init__( self : Union[str, Any] , UpperCAmelCase_ : Any ): SCREAMING_SNAKE_CASE : List[Any] = dict(UpperCAmelCase_ ) # Converts the dictionary of format {index: token} containing the tokens to be forced to an array, where the # index of the array corresponds to the index of the token to be forced, for XLA compatibility. # Indexes without forced tokens will have a negative value. SCREAMING_SNAKE_CASE : Optional[Any] = jnp.ones((max(force_token_map.keys() ) + 1) , dtype=jnp.intaa ) * -1 for index, token in force_token_map.items(): if token is not None: SCREAMING_SNAKE_CASE : Any = force_token_array.at[index].set(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Tuple = jnp.intaa(UpperCAmelCase_ ) def __call__( self : Tuple , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : jnp.ndarray , UpperCAmelCase_ : int ): def _force_token(UpperCAmelCase_ : Tuple ): SCREAMING_SNAKE_CASE : List[str] = scores.shape[0] SCREAMING_SNAKE_CASE : Optional[int] = self.force_token_array[generation_idx] SCREAMING_SNAKE_CASE : Tuple = jnp.ones_like(UpperCAmelCase_ , dtype=scores.dtype ) * -float("inf" ) SCREAMING_SNAKE_CASE : Dict = jnp.zeros((batch_size, 1) , dtype=scores.dtype ) SCREAMING_SNAKE_CASE : Optional[Any] = lax.dynamic_update_slice(UpperCAmelCase_ , UpperCAmelCase_ , (0, current_token) ) return new_scores SCREAMING_SNAKE_CASE : Any = lax.cond( cur_len >= self.force_token_array.shape[0] , lambda: scores , lambda: lax.cond( self.force_token_array[cur_len] >= 0 , lambda: _force_token(UpperCAmelCase_ ) , lambda: scores , ) , ) return scores class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' def __init__( self : int , UpperCAmelCase_ : int , UpperCAmelCase_ : str , UpperCAmelCase_ : Tuple ): SCREAMING_SNAKE_CASE : Union[str, Any] = generate_config.eos_token_id SCREAMING_SNAKE_CASE : Tuple = generate_config.no_timestamps_token_id SCREAMING_SNAKE_CASE : List[Any] = generate_config.no_timestamps_token_id + 1 SCREAMING_SNAKE_CASE : Dict = decoder_input_length + 1 if generate_config.is_multilingual: # room for language token and task token self.begin_index += 2 if hasattr(UpperCAmelCase_ , "max_initial_timestamp_index" ): SCREAMING_SNAKE_CASE : List[Any] = generate_config.max_initial_timestamp_index else: SCREAMING_SNAKE_CASE : List[str] = model_config.vocab_size if self.max_initial_timestamp_index is None: SCREAMING_SNAKE_CASE : List[str] = model_config.vocab_size def __call__( self : List[str] , UpperCAmelCase_ : int , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[int] ): # suppress <|notimestamps|> which is handled by without_timestamps SCREAMING_SNAKE_CASE : int = scores.at[:, self.no_timestamps_token_id].set(-float("inf" ) ) def handle_pairs(UpperCAmelCase_ : int , UpperCAmelCase_ : Optional[int] ): SCREAMING_SNAKE_CASE : Tuple = jnp.where((cur_len - self.begin_index) >= 1 , UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : int = jnp.where( input_ids_k[cur_len - 1] >= self.timestamp_begin , True and last_was_timestamp , UpperCAmelCase_ , ) SCREAMING_SNAKE_CASE : Tuple = jnp.where((cur_len - self.begin_index) < 2 , UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : List[str] = jnp.where( input_ids_k[cur_len - 2] >= self.timestamp_begin , UpperCAmelCase_ , UpperCAmelCase_ , ) return jnp.where( UpperCAmelCase_ , jnp.where( penultimate_was_timestamp > 0 , scores_k.at[self.timestamp_begin :].set(-float("inf" ) ) , scores_k.at[: self.eos_token_id].set(-float("inf" ) ) , ) , UpperCAmelCase_ , ) SCREAMING_SNAKE_CASE : Optional[Any] = jax.vmap(UpperCAmelCase_ )(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Union[str, Any] = jnp.where(cur_len == self.begin_index , UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : List[Any] = jnp.where( self.max_initial_timestamp_index is not None , True and apply_max_initial_timestamp , UpperCAmelCase_ , ) SCREAMING_SNAKE_CASE : List[str] = self.timestamp_begin + self.max_initial_timestamp_index SCREAMING_SNAKE_CASE : Optional[Any] = jnp.where( UpperCAmelCase_ , scores.at[:, last_allowed + 1 :].set(-float("inf" ) ) , UpperCAmelCase_ , ) # if sum of probability over timestamps is above any other token, sample timestamp SCREAMING_SNAKE_CASE : List[Any] = jax.nn.log_softmax(UpperCAmelCase_ , axis=-1 ) def handle_cumulative_probs(UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[int] ): SCREAMING_SNAKE_CASE : Union[str, Any] = jax.nn.logsumexp(logprobs_k[self.timestamp_begin :] , axis=-1 ) SCREAMING_SNAKE_CASE : Union[str, Any] = jnp.max(logprobs_k[: self.timestamp_begin] ) return jnp.where( timestamp_logprob > max_text_token_logprob , scores_k.at[: self.timestamp_begin].set(-float("inf" ) ) , UpperCAmelCase_ , ) SCREAMING_SNAKE_CASE : List[str] = jax.vmap(UpperCAmelCase_ )(UpperCAmelCase_ , UpperCAmelCase_ ) return scores
319
1
'''simple docstring''' from collections import defaultdict from pathlib import Path import pandas as pd from rouge_cli import calculate_rouge_path from utils import calculate_rouge _lowerCamelCase : Dict = [ 'Prosecutor: "No videos were used in the crash investigation" German papers say they saw a cell phone video of the' ' final seconds on board Flight 9525. The Germanwings co-pilot says he had a "previous episode of severe' ' depression\" German airline confirms it knew of Andreas Lubitz\'s depression years before he took control.', 'The Palestinian Authority officially becomes the 123rd member of the International Criminal Court. The formal' ' accession was marked with a ceremony at The Hague, in the Netherlands. The Palestinians signed the ICC\'s' ' founding Rome Statute in January. Israel and the United States opposed the Palestinians\' efforts to join the' ' body.', 'Amnesty International releases its annual report on the death penalty. The report catalogs the use of' ' state-sanctioned killing as a punitive measure across the globe. At least 607 people were executed around the' ' world in 2014, compared to 778 in 2013. The U.S. remains one of the worst offenders for imposing capital' ' punishment.', ] _lowerCamelCase : List[Any] = [ 'Marseille prosecutor says "so far no videos were used in the crash investigation" despite media reports .' ' Journalists at Bild and Paris Match are "very confident" the video clip is real, an editor says . Andreas Lubitz' ' had informed his Lufthansa training school of an episode of severe depression, airline says .', 'Membership gives the ICC jurisdiction over alleged crimes committed in Palestinian territories since last June .' ' Israel and the United States opposed the move, which could open the door to war crimes investigations against' ' Israelis .', 'Amnesty\'s annual death penalty report catalogs encouraging signs, but setbacks in numbers of those sentenced to' ' death . Organization claims that governments around the world are using the threat of terrorism to advance' ' executions . The number of executions worldwide has gone down by almost 22% compared with 2013, but death' ' sentences up by 28% .', ] def __a ( ) ->Dict: """simple docstring""" A = calculate_rouge(UpperCAmelCase , UpperCAmelCase , bootstrap_aggregation=UpperCAmelCase , rouge_keys=["""rouge2""", """rougeL"""] ) assert isinstance(UpperCAmelCase , UpperCAmelCase ) A = calculate_rouge(UpperCAmelCase , UpperCAmelCase , bootstrap_aggregation=UpperCAmelCase , rouge_keys=["""rouge2"""] ) assert ( pd.DataFrame(no_aggregation["""rouge2"""] ).fmeasure.mean() == pd.DataFrame(no_aggregation_just_ra["""rouge2"""] ).fmeasure.mean() ) def __a ( ) ->Union[str, Any]: """simple docstring""" A = """rougeLsum""" A = calculate_rouge(UpperCAmelCase , UpperCAmelCase , newline_sep=UpperCAmelCase , rouge_keys=[k] )[k] A = calculate_rouge(UpperCAmelCase , UpperCAmelCase , newline_sep=UpperCAmelCase , rouge_keys=[k] )[k] assert score > score_no_sep def __a ( ) ->Optional[Any]: """simple docstring""" A = ["""rouge1""", """rouge2""", """rougeL"""] A = calculate_rouge(UpperCAmelCase , UpperCAmelCase , newline_sep=UpperCAmelCase , rouge_keys=UpperCAmelCase ) A = calculate_rouge(UpperCAmelCase , UpperCAmelCase , newline_sep=UpperCAmelCase , rouge_keys=UpperCAmelCase ) assert score_sep == score_no_sep def __a ( ) ->int: """simple docstring""" A = [ """Her older sister, Margot Frank, died in 1945, a month earlier than previously thought.""", """Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports .""", ] A = [ """Margot Frank, died in 1945, a month earlier than previously thought.""", """Prosecutor: \"No videos were used in the crash investigation\" German papers say they saw a cell phone video of""" """ the final seconds on board Flight 9525.""", ] assert calculate_rouge(UpperCAmelCase , UpperCAmelCase , newline_sep=UpperCAmelCase ) == calculate_rouge(UpperCAmelCase , UpperCAmelCase , newline_sep=UpperCAmelCase ) def __a ( ) ->List[str]: """simple docstring""" A = [ """\" \"a person who has such a video needs to immediately give it to the investigators,\" prosecutor says .<n> \"it is a very disturbing scene,\" editor-in-chief of bild online tells \"erin burnett: outfront\" """ ] A = [ """ Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports . Journalists at Bild and Paris Match are \"very confident\" the video clip is real, an editor says . Andreas Lubitz had informed his Lufthansa training school of an episode of severe depression, airline says .""" ] A = calculate_rouge(UpperCAmelCase , UpperCAmelCase , rouge_keys=["""rougeLsum"""] , newline_sep=UpperCAmelCase )["""rougeLsum"""] A = calculate_rouge(UpperCAmelCase , UpperCAmelCase , rouge_keys=["""rougeLsum"""] )["""rougeLsum"""] assert new_score > prev_score def __a ( ) ->Dict: """simple docstring""" A = Path("""examples/seq2seq/test_data/wmt_en_ro""" ) A = calculate_rouge_path(data_dir.joinpath("""test.source""" ) , data_dir.joinpath("""test.target""" ) ) assert isinstance(UpperCAmelCase , UpperCAmelCase ) A = calculate_rouge_path( data_dir.joinpath("""test.source""" ) , data_dir.joinpath("""test.target""" ) , bootstrap_aggregation=UpperCAmelCase ) assert isinstance(UpperCAmelCase , UpperCAmelCase )
258
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __UpperCAmelCase ( A__ , A__ , unittest.TestCase ): '''simple docstring''' __lowerCAmelCase = CycleDiffusionPipeline __lowerCAmelCase = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { '''negative_prompt''', '''height''', '''width''', '''negative_prompt_embeds''', } __lowerCAmelCase = PipelineTesterMixin.required_optional_params - {'''latents'''} __lowerCAmelCase = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'''source_prompt'''} ) __lowerCAmelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS __lowerCAmelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS def A (self : int ): torch.manual_seed(0 ) A = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) A = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , num_train_timesteps=1000 , clip_sample=_lowerCAmelCase , set_alpha_to_one=_lowerCAmelCase , ) torch.manual_seed(0 ) A = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) A = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) A = CLIPTextModel(_lowerCAmelCase ) A = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) A = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def A (self : Dict , _lowerCAmelCase : str , _lowerCAmelCase : List[str]=0 ): A = floats_tensor((1, 3, 32, 32) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) A = image / 2 + 0.5 if str(_lowerCAmelCase ).startswith("""mps""" ): A = torch.manual_seed(_lowerCAmelCase ) else: A = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) A = { """prompt""": """An astronaut riding an elephant""", """source_prompt""": """An astronaut riding a horse""", """image""": image, """generator""": generator, """num_inference_steps""": 2, """eta""": 0.1, """strength""": 0.8, """guidance_scale""": 3, """source_guidance_scale""": 1, """output_type""": """numpy""", } return inputs def A (self : Any ): A = """cpu""" # ensure determinism for the device-dependent torch.Generator A = self.get_dummy_components() A = CycleDiffusionPipeline(**_lowerCAmelCase ) A = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) A = self.get_dummy_inputs(_lowerCAmelCase ) A = pipe(**_lowerCAmelCase ) A = output.images A = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) A = np.array([0.4_459, 0.4_943, 0.4_544, 0.6_643, 0.5_474, 0.4_327, 0.5_701, 0.5_959, 0.5_179] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def A (self : str ): A = self.get_dummy_components() for name, module in components.items(): if hasattr(_lowerCAmelCase , """half""" ): A = module.half() A = CycleDiffusionPipeline(**_lowerCAmelCase ) A = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) A = self.get_dummy_inputs(_lowerCAmelCase ) A = pipe(**_lowerCAmelCase ) A = output.images A = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) A = np.array([0.3_506, 0.4_543, 0.446, 0.4_575, 0.5_195, 0.4_155, 0.5_273, 0.518, 0.4_116] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def A (self : Optional[int] ): return super().test_save_load_local() @unittest.skip("""non-deterministic pipeline""" ) def A (self : Optional[Any] ): return super().test_inference_batch_single_identical() @skip_mps def A (self : Dict ): return super().test_dict_tuple_outputs_equivalent() @skip_mps def A (self : Optional[Any] ): return super().test_save_load_optional_components() @skip_mps def A (self : Optional[int] ): return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class __UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def A (self : Optional[int] ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A (self : int ): A = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/cycle-diffusion/black_colored_car.png""" ) A = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy""" ) A = init_image.resize((512, 512) ) A = """CompVis/stable-diffusion-v1-4""" A = DDIMScheduler.from_pretrained(_lowerCAmelCase , subfolder="""scheduler""" ) A = CycleDiffusionPipeline.from_pretrained( _lowerCAmelCase , scheduler=_lowerCAmelCase , safety_checker=_lowerCAmelCase , torch_dtype=torch.floataa , revision="""fp16""" ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() A = """A black colored car""" A = """A blue colored car""" A = torch.manual_seed(0 ) A = pipe( prompt=_lowerCAmelCase , source_prompt=_lowerCAmelCase , image=_lowerCAmelCase , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=_lowerCAmelCase , output_type="""np""" , ) A = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5e-1 def A (self : int ): A = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/cycle-diffusion/black_colored_car.png""" ) A = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy""" ) A = init_image.resize((512, 512) ) A = """CompVis/stable-diffusion-v1-4""" A = DDIMScheduler.from_pretrained(_lowerCAmelCase , subfolder="""scheduler""" ) A = CycleDiffusionPipeline.from_pretrained(_lowerCAmelCase , scheduler=_lowerCAmelCase , safety_checker=_lowerCAmelCase ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() A = """A black colored car""" A = """A blue colored car""" A = torch.manual_seed(0 ) A = pipe( prompt=_lowerCAmelCase , source_prompt=_lowerCAmelCase , image=_lowerCAmelCase , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=_lowerCAmelCase , output_type="""np""" , ) A = output.images assert np.abs(image - expected_image ).max() < 2e-2
258
1
"""simple docstring""" import argparse import os import re import packaging.version UpperCAmelCase__ = 'examples/' UpperCAmelCase__ = { 'examples': (re.compile(r'^check_min_version\("[^"]+"\)\s*$', re.MULTILINE), 'check_min_version("VERSION")\n'), 'init': (re.compile(r'^__version__\s+=\s+"([^"]+)"\s*$', re.MULTILINE), '__version__ = "VERSION"\n'), 'setup': (re.compile(r'^(\s*)version\s*=\s*"[^"]+",', re.MULTILINE), r'\1version="VERSION",'), 'doc': (re.compile(r'^(\s*)release\s*=\s*"[^"]+"$', re.MULTILINE), 'release = "VERSION"\n'), } UpperCAmelCase__ = { 'init': 'src/diffusers/__init__.py', 'setup': 'setup.py', } UpperCAmelCase__ = 'README.md' def _UpperCAmelCase ( __lowerCamelCase : str , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict ) -> int: with open(__lowerCamelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: _snake_case = f.read() _snake_case , _snake_case = REPLACE_PATTERNS[pattern] _snake_case = replace.replace('''VERSION''' , __lowerCamelCase ) _snake_case = re_pattern.sub(__lowerCamelCase , __lowerCamelCase ) with open(__lowerCamelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(__lowerCamelCase ) def _UpperCAmelCase ( __lowerCamelCase : Any ) -> Any: for folder, directories, fnames in os.walk(__lowerCamelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(__lowerCamelCase , __lowerCamelCase ) , __lowerCamelCase , pattern='''examples''' ) def _UpperCAmelCase ( __lowerCamelCase : int , __lowerCamelCase : Tuple=False ) -> Dict: for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if not patch: update_version_in_examples(__lowerCamelCase ) def _UpperCAmelCase ( ) -> Union[str, Any]: _snake_case = '''🤗 Transformers currently provides the following architectures''' _snake_case = '''1. Want to contribute a new model?''' with open(__lowerCamelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: _snake_case = f.readlines() # Find the start of the list. _snake_case = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 _snake_case = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): _snake_case = lines[index].replace( '''https://huggingface.co/docs/diffusers/main/model_doc''' , '''https://huggingface.co/docs/diffusers/model_doc''' , ) index += 1 with open(__lowerCamelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(__lowerCamelCase ) def _UpperCAmelCase ( ) -> Optional[int]: with open(REPLACE_FILES['''init'''] , '''r''' ) as f: _snake_case = f.read() _snake_case = REPLACE_PATTERNS['''init'''][0].search(__lowerCamelCase ).groups()[0] return packaging.version.parse(__lowerCamelCase ) def _UpperCAmelCase ( __lowerCamelCase : List[Any]=False ) -> int: _snake_case = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: _snake_case = default_version.base_version elif patch: _snake_case = f'''{default_version.major}.{default_version.minor}.{default_version.micro + 1}''' else: _snake_case = f'''{default_version.major}.{default_version.minor + 1}.0''' # Now let's ask nicely if that's the right one. _snake_case = input(f'''Which version are you releasing? [{default_version}]''' ) if len(__lowerCamelCase ) == 0: _snake_case = default_version print(f'''Updating version to {version}.''' ) global_version_update(__lowerCamelCase , patch=__lowerCamelCase ) def _UpperCAmelCase ( ) -> int: _snake_case = get_version() _snake_case = f'''{current_version.major}.{current_version.minor + 1}.0.dev0''' _snake_case = current_version.base_version # Check with the user we got that right. _snake_case = input(f'''Which version are we developing now? [{dev_version}]''' ) if len(__lowerCamelCase ) == 0: _snake_case = dev_version print(f'''Updating version to {version}.''' ) global_version_update(__lowerCamelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": UpperCAmelCase__ = argparse.ArgumentParser() parser.add_argument('--post_release', action='store_true', help='Whether this is pre or post release.') parser.add_argument('--patch', action='store_true', help='Whether or not this is a patch release.') UpperCAmelCase__ = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('Nothing to do after a patch :-)') else: post_release_work()
366
"""simple docstring""" import re from filelock import FileLock try: import nltk UpperCAmelCase__ = True except (ImportError, ModuleNotFoundError): UpperCAmelCase__ = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def _UpperCAmelCase ( __lowerCamelCase : str ) -> str: re.sub('''<n>''' , '''''' , __lowerCamelCase ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(__lowerCamelCase ) )
40
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import warnings from typing import List from unittest.mock import Mock import torch from torch.utils.data import DataLoader, IterableDataset, TensorDataset from accelerate.accelerator import Accelerator from accelerate.utils.dataclasses import DistributedType class lowerCAmelCase ( __UpperCamelCase ): def __init__( self : Optional[Any] , UpperCAmelCase : int ) -> int: lowerCamelCase__ : Tuple = data def __iter__( self : Dict ) -> List[str]: for element in self.data: yield element def SCREAMING_SNAKE_CASE ( _UpperCAmelCase=True ) -> int: lowerCamelCase__ : Optional[int] = Accelerator(even_batches=_UpperCAmelCase ) assert accelerator.num_processes == 2, "this script expects that two GPUs are available" return accelerator def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = False ) -> List[str]: if iterable: lowerCamelCase__ : Optional[Any] = DummyIterableDataset(torch.as_tensor(range(_UpperCAmelCase ) ) ) else: lowerCamelCase__ : Tuple = TensorDataset(torch.as_tensor(range(_UpperCAmelCase ) ) ) lowerCamelCase__ : str = DataLoader(_UpperCAmelCase , batch_size=_UpperCAmelCase ) lowerCamelCase__ : str = accelerator.prepare(_UpperCAmelCase ) return dl def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ) -> str: lowerCamelCase__ : Any = create_dataloader(accelerator=_UpperCAmelCase , dataset_size=_UpperCAmelCase , batch_size=_UpperCAmelCase ) lowerCamelCase__ : Any = [len(batch[0] ) for batch in dl] if accelerator.process_index == 0: assert batch_sizes == process_0_expected_batch_sizes elif accelerator.process_index == 1: assert batch_sizes == process_1_expected_batch_sizes def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: lowerCamelCase__ : List[str] = create_accelerator() # without padding, we would expect a different number of batches verify_dataloader_batch_sizes( _UpperCAmelCase , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1, 1] , ) # without padding, we would expect the same number of batches, but different sizes verify_dataloader_batch_sizes( _UpperCAmelCase , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 2] , ) def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: lowerCamelCase__ : Tuple = create_accelerator(even_batches=_UpperCAmelCase ) verify_dataloader_batch_sizes( _UpperCAmelCase , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1] , ) verify_dataloader_batch_sizes( _UpperCAmelCase , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 1] , ) def SCREAMING_SNAKE_CASE ( ) -> Union[str, Any]: lowerCamelCase__ : Tuple = create_accelerator(even_batches=_UpperCAmelCase ) lowerCamelCase__ : Any = torch.nn.Linear(1 , 1 ) lowerCamelCase__ : List[Any] = accelerator.prepare(_UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = create_dataloader(_UpperCAmelCase , dataset_size=3 , batch_size=1 ) lowerCamelCase__ : List[str] = [] with accelerator.join_uneven_inputs([ddp_model] ): for batch_idx, batch in enumerate(_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = ddp_model(batch[0].float() ) lowerCamelCase__ : int = output.sum() loss.backward() batch_idxs.append(_UpperCAmelCase ) accelerator.wait_for_everyone() if accelerator.process_index == 0: assert batch_idxs == [0, 1] elif accelerator.process_index == 1: assert batch_idxs == [0] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: with warnings.catch_warnings(record=_UpperCAmelCase ) as w: with accelerator.join_uneven_inputs([Mock()] ): pass assert issubclass(w[-1].category , _UpperCAmelCase ) assert "only supported for multi-GPU" in str(w[-1].message ) def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: lowerCamelCase__ : List[Any] = True lowerCamelCase__ : Tuple = False lowerCamelCase__ : List[str] = create_accelerator(even_batches=_UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = torch.nn.Linear(1 , 1 ) lowerCamelCase__ : List[Any] = accelerator.prepare(_UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = create_dataloader(_UpperCAmelCase , dataset_size=3 , batch_size=1 ) lowerCamelCase__ : List[Any] = create_dataloader(_UpperCAmelCase , dataset_size=3 , batch_size=1 ) with accelerator.join_uneven_inputs([ddp_model] , even_batches=_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = train_dl.batch_sampler.even_batches lowerCamelCase__ : Union[str, Any] = valid_dl.batch_sampler.even_batches assert train_dl_overridden_value == overridden_even_batches assert valid_dl_overridden_value == overridden_even_batches assert train_dl.batch_sampler.even_batches == default_even_batches assert valid_dl.batch_sampler.even_batches == default_even_batches def SCREAMING_SNAKE_CASE ( ) -> List[Any]: lowerCamelCase__ : List[Any] = True lowerCamelCase__ : Optional[int] = False lowerCamelCase__ : Any = create_accelerator(even_batches=_UpperCAmelCase ) lowerCamelCase__ : Tuple = torch.nn.Linear(1 , 1 ) lowerCamelCase__ : Tuple = accelerator.prepare(_UpperCAmelCase ) create_dataloader(_UpperCAmelCase , dataset_size=3 , batch_size=1 , iterable=_UpperCAmelCase ) lowerCamelCase__ : int = create_dataloader(_UpperCAmelCase , dataset_size=3 , batch_size=1 ) with warnings.catch_warnings(): warnings.filterwarnings('ignore' ) try: with accelerator.join_uneven_inputs([ddp_model] , even_batches=_UpperCAmelCase ): lowerCamelCase__ : str = batch_dl.batch_sampler.even_batches except AttributeError: # ensure attribute error is not raised when processing iterable dl raise AssertionError assert batch_dl_overridden_value == overridden_even_batches assert batch_dl.batch_sampler.even_batches == default_even_batches def SCREAMING_SNAKE_CASE ( ) -> Dict: lowerCamelCase__ : int = create_accelerator() lowerCamelCase__ : str = torch.nn.Linear(1 , 1 ) lowerCamelCase__ : Any = accelerator.prepare(_UpperCAmelCase ) create_dataloader(_UpperCAmelCase , dataset_size=3 , batch_size=1 , iterable=_UpperCAmelCase ) with warnings.catch_warnings(record=_UpperCAmelCase ) as w: with accelerator.join_uneven_inputs([ddp_model] , even_batches=_UpperCAmelCase ): pass assert issubclass(w[-1].category , _UpperCAmelCase ) assert "only supported for map-style datasets" in str(w[-1].message ) def SCREAMING_SNAKE_CASE ( ) -> List[Any]: lowerCamelCase__ : str = create_accelerator() accelerator.print('Test that even_batches variable ensures uniform batches across processes' ) test_default_ensures_even_batch_sizes() accelerator.print('Run tests with even_batches disabled' ) test_can_disable_even_batches() accelerator.print('Test joining uneven inputs' ) test_can_join_uneven_inputs() accelerator.print('Test overriding even_batches when joining uneven inputs' ) test_join_can_override_even_batches() accelerator.print('Test overriding even_batches for mixed dataloader types' ) test_join_can_override_for_mixed_type_dataloaders() accelerator.print('Test overriding even_batches raises a warning for iterable dataloaders' ) test_join_raises_warning_for_iterable_when_overriding_even_batches() accelerator.print('Test join with non DDP distributed raises warning' ) lowerCamelCase__ : Dict = accelerator.state.distributed_type lowerCamelCase__ : List[Any] = DistributedType.FSDP test_join_raises_warning_for_non_ddp_distributed(_UpperCAmelCase ) lowerCamelCase__ : Dict = original_state if __name__ == "__main__": main()
50
"""simple docstring""" from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING snake_case_ = logging.get_logger(__name__) @add_end_docstrings(SCREAMING_SNAKE_CASE_ ) class A_ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self :Any , *lowercase_ :str , **lowercase_ :List[Any] ) -> Union[str, Any]: super().__init__(*lowercase_ , **lowercase_ ) self.check_model_type(lowercase_ ) def UpperCAmelCase__ ( self :Union[str, Any] , lowercase_ :Any=None , lowercase_ :Optional[int]=None , lowercase_ :Tuple=None , **lowercase_ :Tuple ) -> Dict: UpperCAmelCase , UpperCAmelCase = {}, {} if padding is not None: UpperCAmelCase = padding if truncation is not None: UpperCAmelCase = truncation if top_k is not None: UpperCAmelCase = top_k return preprocess_params, {}, postprocess_params def __call__( self :List[Any] , lowercase_ :Union["Image.Image", str] , lowercase_ :str = None , **lowercase_ :Union[str, Any] ) -> Union[str, Any]: if isinstance(lowercase_ , (Image.Image, str) ) and isinstance(lowercase_ , lowercase_ ): UpperCAmelCase = {'image': image, 'question': question} else: UpperCAmelCase = image UpperCAmelCase = super().__call__(lowercase_ , **lowercase_ ) return results def UpperCAmelCase__ ( self :List[str] , lowercase_ :List[Any] , lowercase_ :int=False , lowercase_ :Optional[int]=False ) -> Union[str, Any]: UpperCAmelCase = load_image(inputs['image'] ) UpperCAmelCase = self.tokenizer( inputs['question'] , return_tensors=self.framework , padding=lowercase_ , truncation=lowercase_ ) UpperCAmelCase = self.image_processor(images=lowercase_ , return_tensors=self.framework ) model_inputs.update(lowercase_ ) return model_inputs def UpperCAmelCase__ ( self :List[Any] , lowercase_ :List[str] ) -> Any: UpperCAmelCase = self.model(**lowercase_ ) return model_outputs def UpperCAmelCase__ ( self :Dict , lowercase_ :Tuple , lowercase_ :List[Any]=5 ) -> Union[str, Any]: if top_k > self.model.config.num_labels: UpperCAmelCase = self.model.config.num_labels if self.framework == "pt": UpperCAmelCase = model_outputs.logits.sigmoid()[0] UpperCAmelCase , UpperCAmelCase = probs.topk(lowercase_ ) else: raise ValueError(f"""Unsupported framework: {self.framework}""" ) UpperCAmelCase = scores.tolist() UpperCAmelCase = ids.tolist() return [{"score": score, "answer": self.model.config.idalabel[_id]} for score, _id in zip(lowercase_ , lowercase_ )]
78
0
import json import os import unittest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __A( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): snake_case_ = CLIPTokenizer snake_case_ = CLIPTokenizerFast snake_case_ = True snake_case_ = {} snake_case_ = False def SCREAMING_SNAKE_CASE_ ( self ) -> List[str]: '''simple docstring''' super().setUp() # fmt: off __a = ['''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''lo''', '''l</w>''', '''w</w>''', '''r</w>''', '''t</w>''', '''low</w>''', '''er</w>''', '''lowest</w>''', '''newer</w>''', '''wider''', '''<unk>''', '''<|startoftext|>''', '''<|endoftext|>'''] # fmt: on __a = dict(zip(A__ , range(len(A__ ) ) ) ) __a = ['''#version: 0.2''', '''l o''', '''lo w</w>''', '''e r</w>'''] __a = {'''unk_token''': '''<unk>'''} __a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) __a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(A__ ) ) def SCREAMING_SNAKE_CASE_ ( self , **_snake_case ) -> int: '''simple docstring''' kwargs.update(self.special_tokens_map ) return CLIPTokenizer.from_pretrained(self.tmpdirname , **A__ ) def SCREAMING_SNAKE_CASE_ ( self , **_snake_case ) -> Optional[int]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **A__ ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> str: '''simple docstring''' __a = '''lower newer''' __a = '''lower newer''' return input_text, output_text def SCREAMING_SNAKE_CASE_ ( self ) -> int: '''simple docstring''' __a = CLIPTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) __a = '''lower newer''' __a = ['''lo''', '''w''', '''er</w>''', '''n''', '''e''', '''w''', '''er</w>'''] __a = tokenizer.tokenize(A__ ) self.assertListEqual(A__ , A__ ) __a = tokens + [tokenizer.unk_token] __a = [10, 2, 16, 9, 3, 2, 16, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(A__ ) , A__ ) @require_ftfy def SCREAMING_SNAKE_CASE_ ( self ) -> str: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __a = self.tokenizer_class.from_pretrained(A__ , **A__ ) __a = self.rust_tokenizer_class.from_pretrained(A__ , **A__ ) __a = '''A\n\'ll 11p223RF☆ho!!to?\'d\'d\'\'d of a cat to-$\'\'d.''' __a = tokenizer_s.tokenize(A__ ) __a = tokenizer_r.tokenize(A__ ) self.assertListEqual(A__ , A__ ) # Test that the tokenization is identical on an example containing a character (Latin Small Letter A # with Tilde) encoded in 2 different ways __a = '''xa\u0303y''' + ''' ''' + '''x\xe3y''' __a = tokenizer_s.tokenize(A__ ) __a = tokenizer_r.tokenize(A__ ) self.assertListEqual(A__ , A__ ) # Test that the tokenization is identical on unicode of space type __a = [ '''\u0009''', # (horizontal tab, '\t') '''\u000B''', # (vertical tab) '''\u000C''', # (form feed) '''\u0020''', # (space, ' ') '''\u200E''', # (left-to-right mark):w '''\u200F''', # (right-to-left mark) ] for unicode_seq in spaces_unicodes: __a = tokenizer_s.tokenize(A__ ) __a = tokenizer_r.tokenize(A__ ) self.assertListEqual(A__ , A__ ) # Test that the tokenization is identical on unicode of line break type __a = [ '''\u000A''', # (line feed, '\n') '''\r\n''', # (carriage return and line feed, '\r\n') '''\u000D''', # (carriage return, '\r') '''\r''', # (carriage return, '\r') '''\u000D''', # (carriage return, '\r') '''\u2028''', # (line separator) '''\u2029''', # (paragraph separator) # "\u0085", # (next line) ] # The tokenization is not identical for the character "\u0085" (next line). The slow version using ftfy transforms # it into the Horizontal Ellipsis character "…" ("\u2026") while the fast version transforms it into a # space (and thus into an empty list). for unicode_seq in line_break_unicodes: __a = tokenizer_s.tokenize(A__ ) __a = tokenizer_r.tokenize(A__ ) self.assertListEqual(A__ , A__ ) def SCREAMING_SNAKE_CASE_ ( self ) -> int: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __a = '''hello''' # `hello` is a token in the vocabulary of `pretrained_name` __a = F"""{text_of_1_token} {text_of_1_token}""" __a = self.rust_tokenizer_class.from_pretrained( A__ , use_fast=A__ , ) __a = tokenizer_r(A__ , return_offsets_mapping=A__ , add_special_tokens=A__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(A__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(A__ ) + 1, len(A__ ) + 1 + len(A__ )) , ) __a = F""" {text}""" __a = self.rust_tokenizer_class.from_pretrained( A__ , use_fast=A__ , ) __a = tokenizer_r(A__ , return_offsets_mapping=A__ , add_special_tokens=A__ ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(A__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(A__ ) + 1, 1 + len(A__ ) + 1 + len(A__ )) , ) def SCREAMING_SNAKE_CASE_ ( self ) -> Any: '''simple docstring''' with self.assertRaises(A__ ) as context: self.rust_tokenizer_class.from_pretrained('''robot-test/old-clip-tokenizer''' ) self.assertTrue( context.exception.args[0].startswith( '''The `backend_tokenizer` provided does not match the expected format.''' ) ) @require_ftfy def SCREAMING_SNAKE_CASE_ ( self ) -> str: '''simple docstring''' super().test_tokenization_python_rust_equals() def SCREAMING_SNAKE_CASE_ ( self ) -> int: '''simple docstring''' pass
358
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A : Optional[int] = { 'configuration_roberta': ['ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RobertaConfig', 'RobertaOnnxConfig'], 'tokenization_roberta': ['RobertaTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : List[str] = ['RobertaTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : str = [ 'ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'RobertaForCausalLM', 'RobertaForMaskedLM', 'RobertaForMultipleChoice', 'RobertaForQuestionAnswering', 'RobertaForSequenceClassification', 'RobertaForTokenClassification', 'RobertaModel', 'RobertaPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : List[Any] = [ 'TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFRobertaForCausalLM', 'TFRobertaForMaskedLM', 'TFRobertaForMultipleChoice', 'TFRobertaForQuestionAnswering', 'TFRobertaForSequenceClassification', 'TFRobertaForTokenClassification', 'TFRobertaMainLayer', 'TFRobertaModel', 'TFRobertaPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Union[str, Any] = [ 'FlaxRobertaForCausalLM', 'FlaxRobertaForMaskedLM', 'FlaxRobertaForMultipleChoice', 'FlaxRobertaForQuestionAnswering', 'FlaxRobertaForSequenceClassification', 'FlaxRobertaForTokenClassification', 'FlaxRobertaModel', 'FlaxRobertaPreTrainedModel', ] if TYPE_CHECKING: from .configuration_roberta import ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaConfig, RobertaOnnxConfig from .tokenization_roberta import RobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roberta_fast import RobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta import ( ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaForCausalLM, RobertaForMaskedLM, RobertaForMultipleChoice, RobertaForQuestionAnswering, RobertaForSequenceClassification, RobertaForTokenClassification, RobertaModel, RobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta import ( TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForMultipleChoice, TFRobertaForQuestionAnswering, TFRobertaForSequenceClassification, TFRobertaForTokenClassification, TFRobertaMainLayer, TFRobertaModel, TFRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, FlaxRobertaPreTrainedModel, ) else: import sys A : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
33
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _UpperCAmelCase : Tuple = logging.get_logger(__name__) class a__ ( __A ): """simple docstring""" def __init__(self , *__lowercase , **__lowercase ): warnings.warn( '''The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use YolosImageProcessor instead.''' , __lowercase , ) super().__init__(*__lowercase , **__lowercase )
174
'''simple docstring''' # Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __magic_name__( lowerCamelCase, lowerCamelCase, lowerCamelCase): __lowerCAmelCase = { '''en''': '''Machine learning is great, isn\'t it?''', '''ru''': '''Машинное обучение - это здорово, не так ли?''', '''de''': '''Maschinelles Lernen ist großartig, oder?''', } # BLUE scores as follows: # "pair": [fairseq, transformers] __lowerCAmelCase = { '''ru-en''': ['''[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)''', '''39.20'''], '''en-ru''': ['''[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)''', '''33.47'''], '''en-de''': ['''[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)''', '''42.83'''], '''de-en''': ['''[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)''', '''41.35'''], } __lowerCAmelCase = F"""{src_lang}-{tgt_lang}""" __lowerCAmelCase = F""" --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR's WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = \"facebook/wmt19-{src_lang}-{tgt_lang}\" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = \"{texts[src_lang]}\" input_ids = tokenizer.encode(input, return_tensors=\"pt\") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn't seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn't support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH=\"src:examples/seq2seq\" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR's WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) """ os.makedirs(lowerCamelCase, exist_ok=lowerCamelCase) __lowerCAmelCase = os.path.join(lowerCamelCase, '''README.md''') print(F"""Generating {path}""") with open(lowerCamelCase, '''w''', encoding='''utf-8''') as f: f.write(lowerCamelCase) # make sure we are under the root of the project _UpperCAmelCase : Dict = Path(__file__).resolve().parent.parent.parent _UpperCAmelCase : Optional[int] = repo_dir / """model_cards""" for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: _UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase : Dict = model_name.split("""-""") _UpperCAmelCase : Union[str, Any] = model_cards_dir / """facebook""" / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
174
1
"""simple docstring""" from ...utils import deprecate from ..controlnet.pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline # noqa: F401 deprecate( '''stable diffusion controlnet''', '''0.22.0''', '''Importing `FlaxStableDiffusionControlNetPipeline` from diffusers.pipelines.stable_diffusion.flax_pipeline_stable_diffusion_controlnet is deprecated. Please import `from diffusers import FlaxStableDiffusionControlNetPipeline` instead.''', standard_warn=False, stacklevel=3, )
363
"""simple docstring""" import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def lowerCamelCase_( _lowerCamelCase ) -> Tuple: '''simple docstring''' _lowerCamelCase : Optional[int] = {} _lowerCamelCase : Optional[int] = tokenizer(example["content"] , truncation=_lowerCamelCase )["input_ids"] _lowerCamelCase : Dict = len(example["content"] ) / len(output["input_ids"] ) return output _lowerCAmelCase : Tuple = HfArgumentParser(PretokenizationArguments) _lowerCAmelCase : Optional[int] = parser.parse_args() if args.num_workers is None: _lowerCAmelCase : Any = multiprocessing.cpu_count() _lowerCAmelCase : List[str] = AutoTokenizer.from_pretrained(args.tokenizer_dir) _lowerCAmelCase : Union[str, Any] = time.time() _lowerCAmelCase : Optional[int] = load_dataset(args.dataset_name, split='''train''') print(f'''Dataset loaded in {time.time()-t_start:.2f}s''') _lowerCAmelCase : Any = time.time() _lowerCAmelCase : Dict = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ '''repo_name''', '''path''', '''copies''', '''size''', '''content''', '''license''', '''hash''', '''line_mean''', '''line_max''', '''alpha_frac''', '''autogenerated''', ], ) print(f'''Dataset tokenized in {time.time()-t_start:.2f}s''') _lowerCAmelCase : str = time.time() ds.push_to_hub(args.tokenized_data_repo) print(f'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
340
0
from manim import * class lowercase ( UpperCamelCase__ ): def a__ ( self ) -> Dict: _A : List[Any] = Rectangle(height=0.5 , width=0.5 ) _A : Optional[int] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) _A : Optional[Any] = [mem.copy() for i in range(6 )] _A : List[Any] = [mem.copy() for i in range(6 )] _A : Optional[Any] = VGroup(*_a ).arrange(_a , buff=0 ) _A : Any = VGroup(*_a ).arrange(_a , buff=0 ) _A : Optional[int] = VGroup(_a , _a ).arrange(_a , buff=0 ) _A : Optional[Any] = Text("""CPU""" , font_size=24 ) _A : List[str] = Group(_a , _a ).arrange(_a , buff=0.5 , aligned_edge=_a ) cpu.move_to([-2.5, -0.5, 0] ) self.add(_a ) _A : Optional[Any] = [mem.copy() for i in range(4 )] _A : Optional[int] = VGroup(*_a ).arrange(_a , buff=0 ) _A : Optional[Any] = Text("""GPU""" , font_size=24 ) _A : Union[str, Any] = Group(_a , _a ).arrange(_a , buff=0.5 , aligned_edge=_a ) gpu.move_to([-1, -1, 0] ) self.add(_a ) _A : str = [mem.copy() for i in range(6 )] _A : int = VGroup(*_a ).arrange(_a , buff=0 ) _A : str = Text("""Model""" , font_size=24 ) _A : List[Any] = Group(_a , _a ).arrange(_a , buff=0.5 , aligned_edge=_a ) model.move_to([3, -1.0, 0] ) self.add(_a ) _A : List[str] = [] for i, rect in enumerate(_a ): rect.set_stroke(_a ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) _A : str = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(_a , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=_a ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=_a , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=_a , buff=0.0 ) self.add(_a ) cpu_targs.append(_a ) _A : Union[str, Any] = [mem.copy() for i in range(6 )] _A : str = VGroup(*_a ).arrange(_a , buff=0 ) _A : List[str] = Text("""Loaded Checkpoint""" , font_size=24 ) _A : Optional[int] = Group(_a , _a ).arrange(_a , aligned_edge=_a , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) _A : List[Any] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _A : int = MarkupText( F'''<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model''' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(_a , _a ) _A : List[str] = MarkupText( F'''<span fgcolor=\'{BLUE}\'>●</span> Checkpoint''' , font_size=18 , ) blue_text.next_to(_a , DOWN * 2.4 , aligned_edge=key_text.get_left() ) _A : Optional[int] = MarkupText( F'''Next, a <i><span fgcolor="{BLUE}">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor="{BLUE}">single shard</span>.''' , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(_a ) , Write(_a ) ) self.play(Write(_a , run_time=1 ) , Create(_a , run_time=1 ) ) _A : Optional[int] = [] _A : Dict = [] for i, rect in enumerate(_a ): _A : int = fill.copy().set_fill(_a , opacity=0.7 ) target.move_to(_a ) first_animations.append(GrowFromCenter(_a , run_time=1 ) ) _A : Optional[Any] = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(_a , run_time=1.5 ) ) self.play(*_a ) self.play(*_a ) self.wait()
26
import os import tempfile import unittest from transformers import FlaubertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FlaubertForMultipleChoice, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertModel, FlaubertWithLMHeadModel, ) from transformers.models.flaubert.modeling_flaubert import FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST class _SCREAMING_SNAKE_CASE ( _a ): def __init__( self : List[str] , __lowerCamelCase : List[str] , __lowerCamelCase : Union[str, Any]=13 , __lowerCamelCase : str=7 , __lowerCamelCase : Tuple=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Any=True , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : str=False , __lowerCamelCase : List[Any]=False , __lowerCamelCase : List[Any]=2 , __lowerCamelCase : Union[str, Any]=99 , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : Tuple=32 , __lowerCamelCase : Any=5 , __lowerCamelCase : Optional[Any]=4 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : List[str]=512 , __lowerCamelCase : List[Any]=12 , __lowerCamelCase : int=2 , __lowerCamelCase : List[str]=0.02 , __lowerCamelCase : Union[str, Any]=3 , __lowerCamelCase : Tuple=4 , __lowerCamelCase : Optional[int]="last" , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : List[str]=None , ): UpperCamelCase :int = parent UpperCamelCase :Optional[int] = batch_size UpperCamelCase :str = seq_length UpperCamelCase :Optional[int] = is_training UpperCamelCase :Optional[int] = use_input_lengths UpperCamelCase :Union[str, Any] = use_token_type_ids UpperCamelCase :List[str] = use_labels UpperCamelCase :Dict = gelu_activation UpperCamelCase :Optional[int] = sinusoidal_embeddings UpperCamelCase :List[Any] = causal UpperCamelCase :Optional[int] = asm UpperCamelCase :List[str] = n_langs UpperCamelCase :int = vocab_size UpperCamelCase :List[Any] = n_special UpperCamelCase :List[Any] = hidden_size UpperCamelCase :List[str] = num_hidden_layers UpperCamelCase :List[Any] = num_attention_heads UpperCamelCase :Tuple = hidden_dropout_prob UpperCamelCase :List[str] = attention_probs_dropout_prob UpperCamelCase :Tuple = max_position_embeddings UpperCamelCase :List[str] = type_vocab_size UpperCamelCase :Union[str, Any] = type_sequence_label_size UpperCamelCase :int = initializer_range UpperCamelCase :List[str] = num_labels UpperCamelCase :Optional[int] = num_choices UpperCamelCase :Optional[Any] = summary_type UpperCamelCase :Tuple = use_proj UpperCamelCase :Optional[Any] = scope def _A ( self : List[str] ): UpperCamelCase :Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase :Any = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase :List[Any] = None if self.use_input_lengths: UpperCamelCase :Dict = ( ids_tensor([self.batch_size] , vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length UpperCamelCase :str = None if self.use_token_type_ids: UpperCamelCase :int = ids_tensor([self.batch_size, self.seq_length] , self.n_langs ) UpperCamelCase :Optional[int] = None UpperCamelCase :int = None UpperCamelCase :List[Any] = None if self.use_labels: UpperCamelCase :Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase :List[str] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase :List[str] = ids_tensor([self.batch_size] , 2 ).float() UpperCamelCase :List[str] = ids_tensor([self.batch_size] , self.num_choices ) UpperCamelCase :Union[str, Any] = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def _A ( self : List[Any] ): return FlaubertConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , ) def _A ( self : Union[str, Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any , __lowerCamelCase : int , ): UpperCamelCase :Tuple = FlaubertModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :int = model(__lowerCamelCase , lengths=__lowerCamelCase , langs=__lowerCamelCase ) UpperCamelCase :List[Any] = model(__lowerCamelCase , langs=__lowerCamelCase ) UpperCamelCase :int = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _A ( self : Optional[int] , __lowerCamelCase : List[str] , __lowerCamelCase : Any , __lowerCamelCase : Tuple , __lowerCamelCase : int , __lowerCamelCase : List[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] , __lowerCamelCase : Any , __lowerCamelCase : Dict , ): UpperCamelCase :Any = FlaubertWithLMHeadModel(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Dict = model(__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _A ( self : List[Any] , __lowerCamelCase : str , __lowerCamelCase : Tuple , __lowerCamelCase : List[str] , __lowerCamelCase : int , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] , ): UpperCamelCase :Any = FlaubertForQuestionAnsweringSimple(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Any = model(__lowerCamelCase ) UpperCamelCase :int = model(__lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _A ( self : str , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Tuple , __lowerCamelCase : Any , __lowerCamelCase : List[str] , __lowerCamelCase : Dict , __lowerCamelCase : str , ): UpperCamelCase :str = FlaubertForQuestionAnswering(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Any = model(__lowerCamelCase ) UpperCamelCase :Optional[int] = model( __lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase , cls_index=__lowerCamelCase , is_impossible=__lowerCamelCase , p_mask=__lowerCamelCase , ) UpperCamelCase :Union[str, Any] = model( __lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase , cls_index=__lowerCamelCase , is_impossible=__lowerCamelCase , ) ((UpperCamelCase) , ) :int = result_with_labels.to_tuple() UpperCamelCase :int = model(__lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase ) ((UpperCamelCase) , ) :List[Any] = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape , () ) self.parent.assertEqual(result.start_top_log_probs.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual(result.start_top_index.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual( result.end_top_log_probs.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual( result.end_top_index.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual(result.cls_logits.shape , (self.batch_size,) ) def _A ( self : List[Any] , __lowerCamelCase : Any , __lowerCamelCase : Any , __lowerCamelCase : int , __lowerCamelCase : Dict , __lowerCamelCase : Any , __lowerCamelCase : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Tuple , ): UpperCamelCase :Optional[int] = FlaubertForSequenceClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Tuple = model(__lowerCamelCase ) UpperCamelCase :List[str] = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _A ( self : Any , __lowerCamelCase : int , __lowerCamelCase : Optional[int] , __lowerCamelCase : str , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : int , __lowerCamelCase : Optional[int] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , ): UpperCamelCase :Dict = self.num_labels UpperCamelCase :Tuple = FlaubertForTokenClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Optional[Any] = model(__lowerCamelCase , attention_mask=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _A ( self : Tuple , __lowerCamelCase : Tuple , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : Tuple , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] , ): UpperCamelCase :Union[str, Any] = self.num_choices UpperCamelCase :List[Any] = FlaubertForMultipleChoice(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Optional[Any] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase :Optional[int] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase :int = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase :Union[str, Any] = model( __lowerCamelCase , attention_mask=__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _A ( self : str ): UpperCamelCase :List[str] = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) :List[Any] = config_and_inputs UpperCamelCase :Union[str, Any] = { """input_ids""": input_ids, """token_type_ids""": token_type_ids, """lengths""": input_lengths, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class _SCREAMING_SNAKE_CASE ( _a , _a , unittest.TestCase ): snake_case__ : Optional[int] = ( ( FlaubertModel, FlaubertWithLMHeadModel, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertForMultipleChoice, ) if is_torch_available() else () ) snake_case__ : Tuple = ( { """feature-extraction""": FlaubertModel, """fill-mask""": FlaubertWithLMHeadModel, """question-answering""": FlaubertForQuestionAnsweringSimple, """text-classification""": FlaubertForSequenceClassification, """token-classification""": FlaubertForTokenClassification, """zero-shot""": FlaubertForSequenceClassification, } if is_torch_available() else {} ) def _A ( self : int , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] ): if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith("""Fast""" ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def _A ( self : Optional[int] , __lowerCamelCase : Dict , __lowerCamelCase : Tuple , __lowerCamelCase : Tuple=False ): UpperCamelCase :Tuple = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if model_class.__name__ == "FlaubertForQuestionAnswering": UpperCamelCase :Tuple = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowerCamelCase ) UpperCamelCase :List[Any] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowerCamelCase ) return inputs_dict def _A ( self : str ): UpperCamelCase :List[Any] = FlaubertModelTester(self ) UpperCamelCase :Any = ConfigTester(self , config_class=__lowerCamelCase , emb_dim=37 ) def _A ( self : Optional[int] ): self.config_tester.run_common_tests() def _A ( self : List[Any] ): UpperCamelCase :List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*__lowerCamelCase ) def _A ( self : Optional[int] ): UpperCamelCase :Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*__lowerCamelCase ) def _A ( self : List[Any] ): UpperCamelCase :Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_simple_qa(*__lowerCamelCase ) def _A ( self : Union[str, Any] ): UpperCamelCase :int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*__lowerCamelCase ) def _A ( self : Optional[Any] ): UpperCamelCase :List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*__lowerCamelCase ) def _A ( self : Tuple ): UpperCamelCase :Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_token_classif(*__lowerCamelCase ) def _A ( self : int ): UpperCamelCase :Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_multiple_choice(*__lowerCamelCase ) @slow def _A ( self : Any ): for model_name in FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase :Optional[int] = FlaubertModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) @slow @require_torch_gpu def _A ( self : Tuple ): UpperCamelCase , UpperCamelCase :Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # FlauBertForMultipleChoice behaves incorrectly in JIT environments. if model_class == FlaubertForMultipleChoice: return UpperCamelCase :Optional[Any] = True UpperCamelCase :Optional[Any] = model_class(config=__lowerCamelCase ) UpperCamelCase :str = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) UpperCamelCase :str = torch.jit.trace( __lowerCamelCase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(__lowerCamelCase , os.path.join(__lowerCamelCase , """traced_model.pt""" ) ) UpperCamelCase :int = torch.jit.load(os.path.join(__lowerCamelCase , """traced_model.pt""" ) , map_location=__lowerCamelCase ) loaded(inputs_dict["""input_ids"""].to(__lowerCamelCase ) , inputs_dict["""attention_mask"""].to(__lowerCamelCase ) ) @require_torch class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _A ( self : Optional[Any] ): UpperCamelCase :Union[str, Any] = FlaubertModel.from_pretrained("""flaubert/flaubert_base_cased""" ) UpperCamelCase :Optional[Any] = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) with torch.no_grad(): UpperCamelCase :Tuple = model(__lowerCamelCase )[0] UpperCamelCase :Union[str, Any] = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , __lowerCamelCase ) UpperCamelCase :int = torch.tensor( [[[-2.6251, -1.4298, -0.0227], [-2.8510, -1.6387, 0.2258], [-2.8114, -1.1832, -0.3066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowerCamelCase , atol=1E-4 ) )
38
0
def lowerCAmelCase__ ( a__ ) ->list: '''simple docstring''' if any(not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or x < 0 for x in sequence ): raise TypeError("Sequence must be list of non-negative integers" ) for _ in range(len(lowerCAmelCase_ ) ): for i, (rod_upper, rod_lower) in enumerate(zip(lowerCAmelCase_ , sequence[1:] ) ): if rod_upper > rod_lower: sequence[i] -= rod_upper - rod_lower sequence[i + 1] += rod_upper - rod_lower return sequence if __name__ == "__main__": assert bead_sort([5, 4, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bead_sort([7, 9, 4, 3, 5]) == [3, 4, 5, 7, 9]
368
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = ['''image_processor''', '''tokenizer'''] __A = '''BridgeTowerImageProcessor''' __A = ('''RobertaTokenizer''', '''RobertaTokenizerFast''') def __init__( self : List[Any] , lowercase_ : Dict , lowercase_ : List[Any]) -> List[str]: """simple docstring""" super().__init__(lowercase_ , lowercase_) def __call__( self : Any , lowercase_ : List[Any] , lowercase_ : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , lowercase_ : bool = True , lowercase_ : Union[bool, str, PaddingStrategy] = False , lowercase_ : Union[bool, str, TruncationStrategy] = None , lowercase_ : Optional[int] = None , lowercase_ : int = 0 , lowercase_ : Optional[int] = None , lowercase_ : Optional[bool] = None , lowercase_ : Optional[bool] = None , lowercase_ : bool = False , lowercase_ : bool = False , lowercase_ : bool = False , lowercase_ : bool = False , lowercase_ : bool = True , lowercase_ : Optional[Union[str, TensorType]] = None , **lowercase_ : str , ) -> BatchEncoding: """simple docstring""" _UpperCamelCase = self.tokenizer( text=lowercase_ , add_special_tokens=lowercase_ , padding=lowercase_ , truncation=lowercase_ , max_length=lowercase_ , stride=lowercase_ , pad_to_multiple_of=lowercase_ , return_token_type_ids=lowercase_ , return_attention_mask=lowercase_ , return_overflowing_tokens=lowercase_ , return_special_tokens_mask=lowercase_ , return_offsets_mapping=lowercase_ , return_length=lowercase_ , verbose=lowercase_ , return_tensors=lowercase_ , **lowercase_ , ) # add pixel_values + pixel_mask _UpperCamelCase = self.image_processor( lowercase_ , return_tensors=lowercase_ , do_normalize=lowercase_ , do_center_crop=lowercase_ , **lowercase_) encoding.update(lowercase_) return encoding def __UpperCAmelCase ( self : Union[str, Any] , *lowercase_ : List[str] , **lowercase_ : int) -> List[Any]: """simple docstring""" return self.tokenizer.batch_decode(*lowercase_ , **lowercase_) def __UpperCAmelCase ( self : Optional[Any] , *lowercase_ : Union[str, Any] , **lowercase_ : Dict) -> List[Any]: """simple docstring""" return self.tokenizer.decode(*lowercase_ , **lowercase_) @property def __UpperCAmelCase ( self : str) -> Dict: """simple docstring""" _UpperCamelCase = self.tokenizer.model_input_names _UpperCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names))
63
0
from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def _A ( SCREAMING_SNAKE_CASE : int ): """simple docstring""" a__ : int =int(number**0.5 ) return number == sq * sq def _A ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int ): """simple docstring""" a__ : int =x_num * y_den * z_den + y_num * x_den * z_den + z_num * x_den * y_den a__ : int =x_den * y_den * z_den a__ : int =gcd(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) top //= hcf bottom //= hcf return top, bottom def _A ( SCREAMING_SNAKE_CASE : int = 35 ): """simple docstring""" a__ : set =set() a__ : int a__ : Fraction =Fraction(0 ) a__ : tuple[int, int] for x_num in range(1 , order + 1 ): for x_den in range(x_num + 1 , order + 1 ): for y_num in range(1 , order + 1 ): for y_den in range(y_num + 1 , order + 1 ): # n=1 a__ : Dict =x_num * y_den + x_den * y_num a__ : Optional[Any] =x_den * y_den a__ : int =gcd(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: a__ : List[Any] =add_three( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) unique_s.add(SCREAMING_SNAKE_CASE ) # n=2 a__ : List[str] =( x_num * x_num * y_den * y_den + x_den * x_den * y_num * y_num ) a__ : int =x_den * x_den * y_den * y_den if is_sq(SCREAMING_SNAKE_CASE ) and is_sq(SCREAMING_SNAKE_CASE ): a__ : Optional[int] =int(sqrt(SCREAMING_SNAKE_CASE ) ) a__ : List[str] =int(sqrt(SCREAMING_SNAKE_CASE ) ) a__ : int =gcd(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: a__ : List[Any] =add_three( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) unique_s.add(SCREAMING_SNAKE_CASE ) # n=-1 a__ : List[str] =x_num * y_num a__ : Dict =x_den * y_num + x_num * y_den a__ : Tuple =gcd(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: a__ : Dict =add_three( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) unique_s.add(SCREAMING_SNAKE_CASE ) # n=2 a__ : List[str] =x_num * x_num * y_num * y_num a__ : Optional[int] =( x_den * x_den * y_num * y_num + x_num * x_num * y_den * y_den ) if is_sq(SCREAMING_SNAKE_CASE ) and is_sq(SCREAMING_SNAKE_CASE ): a__ : List[Any] =int(sqrt(SCREAMING_SNAKE_CASE ) ) a__ : Optional[Any] =int(sqrt(SCREAMING_SNAKE_CASE ) ) a__ : List[Any] =gcd(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: a__ : Tuple =add_three( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) unique_s.add(SCREAMING_SNAKE_CASE ) for num, den in unique_s: total += Fraction(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return total.denominator + total.numerator if __name__ == "__main__": print(F"""{solution() = }""")
95
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class a ( __snake_case ): SCREAMING_SNAKE_CASE : Dict = """Salesforce/blip-image-captioning-base""" SCREAMING_SNAKE_CASE : int = ( """This is a tool that generates a description of an image. It takes an input named `image` which should be the """ """image to caption, and returns a text that contains the description in English.""" ) SCREAMING_SNAKE_CASE : Optional[int] = """image_captioner""" SCREAMING_SNAKE_CASE : Optional[Any] = AutoModelForVisionaSeq SCREAMING_SNAKE_CASE : int = ["""image"""] SCREAMING_SNAKE_CASE : Optional[Any] = ["""text"""] def __init__( self : Optional[Any] , *__SCREAMING_SNAKE_CASE : Optional[Any] , **__SCREAMING_SNAKE_CASE : List[Any] ) -> Optional[Any]: requires_backends(self , ['vision'] ) super().__init__(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) def UpperCamelCase ( self : List[Any] , __SCREAMING_SNAKE_CASE : "Image" ) -> Dict: return self.pre_processor(images=__SCREAMING_SNAKE_CASE , return_tensors='pt' ) def UpperCamelCase ( self : List[str] , __SCREAMING_SNAKE_CASE : Tuple ) -> str: return self.model.generate(**__SCREAMING_SNAKE_CASE ) def UpperCamelCase ( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : List[str] ) -> Optional[Any]: return self.pre_processor.batch_decode(__SCREAMING_SNAKE_CASE , skip_special_tokens=__SCREAMING_SNAKE_CASE )[0].strip()
183
0
'''simple docstring''' import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / "utils")) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def lowerCAmelCase ( self : Any )-> Optional[Any]: # A mock response for an HTTP head request to emulate server down snake_case = mock.Mock() snake_case = 5_00 snake_case = {} snake_case = HTTPError snake_case = {} # Download this model to make sure it's in the cache. snake_case = BertTokenizer.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("""requests.Session.request""" , return_value=__snake_case ) as mock_head: snake_case = BertTokenizer.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def lowerCAmelCase ( self : Dict )-> Tuple: # A mock response for an HTTP head request to emulate server down snake_case = mock.Mock() snake_case = 5_00 snake_case = {} snake_case = HTTPError snake_case = {} # Download this model to make sure it's in the cache. snake_case = GPTaTokenizerFast.from_pretrained("""gpt2""" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("""requests.Session.request""" , return_value=__snake_case ) as mock_head: snake_case = GPTaTokenizerFast.from_pretrained("""gpt2""" ) # This check we did call the fake head request mock_head.assert_called() def lowerCAmelCase ( self : Tuple )-> Union[str, Any]: # This test is for deprecated behavior and can be removed in v5 try: snake_case = tempfile.mktemp() with open(__snake_case , """wb""" ) as f: http_get("""https://huggingface.co/albert-base-v1/resolve/main/spiece.model""" , __snake_case ) snake_case = AlbertTokenizer.from_pretrained(__snake_case ) finally: os.remove(__snake_case ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile("""tokenizer.json""" ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open("""tokenizer.json""" , """wb""" ) as f: http_get("""https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json""" , __snake_case ) snake_case = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 10_00 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove("""tokenizer.json""" ) def lowerCAmelCase ( self : Any )-> Dict: # This test is for deprecated behavior and can be removed in v5 snake_case = AlbertTokenizer.from_pretrained("""https://huggingface.co/albert-base-v1/resolve/main/spiece.model""" ) @is_staging_test class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" snake_case_ = ["[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "bla", "blou"] @classmethod def lowerCAmelCase ( cls : List[str] )-> List[str]: snake_case = TOKEN HfFolder.save_token(__snake_case ) @classmethod def lowerCAmelCase ( cls : Any )-> List[Any]: try: delete_repo(token=cls._token , repo_id="""test-tokenizer""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-tokenizer-org""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""test-dynamic-tokenizer""" ) except HTTPError: pass def lowerCAmelCase ( self : int )-> Any: with tempfile.TemporaryDirectory() as tmp_dir: snake_case = os.path.join(__snake_case , """vocab.txt""" ) with open(__snake_case , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in self.vocab_tokens] ) ) snake_case = BertTokenizer(__snake_case ) tokenizer.push_to_hub("""test-tokenizer""" , use_auth_token=self._token ) snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="""test-tokenizer""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(__snake_case , repo_id="""test-tokenizer""" , push_to_hub=__snake_case , use_auth_token=self._token ) snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def lowerCAmelCase ( self : Union[str, Any] )-> List[Any]: with tempfile.TemporaryDirectory() as tmp_dir: snake_case = os.path.join(__snake_case , """vocab.txt""" ) with open(__snake_case , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in self.vocab_tokens] ) ) snake_case = BertTokenizer(__snake_case ) tokenizer.push_to_hub("""valid_org/test-tokenizer-org""" , use_auth_token=self._token ) snake_case = BertTokenizer.from_pretrained("""valid_org/test-tokenizer-org""" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-tokenizer-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( __snake_case , repo_id="""valid_org/test-tokenizer-org""" , push_to_hub=__snake_case , use_auth_token=self._token ) snake_case = BertTokenizer.from_pretrained("""valid_org/test-tokenizer-org""" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def lowerCAmelCase ( self : Optional[Any] )-> int: CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: snake_case = os.path.join(__snake_case , """vocab.txt""" ) with open(__snake_case , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in self.vocab_tokens] ) ) snake_case = CustomTokenizer(__snake_case ) # No fast custom tokenizer tokenizer.push_to_hub("""test-dynamic-tokenizer""" , use_auth_token=self._token ) snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=__snake_case ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , """CustomTokenizer""" ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: snake_case = os.path.join(__snake_case , """vocab.txt""" ) with open(__snake_case , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in self.vocab_tokens] ) ) snake_case = BertTokenizerFast.from_pretrained(__snake_case ) bert_tokenizer.save_pretrained(__snake_case ) snake_case = CustomTokenizerFast.from_pretrained(__snake_case ) tokenizer.push_to_hub("""test-dynamic-tokenizer""" , use_auth_token=self._token ) snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=__snake_case ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , """CustomTokenizerFast""" ) snake_case = AutoTokenizer.from_pretrained( f'''{USER}/test-dynamic-tokenizer''' , use_fast=__snake_case , trust_remote_code=__snake_case ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , """CustomTokenizer""" ) class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def lowerCAmelCase ( self : Tuple )-> Dict: snake_case = Trie() trie.add("""Hello 友達""" ) self.assertEqual(trie.data , {"""H""": {"""e""": {"""l""": {"""l""": {"""o""": {""" """: {"""友""": {"""達""": {"""""": 1}}}}}}}}} ) trie.add("""Hello""" ) trie.data self.assertEqual(trie.data , {"""H""": {"""e""": {"""l""": {"""l""": {"""o""": {"""""": 1, """ """: {"""友""": {"""達""": {"""""": 1}}}}}}}}} ) def lowerCAmelCase ( self : str )-> Dict: snake_case = Trie() self.assertEqual(trie.split("""[CLS] This is a extra_id_100""" ) , ["""[CLS] This is a extra_id_100"""] ) trie.add("""[CLS]""" ) trie.add("""extra_id_1""" ) trie.add("""extra_id_100""" ) self.assertEqual(trie.split("""[CLS] This is a extra_id_100""" ) , ["""[CLS]""", """ This is a """, """extra_id_100"""] ) def lowerCAmelCase ( self : List[Any] )-> List[str]: snake_case = Trie() trie.add("""A""" ) self.assertEqual(trie.split("""ABC""" ) , ["""A""", """BC"""] ) self.assertEqual(trie.split("""BCA""" ) , ["""BC""", """A"""] ) def lowerCAmelCase ( self : int )-> Optional[int]: snake_case = Trie() trie.add("""TOKEN]""" ) trie.add("""[SPECIAL_TOKEN]""" ) self.assertEqual(trie.split("""This is something [SPECIAL_TOKEN]""" ) , ["""This is something """, """[SPECIAL_TOKEN]"""] ) def lowerCAmelCase ( self : List[str] )-> List[Any]: snake_case = Trie() trie.add("""A""" ) trie.add("""P""" ) trie.add("""[SPECIAL_TOKEN]""" ) self.assertEqual(trie.split("""This is something [SPECIAL_TOKEN]""" ) , ["""This is something """, """[SPECIAL_TOKEN]"""] ) def lowerCAmelCase ( self : Union[str, Any] )-> Dict: snake_case = Trie() trie.add("""AB""" ) trie.add("""B""" ) trie.add("""C""" ) self.assertEqual(trie.split("""ABC""" ) , ["""AB""", """C"""] ) def lowerCAmelCase ( self : Optional[int] )-> Optional[int]: snake_case = Trie() trie.add("""ABC""" ) trie.add("""B""" ) trie.add("""CD""" ) self.assertEqual(trie.split("""ABCD""" ) , ["""ABC""", """D"""] ) def lowerCAmelCase ( self : Optional[int] )-> List[Any]: # Even if the offsets are wrong, we necessarily output correct string # parts. snake_case = Trie() snake_case = trie.cut_text("""ABC""" , [0, 0, 2, 1, 2, 3] ) self.assertEqual(__snake_case , ["""AB""", """C"""] )
3
'''simple docstring''' import json import os import sys import tempfile import unittest from pathlib import Path from shutil import copyfile from huggingface_hub import HfFolder, Repository, create_repo, delete_repo from requests.exceptions import HTTPError import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, PROCESSOR_MAPPING, TOKENIZER_MAPPING, AutoConfig, AutoFeatureExtractor, AutoProcessor, AutoTokenizer, BertTokenizer, ProcessorMixin, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaProcessor, ) from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test from transformers.tokenization_utils import TOKENIZER_CONFIG_FILE from transformers.utils import FEATURE_EXTRACTOR_NAME, is_tokenizers_available sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 from test_module.custom_processing import CustomProcessor # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 _SCREAMING_SNAKE_CASE = get_tests_dir("fixtures/dummy_feature_extractor_config.json") _SCREAMING_SNAKE_CASE = get_tests_dir("fixtures/vocab.json") _SCREAMING_SNAKE_CASE = get_tests_dir("fixtures") class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" snake_case_ = ["[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "bla", "blou"] def lowerCAmelCase ( self : str )-> Any: snake_case = 0 def lowerCAmelCase ( self : Tuple )-> Optional[Any]: snake_case = AutoProcessor.from_pretrained("""facebook/wav2vec2-base-960h""" ) self.assertIsInstance(__snake_case , __snake_case ) def lowerCAmelCase ( self : Dict )-> Union[str, Any]: with tempfile.TemporaryDirectory() as tmpdirname: snake_case = WavaVecaConfig() snake_case = AutoProcessor.from_pretrained("""facebook/wav2vec2-base-960h""" ) # save in new folder model_config.save_pretrained(__snake_case ) processor.save_pretrained(__snake_case ) snake_case = AutoProcessor.from_pretrained(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) def lowerCAmelCase ( self : int )-> str: with tempfile.TemporaryDirectory() as tmpdirname: # copy relevant files copyfile(__snake_case , os.path.join(__snake_case , __snake_case ) ) copyfile(__snake_case , os.path.join(__snake_case , """vocab.json""" ) ) snake_case = AutoProcessor.from_pretrained(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) def lowerCAmelCase ( self : List[Any] )-> str: with tempfile.TemporaryDirectory() as tmpdirname: snake_case = WavaVecaFeatureExtractor() snake_case = AutoTokenizer.from_pretrained("""facebook/wav2vec2-base-960h""" ) snake_case = WavaVecaProcessor(__snake_case , __snake_case ) # save in new folder processor.save_pretrained(__snake_case ) # drop `processor_class` in tokenizer with open(os.path.join(__snake_case , __snake_case ) , """r""" ) as f: snake_case = json.load(__snake_case ) config_dict.pop("""processor_class""" ) with open(os.path.join(__snake_case , __snake_case ) , """w""" ) as f: f.write(json.dumps(__snake_case ) ) snake_case = AutoProcessor.from_pretrained(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) def lowerCAmelCase ( self : Dict )-> Optional[int]: with tempfile.TemporaryDirectory() as tmpdirname: snake_case = WavaVecaFeatureExtractor() snake_case = AutoTokenizer.from_pretrained("""facebook/wav2vec2-base-960h""" ) snake_case = WavaVecaProcessor(__snake_case , __snake_case ) # save in new folder processor.save_pretrained(__snake_case ) # drop `processor_class` in feature extractor with open(os.path.join(__snake_case , __snake_case ) , """r""" ) as f: snake_case = json.load(__snake_case ) config_dict.pop("""processor_class""" ) with open(os.path.join(__snake_case , __snake_case ) , """w""" ) as f: f.write(json.dumps(__snake_case ) ) snake_case = AutoProcessor.from_pretrained(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) def lowerCAmelCase ( self : Optional[int] )-> str: with tempfile.TemporaryDirectory() as tmpdirname: snake_case = WavaVecaConfig(processor_class="""Wav2Vec2Processor""" ) model_config.save_pretrained(__snake_case ) # copy relevant files copyfile(__snake_case , os.path.join(__snake_case , """vocab.json""" ) ) # create emtpy sample processor with open(os.path.join(__snake_case , __snake_case ) , """w""" ) as f: f.write("""{}""" ) snake_case = AutoProcessor.from_pretrained(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) def lowerCAmelCase ( self : int )-> Any: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(__snake_case ): snake_case = AutoProcessor.from_pretrained("""hf-internal-testing/test_dynamic_processor""" ) # If remote code is disabled, we can't load this config. with self.assertRaises(__snake_case ): snake_case = AutoProcessor.from_pretrained( """hf-internal-testing/test_dynamic_processor""" , trust_remote_code=__snake_case ) snake_case = AutoProcessor.from_pretrained("""hf-internal-testing/test_dynamic_processor""" , trust_remote_code=__snake_case ) self.assertTrue(processor.special_attribute_present ) self.assertEqual(processor.__class__.__name__ , """NewProcessor""" ) snake_case = processor.feature_extractor self.assertTrue(feature_extractor.special_attribute_present ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) snake_case = processor.tokenizer self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizerFast""" ) # Test we can also load the slow version snake_case = AutoProcessor.from_pretrained( """hf-internal-testing/test_dynamic_processor""" , trust_remote_code=__snake_case , use_fast=__snake_case ) snake_case = new_processor.tokenizer self.assertTrue(new_tokenizer.special_attribute_present ) self.assertEqual(new_tokenizer.__class__.__name__ , """NewTokenizer""" ) else: self.assertEqual(tokenizer.__class__.__name__ , """NewTokenizer""" ) def lowerCAmelCase ( self : List[Any] )-> List[Any]: try: AutoConfig.register("""custom""" , __snake_case ) AutoFeatureExtractor.register(__snake_case , __snake_case ) AutoTokenizer.register(__snake_case , slow_tokenizer_class=__snake_case ) AutoProcessor.register(__snake_case , __snake_case ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__snake_case ): AutoProcessor.register(__snake_case , __snake_case ) # Now that the config is registered, it can be used as any other config with the auto-API snake_case = CustomFeatureExtractor.from_pretrained(__snake_case ) with tempfile.TemporaryDirectory() as tmp_dir: snake_case = os.path.join(__snake_case , """vocab.txt""" ) with open(__snake_case , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in self.vocab_tokens] ) ) snake_case = CustomTokenizer(__snake_case ) snake_case = CustomProcessor(__snake_case , __snake_case ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained(__snake_case ) snake_case = AutoProcessor.from_pretrained(__snake_case ) self.assertIsInstance(__snake_case , __snake_case ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def lowerCAmelCase ( self : Any )-> Tuple: class _lowerCAmelCase ( A__ ): """simple docstring""" snake_case_ = False class _lowerCAmelCase ( A__ ): """simple docstring""" snake_case_ = False class _lowerCAmelCase ( A__ ): """simple docstring""" snake_case_ = "AutoFeatureExtractor" snake_case_ = "AutoTokenizer" snake_case_ = False try: AutoConfig.register("""custom""" , __snake_case ) AutoFeatureExtractor.register(__snake_case , __snake_case ) AutoTokenizer.register(__snake_case , slow_tokenizer_class=__snake_case ) AutoProcessor.register(__snake_case , __snake_case ) # If remote code is not set, the default is to use local classes. snake_case = AutoProcessor.from_pretrained("""hf-internal-testing/test_dynamic_processor""" ) self.assertEqual(processor.__class__.__name__ , """NewProcessor""" ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote code is disabled, we load the local ones. snake_case = AutoProcessor.from_pretrained( """hf-internal-testing/test_dynamic_processor""" , trust_remote_code=__snake_case ) self.assertEqual(processor.__class__.__name__ , """NewProcessor""" ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub. snake_case = AutoProcessor.from_pretrained( """hf-internal-testing/test_dynamic_processor""" , trust_remote_code=__snake_case ) self.assertEqual(processor.__class__.__name__ , """NewProcessor""" ) self.assertTrue(processor.special_attribute_present ) self.assertTrue(processor.feature_extractor.special_attribute_present ) self.assertTrue(processor.tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def lowerCAmelCase ( self : str )-> Union[str, Any]: snake_case = AutoProcessor.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) self.assertEqual(processor.__class__.__name__ , """BertTokenizerFast""" ) def lowerCAmelCase ( self : Any )-> List[str]: snake_case = AutoProcessor.from_pretrained("""hf-internal-testing/tiny-random-convnext""" ) self.assertEqual(processor.__class__.__name__ , """ConvNextImageProcessor""" ) @is_staging_test class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" snake_case_ = ["[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "bla", "blou"] @classmethod def lowerCAmelCase ( cls : Optional[Any] )-> Tuple: snake_case = TOKEN HfFolder.save_token(__snake_case ) @classmethod def lowerCAmelCase ( cls : Optional[Any] )-> Optional[Any]: try: delete_repo(token=cls._token , repo_id="""test-processor""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-processor-org""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""test-dynamic-processor""" ) except HTTPError: pass def lowerCAmelCase ( self : List[Any] )-> str: snake_case = WavaVecaProcessor.from_pretrained(__snake_case ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(__snake_case , """test-processor""" ) , push_to_hub=__snake_case , use_auth_token=self._token ) snake_case = WavaVecaProcessor.from_pretrained(f'''{USER}/test-processor''' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(__snake_case , getattr(new_processor.feature_extractor , __snake_case ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def lowerCAmelCase ( self : Any )-> Optional[Any]: snake_case = WavaVecaProcessor.from_pretrained(__snake_case ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(__snake_case , """test-processor-org""" ) , push_to_hub=__snake_case , use_auth_token=self._token , organization="""valid_org""" , ) snake_case = WavaVecaProcessor.from_pretrained("""valid_org/test-processor-org""" ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(__snake_case , getattr(new_processor.feature_extractor , __snake_case ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def lowerCAmelCase ( self : List[str] )-> int: CustomFeatureExtractor.register_for_auto_class() CustomTokenizer.register_for_auto_class() CustomProcessor.register_for_auto_class() snake_case = CustomFeatureExtractor.from_pretrained(__snake_case ) with tempfile.TemporaryDirectory() as tmp_dir: snake_case = os.path.join(__snake_case , """vocab.txt""" ) with open(__snake_case , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in self.vocab_tokens] ) ) snake_case = CustomTokenizer(__snake_case ) snake_case = CustomProcessor(__snake_case , __snake_case ) with tempfile.TemporaryDirectory() as tmp_dir: create_repo(f'''{USER}/test-dynamic-processor''' , token=self._token ) snake_case = Repository(__snake_case , clone_from=f'''{USER}/test-dynamic-processor''' , token=self._token ) processor.save_pretrained(__snake_case ) # This has added the proper auto_map field to the feature extractor config self.assertDictEqual( processor.feature_extractor.auto_map , { """AutoFeatureExtractor""": """custom_feature_extraction.CustomFeatureExtractor""", """AutoProcessor""": """custom_processing.CustomProcessor""", } , ) # This has added the proper auto_map field to the tokenizer config with open(os.path.join(__snake_case , """tokenizer_config.json""" ) ) as f: snake_case = json.load(__snake_case ) self.assertDictEqual( tokenizer_config["""auto_map"""] , { """AutoTokenizer""": ["""custom_tokenization.CustomTokenizer""", None], """AutoProcessor""": """custom_processing.CustomProcessor""", } , ) # The code has been copied from fixtures self.assertTrue(os.path.isfile(os.path.join(__snake_case , """custom_feature_extraction.py""" ) ) ) self.assertTrue(os.path.isfile(os.path.join(__snake_case , """custom_tokenization.py""" ) ) ) self.assertTrue(os.path.isfile(os.path.join(__snake_case , """custom_processing.py""" ) ) ) repo.push_to_hub() snake_case = AutoProcessor.from_pretrained(f'''{USER}/test-dynamic-processor''' , trust_remote_code=__snake_case ) # Can't make an isinstance check because the new_processor is from the CustomProcessor class of a dynamic module self.assertEqual(new_processor.__class__.__name__ , """CustomProcessor""" )
3
1
import math from datetime import datetime, timedelta def __UpperCamelCase ( lowerCAmelCase__ : int ): __a : str = year % 1_9 __a : Any = year % 4 __a : List[str] = year % 7 __a : int = math.floor(year / 1_0_0 ) __a : Union[str, Any] = math.floor((1_3 + 8 * leap_day_inhibits) / 2_5 ) __a : Dict = leap_day_inhibits / 4 __a : int = ( 1_5 - lunar_orbit_correction + leap_day_inhibits - leap_day_reinstall_number ) % 3_0 __a : Any = (4 + leap_day_inhibits - leap_day_reinstall_number) % 7 # days to be added to March 21 __a : Tuple = (1_9 * metonic_cycle + secular_moon_shift) % 3_0 # PHM -> Paschal Full Moon __a : Tuple = ( 2 * julian_leap_year + 4 * non_leap_year + 6 * days_to_add + century_starting_point ) % 7 if days_to_add == 2_9 and days_from_phm_to_sunday == 6: return datetime(lowerCAmelCase_ , 4 , 1_9 ) elif days_to_add == 2_8 and days_from_phm_to_sunday == 6: return datetime(lowerCAmelCase_ , 4 , 1_8 ) else: return datetime(lowerCAmelCase_ , 3 , 2_2 ) + timedelta( days=int(days_to_add + days_from_phm_to_sunday ) ) if __name__ == "__main__": for year in (1994, 2000, 2010, 2021, 2023): lowercase__ ='will be' if year > datetime.now().year else 'was' print(F"""Easter in {year} {tense} {gauss_easter(year)}""")
216
import re from filelock import FileLock try: import nltk _snake_case : Any = True except (ImportError, ModuleNotFoundError): _snake_case : Union[str, Any] = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def a_ ( lowerCAmelCase_ : str ): re.sub('<n>', '', lowerCAmelCase_ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(lowerCAmelCase_ ) )
284
0
"""simple docstring""" import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPSegProcessor, ViTImageProcessor @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self : Union[str, Any] ): __lowercase = tempfile.mkdtemp() # fmt: off __lowercase = ["l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "lo", "l</w>", "w</w>", "r</w>", "t</w>", "low</w>", "er</w>", "lowest</w>", "newer</w>", "wider", "<unk>", "<|startoftext|>", "<|endoftext|>"] # fmt: on __lowercase = dict(zip(UpperCAmelCase__, range(len(UpperCAmelCase__ ) ) ) ) __lowercase = ["#version: 0.2", "l o", "lo w</w>", "e r</w>", ""] __lowercase = {"unk_token": "<unk>"} __lowercase = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["vocab_file"] ) __lowercase = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file, "w", encoding="utf-8" ) as fp: fp.write(json.dumps(UpperCAmelCase__ ) + "\n" ) with open(self.merges_file, "w", encoding="utf-8" ) as fp: fp.write("\n".join(UpperCAmelCase__ ) ) __lowercase = { "do_resize": True, "size": 2_0, "do_center_crop": True, "crop_size": 1_8, "do_normalize": True, "image_mean": [0.48_145_466, 0.4_578_275, 0.40_821_073], "image_std": [0.26_862_954, 0.26_130_258, 0.27_577_711], } __lowercase = os.path.join(self.tmpdirname, UpperCAmelCase__ ) with open(self.image_processor_file, "w", encoding="utf-8" ) as fp: json.dump(UpperCAmelCase__, UpperCAmelCase__ ) def _lowercase ( self : List[Any], **UpperCAmelCase__ : List[str] ): return CLIPTokenizer.from_pretrained(self.tmpdirname, **UpperCAmelCase__ ) def _lowercase ( self : List[str], **UpperCAmelCase__ : Tuple ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname, **UpperCAmelCase__ ) def _lowercase ( self : Dict, **UpperCAmelCase__ : List[Any] ): return ViTImageProcessor.from_pretrained(self.tmpdirname, **UpperCAmelCase__ ) def _lowercase ( self : Optional[Any] ): shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Tuple ): __lowercase = [np.random.randint(2_5_5, size=(3, 3_0, 4_0_0), dtype=np.uinta )] __lowercase = [Image.fromarray(np.moveaxis(UpperCAmelCase__, 0, -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : Tuple ): __lowercase = self.get_tokenizer() __lowercase = self.get_rust_tokenizer() __lowercase = self.get_image_processor() __lowercase = CLIPSegProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) processor_slow.save_pretrained(self.tmpdirname ) __lowercase = CLIPSegProcessor.from_pretrained(self.tmpdirname, use_fast=UpperCAmelCase__ ) __lowercase = CLIPSegProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) processor_fast.save_pretrained(self.tmpdirname ) __lowercase = CLIPSegProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab(), tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab(), tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab(), tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer, UpperCAmelCase__ ) self.assertIsInstance(processor_fast.tokenizer, UpperCAmelCase__ ) self.assertEqual(processor_slow.image_processor.to_json_string(), image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string(), image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor, UpperCAmelCase__ ) self.assertIsInstance(processor_fast.image_processor, UpperCAmelCase__ ) def _lowercase ( self : List[str] ): __lowercase = CLIPSegProcessor(tokenizer=self.get_tokenizer(), image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __lowercase = self.get_tokenizer(bos_token="(BOS)", eos_token="(EOS)" ) __lowercase = self.get_image_processor(do_normalize=UpperCAmelCase__, padding_value=1.0 ) __lowercase = CLIPSegProcessor.from_pretrained( self.tmpdirname, bos_token="(BOS)", eos_token="(EOS)", do_normalize=UpperCAmelCase__, padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab(), tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer, UpperCAmelCase__ ) self.assertEqual(processor.image_processor.to_json_string(), image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor, UpperCAmelCase__ ) def _lowercase ( self : Tuple ): __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = CLIPSegProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) __lowercase = self.prepare_image_inputs() __lowercase = image_processor(UpperCAmelCase__, return_tensors="np" ) __lowercase = processor(images=UpperCAmelCase__, return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum(), input_processor[key].sum(), delta=1E-2 ) def _lowercase ( self : int ): __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = CLIPSegProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) __lowercase = "lower newer" __lowercase = processor(text=UpperCAmelCase__ ) __lowercase = tokenizer(UpperCAmelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key], encoded_processor[key] ) def _lowercase ( self : int ): __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = CLIPSegProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) __lowercase = "lower newer" __lowercase = self.prepare_image_inputs() __lowercase = processor(text=UpperCAmelCase__, images=UpperCAmelCase__ ) self.assertListEqual(list(inputs.keys() ), ["input_ids", "attention_mask", "pixel_values"] ) # test if it raises when no input is passed with pytest.raises(UpperCAmelCase__ ): processor() def _lowercase ( self : int ): __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = CLIPSegProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) __lowercase = self.prepare_image_inputs() __lowercase = self.prepare_image_inputs() __lowercase = processor(images=UpperCAmelCase__, visual_prompt=UpperCAmelCase__ ) self.assertListEqual(list(inputs.keys() ), ["pixel_values", "conditional_pixel_values"] ) # test if it raises when no input is passed with pytest.raises(UpperCAmelCase__ ): processor() def _lowercase ( self : Dict ): __lowercase = self.get_image_processor() __lowercase = self.get_tokenizer() __lowercase = CLIPSegProcessor(tokenizer=UpperCAmelCase__, image_processor=UpperCAmelCase__ ) __lowercase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __lowercase = processor.batch_decode(UpperCAmelCase__ ) __lowercase = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__, UpperCAmelCase__ )
356
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { 'caidas/swin2sr-classicalsr-x2-64': ( 'https://huggingface.co/caidas/swin2sr-classicalsr-x2-64/resolve/main/config.json' ), } class _lowerCAmelCase ( lowercase ): """simple docstring""" __UpperCAmelCase : List[Any] = "swin2sr" __UpperCAmelCase : List[Any] = { "hidden_size": "embed_dim", "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self : Any, UpperCAmelCase__ : Dict=6_4, UpperCAmelCase__ : List[Any]=1, UpperCAmelCase__ : Dict=3, UpperCAmelCase__ : Optional[Any]=1_8_0, UpperCAmelCase__ : Any=[6, 6, 6, 6, 6, 6], UpperCAmelCase__ : Dict=[6, 6, 6, 6, 6, 6], UpperCAmelCase__ : Tuple=8, UpperCAmelCase__ : Optional[int]=2.0, UpperCAmelCase__ : List[str]=True, UpperCAmelCase__ : Tuple=0.0, UpperCAmelCase__ : Optional[Any]=0.0, UpperCAmelCase__ : List[str]=0.1, UpperCAmelCase__ : Dict="gelu", UpperCAmelCase__ : Dict=False, UpperCAmelCase__ : Dict=0.02, UpperCAmelCase__ : Tuple=1E-5, UpperCAmelCase__ : str=2, UpperCAmelCase__ : str=1.0, UpperCAmelCase__ : Optional[int]="1conv", UpperCAmelCase__ : Dict="pixelshuffle", **UpperCAmelCase__ : List[Any], ): super().__init__(**UpperCAmelCase__ ) __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = embed_dim __lowercase = depths __lowercase = len(UpperCAmelCase__ ) __lowercase = num_heads __lowercase = window_size __lowercase = mlp_ratio __lowercase = qkv_bias __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = drop_path_rate __lowercase = hidden_act __lowercase = use_absolute_embeddings __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = upscale __lowercase = img_range __lowercase = resi_connection __lowercase = upsampler
144
0
"""simple docstring""" from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
61
"""simple docstring""" from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = 1 / sqrt(2 ) ): UpperCAmelCase_ : int = tau * frequency / samplerate UpperCAmelCase_ : List[str] = sin(__lowerCamelCase ) UpperCAmelCase_ : int = cos(__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = _sin / (2 * q_factor) UpperCAmelCase_ : int = (1 - _cos) / 2 UpperCAmelCase_ : Optional[Any] = 1 - _cos UpperCAmelCase_ : int = 1 + alpha UpperCAmelCase_ : Dict = -2 * _cos UpperCAmelCase_ : Tuple = 1 - alpha UpperCAmelCase_ : Dict = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa], [ba, ba, ba] ) return filt def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = 1 / sqrt(2 ) ): UpperCAmelCase_ : Dict = tau * frequency / samplerate UpperCAmelCase_ : Tuple = sin(__lowerCamelCase ) UpperCAmelCase_ : Any = cos(__lowerCamelCase ) UpperCAmelCase_ : List[str] = _sin / (2 * q_factor) UpperCAmelCase_ : List[Any] = (1 + _cos) / 2 UpperCAmelCase_ : Optional[int] = -1 - _cos UpperCAmelCase_ : Union[str, Any] = 1 + alpha UpperCAmelCase_ : Optional[int] = -2 * _cos UpperCAmelCase_ : Tuple = 1 - alpha UpperCAmelCase_ : List[str] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa], [ba, ba, ba] ) return filt def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = 1 / sqrt(2 ) ): UpperCAmelCase_ : Union[str, Any] = tau * frequency / samplerate UpperCAmelCase_ : str = sin(__lowerCamelCase ) UpperCAmelCase_ : Tuple = cos(__lowerCamelCase ) UpperCAmelCase_ : List[Any] = _sin / (2 * q_factor) UpperCAmelCase_ : Any = _sin / 2 UpperCAmelCase_ : Union[str, Any] = 0 UpperCAmelCase_ : Tuple = -ba UpperCAmelCase_ : Optional[Any] = 1 + alpha UpperCAmelCase_ : Dict = -2 * _cos UpperCAmelCase_ : Optional[int] = 1 - alpha UpperCAmelCase_ : List[str] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa], [ba, ba, ba] ) return filt def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = 1 / sqrt(2 ) ): UpperCAmelCase_ : Any = tau * frequency / samplerate UpperCAmelCase_ : Any = sin(__lowerCamelCase ) UpperCAmelCase_ : Optional[int] = cos(__lowerCamelCase ) UpperCAmelCase_ : str = _sin / (2 * q_factor) UpperCAmelCase_ : List[str] = 1 - alpha UpperCAmelCase_ : str = -2 * _cos UpperCAmelCase_ : Any = 1 + alpha UpperCAmelCase_ : Tuple = IIRFilter(2 ) filt.set_coefficients([ba, ba, ba], [ba, ba, ba] ) return filt def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = 1 / sqrt(2 ), ): UpperCAmelCase_ : Dict = tau * frequency / samplerate UpperCAmelCase_ : Union[str, Any] = sin(__lowerCamelCase ) UpperCAmelCase_ : int = cos(__lowerCamelCase ) UpperCAmelCase_ : Optional[int] = _sin / (2 * q_factor) UpperCAmelCase_ : List[str] = 10 ** (gain_db / 40) UpperCAmelCase_ : List[Any] = 1 + alpha * big_a UpperCAmelCase_ : Tuple = -2 * _cos UpperCAmelCase_ : Tuple = 1 - alpha * big_a UpperCAmelCase_ : str = 1 + alpha / big_a UpperCAmelCase_ : List[str] = -2 * _cos UpperCAmelCase_ : List[str] = 1 - alpha / big_a UpperCAmelCase_ : Tuple = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa], [ba, ba, ba] ) return filt def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = 1 / sqrt(2 ), ): UpperCAmelCase_ : str = tau * frequency / samplerate UpperCAmelCase_ : int = sin(__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = cos(__lowerCamelCase ) UpperCAmelCase_ : Tuple = _sin / (2 * q_factor) UpperCAmelCase_ : List[Any] = 10 ** (gain_db / 40) UpperCAmelCase_ : Tuple = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase_ : int = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase_ : Optional[Any] = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase_ : Optional[int] = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase_ : Dict = 2 * sqrt(__lowerCamelCase ) * alpha UpperCAmelCase_ : List[str] = big_a * (pmc + aaa) UpperCAmelCase_ : int = 2 * big_a * mpc UpperCAmelCase_ : int = big_a * (pmc - aaa) UpperCAmelCase_ : Dict = ppmc + aaa UpperCAmelCase_ : Any = -2 * pmpc UpperCAmelCase_ : List[str] = ppmc - aaa UpperCAmelCase_ : List[Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa], [ba, ba, ba] ) return filt def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = 1 / sqrt(2 ), ): UpperCAmelCase_ : int = tau * frequency / samplerate UpperCAmelCase_ : Optional[Any] = sin(__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = cos(__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = _sin / (2 * q_factor) UpperCAmelCase_ : Tuple = 10 ** (gain_db / 40) UpperCAmelCase_ : Tuple = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase_ : Optional[Any] = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase_ : List[Any] = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase_ : Any = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase_ : Dict = 2 * sqrt(__lowerCamelCase ) * alpha UpperCAmelCase_ : Any = big_a * (ppmc + aaa) UpperCAmelCase_ : Union[str, Any] = -2 * big_a * pmpc UpperCAmelCase_ : Dict = big_a * (ppmc - aaa) UpperCAmelCase_ : Optional[int] = pmc + aaa UpperCAmelCase_ : Union[str, Any] = 2 * mpc UpperCAmelCase_ : int = pmc - aaa UpperCAmelCase_ : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa], [ba, ba, ba] ) return filt
61
1
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record lowerCAmelCase : Optional[int] = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ lowerCAmelCase : str = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ lowerCAmelCase : Optional[int] = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def A_ ( _UpperCAmelCase , _UpperCAmelCase ): return float((preds == labels).mean() ) def A_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase="binary" ): SCREAMING_SNAKE_CASE_: List[Any] = simple_accuracy(_UpperCAmelCase , _UpperCAmelCase ) SCREAMING_SNAKE_CASE_: str = float(fa_score(y_true=_UpperCAmelCase , y_pred=_UpperCAmelCase , average=_UpperCAmelCase ) ) return { "accuracy": acc, "f1": fa, } def A_ ( _UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: str = {} for id_pred, label in zip(_UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Any = f"{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}" SCREAMING_SNAKE_CASE_: Optional[Any] = id_pred["prediction"] if question_id in question_map: question_map[question_id].append((pred, label) ) else: SCREAMING_SNAKE_CASE_: int = [(pred, label)] SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[int] = [], [] for question, preds_labels in question_map.items(): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: int = zip(*_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: int = fa_score(y_true=_UpperCAmelCase , y_pred=_UpperCAmelCase , average="macro" ) fas.append(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: List[Any] = int(sum(pred == label for pred, label in preds_labels ) == len(_UpperCAmelCase ) ) ems.append(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: int = float(sum(_UpperCAmelCase ) / len(_UpperCAmelCase ) ) SCREAMING_SNAKE_CASE_: Dict = sum(_UpperCAmelCase ) / len(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Tuple = float(fa_score(y_true=_UpperCAmelCase , y_pred=[id_pred["prediction"] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): """simple docstring""" def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]") return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types()) , codebase_urls=[] , reference_urls=[] , format="numpy" if not self.config_name == "record" and not self.config_name == "multirc" else None , ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("int64"), "query": datasets.Value("int64"), }, "prediction_text": datasets.Value("string"), }, "references": { "idx": { "passage": datasets.Value("int64"), "query": datasets.Value("int64"), }, "answers": datasets.Sequence(datasets.Value("string")), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("int64"), "paragraph": datasets.Value("int64"), "question": datasets.Value("int64"), }, "prediction": datasets.Value("int64"), }, "references": datasets.Value("int64"), } else: return { "predictions": datasets.Value("int64"), "references": datasets.Value("int64"), } def _SCREAMING_SNAKE_CASE ( self : List[Any] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Optional[int]): if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(lowerCAmelCase__ , lowerCAmelCase__)} elif self.config_name == "cb": return acc_and_fa(lowerCAmelCase__ , lowerCAmelCase__ , fa_avg="macro") elif self.config_name == "record": SCREAMING_SNAKE_CASE_: Dict = [ { "qas": [ {"id": ref["idx"]["query"], "answers": [{"text": ans} for ans in ref["answers"]]} for ref in references ] } ] SCREAMING_SNAKE_CASE_: Tuple = {pred["idx"]["query"]: pred["prediction_text"] for pred in predictions} return evaluate_record(lowerCAmelCase__ , lowerCAmelCase__)[0] elif self.config_name == "multirc": return evaluate_multirc(lowerCAmelCase__ , lowerCAmelCase__) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(lowerCAmelCase__ , lowerCAmelCase__)} else: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]")
127
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __lowercase ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): """simple docstring""" _UpperCAmelCase : Dict = StableDiffusionInpaintPipeline _UpperCAmelCase : Tuple = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS _UpperCAmelCase : List[str] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _UpperCAmelCase : Tuple = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _UpperCAmelCase : Optional[int] = frozenset([] ) def _SCREAMING_SNAKE_CASE ( self : int): torch.manual_seed(0) SCREAMING_SNAKE_CASE_: Optional[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=9 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=lowerCAmelCase__ , ) SCREAMING_SNAKE_CASE_: Optional[Any] = PNDMScheduler(skip_prk_steps=lowerCAmelCase__) torch.manual_seed(0) SCREAMING_SNAKE_CASE_: Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0) SCREAMING_SNAKE_CASE_: Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="gelu" , projection_dim=512 , ) SCREAMING_SNAKE_CASE_: Optional[int] = CLIPTextModel(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[int] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip") SCREAMING_SNAKE_CASE_: List[str] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Optional[int]=0): # TODO: use tensor inputs instead of PIL, this is here just to leave the old expected_slices untouched SCREAMING_SNAKE_CASE_: Dict = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCAmelCase__)).to(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: str = image.cpu().permute(0 , 2 , 3 , 1)[0] SCREAMING_SNAKE_CASE_: Tuple = Image.fromarray(np.uinta(lowerCAmelCase__)).convert("RGB").resize((64, 64)) SCREAMING_SNAKE_CASE_: List[str] = Image.fromarray(np.uinta(image + 4)).convert("RGB").resize((64, 64)) if str(lowerCAmelCase__).startswith("mps"): SCREAMING_SNAKE_CASE_: Tuple = torch.manual_seed(lowerCAmelCase__) else: SCREAMING_SNAKE_CASE_: Any = torch.Generator(device=lowerCAmelCase__).manual_seed(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[Any] = { "prompt": "A painting of a squirrel eating a burger", "image": init_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def _SCREAMING_SNAKE_CASE ( self : Tuple): SCREAMING_SNAKE_CASE_: Tuple = "cpu" # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE_: int = self.get_dummy_components() SCREAMING_SNAKE_CASE_: int = StableDiffusionInpaintPipeline(**lowerCAmelCase__) SCREAMING_SNAKE_CASE_: int = sd_pipe.to(lowerCAmelCase__) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Any = self.get_dummy_inputs(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[Any] = sd_pipe(**lowerCAmelCase__).images SCREAMING_SNAKE_CASE_: Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) SCREAMING_SNAKE_CASE_: Tuple = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def _SCREAMING_SNAKE_CASE ( self : List[str]): super().test_inference_batch_single_identical(expected_max_diff=3E-3) @slow @require_torch_gpu class __lowercase ( unittest.TestCase ): """simple docstring""" def _SCREAMING_SNAKE_CASE ( self : str): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _SCREAMING_SNAKE_CASE ( self : List[Any]): SCREAMING_SNAKE_CASE_: Any = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png") SCREAMING_SNAKE_CASE_: int = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png") SCREAMING_SNAKE_CASE_: Optional[int] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint" "/yellow_cat_sitting_on_a_park_bench.npy") SCREAMING_SNAKE_CASE_: List[str] = "stabilityai/stable-diffusion-2-inpainting" SCREAMING_SNAKE_CASE_: Any = StableDiffusionInpaintPipeline.from_pretrained(lowerCAmelCase__ , safety_checker=lowerCAmelCase__) pipe.to(lowerCAmelCase__) pipe.set_progress_bar_config(disable=lowerCAmelCase__) pipe.enable_attention_slicing() SCREAMING_SNAKE_CASE_: str = "Face of a yellow cat, high resolution, sitting on a park bench" SCREAMING_SNAKE_CASE_: Optional[int] = torch.manual_seed(0) SCREAMING_SNAKE_CASE_: Optional[int] = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , output_type="np" , ) SCREAMING_SNAKE_CASE_: Optional[Any] = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image).max() < 9E-3 def _SCREAMING_SNAKE_CASE ( self : Tuple): SCREAMING_SNAKE_CASE_: Union[str, Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png") SCREAMING_SNAKE_CASE_: List[Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png") SCREAMING_SNAKE_CASE_: Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint" "/yellow_cat_sitting_on_a_park_bench_fp16.npy") SCREAMING_SNAKE_CASE_: str = "stabilityai/stable-diffusion-2-inpainting" SCREAMING_SNAKE_CASE_: Dict = StableDiffusionInpaintPipeline.from_pretrained( lowerCAmelCase__ , torch_dtype=torch.floataa , safety_checker=lowerCAmelCase__ , ) pipe.to(lowerCAmelCase__) pipe.set_progress_bar_config(disable=lowerCAmelCase__) pipe.enable_attention_slicing() SCREAMING_SNAKE_CASE_: List[str] = "Face of a yellow cat, high resolution, sitting on a park bench" SCREAMING_SNAKE_CASE_: Tuple = torch.manual_seed(0) SCREAMING_SNAKE_CASE_: Dict = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , output_type="np" , ) SCREAMING_SNAKE_CASE_: Any = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image).max() < 5E-1 def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() SCREAMING_SNAKE_CASE_: Tuple = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png") SCREAMING_SNAKE_CASE_: Union[str, Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png") SCREAMING_SNAKE_CASE_: List[str] = "stabilityai/stable-diffusion-2-inpainting" SCREAMING_SNAKE_CASE_: Tuple = PNDMScheduler.from_pretrained(lowerCAmelCase__ , subfolder="scheduler") SCREAMING_SNAKE_CASE_: Any = StableDiffusionInpaintPipeline.from_pretrained( lowerCAmelCase__ , safety_checker=lowerCAmelCase__ , scheduler=lowerCAmelCase__ , torch_dtype=torch.floataa , ) pipe.to(lowerCAmelCase__) pipe.set_progress_bar_config(disable=lowerCAmelCase__) pipe.enable_attention_slicing(1) pipe.enable_sequential_cpu_offload() SCREAMING_SNAKE_CASE_: Any = "Face of a yellow cat, high resolution, sitting on a park bench" SCREAMING_SNAKE_CASE_: Any = torch.manual_seed(0) SCREAMING_SNAKE_CASE_: Union[str, Any] = pipe( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , mask_image=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="np" , ) SCREAMING_SNAKE_CASE_: Optional[Any] = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 10**9
127
1
import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class A_ ( unittest.TestCase ): '''simple docstring''' def __init__(self , lowercase__ , lowercase__=13 , lowercase__=7 , lowercase__=True , lowercase__=True , lowercase__=True , lowercase__=True , lowercase__=99 , lowercase__=32 , lowercase__=5 , lowercase__=4 , lowercase__=37 , lowercase__="gelu" , lowercase__=0.1 , lowercase__=0.1 , lowercase__=512 , lowercase__=16 , lowercase__=2 , lowercase__=0.02 , lowercase__=4 , ) -> Optional[Any]: __UpperCAmelCase = parent __UpperCAmelCase = batch_size __UpperCAmelCase = seq_length __UpperCAmelCase = is_training __UpperCAmelCase = use_attention_mask __UpperCAmelCase = use_token_type_ids __UpperCAmelCase = use_labels __UpperCAmelCase = vocab_size __UpperCAmelCase = hidden_size __UpperCAmelCase = num_hidden_layers __UpperCAmelCase = num_attention_heads __UpperCAmelCase = intermediate_size __UpperCAmelCase = hidden_act __UpperCAmelCase = hidden_dropout_prob __UpperCAmelCase = attention_probs_dropout_prob __UpperCAmelCase = max_position_embeddings __UpperCAmelCase = type_vocab_size __UpperCAmelCase = type_sequence_label_size __UpperCAmelCase = initializer_range __UpperCAmelCase = num_choices def lowerCAmelCase_ (self ) -> int: __UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCAmelCase = None if self.use_attention_mask: __UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase = None if self.use_token_type_ids: __UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCAmelCase = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCamelCase_ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def lowerCAmelCase_ (self ) -> Dict: __UpperCAmelCase = self.prepare_config_and_inputs() __UpperCAmelCase = config_and_inputs __UpperCAmelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_flax class A_ ( a__ , unittest.TestCase ): '''simple docstring''' a__ = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def lowerCAmelCase_ (self ) -> str: __UpperCAmelCase = FlaxAlbertModelTester(self ) @slow def lowerCAmelCase_ (self ) -> Optional[int]: for model_class_name in self.all_model_classes: __UpperCAmelCase = model_class_name.from_pretrained('''albert-base-v2''' ) __UpperCAmelCase = model(np.ones((1, 1) ) ) self.assertIsNotNone(UpperCamelCase_ ) @require_flax class A_ ( unittest.TestCase ): '''simple docstring''' @slow def lowerCAmelCase_ (self ) -> str: __UpperCAmelCase = FlaxAlbertModel.from_pretrained('''albert-base-v2''' ) __UpperCAmelCase = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) __UpperCAmelCase = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) __UpperCAmelCase = model(UpperCamelCase_ , attention_mask=UpperCamelCase_ )[0] __UpperCAmelCase = (1, 11, 768) self.assertEqual(output.shape , UpperCamelCase_ ) __UpperCAmelCase = np.array( [[[-0.6513, 1.5035, -0.2766], [-0.6515, 1.5046, -0.2780], [-0.6512, 1.5049, -0.2784]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , UpperCamelCase_ , atol=1E-4 ) )
333
"""simple docstring""" import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class snake_case_( unittest.TestCase ): def __init__( self : Union[str, Any] , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : bool = True , UpperCamelCase_ : Dict[str, int] = None , UpperCamelCase_ : int = 3_2 , UpperCamelCase_ : bool = True , UpperCamelCase_ : Union[int, float] = 1 / 2_5_5 , UpperCamelCase_ : bool = True , UpperCamelCase_ : bool = True , UpperCamelCase_ : Optional[Union[float, List[float]]] = [0.48_145_466, 0.4_578_275, 0.40_821_073] , UpperCamelCase_ : Optional[Union[float, List[float]]] = [0.26_862_954, 0.26_130_258, 0.27_577_711] , UpperCamelCase_ : bool = True , UpperCamelCase_ : Optional[int]=7 , UpperCamelCase_ : int=3_0 , UpperCamelCase_ : str=4_0_0 , UpperCamelCase_ : List[Any]=3 , ): lowerCAmelCase : Union[str, Any] = parent lowerCAmelCase : Union[str, Any] = do_resize lowerCAmelCase : List[str] = size if size is not None else {'''shortest_edge''': 2_8_8} lowerCAmelCase : int = size_divisor lowerCAmelCase : List[str] = do_rescale lowerCAmelCase : Optional[Any] = rescale_factor lowerCAmelCase : Dict = do_normalize lowerCAmelCase : Any = do_center_crop lowerCAmelCase : Union[str, Any] = image_mean lowerCAmelCase : Optional[Any] = image_std lowerCAmelCase : Union[str, Any] = do_pad lowerCAmelCase : Union[str, Any] = batch_size lowerCAmelCase : Any = num_channels lowerCAmelCase : Union[str, Any] = min_resolution lowerCAmelCase : int = max_resolution def lowerCamelCase__ ( self : Dict ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def lowerCamelCase__ ( self : Any , UpperCamelCase_ : int , UpperCamelCase_ : List[str]=False ): if not batched: lowerCAmelCase : Dict = self.size['''shortest_edge'''] lowerCAmelCase : Dict = image_inputs[0] if isinstance(UpperCamelCase_ , Image.Image ): lowerCAmelCase, lowerCAmelCase : Optional[int] = image.size else: lowerCAmelCase, lowerCAmelCase : List[Any] = image.shape[1], image.shape[2] lowerCAmelCase : Union[str, Any] = size / min(UpperCamelCase_ , UpperCamelCase_ ) if h < w: lowerCAmelCase, lowerCAmelCase : Dict = size, scale * w else: lowerCAmelCase, lowerCAmelCase : Optional[int] = scale * h, size lowerCAmelCase : List[Any] = int((1_3_3_3 / 8_0_0) * size ) if max(UpperCamelCase_ , UpperCamelCase_ ) > max_size: lowerCAmelCase : int = max_size / max(UpperCamelCase_ , UpperCamelCase_ ) lowerCAmelCase : str = newh * scale lowerCAmelCase : Tuple = neww * scale lowerCAmelCase, lowerCAmelCase : List[str] = int(newh + 0.5 ), int(neww + 0.5 ) lowerCAmelCase, lowerCAmelCase : Tuple = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: lowerCAmelCase : Optional[int] = [] for image in image_inputs: lowerCAmelCase, lowerCAmelCase : List[str] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) lowerCAmelCase : Union[str, Any] = max(UpperCamelCase_ , key=lambda UpperCamelCase_ : item[0] )[0] lowerCAmelCase : Union[str, Any] = max(UpperCamelCase_ , key=lambda UpperCamelCase_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class snake_case_( a__ , unittest.TestCase ): __UpperCamelCase = BridgeTowerImageProcessor if is_vision_available() else None def lowerCamelCase__ ( self : Optional[int] ): lowerCAmelCase : Optional[int] = BridgeTowerImageProcessingTester(self ) @property def lowerCamelCase__ ( self : List[str] ): return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase__ ( self : List[str] ): lowerCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCamelCase_ , '''image_mean''' ) ) self.assertTrue(hasattr(UpperCamelCase_ , '''image_std''' ) ) self.assertTrue(hasattr(UpperCamelCase_ , '''do_normalize''' ) ) self.assertTrue(hasattr(UpperCamelCase_ , '''do_resize''' ) ) self.assertTrue(hasattr(UpperCamelCase_ , '''size''' ) ) self.assertTrue(hasattr(UpperCamelCase_ , '''size_divisor''' ) ) def lowerCamelCase__ ( self : int ): pass def lowerCamelCase__ ( self : Optional[Any] ): # Initialize image processor lowerCAmelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , Image.Image ) # Test not batched input lowerCAmelCase : Optional[int] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowerCAmelCase, lowerCAmelCase : List[Any] = self.image_processor_tester.get_expected_values(UpperCamelCase_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase : Dict = image_processing(UpperCamelCase_ , return_tensors='''pt''' ).pixel_values lowerCAmelCase, lowerCAmelCase : int = self.image_processor_tester.get_expected_values(UpperCamelCase_ , batched=UpperCamelCase_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCamelCase__ ( self : Optional[Any] ): # Initialize image processor lowerCAmelCase : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , numpify=UpperCamelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , np.ndarray ) # Test not batched input lowerCAmelCase : Any = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowerCAmelCase, lowerCAmelCase : Optional[Any] = self.image_processor_tester.get_expected_values(UpperCamelCase_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase : Tuple = image_processing(UpperCamelCase_ , return_tensors='''pt''' ).pixel_values lowerCAmelCase, lowerCAmelCase : str = self.image_processor_tester.get_expected_values(UpperCamelCase_ , batched=UpperCamelCase_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCamelCase__ ( self : Optional[int] ): # Initialize image processor lowerCAmelCase : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , torchify=UpperCamelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , torch.Tensor ) # Test not batched input lowerCAmelCase : Any = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values lowerCAmelCase, lowerCAmelCase : Tuple = self.image_processor_tester.get_expected_values(UpperCamelCase_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase : str = image_processing(UpperCamelCase_ , return_tensors='''pt''' ).pixel_values lowerCAmelCase, lowerCAmelCase : str = self.image_processor_tester.get_expected_values(UpperCamelCase_ , batched=UpperCamelCase_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
60
0
import math import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from .attention_processor import Attention from .embeddings import get_timestep_embedding from .modeling_utils import ModelMixin class SCREAMING_SNAKE_CASE__ ( UpperCamelCase_ , UpperCamelCase_ ): """simple docstring""" @register_to_config def __init__( self : Optional[Any] , __A : int = 1_2_8 , __A : int = 2_5_6 , __A : float = 2_0_0_0.0 , __A : int = 7_6_8 , __A : int = 1_2 , __A : int = 1_2 , __A : int = 6_4 , __A : int = 2_0_4_8 , __A : float = 0.1 , ): super().__init__() snake_case__ : Any = nn.Sequential( nn.Linear(__A , d_model * 4 , bias=__A ) , nn.SiLU() , nn.Linear(d_model * 4 , d_model * 4 , bias=__A ) , nn.SiLU() , ) snake_case__ : int = nn.Embedding(__A , __A ) snake_case__ : List[str] = False snake_case__ : List[str] = nn.Linear(__A , __A , bias=__A ) snake_case__ : Dict = nn.Dropout(p=__A ) snake_case__ : int = nn.ModuleList() for lyr_num in range(__A ): # FiLM conditional T5 decoder snake_case__ : Any = DecoderLayer(d_model=__A , d_kv=__A , num_heads=__A , d_ff=__A , dropout_rate=__A ) self.decoders.append(__A ) snake_case__ : str = TaLayerNorm(__A ) snake_case__ : Union[str, Any] = nn.Dropout(p=__A ) snake_case__ : Optional[int] = nn.Linear(__A , __A , bias=__A ) def _lowercase ( self : Tuple , __A : Tuple , __A : Optional[Any] ): snake_case__ : List[str] = torch.mul(query_input.unsqueeze(-1 ) , key_input.unsqueeze(-2 ) ) return mask.unsqueeze(-3 ) def _lowercase ( self : List[str] , __A : Tuple , __A : Union[str, Any] , __A : Dict ): snake_case__, snake_case__, snake_case__ : Optional[int] = decoder_input_tokens.shape assert decoder_noise_time.shape == (batch,) # decoder_noise_time is in [0, 1), so rescale to expected timing range. snake_case__ : Tuple = get_timestep_embedding( decoder_noise_time * self.config.max_decoder_noise_time , embedding_dim=self.config.d_model , max_period=self.config.max_decoder_noise_time , ).to(dtype=self.dtype ) snake_case__ : int = self.conditioning_emb(__A ).unsqueeze(1 ) assert conditioning_emb.shape == (batch, 1, self.config.d_model * 4) snake_case__ : Any = decoder_input_tokens.shape[1] # If we want to use relative positions for audio context, we can just offset # this sequence by the length of encodings_and_masks. snake_case__ : Dict = torch.broadcast_to( torch.arange(__A , device=decoder_input_tokens.device ) , (batch, seq_length) , ) snake_case__ : Union[str, Any] = self.position_encoding(__A ) snake_case__ : Dict = self.continuous_inputs_projection(__A ) inputs += position_encodings snake_case__ : List[Any] = self.dropout(__A ) # decoder: No padding present. snake_case__ : Tuple = torch.ones( decoder_input_tokens.shape[:2] , device=decoder_input_tokens.device , dtype=inputs.dtype ) # Translate encoding masks to encoder-decoder masks. snake_case__ : List[str] = [(x, self.encoder_decoder_mask(__A , __A )) for x, y in encodings_and_masks] # cross attend style: concat encodings snake_case__ : Optional[int] = torch.cat([x[0] for x in encodings_and_encdec_masks] , dim=1 ) snake_case__ : List[str] = torch.cat([x[1] for x in encodings_and_encdec_masks] , dim=-1 ) for lyr in self.decoders: snake_case__ : Optional[Any] = lyr( __A , conditioning_emb=__A , encoder_hidden_states=__A , encoder_attention_mask=__A , )[0] snake_case__ : int = self.decoder_norm(__A ) snake_case__ : List[str] = self.post_dropout(__A ) snake_case__ : List[Any] = self.spec_out(__A ) return spec_out class SCREAMING_SNAKE_CASE__ ( nn.Module ): """simple docstring""" def __init__( self : Optional[int] , __A : Dict , __A : Optional[int] , __A : Tuple , __A : Dict , __A : Optional[int] , __A : List[str]=1e-6 ): super().__init__() snake_case__ : Optional[Any] = nn.ModuleList() # cond self attention: layer 0 self.layer.append( TaLayerSelfAttentionCond(d_model=__A , d_kv=__A , num_heads=__A , dropout_rate=__A ) ) # cross attention: layer 1 self.layer.append( TaLayerCrossAttention( d_model=__A , d_kv=__A , num_heads=__A , dropout_rate=__A , layer_norm_epsilon=__A , ) ) # Film Cond MLP + dropout: last layer self.layer.append( TaLayerFFCond(d_model=__A , d_ff=__A , dropout_rate=__A , layer_norm_epsilon=__A ) ) def _lowercase ( self : Optional[int] , __A : Dict , __A : int=None , __A : List[str]=None , __A : str=None , __A : Any=None , __A : str=None , ): snake_case__ : List[Any] = self.layer[0]( __A , conditioning_emb=__A , attention_mask=__A , ) if encoder_hidden_states is not None: snake_case__ : Optional[int] = torch.where(encoder_attention_mask > 0 , 0 , -1e1_0 ).to( encoder_hidden_states.dtype ) snake_case__ : Any = self.layer[1]( __A , key_value_states=__A , attention_mask=__A , ) # Apply Film Conditional Feed Forward layer snake_case__ : Tuple = self.layer[-1](__A , __A ) return (hidden_states,) class SCREAMING_SNAKE_CASE__ ( nn.Module ): """simple docstring""" def __init__( self : Tuple , __A : Any , __A : List[str] , __A : Optional[int] , __A : Any ): super().__init__() snake_case__ : Optional[int] = TaLayerNorm(__A ) snake_case__ : List[Any] = TaFiLMLayer(in_features=d_model * 4 , out_features=__A ) snake_case__ : Optional[Any] = Attention(query_dim=__A , heads=__A , dim_head=__A , out_bias=__A , scale_qk=__A ) snake_case__ : List[Any] = nn.Dropout(__A ) def _lowercase ( self : str , __A : Tuple , __A : Union[str, Any]=None , __A : Union[str, Any]=None , ): # pre_self_attention_layer_norm snake_case__ : Optional[Any] = self.layer_norm(__A ) if conditioning_emb is not None: snake_case__ : Optional[Any] = self.FiLMLayer(__A , __A ) # Self-attention block snake_case__ : Union[str, Any] = self.attention(__A ) snake_case__ : Any = hidden_states + self.dropout(__A ) return hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): """simple docstring""" def __init__( self : Dict , __A : Any , __A : List[str] , __A : List[str] , __A : Optional[Any] , __A : int ): super().__init__() snake_case__ : Tuple = Attention(query_dim=__A , heads=__A , dim_head=__A , out_bias=__A , scale_qk=__A ) snake_case__ : Optional[int] = TaLayerNorm(__A , eps=__A ) snake_case__ : str = nn.Dropout(__A ) def _lowercase ( self : Union[str, Any] , __A : str , __A : Union[str, Any]=None , __A : List[str]=None , ): snake_case__ : Union[str, Any] = self.layer_norm(__A ) snake_case__ : Union[str, Any] = self.attention( __A , encoder_hidden_states=__A , attention_mask=attention_mask.squeeze(1 ) , ) snake_case__ : Optional[Any] = hidden_states + self.dropout(__A ) return layer_output class SCREAMING_SNAKE_CASE__ ( nn.Module ): """simple docstring""" def __init__( self : Optional[int] , __A : Optional[int] , __A : List[Any] , __A : List[Any] , __A : Optional[int] ): super().__init__() snake_case__ : Union[str, Any] = TaDenseGatedActDense(d_model=__A , d_ff=__A , dropout_rate=__A ) snake_case__ : Tuple = TaFiLMLayer(in_features=d_model * 4 , out_features=__A ) snake_case__ : List[Any] = TaLayerNorm(__A , eps=__A ) snake_case__ : Tuple = nn.Dropout(__A ) def _lowercase ( self : int , __A : int , __A : Any=None ): snake_case__ : str = self.layer_norm(__A ) if conditioning_emb is not None: snake_case__ : int = self.film(__A , __A ) snake_case__ : List[Any] = self.DenseReluDense(__A ) snake_case__ : Dict = hidden_states + self.dropout(__A ) return hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): """simple docstring""" def __init__( self : int , __A : Dict , __A : str , __A : Tuple ): super().__init__() snake_case__ : List[str] = nn.Linear(__A , __A , bias=__A ) snake_case__ : int = nn.Linear(__A , __A , bias=__A ) snake_case__ : Union[str, Any] = nn.Linear(__A , __A , bias=__A ) snake_case__ : Union[str, Any] = nn.Dropout(__A ) snake_case__ : List[Any] = NewGELUActivation() def _lowercase ( self : List[Any] , __A : Any ): snake_case__ : Dict = self.act(self.wi_a(__A ) ) snake_case__ : Union[str, Any] = self.wi_a(__A ) snake_case__ : Union[str, Any] = hidden_gelu * hidden_linear snake_case__ : Union[str, Any] = self.dropout(__A ) snake_case__ : Optional[int] = self.wo(__A ) return hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): """simple docstring""" def __init__( self : str , __A : Dict , __A : Dict=1e-6 ): super().__init__() snake_case__ : List[Any] = nn.Parameter(torch.ones(__A ) ) snake_case__ : List[str] = eps def _lowercase ( self : Any , __A : Union[str, Any] ): # T5 uses a layer_norm which only scales and doesn't shift, which is also known as Root Mean # Square Layer Normalization https://arxiv.org/abs/1910.07467 thus variance is calculated # w/o mean and there is no bias. Additionally we want to make sure that the accumulation for # half-precision inputs is done in fp32 snake_case__ : Union[str, Any] = hidden_states.to(torch.floataa ).pow(2 ).mean(-1 , keepdim=__A ) snake_case__ : List[Any] = hidden_states * torch.rsqrt(variance + self.variance_epsilon ) # convert into half-precision if necessary if self.weight.dtype in [torch.floataa, torch.bfloataa]: snake_case__ : Optional[Any] = hidden_states.to(self.weight.dtype ) return self.weight * hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): """simple docstring""" def _lowercase ( self : int , __A : torch.Tensor ): return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi ) * (input + 0.0_4_4_7_1_5 * torch.pow(__A , 3.0 )) )) class SCREAMING_SNAKE_CASE__ ( nn.Module ): """simple docstring""" def __init__( self : List[str] , __A : Dict , __A : Optional[int] ): super().__init__() snake_case__ : Tuple = nn.Linear(__A , out_features * 2 , bias=__A ) def _lowercase ( self : Optional[int] , __A : Any , __A : List[str] ): snake_case__ : List[str] = self.scale_bias(__A ) snake_case__, snake_case__ : Optional[Any] = torch.chunk(__A , 2 , -1 ) snake_case__ : str = x * (1 + scale) + shift return x
286
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCamelCase : Union[str, Any] = { """configuration_time_series_transformer""": [ """TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TimeSeriesTransformerConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Tuple = [ """TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TimeSeriesTransformerForPrediction""", """TimeSeriesTransformerModel""", """TimeSeriesTransformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys __lowerCamelCase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
286
1
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase_ ): def UpperCamelCase ( self,__lowerCamelCase ): with open(__lowercase,encoding='''utf-8''' ) as input_file: A__ = re.compile(r'''(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)''' ) A__ = input_file.read() A__ = regexp.search(__lowercase ) return match def UpperCamelCase ( self,__lowerCamelCase ): with open(__lowercase,encoding='''utf-8''' ) as input_file: A__ = re.compile(r'''#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()''',re.DOTALL ) A__ = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` A__ = regexp.finditer(__lowercase ) A__ = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def UpperCamelCase ( self ): A__ = Path('''./datasets''' ) A__ = list(dataset_paths.absolute().glob('''**/*.py''' ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(__lowercase ) ): raise AssertionError(f"open(...) must use utf-8 encoding in {dataset}" ) def UpperCamelCase ( self ): A__ = Path('''./datasets''' ) A__ = list(dataset_paths.absolute().glob('''**/*.py''' ) ) for dataset in dataset_files: if self._no_print_statements(str(__lowercase ) ): raise AssertionError(f"print statement found in {dataset}. Use datasets.logger/logging instead." )
193
from typing import List, Optional, Union import numpy as np import PIL.Image from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, PILImageResampling, get_image_size, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __lowercase = logging.get_logger(__name__) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[Any] = ["""pixel_values"""] def __init__( self , __lowercase = True , __lowercase = 32 , __lowercase=PILImageResampling.BILINEAR , __lowercase = True , **__lowercase , ) -> None: __UpperCamelCase :Optional[int] = do_resize __UpperCamelCase :Any = do_rescale __UpperCamelCase :str = size_divisor __UpperCamelCase :Dict = resample super().__init__(**__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase = None , **__lowercase) -> np.ndarray: __UpperCamelCase , __UpperCamelCase :int = get_image_size(__lowercase) # Rounds the height and width down to the closest multiple of size_divisor __UpperCamelCase :List[Any] = height // size_divisor * size_divisor __UpperCamelCase :List[str] = width // size_divisor * size_divisor __UpperCamelCase :str = resize(__lowercase , (new_h, new_w) , resample=__lowercase , data_format=__lowercase , **__lowercase) return image def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase = None , **__lowercase) -> np.ndarray: return rescale(image=__lowercase , scale=__lowercase , data_format=__lowercase , **__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase = None , __lowercase = None , __lowercase=None , __lowercase = None , __lowercase = None , __lowercase = ChannelDimension.FIRST , **__lowercase , ) -> BatchFeature: __UpperCamelCase :Union[str, Any] = do_resize if do_resize is not None else self.do_resize __UpperCamelCase :Tuple = do_rescale if do_rescale is not None else self.do_rescale __UpperCamelCase :List[str] = size_divisor if size_divisor is not None else self.size_divisor __UpperCamelCase :List[Any] = resample if resample is not None else self.resample if do_resize and size_divisor is None: raise ValueError('''size_divisor is required for resizing''') __UpperCamelCase :List[Any] = make_list_of_images(__lowercase) if not valid_images(__lowercase): raise ValueError('''Invalid image(s)''') # All transformations expect numpy arrays. __UpperCamelCase :Optional[Any] = [to_numpy_array(__lowercase) for img in images] if do_resize: __UpperCamelCase :List[str] = [self.resize(__lowercase , size_divisor=__lowercase , resample=__lowercase) for image in images] if do_rescale: __UpperCamelCase :Dict = [self.rescale(__lowercase , scale=1 / 255) for image in images] __UpperCamelCase :str = [to_channel_dimension_format(__lowercase , __lowercase) for image in images] __UpperCamelCase :int = {'''pixel_values''': images} return BatchFeature(data=__lowercase , tensor_type=__lowercase)
43
0
'''simple docstring''' import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse('''0.8.3'''): raise Exception('''requires gluonnlp == 0.8.3''') if version.parse(mx.__version__) != version.parse('''1.5.0'''): raise Exception('''requires mxnet == 1.5.0''') logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = '''The Nymphenburg Palace is a beautiful palace in Munich!''' def lowercase_ ( lowerCAmelCase__ : str , lowerCAmelCase__ : str ): """simple docstring""" __UpperCAmelCase : Dict = { """attention_cell""": """multi_head""", """num_layers""": 4, """units""": 1024, """hidden_size""": 768, """max_length""": 512, """num_heads""": 8, """scaled""": True, """dropout""": 0.1, """use_residual""": True, """embed_size""": 1024, """embed_dropout""": 0.1, """word_embed""": None, """layer_norm_eps""": 1E-5, """token_type_vocab_size""": 2, } __UpperCAmelCase : Union[str, Any] = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py __UpperCAmelCase : List[Any] = BERTEncoder( attention_cell=predefined_args["""attention_cell"""] , num_layers=predefined_args["""num_layers"""] , units=predefined_args["""units"""] , hidden_size=predefined_args["""hidden_size"""] , max_length=predefined_args["""max_length"""] , num_heads=predefined_args["""num_heads"""] , scaled=predefined_args["""scaled"""] , dropout=predefined_args["""dropout"""] , output_attention=lowerCAmelCase__ , output_all_encodings=lowerCAmelCase__ , use_residual=predefined_args["""use_residual"""] , activation=predefined_args.get("""activation""" , """gelu""" ) , layer_norm_eps=predefined_args.get("""layer_norm_eps""" , lowerCAmelCase__ ) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later __UpperCAmelCase : Any = """openwebtext_ccnews_stories_books_cased""" # Specify download folder to Gluonnlp's vocab __UpperCAmelCase : List[str] = os.path.join(get_home_dir() , """models""" ) __UpperCAmelCase : Union[str, Any] = _load_vocab(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , cls=lowerCAmelCase__ ) __UpperCAmelCase : List[str] = nlp.model.BERTModel( lowerCAmelCase__ , len(lowerCAmelCase__ ) , units=predefined_args["""units"""] , embed_size=predefined_args["""embed_size"""] , embed_dropout=predefined_args["""embed_dropout"""] , word_embed=predefined_args["""word_embed"""] , use_pooler=lowerCAmelCase__ , use_token_type_embed=lowerCAmelCase__ , token_type_vocab_size=predefined_args["""token_type_vocab_size"""] , use_classifier=lowerCAmelCase__ , use_decoder=lowerCAmelCase__ , ) original_bort.load_parameters(lowerCAmelCase__ , cast_dtype=lowerCAmelCase__ , ignore_extra=lowerCAmelCase__ ) __UpperCAmelCase : Union[str, Any] = original_bort._collect_params_with_prefix() # Build our config 🤗 __UpperCAmelCase : str = { """architectures""": ["""BertForMaskedLM"""], """attention_probs_dropout_prob""": predefined_args["""dropout"""], """hidden_act""": """gelu""", """hidden_dropout_prob""": predefined_args["""dropout"""], """hidden_size""": predefined_args["""embed_size"""], """initializer_range""": 0.02, """intermediate_size""": predefined_args["""hidden_size"""], """layer_norm_eps""": predefined_args["""layer_norm_eps"""], """max_position_embeddings""": predefined_args["""max_length"""], """model_type""": """bort""", """num_attention_heads""": predefined_args["""num_heads"""], """num_hidden_layers""": predefined_args["""num_layers"""], """pad_token_id""": 1, # 2 = BERT, 1 = RoBERTa """type_vocab_size""": 1, # 2 = BERT, 1 = RoBERTa """vocab_size""": len(lowerCAmelCase__ ), } __UpperCAmelCase : Tuple = BertConfig.from_dict(lowerCAmelCase__ ) __UpperCAmelCase : str = BertForMaskedLM(lowerCAmelCase__ ) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(lowerCAmelCase__ : List[str] ) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy() ) ) # Check param shapes and map new HF param back def check_and_map_params(lowerCAmelCase__ : str , lowerCAmelCase__ : Optional[int] ): __UpperCAmelCase : Union[str, Any] = hf_param.shape __UpperCAmelCase : Optional[int] = to_torch(params[gluon_param] ) __UpperCAmelCase : List[Any] = gluon_param.shape assert ( shape_hf == shape_gluon ), f'The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers' return gluon_param __UpperCAmelCase : List[str] = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , """word_embed.0.weight""" ) __UpperCAmelCase : Any = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , """encoder.position_weight""" ) __UpperCAmelCase : List[str] = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , """encoder.layer_norm.beta""" ) __UpperCAmelCase : List[Any] = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , """encoder.layer_norm.gamma""" ) # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) __UpperCAmelCase : List[Any] = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data ) for i in range(hf_bort_config.num_hidden_layers ): __UpperCAmelCase : BertLayer = hf_bort_model.bert.encoder.layer[i] # self attention __UpperCAmelCase : BertSelfAttention = layer.attention.self __UpperCAmelCase : Any = check_and_map_params( self_attn.key.bias.data , f'encoder.transformer_cells.{i}.attention_cell.proj_key.bias' ) __UpperCAmelCase : Any = check_and_map_params( self_attn.key.weight.data , f'encoder.transformer_cells.{i}.attention_cell.proj_key.weight' ) __UpperCAmelCase : Optional[int] = check_and_map_params( self_attn.query.bias.data , f'encoder.transformer_cells.{i}.attention_cell.proj_query.bias' ) __UpperCAmelCase : Dict = check_and_map_params( self_attn.query.weight.data , f'encoder.transformer_cells.{i}.attention_cell.proj_query.weight' ) __UpperCAmelCase : Optional[Any] = check_and_map_params( self_attn.value.bias.data , f'encoder.transformer_cells.{i}.attention_cell.proj_value.bias' ) __UpperCAmelCase : List[Any] = check_and_map_params( self_attn.value.weight.data , f'encoder.transformer_cells.{i}.attention_cell.proj_value.weight' ) # self attention output __UpperCAmelCase : BertSelfOutput = layer.attention.output __UpperCAmelCase : str = check_and_map_params( self_output.dense.bias , f'encoder.transformer_cells.{i}.proj.bias' ) __UpperCAmelCase : List[str] = check_and_map_params( self_output.dense.weight , f'encoder.transformer_cells.{i}.proj.weight' ) __UpperCAmelCase : Optional[int] = check_and_map_params( self_output.LayerNorm.bias , f'encoder.transformer_cells.{i}.layer_norm.beta' ) __UpperCAmelCase : str = check_and_map_params( self_output.LayerNorm.weight , f'encoder.transformer_cells.{i}.layer_norm.gamma' ) # intermediate __UpperCAmelCase : BertIntermediate = layer.intermediate __UpperCAmelCase : Union[str, Any] = check_and_map_params( intermediate.dense.bias , f'encoder.transformer_cells.{i}.ffn.ffn_1.bias' ) __UpperCAmelCase : List[str] = check_and_map_params( intermediate.dense.weight , f'encoder.transformer_cells.{i}.ffn.ffn_1.weight' ) # output __UpperCAmelCase : BertOutput = layer.output __UpperCAmelCase : List[Any] = check_and_map_params( bert_output.dense.bias , f'encoder.transformer_cells.{i}.ffn.ffn_2.bias' ) __UpperCAmelCase : Any = check_and_map_params( bert_output.dense.weight , f'encoder.transformer_cells.{i}.ffn.ffn_2.weight' ) __UpperCAmelCase : Tuple = check_and_map_params( bert_output.LayerNorm.bias , f'encoder.transformer_cells.{i}.ffn.layer_norm.beta' ) __UpperCAmelCase : Tuple = check_and_map_params( bert_output.LayerNorm.weight , f'encoder.transformer_cells.{i}.ffn.layer_norm.gamma' ) # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models __UpperCAmelCase : List[str] = RobertaTokenizer.from_pretrained("""roberta-base""" ) __UpperCAmelCase : Dict = tokenizer.encode_plus(lowerCAmelCase__ )["""input_ids"""] # Get gluon output __UpperCAmelCase : int = mx.nd.array([input_ids] ) __UpperCAmelCase : int = original_bort(inputs=lowerCAmelCase__ , token_types=[] ) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(lowerCAmelCase__ ) __UpperCAmelCase : Optional[Any] = BertModel.from_pretrained(lowerCAmelCase__ ) hf_bort_model.eval() __UpperCAmelCase : Optional[int] = tokenizer.encode_plus(lowerCAmelCase__ , return_tensors="""pt""" ) __UpperCAmelCase : Dict = hf_bort_model(**lowerCAmelCase__ )[0] __UpperCAmelCase : int = output_gluon[0].asnumpy() __UpperCAmelCase : str = output_hf[0].detach().numpy() __UpperCAmelCase : Optional[Any] = np.max(np.abs(hf_layer - gluon_layer ) ).item() __UpperCAmelCase : Any = np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-3 ) if success: print("""✔️ Both model do output the same tensors""" ) else: print("""❌ Both model do **NOT** output the same tensors""" ) print("""Absolute difference is:""" , lowerCAmelCase__ ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bort_checkpoint_path''', default=None, type=str, required=True, help='''Path the official Bort params file.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) _UpperCamelCase = parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
16
'''simple docstring''' import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor _UpperCamelCase = logging.get_logger(__name__) class _A ( __SCREAMING_SNAKE_CASE ): def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> None: '''simple docstring''' warnings.warn( """The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use LayoutLMv2ImageProcessor instead.""" , __UpperCAmelCase , ) super().__init__(*__UpperCAmelCase , **__UpperCAmelCase )
16
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_lxmert import LxmertTokenizer A_ :Dict = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ :Dict = { '''vocab_file''': { '''unc-nlp/lxmert-base-uncased''': '''https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/vocab.txt''', }, '''tokenizer_file''': { '''unc-nlp/lxmert-base-uncased''': ( '''https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/tokenizer.json''' ), }, } A_ :str = { '''unc-nlp/lxmert-base-uncased''': 512, } A_ :Dict = { '''unc-nlp/lxmert-base-uncased''': {'''do_lower_case''': True}, } class __A ( a ): """simple docstring""" UpperCamelCase__ : Optional[Any] =VOCAB_FILES_NAMES UpperCamelCase__ : List[str] =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : Union[str, Any] =PRETRAINED_INIT_CONFIGURATION UpperCamelCase__ : int =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : str =LxmertTokenizer def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__="[UNK]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ): """simple docstring""" super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) __UpperCamelCase : Dict =json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , lowerCamelCase__ ) != do_lower_case or normalizer_state.get('strip_accents' , lowerCamelCase__ ) != strip_accents or normalizer_state.get('handle_chinese_chars' , lowerCamelCase__ ) != tokenize_chinese_chars ): __UpperCamelCase : str =getattr(lowerCamelCase__ , normalizer_state.pop('type' ) ) __UpperCamelCase : Any =do_lower_case __UpperCamelCase : Dict =strip_accents __UpperCamelCase : List[str] =tokenize_chinese_chars __UpperCamelCase : Union[str, Any] =normalizer_class(**lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =do_lower_case def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=None ): """simple docstring""" __UpperCamelCase : List[str] =[self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : str =[self.sep_token_id] __UpperCamelCase : Union[str, Any] =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : int =self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
71
from __future__ import annotations import math lowerCamelCase : List[Any] = '''2020.9.26''' lowerCamelCase : str = '''xcodz-dot, cclaus, dhruvmanila''' def snake_case_ ( lowerCAmelCase_ : float , lowerCAmelCase_ : float , lowerCAmelCase_ : float , lowerCAmelCase_ : float , lowerCAmelCase_ : float ): if not all(isinstance(lowerCAmelCase_ , (float, int) ) for val in locals().values() ): __lowercase : str = F"Input values must either be float or int: {list(locals().values() )}" raise TypeError(lowerCAmelCase_ ) __lowercase : List[Any] = ((x * distance) / (z + distance)) * scale __lowercase : Tuple = ((y * distance) / (z + distance)) * scale return projected_x, projected_y def snake_case_ ( lowerCAmelCase_ : float , lowerCAmelCase_ : float , lowerCAmelCase_ : float , lowerCAmelCase_ : str , lowerCAmelCase_ : float ): if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("""Axis must be a str""" ) __lowercase : Optional[int] = locals() del input_variables["axis"] if not all(isinstance(lowerCAmelCase_ , (float, int) ) for val in input_variables.values() ): __lowercase : List[str] = ( """Input values except axis must either be float or int: """ F"{list(input_variables.values() )}" ) raise TypeError(lowerCAmelCase_ ) __lowercase : Tuple = (angle % 360) / 450 * 180 / math.pi if axis == "z": __lowercase : int = x * math.cos(lowerCAmelCase_ ) - y * math.sin(lowerCAmelCase_ ) __lowercase : Tuple = y * math.cos(lowerCAmelCase_ ) + x * math.sin(lowerCAmelCase_ ) __lowercase : Union[str, Any] = z elif axis == "x": __lowercase : str = y * math.cos(lowerCAmelCase_ ) - z * math.sin(lowerCAmelCase_ ) __lowercase : Dict = z * math.cos(lowerCAmelCase_ ) + y * math.sin(lowerCAmelCase_ ) __lowercase : List[str] = x elif axis == "y": __lowercase : List[str] = x * math.cos(lowerCAmelCase_ ) - z * math.sin(lowerCAmelCase_ ) __lowercase : List[str] = z * math.cos(lowerCAmelCase_ ) + x * math.sin(lowerCAmelCase_ ) __lowercase : List[Any] = y else: raise ValueError("""not a valid axis, choose one of 'x', 'y', 'z'""" ) return new_x, new_y, new_z if __name__ == "__main__": import doctest doctest.testmod() print(f'''{convert_to_ad(1.0, 2.0, 3.0, 10.0, 10.0) = }''') print(f'''{rotate(1.0, 2.0, 3.0, "y", 90.0) = }''')
233
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class __snake_case : lowerCAmelCase_ = LEDConfig lowerCAmelCase_ = {} lowerCAmelCase_ = "gelu" def __init__( self : Any , _lowercase : Tuple , _lowercase : str=13 , _lowercase : Optional[int]=7 , _lowercase : Optional[Any]=True , _lowercase : Dict=False , _lowercase : Union[str, Any]=99 , _lowercase : Any=32 , _lowercase : int=2 , _lowercase : List[str]=4 , _lowercase : Optional[int]=37 , _lowercase : Union[str, Any]=0.1 , _lowercase : str=0.1 , _lowercase : Union[str, Any]=20 , _lowercase : List[str]=2 , _lowercase : Optional[int]=1 , _lowercase : Dict=0 , _lowercase : List[str]=4 , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = seq_length SCREAMING_SNAKE_CASE__ = is_training SCREAMING_SNAKE_CASE__ = use_labels SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = eos_token_id SCREAMING_SNAKE_CASE__ = pad_token_id SCREAMING_SNAKE_CASE__ = bos_token_id SCREAMING_SNAKE_CASE__ = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after SCREAMING_SNAKE_CASE__ = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests SCREAMING_SNAKE_CASE__ = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def __a ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) SCREAMING_SNAKE_CASE__ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) SCREAMING_SNAKE_CASE__ = tf.concat([input_ids, eos_tensor] , axis=1 ) SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE__ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) SCREAMING_SNAKE_CASE__ = prepare_led_inputs_dict(_lowercase , _lowercase , _lowercase ) SCREAMING_SNAKE_CASE__ = tf.concat( [tf.zeros_like(_lowercase )[:, :-1], tf.ones_like(_lowercase )[:, -1:]] , axis=-1 , ) SCREAMING_SNAKE_CASE__ = global_attention_mask return config, inputs_dict def __a ( self : Tuple , _lowercase : int , _lowercase : int ): """simple docstring""" SCREAMING_SNAKE_CASE__ = TFLEDModel(config=_lowercase ).get_decoder() SCREAMING_SNAKE_CASE__ = inputs_dict["""input_ids"""] SCREAMING_SNAKE_CASE__ = input_ids[:1, :] SCREAMING_SNAKE_CASE__ = inputs_dict["""attention_mask"""][:1, :] SCREAMING_SNAKE_CASE__ = 1 # first forward pass SCREAMING_SNAKE_CASE__ = model(_lowercase , attention_mask=_lowercase , use_cache=_lowercase ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids SCREAMING_SNAKE_CASE__ = ids_tensor((self.batch_size, 3) , config.vocab_size ) SCREAMING_SNAKE_CASE__ = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and SCREAMING_SNAKE_CASE__ = tf.concat([input_ids, next_tokens] , axis=-1 ) SCREAMING_SNAKE_CASE__ = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) SCREAMING_SNAKE_CASE__ = model(_lowercase , attention_mask=_lowercase )[0] SCREAMING_SNAKE_CASE__ = model(_lowercase , attention_mask=_lowercase , past_key_values=_lowercase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice SCREAMING_SNAKE_CASE__ = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) SCREAMING_SNAKE_CASE__ = output_from_no_past[:, -3:, random_slice_idx] SCREAMING_SNAKE_CASE__ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_lowercase , _lowercase , rtol=1E-3 ) def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : Dict , __UpperCamelCase : int , __UpperCamelCase : List[str]=None , __UpperCamelCase : Union[str, Any]=None , __UpperCamelCase : Dict=None , __UpperCamelCase : Optional[Any]=None , ) -> int: """simple docstring""" if attention_mask is None: SCREAMING_SNAKE_CASE__ = tf.cast(tf.math.not_equal(__UpperCamelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: SCREAMING_SNAKE_CASE__ = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: SCREAMING_SNAKE_CASE__ = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: SCREAMING_SNAKE_CASE__ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class __snake_case ( lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): lowerCAmelCase_ = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () lowerCAmelCase_ = (TFLEDForConditionalGeneration,) if is_tf_available() else () lowerCAmelCase_ = ( { "conversational": TFLEDForConditionalGeneration, "feature-extraction": TFLEDModel, "summarization": TFLEDForConditionalGeneration, "text2text-generation": TFLEDForConditionalGeneration, "translation": TFLEDForConditionalGeneration, } if is_tf_available() else {} ) lowerCAmelCase_ = True lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def __a ( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE__ = TFLEDModelTester(self ) SCREAMING_SNAKE_CASE__ = ConfigTester(self , config_class=_lowercase ) def __a ( self : str ): """simple docstring""" self.config_tester.run_common_tests() def __a ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_lowercase ) def __a ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE__ = tf.zeros_like(inputs_dict["""attention_mask"""] ) SCREAMING_SNAKE_CASE__ = 2 SCREAMING_SNAKE_CASE__ = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict["""global_attention_mask"""] , ) SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = self.model_tester.seq_length SCREAMING_SNAKE_CASE__ = self.model_tester.encoder_seq_length def check_decoder_attentions_output(_lowercase : List[str] ): SCREAMING_SNAKE_CASE__ = outputs.decoder_attentions self.assertEqual(len(_lowercase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(_lowercase : List[str] ): SCREAMING_SNAKE_CASE__ = [t.numpy() for t in outputs.encoder_attentions] SCREAMING_SNAKE_CASE__ = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(_lowercase ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(_lowercase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = model_class(_lowercase ) SCREAMING_SNAKE_CASE__ = model(self._prepare_for_class(_lowercase , _lowercase ) ) SCREAMING_SNAKE_CASE__ = len(_lowercase ) self.assertEqual(config.output_hidden_states , _lowercase ) check_encoder_attentions_output(_lowercase ) if self.is_encoder_decoder: SCREAMING_SNAKE_CASE__ = model_class(_lowercase ) SCREAMING_SNAKE_CASE__ = model(self._prepare_for_class(_lowercase , _lowercase ) ) self.assertEqual(config.output_hidden_states , _lowercase ) check_decoder_attentions_output(_lowercase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = model_class(_lowercase ) SCREAMING_SNAKE_CASE__ = model(self._prepare_for_class(_lowercase , _lowercase ) ) self.assertEqual(config.output_hidden_states , _lowercase ) check_encoder_attentions_output(_lowercase ) # Check attention is always last and order is fine SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = model_class(_lowercase ) SCREAMING_SNAKE_CASE__ = model(self._prepare_for_class(_lowercase , _lowercase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(_lowercase ) ) self.assertEqual(model.config.output_hidden_states , _lowercase ) check_encoder_attentions_output(_lowercase ) @unittest.skip("""LED keeps using potentially symbolic tensors in conditionals and breaks tracing.""" ) def __a ( self : List[Any] ): """simple docstring""" pass def __a ( self : List[str] ): """simple docstring""" pass def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : List[str] ) -> Dict: """simple docstring""" return tf.constant(__UpperCamelCase , dtype=tf.intaa ) __lowerCamelCase : List[str] = 1e-4 @slow @require_tf class __snake_case ( unittest.TestCase ): def __a ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ = TFLEDForConditionalGeneration.from_pretrained("""allenai/led-base-16384""" ).led # change to intended input here SCREAMING_SNAKE_CASE__ = _long_tensor([5_12 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) SCREAMING_SNAKE_CASE__ = _long_tensor([1_28 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) SCREAMING_SNAKE_CASE__ = prepare_led_inputs_dict(model.config , _lowercase , _lowercase ) SCREAMING_SNAKE_CASE__ = model(**_lowercase )[0] SCREAMING_SNAKE_CASE__ = (1, 10_24, 7_68) self.assertEqual(output.shape , _lowercase ) # change to expected output here SCREAMING_SNAKE_CASE__ = tf.convert_to_tensor( [[2.30_50, 2.82_79, 0.65_31], [-1.84_57, -0.14_55, -3.56_61], [-1.01_86, 0.45_86, -2.20_43]] , ) tf.debugging.assert_near(output[:, :3, :3] , _lowercase , atol=1E-3 ) def __a ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = TFLEDForConditionalGeneration.from_pretrained("""allenai/led-base-16384""" ) # change to intended input here SCREAMING_SNAKE_CASE__ = _long_tensor([5_12 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) SCREAMING_SNAKE_CASE__ = _long_tensor([1_28 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) SCREAMING_SNAKE_CASE__ = prepare_led_inputs_dict(model.config , _lowercase , _lowercase ) SCREAMING_SNAKE_CASE__ = model(**_lowercase )[0] SCREAMING_SNAKE_CASE__ = (1, 10_24, model.config.vocab_size) self.assertEqual(output.shape , _lowercase ) # change to expected output here SCREAMING_SNAKE_CASE__ = tf.convert_to_tensor( [[33.65_07, 6.45_72, 16.80_89], [5.87_39, -2.42_38, 11.29_02], [-3.21_39, -4.31_49, 4.27_83]] , ) tf.debugging.assert_near(output[:, :3, :3] , _lowercase , atol=1E-3 , rtol=1E-3 )
361
def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : List[Any] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = [0] * len(__UpperCamelCase ) SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [1] * len(__UpperCamelCase ) for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(__UpperCamelCase ) ): if indegree[i] == 0: queue.append(__UpperCamelCase ) while queue: SCREAMING_SNAKE_CASE__ = queue.pop(0 ) for x in graph[vertex]: indegree[x] -= 1 if long_dist[vertex] + 1 > long_dist[x]: SCREAMING_SNAKE_CASE__ = long_dist[vertex] + 1 if indegree[x] == 0: queue.append(__UpperCamelCase ) print(max(__UpperCamelCase ) ) # Adjacency list of Graph __lowerCamelCase : int = {0: [2, 3, 4], 1: [2, 7], 2: [5], 3: [5, 7], 4: [7], 5: [6], 6: [7], 7: []} longest_distance(graph)
204
0
'''simple docstring''' from ..utils import DummyObject, requires_backends class A ( metaclass=SCREAMING_SNAKE_CASE__ ): lowerCamelCase : str = ["""transformers""", """torch""", """note_seq"""] def __init__( self , *lowerCamelCase__ , **lowerCamelCase__ ) -> List[str]: '''simple docstring''' requires_backends(self , ["""transformers""", """torch""", """note_seq"""] ) @classmethod def A__ ( cls , *lowerCamelCase__ , **lowerCamelCase__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ["""transformers""", """torch""", """note_seq"""] ) @classmethod def A__ ( cls , *lowerCamelCase__ , **lowerCamelCase__ ) -> List[Any]: '''simple docstring''' requires_backends(cls , ["""transformers""", """torch""", """note_seq"""] )
164
"""simple docstring""" import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration a_ = 50000 a_ = 5000 a_ , a_ = os.path.split(__file__) a_ = os.path.join(RESULTS_BASEPATH, """results""", RESULTS_FILENAME.replace(""".py""", """.json""")) @get_duration def __lowercase ( snake_case_ : datasets.Dataset ,snake_case_ : Tuple ) ->Tuple: '''simple docstring''' for i in range(snake_case_ ): __A : int = dataset[i] @get_duration def __lowercase ( snake_case_ : datasets.Dataset ,snake_case_ : Optional[Any] ,snake_case_ : int ) ->Tuple: '''simple docstring''' for i in range(0 ,len(snake_case_ ) ,snake_case_ ): __A : List[str] = dataset[i : i + batch_size] @get_duration def __lowercase ( snake_case_ : datasets.Dataset ,snake_case_ : List[Any] ,snake_case_ : Any ) ->int: '''simple docstring''' with dataset.formatted_as(type=snake_case_ ): for i in range(snake_case_ ): __A : Union[str, Any] = dataset[i] @get_duration def __lowercase ( snake_case_ : datasets.Dataset ,snake_case_ : Any ,snake_case_ : Union[str, Any] ,snake_case_ : Optional[int] ) ->Union[str, Any]: '''simple docstring''' with dataset.formatted_as(type=snake_case_ ): for i in range(0 ,snake_case_ ,snake_case_ ): __A : Dict = dataset[i : i + batch_size] def __lowercase ( ) ->Optional[int]: '''simple docstring''' __A : int = {'''num examples''': SPEED_TEST_N_EXAMPLES} __A : Optional[int] = [ (read, {'''length''': SMALL_TEST}), (read, {'''length''': SPEED_TEST_N_EXAMPLES}), (read_batch, {'''length''': SPEED_TEST_N_EXAMPLES, '''batch_size''': 10}), (read_batch, {'''length''': SPEED_TEST_N_EXAMPLES, '''batch_size''': 100}), (read_batch, {'''length''': SPEED_TEST_N_EXAMPLES, '''batch_size''': 1000}), (read_formatted, {'''type''': '''numpy''', '''length''': SMALL_TEST}), (read_formatted, {'''type''': '''pandas''', '''length''': SMALL_TEST}), (read_formatted, {'''type''': '''torch''', '''length''': SMALL_TEST}), (read_formatted, {'''type''': '''tensorflow''', '''length''': SMALL_TEST}), (read_formatted_batch, {'''type''': '''numpy''', '''length''': SMALL_TEST, '''batch_size''': 10}), (read_formatted_batch, {'''type''': '''numpy''', '''length''': SMALL_TEST, '''batch_size''': 1000}), ] __A : int = [ (read, {'''length''': SMALL_TEST}), (read, {'''length''': SPEED_TEST_N_EXAMPLES}), (read_batch, {'''length''': SPEED_TEST_N_EXAMPLES, '''batch_size''': 10}), (read_batch, {'''length''': SPEED_TEST_N_EXAMPLES, '''batch_size''': 100}), (read_batch, {'''length''': SPEED_TEST_N_EXAMPLES, '''batch_size''': 1000}), (read_formatted, {'''type''': '''numpy''', '''length''': SMALL_TEST}), (read_formatted_batch, {'''type''': '''numpy''', '''length''': SMALL_TEST, '''batch_size''': 10}), (read_formatted_batch, {'''type''': '''numpy''', '''length''': SMALL_TEST, '''batch_size''': 1000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print('''generating dataset''' ) __A : Any = datasets.Features( {'''list''': datasets.Sequence(datasets.Value('''float32''' ) ), '''numbers''': datasets.Value('''float32''' )} ) __A : List[Any] = generate_example_dataset( os.path.join(snake_case_ ,'''dataset.arrow''' ) ,snake_case_ ,num_examples=snake_case_ ,seq_shapes={'''list''': (100,)} ,) print('''first set of iterations''' ) for func, kwargs in functions: print(func.__name__ ,str(snake_case_ ) ) __A : Dict = func(snake_case_ ,**snake_case_ ) print('''shuffling dataset''' ) __A : int = dataset.shuffle() print('''Second set of iterations (after shuffling''' ) for func, kwargs in functions_shuffled: print('''shuffled ''' ,func.__name__ ,str(snake_case_ ) ) __A : Optional[Any] = func( snake_case_ ,**snake_case_ ) with open(snake_case_ ,'''wb''' ) as f: f.write(json.dumps(snake_case_ ).encode('''utf-8''' ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
179
0
'''simple docstring''' # DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class _UpperCAmelCase : a : CommonSchedulerState # setable values a : jnp.ndarray a : jnp.ndarray a : Optional[int] =None @classmethod def lowerCamelCase__ ( cls,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): '''simple docstring''' return cls(common=__SCREAMING_SNAKE_CASE,init_noise_sigma=__SCREAMING_SNAKE_CASE,timesteps=__SCREAMING_SNAKE_CASE ) @dataclass class _UpperCAmelCase ( lowerCAmelCase_ ): a : DDPMSchedulerState class _UpperCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ ): a : List[str] =[e.name for e in FlaxKarrasDiffusionSchedulers] a : jnp.dtype @property def lowerCamelCase__ ( self ): '''simple docstring''' return True @register_to_config def __init__( self,__SCREAMING_SNAKE_CASE = 10_00,__SCREAMING_SNAKE_CASE = 0.0001,__SCREAMING_SNAKE_CASE = 0.02,__SCREAMING_SNAKE_CASE = "linear",__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = "fixed_small",__SCREAMING_SNAKE_CASE = True,__SCREAMING_SNAKE_CASE = "epsilon",__SCREAMING_SNAKE_CASE = jnp.floataa,): '''simple docstring''' __lowerCAmelCase = dtype def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE = None ): '''simple docstring''' if common is None: __lowerCAmelCase = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution __lowerCAmelCase = jnp.array(1.0,dtype=self.dtype ) __lowerCAmelCase = jnp.arange(0,self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=__SCREAMING_SNAKE_CASE,init_noise_sigma=__SCREAMING_SNAKE_CASE,timesteps=__SCREAMING_SNAKE_CASE,) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE = None ): '''simple docstring''' return sample def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE = () ): '''simple docstring''' __lowerCAmelCase = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 __lowerCAmelCase = (jnp.arange(0,__SCREAMING_SNAKE_CASE ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=__SCREAMING_SNAKE_CASE,timesteps=__SCREAMING_SNAKE_CASE,) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE=None,__SCREAMING_SNAKE_CASE=None ): '''simple docstring''' __lowerCAmelCase = state.common.alphas_cumprod[t] __lowerCAmelCase = jnp.where(t > 0,state.common.alphas_cumprod[t - 1],jnp.array(1.0,dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample __lowerCAmelCase = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: __lowerCAmelCase = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": __lowerCAmelCase = jnp.clip(__SCREAMING_SNAKE_CASE,a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": __lowerCAmelCase = jnp.log(jnp.clip(__SCREAMING_SNAKE_CASE,a_min=1e-20 ) ) elif variance_type == "fixed_large": __lowerCAmelCase = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log __lowerCAmelCase = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": __lowerCAmelCase = variance __lowerCAmelCase = state.common.betas[t] __lowerCAmelCase = (predicted_variance + 1) / 2 __lowerCAmelCase = frac * max_log + (1 - frac) * min_log return variance def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = True,): '''simple docstring''' __lowerCAmelCase = timestep if key is None: __lowerCAmelCase = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: __lowerCAmelCase , __lowerCAmelCase = jnp.split(__SCREAMING_SNAKE_CASE,sample.shape[1],axis=1 ) else: __lowerCAmelCase = None # 1. compute alphas, betas __lowerCAmelCase = state.common.alphas_cumprod[t] __lowerCAmelCase = jnp.where(t > 0,state.common.alphas_cumprod[t - 1],jnp.array(1.0,dtype=self.dtype ) ) __lowerCAmelCase = 1 - alpha_prod_t __lowerCAmelCase = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": __lowerCAmelCase = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": __lowerCAmelCase = model_output elif self.config.prediction_type == "v_prediction": __lowerCAmelCase = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( f'prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ' """ for the FlaxDDPMScheduler.""" ) # 3. Clip "predicted x_0" if self.config.clip_sample: __lowerCAmelCase = jnp.clip(__SCREAMING_SNAKE_CASE,-1,1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf __lowerCAmelCase = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t __lowerCAmelCase = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf __lowerCAmelCase = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): __lowerCAmelCase = jax.random.split(__SCREAMING_SNAKE_CASE,num=1 ) __lowerCAmelCase = jax.random.normal(__SCREAMING_SNAKE_CASE,shape=model_output.shape,dtype=self.dtype ) return (self._get_variance(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,predicted_variance=__SCREAMING_SNAKE_CASE ) ** 0.5) * noise __lowerCAmelCase = jnp.where(t > 0,random_variance(),jnp.zeros(model_output.shape,dtype=self.dtype ) ) __lowerCAmelCase = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=__SCREAMING_SNAKE_CASE,state=__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,): '''simple docstring''' return add_noise_common(state.common,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,): '''simple docstring''' return get_velocity_common(state.common,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ) def __len__( self ): '''simple docstring''' return self.config.num_train_timesteps
46
'''simple docstring''' import warnings from .generation import TFGenerationMixin class _UpperCAmelCase ( lowerCAmelCase_ ): # warning at import time warnings.warn( """Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will """ """be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.""" , lowerCAmelCase_ , )
46
1
'''simple docstring''' from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput lowerCamelCase_ = logging.get_logger(__name__) # pylint: disable=invalid-name class _UpperCAmelCase ( snake_case_ , snake_case_ ): """simple docstring""" @register_to_config def __init__( self : Union[str, Any] , __UpperCAmelCase : bool , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : Optional[int] = None ): '''simple docstring''' super().__init__() _A = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" _A = torch.zeros(__UpperCAmelCase , __UpperCAmelCase ) else: _A = None _A = torch.nn.Parameter(__UpperCAmelCase ) class _UpperCAmelCase ( snake_case_ ): """simple docstring""" snake_case = 42 snake_case = 42 snake_case = 42 snake_case = 42 snake_case = 42 snake_case = 42 def __init__( self : Any , __UpperCAmelCase : VQModel , __UpperCAmelCase : CLIPTextModel , __UpperCAmelCase : CLIPTokenizer , __UpperCAmelCase : TransformeraDModel , __UpperCAmelCase : VQDiffusionScheduler , __UpperCAmelCase : LearnedClassifierFreeSamplingEmbeddings , ): '''simple docstring''' super().__init__() self.register_modules( vqvae=__UpperCAmelCase , transformer=__UpperCAmelCase , text_encoder=__UpperCAmelCase , tokenizer=__UpperCAmelCase , scheduler=__UpperCAmelCase , learned_classifier_free_sampling_embeddings=__UpperCAmelCase , ) def lowerCAmelCase ( self : int , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : Any ): '''simple docstring''' _A = len(__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else 1 # get prompt text embeddings _A = self.tokenizer( __UpperCAmelCase , padding="max_length" , max_length=self.tokenizer.model_max_length , return_tensors="pt" , ) _A = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: _A = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( "The following part of your input was truncated because CLIP can only handle sequences up to" f''' {self.tokenizer.model_max_length} tokens: {removed_text}''' ) _A = text_input_ids[:, : self.tokenizer.model_max_length] _A = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 _A = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=__UpperCAmelCase ) # duplicate text embeddings for each generation per prompt _A = prompt_embeds.repeat_interleave(__UpperCAmelCase , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: _A = self.learned_classifier_free_sampling_embeddings.embeddings _A = negative_prompt_embeds.unsqueeze(0 ).repeat(__UpperCAmelCase , 1 , 1 ) else: _A = [""] * batch_size _A = text_input_ids.shape[-1] _A = self.tokenizer( __UpperCAmelCase , padding="max_length" , max_length=__UpperCAmelCase , truncation=__UpperCAmelCase , return_tensors="pt" , ) _A = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings _A = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=__UpperCAmelCase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method _A = negative_prompt_embeds.shape[1] _A = negative_prompt_embeds.repeat(1 , __UpperCAmelCase , 1 ) _A = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __UpperCAmelCase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes _A = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self : Optional[Any] , __UpperCAmelCase : Union[str, List[str]] , __UpperCAmelCase : int = 100 , __UpperCAmelCase : float = 5.0 , __UpperCAmelCase : float = 1.0 , __UpperCAmelCase : int = 1 , __UpperCAmelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __UpperCAmelCase : Optional[torch.FloatTensor] = None , __UpperCAmelCase : Optional[str] = "pil" , __UpperCAmelCase : bool = True , __UpperCAmelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , __UpperCAmelCase : int = 1 , ): '''simple docstring''' if isinstance(__UpperCAmelCase , __UpperCAmelCase ): _A = 1 elif isinstance(__UpperCAmelCase , __UpperCAmelCase ): _A = len(__UpperCAmelCase ) else: raise ValueError(f'''`prompt` has to be of type `str` or `list` but is {type(__UpperCAmelCase )}''' ) _A = batch_size * num_images_per_prompt _A = guidance_scale > 1.0 _A = self._encode_prompt(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or callback_steps <= 0) ): raise ValueError( f'''`callback_steps` has to be a positive integer but is {callback_steps} of type''' f''' {type(__UpperCAmelCase )}.''' ) # get the initial completely masked latents unless the user supplied it _A = (batch_size, self.transformer.num_latent_pixels) if latents is None: _A = self.transformer.num_vector_embeds - 1 _A = torch.full(__UpperCAmelCase , __UpperCAmelCase ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( "Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0," f''' {self.transformer.num_vector_embeds - 1} (inclusive).''' ) _A = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(__UpperCAmelCase , device=self.device ) _A = self.scheduler.timesteps.to(self.device ) _A = latents for i, t in enumerate(self.progress_bar(__UpperCAmelCase ) ): # expand the sample if we are doing classifier free guidance _A = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` _A = self.transformer(__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , timestep=__UpperCAmelCase ).sample if do_classifier_free_guidance: _A , _A = model_output.chunk(2 ) _A = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(__UpperCAmelCase , dim=1 , keepdim=__UpperCAmelCase ) _A = self.truncate(__UpperCAmelCase , __UpperCAmelCase ) # remove `log(0)`'s (`-inf`s) _A = model_output.clamp(-70 ) # compute the previous noisy sample x_t -> x_t-1 _A = self.scheduler.step(__UpperCAmelCase , timestep=__UpperCAmelCase , sample=__UpperCAmelCase , generator=__UpperCAmelCase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) _A = self.vqvae.config.vq_embed_dim _A = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) _A = self.vqvae.quantize.get_codebook_entry(__UpperCAmelCase , shape=__UpperCAmelCase ) _A = self.vqvae.decode(__UpperCAmelCase , force_not_quantize=__UpperCAmelCase ).sample _A = (image / 2 + 0.5).clamp(0 , 1 ) _A = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _A = self.numpy_to_pil(__UpperCAmelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__UpperCAmelCase ) def lowerCAmelCase ( self : Union[str, Any] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : float ): '''simple docstring''' _A , _A = torch.sort(__UpperCAmelCase , 1 , descending=__UpperCAmelCase ) _A = torch.exp(__UpperCAmelCase ) _A = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out _A = torch.full_like(keep_mask[:, 0:1, :] , __UpperCAmelCase ) _A = torch.cat((all_true, keep_mask) , dim=1 ) _A = keep_mask[:, :-1, :] _A = keep_mask.gather(1 , indices.argsort(1 ) ) _A = log_p_x_0.clone() _A = -torch.inf # -inf = log(0) return rv
79
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class _snake_case ( _snake_case ): SCREAMING_SNAKE_CASE__ = 'ClapFeatureExtractor' SCREAMING_SNAKE_CASE__ = ('RobertaTokenizer', 'RobertaTokenizerFast') def __init__( self , _lowerCamelCase , _lowerCamelCase ): super().__init__(_lowerCamelCase , _lowerCamelCase ) def __call__( self , _lowerCamelCase=None , _lowerCamelCase=None , _lowerCamelCase=None , **_lowerCamelCase ): a :Dict = kwargs.pop('''sampling_rate''' , _lowerCamelCase ) if text is None and audios is None: raise ValueError('''You have to specify either text or audios. Both cannot be none.''' ) if text is not None: a :Optional[int] = self.tokenizer(_lowerCamelCase , return_tensors=_lowerCamelCase , **_lowerCamelCase ) if audios is not None: a :Tuple = self.feature_extractor( _lowerCamelCase , sampling_rate=_lowerCamelCase , return_tensors=_lowerCamelCase , **_lowerCamelCase ) if text is not None and audios is not None: a :Union[str, Any] = audio_features.input_features return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCamelCase ) , tensor_type=_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self , *_lowerCamelCase , **_lowerCamelCase ): return self.tokenizer.batch_decode(*_lowerCamelCase , **_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self , *_lowerCamelCase , **_lowerCamelCase ): return self.tokenizer.decode(*_lowerCamelCase , **_lowerCamelCase ) @property def SCREAMING_SNAKE_CASE__ ( self ): a :List[Any] = self.tokenizer.model_input_names a :str = self.feature_extractor.model_input_names return list(dict.fromkeys(tokenizer_input_names + feature_extractor_input_names ) )
94
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCAmelCase : Union[str, Any] = logging.get_logger(__name__) _UpperCAmelCase : str = { """facebook/levit-128S""": """https://huggingface.co/facebook/levit-128S/resolve/main/config.json""", # See all LeViT models at https://huggingface.co/models?filter=levit } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """levit""" def __init__( self : Dict , UpperCAmelCase : List[Any]=224 , UpperCAmelCase : Dict=3 , UpperCAmelCase : Any=3 , UpperCAmelCase : Optional[Any]=2 , UpperCAmelCase : Any=1 , UpperCAmelCase : Optional[int]=16 , UpperCAmelCase : Dict=[128, 256, 384] , UpperCAmelCase : List[str]=[4, 8, 12] , UpperCAmelCase : Optional[int]=[4, 4, 4] , UpperCAmelCase : Optional[Any]=[16, 16, 16] , UpperCAmelCase : Tuple=0 , UpperCAmelCase : str=[2, 2, 2] , UpperCAmelCase : List[str]=[2, 2, 2] , UpperCAmelCase : Any=0.0_2 , **UpperCAmelCase : int , ) -> Dict: super().__init__(**UpperCAmelCase ) lowerCamelCase__ : Tuple = image_size lowerCamelCase__ : Optional[Any] = num_channels lowerCamelCase__ : Union[str, Any] = kernel_size lowerCamelCase__ : Optional[int] = stride lowerCamelCase__ : str = padding lowerCamelCase__ : int = hidden_sizes lowerCamelCase__ : Union[str, Any] = num_attention_heads lowerCamelCase__ : Union[str, Any] = depths lowerCamelCase__ : Optional[int] = key_dim lowerCamelCase__ : Dict = drop_path_rate lowerCamelCase__ : Dict = patch_size lowerCamelCase__ : str = attention_ratio lowerCamelCase__ : Tuple = mlp_ratio lowerCamelCase__ : Tuple = initializer_range lowerCamelCase__ : str = [ ['Subsample', key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ['Subsample', key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = version.parse("""1.11""" ) @property def A_ ( self : str ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def A_ ( self : Union[str, Any] ) -> float: return 1e-4
45
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _UpperCAmelCase : Tuple = [ """openmmlab/upernet-convnext-tiny""", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _UpperCAmelCase : List[str] = """UperNetConfig""" class lowerCAmelCase ( nn.Module ): def __init__( self : int , UpperCAmelCase : int , UpperCAmelCase : int , UpperCAmelCase : Union[int, Tuple[int, int]] , UpperCAmelCase : Union[int, Tuple[int, int], str] = 0 , UpperCAmelCase : bool = False , UpperCAmelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() lowerCamelCase__ : Any = nn.Convad( in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , kernel_size=UpperCAmelCase , padding=UpperCAmelCase , bias=UpperCAmelCase , dilation=UpperCAmelCase , ) lowerCamelCase__ : str = nn.BatchNormad(UpperCAmelCase ) lowerCamelCase__ : Tuple = nn.ReLU() def A_ ( self : Tuple , UpperCAmelCase : torch.Tensor ) -> torch.Tensor: lowerCamelCase__ : Tuple = self.conv(UpperCAmelCase ) lowerCamelCase__ : int = self.batch_norm(UpperCAmelCase ) lowerCamelCase__ : List[Any] = self.activation(UpperCAmelCase ) return output class lowerCAmelCase ( nn.Module ): def __init__( self : Optional[Any] , UpperCAmelCase : int , UpperCAmelCase : int , UpperCAmelCase : int ) -> None: super().__init__() lowerCamelCase__ : int = [ nn.AdaptiveAvgPoolad(UpperCAmelCase ), UperNetConvModule(UpperCAmelCase , UpperCAmelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(UpperCAmelCase ) , UpperCAmelCase ) def A_ ( self : Union[str, Any] , UpperCAmelCase : torch.Tensor ) -> torch.Tensor: lowerCamelCase__ : Dict = input for layer in self.layers: lowerCamelCase__ : Tuple = layer(UpperCAmelCase ) return hidden_state class lowerCAmelCase ( nn.Module ): def __init__( self : Tuple , UpperCAmelCase : Tuple[int, ...] , UpperCAmelCase : int , UpperCAmelCase : int , UpperCAmelCase : bool ) -> None: super().__init__() lowerCamelCase__ : int = pool_scales lowerCamelCase__ : Tuple = align_corners lowerCamelCase__ : Union[str, Any] = in_channels lowerCamelCase__ : List[Any] = channels lowerCamelCase__ : Tuple = [] for i, pool_scale in enumerate(UpperCAmelCase ): lowerCamelCase__ : Dict = UperNetPyramidPoolingBlock(pool_scale=UpperCAmelCase , in_channels=UpperCAmelCase , channels=UpperCAmelCase ) self.blocks.append(UpperCAmelCase ) self.add_module(str(UpperCAmelCase ) , UpperCAmelCase ) def A_ ( self : Optional[int] , UpperCAmelCase : torch.Tensor ) -> List[torch.Tensor]: lowerCamelCase__ : Tuple = [] for ppm in self.blocks: lowerCamelCase__ : Union[str, Any] = ppm(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = nn.functional.interpolate( UpperCAmelCase , size=x.size()[2:] , mode='bilinear' , align_corners=self.align_corners ) ppm_outs.append(UpperCAmelCase ) return ppm_outs class lowerCAmelCase ( nn.Module ): def __init__( self : str , UpperCAmelCase : int , UpperCAmelCase : Any ) -> int: super().__init__() lowerCamelCase__ : Tuple = config lowerCamelCase__ : Optional[Any] = config.pool_scales # e.g. (1, 2, 3, 6) lowerCamelCase__ : List[Any] = in_channels lowerCamelCase__ : Optional[int] = config.hidden_size lowerCamelCase__ : Dict = False lowerCamelCase__ : List[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module lowerCamelCase__ : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) lowerCamelCase__ : int = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module lowerCamelCase__ : str = nn.ModuleList() lowerCamelCase__ : str = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer lowerCamelCase__ : str = UperNetConvModule(UpperCAmelCase , self.channels , kernel_size=1 ) lowerCamelCase__ : int = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(UpperCAmelCase ) self.fpn_convs.append(UpperCAmelCase ) lowerCamelCase__ : List[Any] = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def A_ ( self : Tuple ) -> List[Any]: self.apply(self._init_weights ) def A_ ( self : Tuple , UpperCAmelCase : Dict ) -> str: if isinstance(UpperCAmelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def A_ ( self : Optional[int] , UpperCAmelCase : Optional[int] ) -> Optional[int]: lowerCamelCase__ : str = inputs[-1] lowerCamelCase__ : List[str] = [x] psp_outs.extend(self.psp_modules(UpperCAmelCase ) ) lowerCamelCase__ : Tuple = torch.cat(UpperCAmelCase , dim=1 ) lowerCamelCase__ : Optional[Any] = self.bottleneck(UpperCAmelCase ) return output def A_ ( self : str , UpperCAmelCase : torch.Tensor ) -> torch.Tensor: # build laterals lowerCamelCase__ : Union[str, Any] = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(UpperCAmelCase ) ) # build top-down path lowerCamelCase__ : Tuple = len(UpperCAmelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): lowerCamelCase__ : Optional[Any] = laterals[i - 1].shape[2:] lowerCamelCase__ : Any = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=UpperCAmelCase , mode='bilinear' , align_corners=self.align_corners ) # build outputs lowerCamelCase__ : Any = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): lowerCamelCase__ : Optional[Any] = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode='bilinear' , align_corners=self.align_corners ) lowerCamelCase__ : Dict = torch.cat(UpperCAmelCase , dim=1 ) lowerCamelCase__ : List[str] = self.fpn_bottleneck(UpperCAmelCase ) lowerCamelCase__ : int = self.classifier(UpperCAmelCase ) return output class lowerCAmelCase ( nn.Module ): def __init__( self : Optional[int] , UpperCAmelCase : Optional[int] , UpperCAmelCase : int = 2 , UpperCAmelCase : int = 3 , UpperCAmelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() lowerCamelCase__ : Any = config lowerCamelCase__ : Optional[Any] = config.auxiliary_in_channels lowerCamelCase__ : str = config.auxiliary_channels lowerCamelCase__ : Optional[Any] = config.auxiliary_num_convs lowerCamelCase__ : str = config.auxiliary_concat_input lowerCamelCase__ : List[Any] = in_index lowerCamelCase__ : List[str] = (kernel_size // 2) * dilation lowerCamelCase__ : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=UpperCAmelCase , padding=UpperCAmelCase , dilation=UpperCAmelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=UpperCAmelCase , padding=UpperCAmelCase , dilation=UpperCAmelCase ) ) if self.num_convs == 0: lowerCamelCase__ : Optional[Any] = nn.Identity() else: lowerCamelCase__ : Optional[Any] = nn.Sequential(*UpperCAmelCase ) if self.concat_input: lowerCamelCase__ : Any = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=UpperCAmelCase , padding=kernel_size // 2 ) lowerCamelCase__ : Dict = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def A_ ( self : Tuple ) -> Tuple: self.apply(self._init_weights ) def A_ ( self : Union[str, Any] , UpperCAmelCase : List[Any] ) -> List[str]: if isinstance(UpperCAmelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def A_ ( self : Tuple , UpperCAmelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps lowerCamelCase__ : str = encoder_hidden_states[self.in_index] lowerCamelCase__ : Union[str, Any] = self.convs(UpperCAmelCase ) if self.concat_input: lowerCamelCase__ : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) lowerCamelCase__ : Optional[int] = self.classifier(UpperCAmelCase ) return output class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = UperNetConfig UpperCAmelCase__ = """pixel_values""" UpperCAmelCase__ = True def A_ ( self : Optional[Any] , UpperCAmelCase : Union[str, Any] ) -> Optional[int]: if isinstance(UpperCAmelCase , UpperCAmelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def A_ ( self : str ) -> Tuple: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def A_ ( self : Any , UpperCAmelCase : Dict , UpperCAmelCase : Optional[int]=False ) -> str: if isinstance(UpperCAmelCase , UpperCAmelCase ): lowerCamelCase__ : Any = value _UpperCAmelCase : List[Any] = R""" Parameters: This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. config ([`UperNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. """ _UpperCAmelCase : Union[str, Any] = R""" Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details. output_attentions (`bool`, *optional*): Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See `attentions` under returned tensors for more detail. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. """ @add_start_docstrings( """UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes.""", __UpperCamelCase, ) class lowerCAmelCase ( __UpperCamelCase ): def __init__( self : Optional[int] , UpperCAmelCase : Union[str, Any] ) -> Dict: super().__init__(UpperCAmelCase ) lowerCamelCase__ : List[str] = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) lowerCamelCase__ : List[Any] = UperNetHead(UpperCAmelCase , in_channels=self.backbone.channels ) lowerCamelCase__ : int = UperNetFCNHead(UpperCAmelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format('batch_size, sequence_length' ) ) @replace_return_docstrings(output_type=UpperCAmelCase , config_class=_CONFIG_FOR_DOC ) def A_ ( self : Union[str, Any] , UpperCAmelCase : Optional[torch.Tensor] = None , UpperCAmelCase : Optional[bool] = None , UpperCAmelCase : Optional[bool] = None , UpperCAmelCase : Optional[torch.Tensor] = None , UpperCAmelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: lowerCamelCase__ : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict lowerCamelCase__ : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowerCamelCase__ : str = output_attentions if output_attentions is not None else self.config.output_attentions lowerCamelCase__ : Optional[Any] = self.backbone.forward_with_filtered_kwargs( UpperCAmelCase , output_hidden_states=UpperCAmelCase , output_attentions=UpperCAmelCase ) lowerCamelCase__ : List[str] = outputs.feature_maps lowerCamelCase__ : Optional[int] = self.decode_head(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = nn.functional.interpolate(UpperCAmelCase , size=pixel_values.shape[2:] , mode='bilinear' , align_corners=UpperCAmelCase ) lowerCamelCase__ : List[str] = None if self.auxiliary_head is not None: lowerCamelCase__ : List[Any] = self.auxiliary_head(UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = nn.functional.interpolate( UpperCAmelCase , size=pixel_values.shape[2:] , mode='bilinear' , align_corners=UpperCAmelCase ) lowerCamelCase__ : List[str] = None if labels is not None: if self.config.num_labels == 1: raise ValueError('The number of labels should be greater than one' ) else: # compute weighted loss lowerCamelCase__ : Optional[Any] = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) lowerCamelCase__ : str = loss_fct(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Tuple = loss_fct(UpperCAmelCase , UpperCAmelCase ) lowerCamelCase__ : Tuple = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: lowerCamelCase__ : List[str] = (logits,) + outputs[1:] else: lowerCamelCase__ : int = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=UpperCAmelCase , logits=UpperCAmelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
45
1
"""simple docstring""" import torch from diffusers import StableDiffusionPipeline a = '''path-to-your-trained-model''' a = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to('''cuda''') a = '''A photo of sks dog in a bucket''' a = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save('''dog-bucket.png''')
315
"""simple docstring""" import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class snake_case ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): @register_to_config def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = False , ) ->int: super().__init__() a_ = nn.Embedding(__UpperCAmelCase , __UpperCAmelCase) a_ = nn.Embedding(__UpperCAmelCase , __UpperCAmelCase) a_ = False a_ = nn.Dropout(p=__UpperCAmelCase) a_ = TaConfig( vocab_size=__UpperCAmelCase , d_model=__UpperCAmelCase , num_heads=__UpperCAmelCase , d_kv=__UpperCAmelCase , d_ff=__UpperCAmelCase , dropout_rate=__UpperCAmelCase , feed_forward_proj=__UpperCAmelCase , is_decoder=__UpperCAmelCase , is_encoder_decoder=__UpperCAmelCase , ) a_ = nn.ModuleList() for lyr_num in range(__UpperCAmelCase): a_ = TaBlock(__UpperCAmelCase) self.encoders.append(__UpperCAmelCase) a_ = TaLayerNorm(__UpperCAmelCase) a_ = nn.Dropout(p=__UpperCAmelCase) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase) ->Union[str, Any]: a_ = self.token_embedder(__UpperCAmelCase) a_ = encoder_input_tokens.shape[1] a_ = torch.arange(__UpperCAmelCase , device=encoder_input_tokens.device) x += self.position_encoding(__UpperCAmelCase) a_ = self.dropout_pre(__UpperCAmelCase) # inverted the attention mask a_ = encoder_input_tokens.size() a_ = self.get_extended_attention_mask(__UpperCAmelCase , __UpperCAmelCase) for lyr in self.encoders: a_ = lyr(__UpperCAmelCase , __UpperCAmelCase)[0] a_ = self.layer_norm(__UpperCAmelCase) return self.dropout_post(__UpperCAmelCase), encoder_inputs_mask
243
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase : List[str] = {'''configuration_ibert''': ['''IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''IBertConfig''', '''IBertOnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Optional[Any] = [ '''IBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''IBertForMaskedLM''', '''IBertForMultipleChoice''', '''IBertForQuestionAnswering''', '''IBertForSequenceClassification''', '''IBertForTokenClassification''', '''IBertModel''', '''IBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys _lowerCamelCase : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
368
import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class lowerCamelCase : """simple docstring""" def __init__( self : Dict, _UpperCAmelCase : Any, _UpperCAmelCase : List[Any]=sys.maxsize ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = "bilinear" SCREAMING_SNAKE_CASE__ : Optional[int] = max_size SCREAMING_SNAKE_CASE__ : Optional[int] = short_edge_length def __call__( self : Optional[int], _UpperCAmelCase : Union[str, Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = [] for img in imgs: SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ : Tuple = img.shape[:2] # later: provide list and randomly choose index for resize SCREAMING_SNAKE_CASE__ : List[str] = np.random.randint(self.short_edge_length[0], self.short_edge_length[1] + 1 ) if size == 0: return img SCREAMING_SNAKE_CASE__ : int = size * 1.0 / min(_UpperCAmelCase, _UpperCAmelCase ) if h < w: SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ : Optional[int] = size, scale * w else: SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ : List[str] = scale * h, size if max(_UpperCAmelCase, _UpperCAmelCase ) > self.max_size: SCREAMING_SNAKE_CASE__ : str = self.max_size * 1.0 / max(_UpperCAmelCase, _UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Any = newh * scale SCREAMING_SNAKE_CASE__ : List[str] = neww * scale SCREAMING_SNAKE_CASE__ : Any = int(neww + 0.5 ) SCREAMING_SNAKE_CASE__ : List[Any] = int(newh + 0.5 ) if img.dtype == np.uinta: SCREAMING_SNAKE_CASE__ : List[Any] = Image.fromarray(_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = pil_image.resize((neww, newh), PILImageResampling.BILINEAR ) SCREAMING_SNAKE_CASE__ : List[str] = np.asarray(_UpperCAmelCase ) else: SCREAMING_SNAKE_CASE__ : str = img.permute(2, 0, 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw SCREAMING_SNAKE_CASE__ : Tuple = nn.functional.interpolate( _UpperCAmelCase, (newh, neww), mode=self.interp_method, align_corners=_UpperCAmelCase ).squeeze(0 ) img_augs.append(_UpperCAmelCase ) return img_augs class lowerCamelCase : """simple docstring""" def __init__( self : Dict, _UpperCAmelCase : Optional[Any] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST], cfg.INPUT.MAX_SIZE_TEST ) SCREAMING_SNAKE_CASE__ : Any = cfg.INPUT.FORMAT SCREAMING_SNAKE_CASE__ : List[str] = cfg.SIZE_DIVISIBILITY SCREAMING_SNAKE_CASE__ : List[Any] = cfg.PAD_VALUE SCREAMING_SNAKE_CASE__ : Dict = cfg.INPUT.MAX_SIZE_TEST SCREAMING_SNAKE_CASE__ : Optional[int] = cfg.MODEL.DEVICE SCREAMING_SNAKE_CASE__ : int = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ), 1, 1 ) SCREAMING_SNAKE_CASE__ : str = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ), 1, 1 ) SCREAMING_SNAKE_CASE__ : Optional[Any] = lambda _UpperCAmelCase : (x - self.pixel_mean) / self.pixel_std def A_ ( self : str, _UpperCAmelCase : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = tuple(max(_UpperCAmelCase ) for s in zip(*[img.shape for img in images] ) ) SCREAMING_SNAKE_CASE__ : Optional[Any] = [im.shape[-2:] for im in images] SCREAMING_SNAKE_CASE__ : Union[str, Any] = [ nn.functional.pad( _UpperCAmelCase, [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]], value=self.pad_value, ) for size, im in zip(_UpperCAmelCase, _UpperCAmelCase ) ] return torch.stack(_UpperCAmelCase ), torch.tensor(_UpperCAmelCase ) def __call__( self : Any, _UpperCAmelCase : Dict, _UpperCAmelCase : List[str]=False ) -> Optional[Any]: """simple docstring""" with torch.no_grad(): if not isinstance(_UpperCAmelCase, _UpperCAmelCase ): SCREAMING_SNAKE_CASE__ : str = [images] if single_image: assert len(_UpperCAmelCase ) == 1 for i in range(len(_UpperCAmelCase ) ): if isinstance(images[i], torch.Tensor ): images.insert(_UpperCAmelCase, images.pop(_UpperCAmelCase ).to(self.device ).float() ) elif not isinstance(images[i], torch.Tensor ): images.insert( _UpperCAmelCase, torch.as_tensor(img_tensorize(images.pop(_UpperCAmelCase ), input_format=self.input_format ) ) .to(self.device ) .float(), ) # resize smallest edge SCREAMING_SNAKE_CASE__ : List[str] = torch.tensor([im.shape[:2] for im in images] ) SCREAMING_SNAKE_CASE__ : Tuple = self.aug(_UpperCAmelCase ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic SCREAMING_SNAKE_CASE__ : List[Any] = [self.normalizer(_UpperCAmelCase ) for x in images] # now pad them to do the following operations SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ : List[str] = self.pad(_UpperCAmelCase ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad SCREAMING_SNAKE_CASE__ : Union[str, Any] = torch.true_divide(_UpperCAmelCase, _UpperCAmelCase ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def _a ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Dict ) -> Union[str, Any]: '''simple docstring''' boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def _a ( SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Tuple[int, int] ) -> List[Any]: '''simple docstring''' assert torch.isfinite(SCREAMING_SNAKE_CASE__ ).all(), "Box tensor contains infinite or NaN!" SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ : int = box_size tensor[:, 0].clamp_(min=0 , max=SCREAMING_SNAKE_CASE__ ) tensor[:, 1].clamp_(min=0 , max=SCREAMING_SNAKE_CASE__ ) tensor[:, 2].clamp_(min=0 , max=SCREAMING_SNAKE_CASE__ ) tensor[:, 3].clamp_(min=0 , max=SCREAMING_SNAKE_CASE__ )
191
0
'''simple docstring''' def lowerCamelCase ( __lowerCamelCase : int ) ->int: return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def lowerCamelCase ( __lowerCamelCase : int ) ->bool: _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = number while duplicate > 0: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = divmod(__lowerCamelCase , 10 ) fact_sum += factorial(__lowerCamelCase ) return fact_sum == number if __name__ == "__main__": print("""Program to check whether a number is a Krisnamurthy Number or not.""") lowercase_ = int(input("""Enter number: """).strip()) print( f"""{number} is {"" if krishnamurthy(number) else "not "}a Krishnamurthy Number.""" )
58
'''simple docstring''' def _A ( lowercase__ = 1000000 ): lowercase__ = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , lowercase__ ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
164
0
import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch __A ='sshleifer/bart-tiny-random' __A ='patrickvonplaten/t5-tiny-random' @require_torch class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): @cached_property def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: return AutoConfig.from_pretrained(lowercase_ ) def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: lowerCamelCase_ , *lowerCamelCase_ = create_student_by_copying_alternating_layers(lowercase_ , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.num_hidden_layers , 1 ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: lowerCamelCase_ , *lowerCamelCase_ = create_student_by_copying_alternating_layers(lowercase_ , tempfile.mkdtemp() , e=1 , d=lowercase_ ) def SCREAMING_SNAKE_CASE_( self ) -> Dict: lowerCamelCase_ , *lowerCamelCase_ = create_student_by_copying_alternating_layers(lowercase_ , tempfile.mkdtemp() , e=1 , d=lowercase_ ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , self.teacher_config.encoder_layers ) def SCREAMING_SNAKE_CASE_( self ) -> int: lowerCamelCase_ , *lowerCamelCase_ = create_student_by_copying_alternating_layers(lowercase_ , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , 1 ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: with self.assertRaises(lowercase_ ): create_student_by_copying_alternating_layers(lowercase_ , tempfile.mkdtemp() , e=lowercase_ , d=lowercase_ )
355
import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": __A =pd.read_csv('''sample_data.csv''', header=None) __A =df.shape[:1][0] # If you're using some other dataset input the target column __A =df.iloc[:, 1:2] __A =actual_data.values.reshape(len_data, 1) __A =MinMaxScaler().fit_transform(actual_data) __A =1_0 __A =5 __A =2_0 __A =len_data - periods * look_back __A =actual_data[:division] __A =actual_data[division - look_back :] __A, __A =[], [] __A, __A =[], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) __A =np.array(train_x) __A =np.array(test_x) __A =np.array([list(i.ravel()) for i in train_y]) __A =np.array([list(i.ravel()) for i in test_y]) __A =Sequential() model.add(LSTM(1_2_8, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(6_4, input_shape=(1_2_8, 1))) model.add(Dense(forward_days)) model.compile(loss='''mean_squared_error''', optimizer='''adam''') __A =model.fit( x_train, y_train, epochs=1_5_0, verbose=1, shuffle=True, batch_size=4 ) __A =model.predict(x_test)
47
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase: Dict = { 'configuration_mobilebert': [ 'MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MobileBertConfig', 'MobileBertOnnxConfig', ], 'tokenization_mobilebert': ['MobileBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase: Tuple = ['MobileBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase: int = [ 'MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'MobileBertForMaskedLM', 'MobileBertForMultipleChoice', 'MobileBertForNextSentencePrediction', 'MobileBertForPreTraining', 'MobileBertForQuestionAnswering', 'MobileBertForSequenceClassification', 'MobileBertForTokenClassification', 'MobileBertLayer', 'MobileBertModel', 'MobileBertPreTrainedModel', 'load_tf_weights_in_mobilebert', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase: List[Any] = [ 'TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFMobileBertForMaskedLM', 'TFMobileBertForMultipleChoice', 'TFMobileBertForNextSentencePrediction', 'TFMobileBertForPreTraining', 'TFMobileBertForQuestionAnswering', 'TFMobileBertForSequenceClassification', 'TFMobileBertForTokenClassification', 'TFMobileBertMainLayer', 'TFMobileBertModel', 'TFMobileBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mobilebert import ( MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileBertConfig, MobileBertOnnxConfig, ) from .tokenization_mobilebert import MobileBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mobilebert_fast import MobileBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilebert import ( MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertLayer, MobileBertModel, MobileBertPreTrainedModel, load_tf_weights_in_mobilebert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilebert import ( TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertMainLayer, TFMobileBertModel, TFMobileBertPreTrainedModel, ) else: import sys lowerCAmelCase: str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
297
'''simple docstring''' def lowerCamelCase__ ( _A ): return 10 - x * x def lowerCamelCase__ ( _A , _A ): # Bolzano theory in order to find if there is a root between a and b if equation(_A ) * equation(_A ) >= 0: raise ValueError('Wrong space!' ) a : Tuple = a while (b - a) >= 0.01: # Find middle point a : Tuple = (a + b) / 2 # Check if middle point is root if equation(_A ) == 0.0: break # Decide the side to repeat the steps if equation(_A ) * equation(_A ) < 0: a : List[str] = c else: a : Tuple = c return c if __name__ == "__main__": import doctest doctest.testmod() print(bisection(-2, 5)) print(bisection(0, 6))
297
1
"""simple docstring""" import numpy as np def UpperCAmelCase__ (snake_case__ : np.ndarray ): """simple docstring""" return 1 / (1 + np.exp(-vector )) def UpperCAmelCase__ (snake_case__ : np.ndarray ): """simple docstring""" return vector * sigmoid(snake_case__ ) if __name__ == "__main__": import doctest doctest.testmod()
364
"""simple docstring""" import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger A_ = '''<<<<<<< This should probably be modified because it mentions: ''' A_ = '''======= >>>>>>> ''' A_ = [ '''TextEncoderConfig''', '''ByteTextEncoder''', '''SubwordTextEncoder''', '''encoder_config''', '''maybe_build_from_corpus''', '''manual_dir''', ] A_ = [ # (pattern, replacement) # Order is important here for some replacements (r'''tfds\.core''', r'''datasets'''), (r'''tf\.io\.gfile\.GFile''', r'''open'''), (r'''tf\.([\w\d]+)''', r'''datasets.Value(\'\1\')'''), (r'''tfds\.features\.Text\(\)''', r'''datasets.Value(\'string\')'''), (r'''tfds\.features\.Text\(''', r'''datasets.Value(\'string\'),'''), (r'''features\s*=\s*tfds.features.FeaturesDict\(''', r'''features=datasets.Features('''), (r'''tfds\.features\.FeaturesDict\(''', r'''dict('''), (r'''The TensorFlow Datasets Authors''', r'''The TensorFlow Datasets Authors and the HuggingFace Datasets Authors'''), (r'''tfds\.''', r'''datasets.'''), (r'''dl_manager\.manual_dir''', r'''self.config.data_dir'''), (r'''self\.builder_config''', r'''self.config'''), ] def UpperCAmelCase__ (snake_case__ : Namespace ): """simple docstring""" return ConvertCommand(args.tfds_path , args.datasets_directory ) class lowercase( __a ): '''simple docstring''' @staticmethod def UpperCamelCase_ ( a_: ArgumentParser ): '''simple docstring''' _snake_case : Tuple = parser.add_parser( """convert""", help="""Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.""", ) train_parser.add_argument( """--tfds_path""", type=a_, required=a_, help="""Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.""", ) train_parser.add_argument( """--datasets_directory""", type=a_, required=a_, help="""Path to the HuggingFace Datasets folder.""" ) train_parser.set_defaults(func=a_ ) def __init__( self: List[str], a_: str, a_: str, *a_: str ): '''simple docstring''' _snake_case : Optional[Any] = get_logger("""datasets-cli/converting""" ) _snake_case : Any = tfds_path _snake_case : Optional[Any] = datasets_directory def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' if os.path.isdir(self._tfds_path ): _snake_case : int = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): _snake_case : Any = os.path.dirname(self._tfds_path ) else: raise ValueError("""--tfds_path is neither a directory nor a file. Please check path.""" ) _snake_case : Union[str, Any] = os.path.abspath(self._datasets_directory ) self._logger.info(f"Converting datasets from {abs_tfds_path} to {abs_datasets_path}" ) _snake_case : Tuple = [] _snake_case : Dict = [] _snake_case : Optional[Any] = {} if os.path.isdir(self._tfds_path ): _snake_case : List[str] = os.listdir(a_ ) else: _snake_case : int = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(f"Looking at file {f_name}" ) _snake_case : Dict = os.path.join(a_, a_ ) _snake_case : Union[str, Any] = os.path.join(a_, a_ ) if not os.path.isfile(a_ ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info("""Skipping file""" ) continue with open(a_, encoding="""utf-8""" ) as f: _snake_case : str = f.readlines() _snake_case : List[str] = [] _snake_case : Any = False _snake_case : Union[str, Any] = False _snake_case : Optional[Any] = [] for line in lines: _snake_case : Optional[int] = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: _snake_case : Optional[Any] = """import datasets\n""" elif "import tensorflow" in out_line: # order is important here _snake_case : Optional[int] = """""" continue elif "from absl import logging" in out_line: _snake_case : int = """from datasets import logging\n""" elif "getLogger" in out_line: _snake_case : Any = out_line.replace("""getLogger""", """get_logger""" ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): _snake_case : Union[str, Any] = True _snake_case : Optional[Any] = list(filter(lambda a_ : e in out_line, a_ ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(a_ ) + """\n""" ) out_lines.append(a_ ) out_lines.append(a_ ) continue else: for pattern, replacement in TO_CONVERT: _snake_case : List[str] = re.sub(a_, a_, a_ ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: _snake_case : Dict = re.match(r"""from\stensorflow_datasets.*import\s([^\.\r\n]+)""", a_ ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split(""",""" ) ) _snake_case : Optional[Any] = """from . import """ + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(f"Error converting {out_line.strip()}" ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: _snake_case : Tuple = True out_lines.append(a_ ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset _snake_case : List[str] = f_name.replace(""".py""", """""" ) _snake_case : str = os.path.join(a_, a_ ) _snake_case : str = os.path.join(a_, a_ ) os.makedirs(a_, exist_ok=a_ ) self._logger.info(f"Adding directory {output_dir}" ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(a_ ) if needs_manual_update: with_manual_update.append(a_ ) with open(a_, """w""", encoding="""utf-8""" ) as f: f.writelines(a_ ) self._logger.info(f"Converted in {output_file}" ) for utils_file in utils_files: try: _snake_case : Optional[int] = os.path.basename(a_ ) _snake_case : Optional[Any] = imports_to_builder_map[f_name.replace(""".py""", """""" )] self._logger.info(f"Moving {dest_folder} to {utils_file}" ) shutil.copy(a_, a_ ) except KeyError: self._logger.error(f"Cannot find destination folder for {utils_file}. Please copy manually." ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( f"You need to manually update file {file_path} to remove configurations using 'TextEncoderConfig'." )
132
0
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig lowerCamelCase__ : List[str] = logging.get_logger(__name__) class lowerCamelCase_ : '''simple docstring''' def __init__( self : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] ): SCREAMING_SNAKE_CASE_ = question_encoder SCREAMING_SNAKE_CASE_ = generator SCREAMING_SNAKE_CASE_ = self.question_encoder def lowerCAmelCase_ ( self : Tuple , _lowerCAmelCase : List[str] ): if os.path.isfile(_lowerCAmelCase ): raise ValueError(F"Provided path ({save_directory}) should be a directory, not a file" ) os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = os.path.join(_lowerCAmelCase , 'question_encoder_tokenizer' ) SCREAMING_SNAKE_CASE_ = os.path.join(_lowerCAmelCase , 'generator_tokenizer' ) self.question_encoder.save_pretrained(_lowerCAmelCase ) self.generator.save_pretrained(_lowerCAmelCase ) @classmethod def lowerCAmelCase_ ( cls : Union[str, Any] , _lowerCAmelCase : Tuple , **_lowerCAmelCase : Optional[int] ): # dynamically import AutoTokenizer from ..auto.tokenization_auto import AutoTokenizer SCREAMING_SNAKE_CASE_ = kwargs.pop('config' , _lowerCAmelCase ) if config is None: SCREAMING_SNAKE_CASE_ = RagConfig.from_pretrained(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = AutoTokenizer.from_pretrained( _lowerCAmelCase , config=config.question_encoder , subfolder='question_encoder_tokenizer' ) SCREAMING_SNAKE_CASE_ = AutoTokenizer.from_pretrained( _lowerCAmelCase , config=config.generator , subfolder='generator_tokenizer' ) return cls(question_encoder=_lowerCAmelCase , generator=_lowerCAmelCase ) def __call__( self : int , *_lowerCAmelCase : List[Any] , **_lowerCAmelCase : Union[str, Any] ): return self.current_tokenizer(*_lowerCAmelCase , **_lowerCAmelCase ) def lowerCAmelCase_ ( self : Any , *_lowerCAmelCase : str , **_lowerCAmelCase : Dict ): return self.generator.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def lowerCAmelCase_ ( self : str , *_lowerCAmelCase : List[str] , **_lowerCAmelCase : Optional[int] ): return self.generator.decode(*_lowerCAmelCase , **_lowerCAmelCase ) def lowerCAmelCase_ ( self : int ): SCREAMING_SNAKE_CASE_ = self.question_encoder def lowerCAmelCase_ ( self : str ): SCREAMING_SNAKE_CASE_ = self.generator def lowerCAmelCase_ ( self : List[str] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[List[str]] = None , _lowerCAmelCase : Optional[int] = None , _lowerCAmelCase : Optional[int] = None , _lowerCAmelCase : str = "longest" , _lowerCAmelCase : str = None , _lowerCAmelCase : bool = True , **_lowerCAmelCase : Tuple , ): warnings.warn( '`prepare_seq2seq_batch` is deprecated and will be removed in version 5 of 🤗 Transformers. Use the ' 'regular `__call__` method to prepare your inputs and the tokenizer under the `with_target_tokenizer` ' 'context manager to prepare your targets. See the documentation of your specific tokenizer for more ' 'details' , _lowerCAmelCase , ) if max_length is None: SCREAMING_SNAKE_CASE_ = self.current_tokenizer.model_max_length SCREAMING_SNAKE_CASE_ = self( _lowerCAmelCase , add_special_tokens=_lowerCAmelCase , return_tensors=_lowerCAmelCase , max_length=_lowerCAmelCase , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , **_lowerCAmelCase , ) if tgt_texts is None: return model_inputs # Process tgt_texts if max_target_length is None: SCREAMING_SNAKE_CASE_ = self.current_tokenizer.model_max_length SCREAMING_SNAKE_CASE_ = self( text_target=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , return_tensors=_lowerCAmelCase , padding=_lowerCAmelCase , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase , **_lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ = labels['input_ids'] return model_inputs
225
import re from filelock import FileLock try: import nltk lowerCamelCase__ : str = True except (ImportError, ModuleNotFoundError): lowerCamelCase__ : Union[str, Any] = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def UpperCAmelCase_ ( __UpperCAmelCase : str ) -> str: re.sub('<n>' , '' , __UpperCAmelCase ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(__UpperCAmelCase ) )
225
1
A : Dict = {0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} A : List[Any] = {0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def a__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = [] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) order.append(__UpperCamelCase ) return order def a__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = [vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) return component def a__ ( __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = len(__UpperCamelCase ) * [False] SCREAMING_SNAKE_CASE_ = {vert: [] for vert in range(len(__UpperCamelCase ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(__UpperCamelCase ) SCREAMING_SNAKE_CASE_ = [] for i, was_visited in enumerate(__UpperCamelCase ): if not was_visited: order += topology_sort(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = len(__UpperCamelCase ) * [False] for i in range(len(__UpperCamelCase ) ): SCREAMING_SNAKE_CASE_ = order[len(__UpperCamelCase ) - i - 1] if not visited[vert]: SCREAMING_SNAKE_CASE_ = find_components(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) components_list.append(__UpperCamelCase ) return components_list
358
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BlipaProcessor, BlipImageProcessor, GPTaTokenizer, PreTrainedTokenizerFast @require_vision class lowerCamelCase (unittest.TestCase ): """simple docstring""" def __A ( self : int ) -> Any: SCREAMING_SNAKE_CASE_ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ = BlipImageProcessor() SCREAMING_SNAKE_CASE_ = GPTaTokenizer.from_pretrained("hf-internal-testing/tiny-random-GPT2Model" ) SCREAMING_SNAKE_CASE_ = BlipaProcessor(__magic_name__ , __magic_name__ ) processor.save_pretrained(self.tmpdirname ) def __A ( self : str , **__magic_name__ : int ) -> Union[str, Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **__magic_name__ ).tokenizer def __A ( self : Dict , **__magic_name__ : List[Any] ) -> int: return AutoProcessor.from_pretrained(self.tmpdirname , **__magic_name__ ).image_processor def __A ( self : int ) -> Any: shutil.rmtree(self.tmpdirname ) def __A ( self : Dict ) -> Dict: SCREAMING_SNAKE_CASE_ = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] SCREAMING_SNAKE_CASE_ = [Image.fromarray(np.moveaxis(__magic_name__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def __A ( self : List[Any] ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ = BlipaProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) SCREAMING_SNAKE_CASE_ = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) SCREAMING_SNAKE_CASE_ = self.get_image_processor(do_normalize=__magic_name__ , padding_value=1.0 ) SCREAMING_SNAKE_CASE_ = BlipaProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=__magic_name__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __magic_name__ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __magic_name__ ) def __A ( self : Tuple ) -> int: SCREAMING_SNAKE_CASE_ = self.get_image_processor() SCREAMING_SNAKE_CASE_ = self.get_tokenizer() SCREAMING_SNAKE_CASE_ = BlipaProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) SCREAMING_SNAKE_CASE_ = self.prepare_image_inputs() SCREAMING_SNAKE_CASE_ = image_processor(__magic_name__ , return_tensors="np" ) SCREAMING_SNAKE_CASE_ = processor(images=__magic_name__ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def __A ( self : str ) -> Tuple: SCREAMING_SNAKE_CASE_ = self.get_image_processor() SCREAMING_SNAKE_CASE_ = self.get_tokenizer() SCREAMING_SNAKE_CASE_ = BlipaProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) SCREAMING_SNAKE_CASE_ = "lower newer" SCREAMING_SNAKE_CASE_ = processor(text=__magic_name__ ) SCREAMING_SNAKE_CASE_ = tokenizer(__magic_name__ , return_token_type_ids=__magic_name__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __A ( self : Dict ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ = self.get_image_processor() SCREAMING_SNAKE_CASE_ = self.get_tokenizer() SCREAMING_SNAKE_CASE_ = BlipaProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) SCREAMING_SNAKE_CASE_ = "lower newer" SCREAMING_SNAKE_CASE_ = self.prepare_image_inputs() SCREAMING_SNAKE_CASE_ = processor(text=__magic_name__ , images=__magic_name__ ) self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "input_ids", "attention_mask"] ) # test if it raises when no input is passed with pytest.raises(__magic_name__ ): processor() def __A ( self : Dict ) -> Tuple: SCREAMING_SNAKE_CASE_ = self.get_image_processor() SCREAMING_SNAKE_CASE_ = self.get_tokenizer() SCREAMING_SNAKE_CASE_ = BlipaProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) SCREAMING_SNAKE_CASE_ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] SCREAMING_SNAKE_CASE_ = processor.batch_decode(__magic_name__ ) SCREAMING_SNAKE_CASE_ = tokenizer.batch_decode(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) def __A ( self : List[str] ) -> int: SCREAMING_SNAKE_CASE_ = self.get_image_processor() SCREAMING_SNAKE_CASE_ = self.get_tokenizer() SCREAMING_SNAKE_CASE_ = BlipaProcessor(tokenizer=__magic_name__ , image_processor=__magic_name__ ) SCREAMING_SNAKE_CASE_ = "lower newer" SCREAMING_SNAKE_CASE_ = self.prepare_image_inputs() SCREAMING_SNAKE_CASE_ = processor(text=__magic_name__ , images=__magic_name__ ) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "input_ids", "attention_mask"] )
305
0
'''simple docstring''' import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler SCREAMING_SNAKE_CASE_: Tuple =16 SCREAMING_SNAKE_CASE_: Tuple =32 def lowerCAmelCase_ ( snake_case_ : Accelerator , snake_case_ : int = 16 , snake_case_ : str = "bert-base-cased" ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = AutoTokenizer.from_pretrained(snake_case_ ) UpperCAmelCase_ = load_dataset("glue" , "mrpc" ) def tokenize_function(snake_case_ : Optional[Any] ): # max_length=None => use the model max length (it's actually the default) UpperCAmelCase_ = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=snake_case_ , max_length=snake_case_ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset UpperCAmelCase_ = datasets.map( snake_case_ , batched=snake_case_ , remove_columns=["idx", "sentence1", "sentence2"] , load_from_cache_file=snake_case_ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library UpperCAmelCase_ = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(snake_case_ : str ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(snake_case_ , padding="max_length" , max_length=1_28 , return_tensors="pt" ) return tokenizer.pad(snake_case_ , padding="longest" , return_tensors="pt" ) # Instantiate dataloaders. UpperCAmelCase_ = DataLoader( tokenized_datasets["train"] , shuffle=snake_case_ , collate_fn=snake_case_ , batch_size=snake_case_ ) UpperCAmelCase_ = DataLoader( tokenized_datasets["validation"] , shuffle=snake_case_ , collate_fn=snake_case_ , batch_size=snake_case_ ) return train_dataloader, eval_dataloader def lowerCAmelCase_ ( snake_case_ : int , snake_case_ : Union[str, Any] , snake_case_ : Optional[int] , snake_case_ : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' model.eval() UpperCAmelCase_ = 0 for step, batch in enumerate(snake_case_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): UpperCAmelCase_ = model(**snake_case_ ) UpperCAmelCase_ = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times UpperCAmelCase_ , UpperCAmelCase_ = accelerator.gather( (predictions, batch["labels"]) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(snake_case_ ) - 1: UpperCAmelCase_ = predictions[: len(eval_dataloader.dataset ) - samples_seen] UpperCAmelCase_ = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=snake_case_ , references=snake_case_ , ) UpperCAmelCase_ = metric.compute() return eval_metric["accuracy"] def lowerCAmelCase_ ( snake_case_ : Tuple , snake_case_ : Optional[int] ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs UpperCAmelCase_ = config["lr"] UpperCAmelCase_ = int(config["num_epochs"] ) UpperCAmelCase_ = int(config["seed"] ) UpperCAmelCase_ = int(config["batch_size"] ) UpperCAmelCase_ = args.model_name_or_path set_seed(snake_case_ ) UpperCAmelCase_ , UpperCAmelCase_ = get_dataloaders(snake_case_ , snake_case_ , snake_case_ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) UpperCAmelCase_ = AutoModelForSequenceClassification.from_pretrained(snake_case_ , return_dict=snake_case_ ) # Instantiate optimizer UpperCAmelCase_ = ( AdamW if accelerator.state.deepspeed_plugin is None or "optimizer" not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) UpperCAmelCase_ = optimizer_cls(params=model.parameters() , lr=snake_case_ ) if accelerator.state.deepspeed_plugin is not None: UpperCAmelCase_ = accelerator.state.deepspeed_plugin.deepspeed_config[ "gradient_accumulation_steps" ] else: UpperCAmelCase_ = 1 UpperCAmelCase_ = (len(snake_case_ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): UpperCAmelCase_ = get_linear_schedule_with_warmup( optimizer=snake_case_ , num_warmup_steps=0 , num_training_steps=snake_case_ , ) else: UpperCAmelCase_ = DummyScheduler(snake_case_ , total_num_steps=snake_case_ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = accelerator.prepare( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # We need to keep track of how many total steps we have iterated over UpperCAmelCase_ = 0 # We also need to keep track of the stating epoch so files are named properly UpperCAmelCase_ = 0 UpperCAmelCase_ = evaluate.load("glue" , "mrpc" ) UpperCAmelCase_ = num_epochs if args.partial_train_epoch is not None: UpperCAmelCase_ = args.partial_train_epoch if args.resume_from_checkpoint: accelerator.load_state(args.resume_from_checkpoint ) UpperCAmelCase_ = args.resume_from_checkpoint.split("epoch_" )[1] UpperCAmelCase_ = "" for char in epoch_string: if char.isdigit(): state_epoch_num += char else: break UpperCAmelCase_ = int(snake_case_ ) + 1 UpperCAmelCase_ = evaluation_loop(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) accelerator.print("resumed checkpoint performance:" , snake_case_ ) accelerator.print("resumed checkpoint's scheduler's lr:" , lr_scheduler.get_lr()[0] ) accelerator.print("resumed optimizers's lr:" , optimizer.param_groups[0]["lr"] ) with open(os.path.join(args.output_dir , f"""state_{starting_epoch-1}.json""" ) , "r" ) as f: UpperCAmelCase_ = json.load(snake_case_ ) assert resumed_state["accuracy"] == accuracy, "Accuracy mismatch, loading from checkpoint failed" assert ( resumed_state["lr"] == lr_scheduler.get_lr()[0] ), "Scheduler learning rate mismatch, loading from checkpoint failed" assert ( resumed_state["optimizer_lr"] == optimizer.param_groups[0]["lr"] ), "Optimizer learning rate mismatch, loading from checkpoint failed" assert resumed_state["epoch"] == starting_epoch - 1, "Epoch mismatch, loading from checkpoint failed" return # Now we train the model UpperCAmelCase_ = {} for epoch in range(snake_case_ , snake_case_ ): model.train() for step, batch in enumerate(snake_case_ ): UpperCAmelCase_ = model(**snake_case_ ) UpperCAmelCase_ = outputs.loss UpperCAmelCase_ = loss / gradient_accumulation_steps accelerator.backward(snake_case_ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 UpperCAmelCase_ = f"""epoch_{epoch}""" UpperCAmelCase_ = os.path.join(args.output_dir , snake_case_ ) accelerator.save_state(snake_case_ ) UpperCAmelCase_ = evaluation_loop(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) UpperCAmelCase_ = accuracy UpperCAmelCase_ = lr_scheduler.get_lr()[0] UpperCAmelCase_ = optimizer.param_groups[0]["lr"] UpperCAmelCase_ = epoch UpperCAmelCase_ = overall_step accelerator.print(f"""epoch {epoch}:""" , snake_case_ ) accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , f"""state_{epoch}.json""" ) , "w" ) as f: json.dump(snake_case_ , snake_case_ ) def lowerCAmelCase_ ( ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = argparse.ArgumentParser(description="Simple example of training script tracking peak GPU memory usage." ) parser.add_argument( "--model_name_or_path" , type=snake_case_ , default="bert-base-cased" , help="Path to pretrained model or model identifier from huggingface.co/models." , required=snake_case_ , ) parser.add_argument( "--output_dir" , type=snake_case_ , default="." , help="Optional save directory where all checkpoint folders will be stored. Default is the current working directory." , ) parser.add_argument( "--resume_from_checkpoint" , type=snake_case_ , default=snake_case_ , help="If the training should continue from a checkpoint folder." , ) parser.add_argument( "--partial_train_epoch" , type=snake_case_ , default=snake_case_ , help="If passed, the training will stop after this number of epochs." , ) parser.add_argument( "--num_epochs" , type=snake_case_ , default=2 , help="Number of train epochs." , ) UpperCAmelCase_ = parser.parse_args() UpperCAmelCase_ = {"lr": 2E-5, "num_epochs": args.num_epochs, "seed": 42, "batch_size": 16} training_function(snake_case_ , snake_case_ ) if __name__ == "__main__": main()
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class _SCREAMING_SNAKE_CASE ( unittest.TestCase): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=7 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=18 , _SCREAMING_SNAKE_CASE=30 , _SCREAMING_SNAKE_CASE=400 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=True , )-> Union[str, Any]: lowerCamelCase_ =size if size is not None else {"""shortest_edge""": 20} lowerCamelCase_ =crop_size if crop_size is not None else {"""height""": 18, """width""": 18} lowerCamelCase_ =parent lowerCamelCase_ =batch_size lowerCamelCase_ =num_channels lowerCamelCase_ =image_size lowerCamelCase_ =min_resolution lowerCamelCase_ =max_resolution lowerCamelCase_ =do_resize lowerCamelCase_ =size lowerCamelCase_ =do_center_crop lowerCamelCase_ =crop_size lowerCamelCase_ =do_flip_channel_order def _snake_case ( self )-> List[Any]: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class _SCREAMING_SNAKE_CASE ( lowerCAmelCase__ , unittest.TestCase): _UpperCamelCase:Tuple = MobileViTImageProcessor if is_vision_available() else None def _snake_case ( self )-> List[str]: lowerCamelCase_ =MobileViTImageProcessingTester(self ) @property def _snake_case ( self )-> Optional[int]: return self.image_processor_tester.prepare_image_processor_dict() def _snake_case ( self )-> Any: lowerCamelCase_ =self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , """do_resize""" ) ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , """size""" ) ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , """do_center_crop""" ) ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , """center_crop""" ) ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , """do_flip_channel_order""" ) ) def _snake_case ( self )-> Optional[Any]: lowerCamelCase_ =self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 20} ) self.assertEqual(image_processor.crop_size , {"""height""": 18, """width""": 18} ) lowerCamelCase_ =self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) def _snake_case ( self )-> Union[str, Any]: pass def _snake_case ( self )-> Dict: # Initialize image_processing lowerCamelCase_ =self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase_ =prepare_image_inputs(self.image_processor_tester , equal_resolution=_SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(_SCREAMING_SNAKE_CASE , Image.Image ) # Test not batched input lowerCamelCase_ =image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched lowerCamelCase_ =image_processing(_SCREAMING_SNAKE_CASE , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def _snake_case ( self )-> str: # Initialize image_processing lowerCamelCase_ =self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase_ =prepare_image_inputs(self.image_processor_tester , equal_resolution=_SCREAMING_SNAKE_CASE , numpify=_SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(_SCREAMING_SNAKE_CASE , np.ndarray ) # Test not batched input lowerCamelCase_ =image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched lowerCamelCase_ =image_processing(_SCREAMING_SNAKE_CASE , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def _snake_case ( self )-> List[Any]: # Initialize image_processing lowerCamelCase_ =self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase_ =prepare_image_inputs(self.image_processor_tester , equal_resolution=_SCREAMING_SNAKE_CASE , torchify=_SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(_SCREAMING_SNAKE_CASE , torch.Tensor ) # Test not batched input lowerCamelCase_ =image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched lowerCamelCase_ =image_processing(_SCREAMING_SNAKE_CASE , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , )
154
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __A = {"configuration_swin": ["SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP", "SwinConfig", "SwinOnnxConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "SWIN_PRETRAINED_MODEL_ARCHIVE_LIST", "SwinForImageClassification", "SwinForMaskedImageModeling", "SwinModel", "SwinPreTrainedModel", "SwinBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST", "TFSwinForImageClassification", "TFSwinForMaskedImageModeling", "TFSwinModel", "TFSwinPreTrainedModel", ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
273
import shutil import tempfile import unittest from unittest.mock import patch from transformers import ( DefaultFlowCallback, IntervalStrategy, PrinterCallback, ProgressCallback, Trainer, TrainerCallback, TrainingArguments, is_torch_available, ) from transformers.testing_utils import require_torch if is_torch_available(): from transformers.trainer import DEFAULT_CALLBACKS from .test_trainer import RegressionDataset, RegressionModelConfig, RegressionPreTrainedModel class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__(self : str) ->List[Any]: '''simple docstring''' lowerCamelCase__: List[Any] =[] def SCREAMING_SNAKE_CASE_ (self : Any , UpperCAmelCase_ : int , UpperCAmelCase_ : str , UpperCAmelCase_ : Dict , **UpperCAmelCase_ : Any) ->Dict: '''simple docstring''' self.events.append("on_init_end") def SCREAMING_SNAKE_CASE_ (self : Union[str, Any] , UpperCAmelCase_ : Any , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Dict , **UpperCAmelCase_ : str) ->List[str]: '''simple docstring''' self.events.append("on_train_begin") def SCREAMING_SNAKE_CASE_ (self : List[str] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : List[Any] , **UpperCAmelCase_ : str) ->int: '''simple docstring''' self.events.append("on_train_end") def SCREAMING_SNAKE_CASE_ (self : Tuple , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Any , UpperCAmelCase_ : List[str] , **UpperCAmelCase_ : Optional[int]) ->List[Any]: '''simple docstring''' self.events.append("on_epoch_begin") def SCREAMING_SNAKE_CASE_ (self : Optional[int] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : int , **UpperCAmelCase_ : Any) ->Tuple: '''simple docstring''' self.events.append("on_epoch_end") def SCREAMING_SNAKE_CASE_ (self : List[Any] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Any , UpperCAmelCase_ : Any , **UpperCAmelCase_ : List[Any]) ->Optional[int]: '''simple docstring''' self.events.append("on_step_begin") def SCREAMING_SNAKE_CASE_ (self : List[str] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Optional[int] , **UpperCAmelCase_ : List[str]) ->Tuple: '''simple docstring''' self.events.append("on_step_end") def SCREAMING_SNAKE_CASE_ (self : Union[str, Any] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[Any] , **UpperCAmelCase_ : str) ->Optional[int]: '''simple docstring''' self.events.append("on_evaluate") def SCREAMING_SNAKE_CASE_ (self : Tuple , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Union[str, Any] , **UpperCAmelCase_ : Any) ->int: '''simple docstring''' self.events.append("on_predict") def SCREAMING_SNAKE_CASE_ (self : Union[str, Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Tuple , **UpperCAmelCase_ : List[Any]) ->Any: '''simple docstring''' self.events.append("on_save") def SCREAMING_SNAKE_CASE_ (self : str , UpperCAmelCase_ : Any , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Any , **UpperCAmelCase_ : Optional[Any]) ->str: '''simple docstring''' self.events.append("on_log") def SCREAMING_SNAKE_CASE_ (self : Optional[Any] , UpperCAmelCase_ : Any , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : str , **UpperCAmelCase_ : List[str]) ->Optional[int]: '''simple docstring''' self.events.append("on_prediction_step") @require_torch class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE_ (self : List[Any]) ->int: '''simple docstring''' lowerCamelCase__: Tuple =tempfile.mkdtemp() def SCREAMING_SNAKE_CASE_ (self : List[str]) ->Tuple: '''simple docstring''' shutil.rmtree(self.output_dir) def SCREAMING_SNAKE_CASE_ (self : List[Any] , UpperCAmelCase_ : int=0 , UpperCAmelCase_ : Tuple=0 , UpperCAmelCase_ : str=64 , UpperCAmelCase_ : List[Any]=64 , UpperCAmelCase_ : Union[str, Any]=None , UpperCAmelCase_ : List[str]=False , **UpperCAmelCase_ : Tuple) ->Any: '''simple docstring''' lowerCamelCase__: Union[str, Any] =RegressionDataset(length=UpperCAmelCase_) lowerCamelCase__: int =RegressionDataset(length=UpperCAmelCase_) lowerCamelCase__: str =RegressionModelConfig(a=UpperCAmelCase_ , b=UpperCAmelCase_) lowerCamelCase__: Union[str, Any] =RegressionPreTrainedModel(UpperCAmelCase_) lowerCamelCase__: int =TrainingArguments(self.output_dir , disable_tqdm=UpperCAmelCase_ , report_to=[] , **UpperCAmelCase_) return Trainer( UpperCAmelCase_ , UpperCAmelCase_ , train_dataset=UpperCAmelCase_ , eval_dataset=UpperCAmelCase_ , callbacks=UpperCAmelCase_ , ) def SCREAMING_SNAKE_CASE_ (self : List[str] , UpperCAmelCase_ : str , UpperCAmelCase_ : List[str]) ->Dict: '''simple docstring''' self.assertEqual(len(UpperCAmelCase_) , len(UpperCAmelCase_)) # Order doesn't matter lowerCamelCase__: Dict =sorted(UpperCAmelCase_ , key=lambda UpperCAmelCase_: cb.__name__ if isinstance(UpperCAmelCase_ , UpperCAmelCase_) else cb.__class__.__name__) lowerCamelCase__: Optional[int] =sorted(UpperCAmelCase_ , key=lambda UpperCAmelCase_: cb.__name__ if isinstance(UpperCAmelCase_ , UpperCAmelCase_) else cb.__class__.__name__) for cba, cba in zip(UpperCAmelCase_ , UpperCAmelCase_): if isinstance(UpperCAmelCase_ , UpperCAmelCase_) and isinstance(UpperCAmelCase_ , UpperCAmelCase_): self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_) elif isinstance(UpperCAmelCase_ , UpperCAmelCase_) and not isinstance(UpperCAmelCase_ , UpperCAmelCase_): self.assertEqual(UpperCAmelCase_ , cba.__class__) elif not isinstance(UpperCAmelCase_ , UpperCAmelCase_) and isinstance(UpperCAmelCase_ , UpperCAmelCase_): self.assertEqual(cba.__class__ , UpperCAmelCase_) else: self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : List[Any] , UpperCAmelCase_ : Optional[Any]) ->Optional[int]: '''simple docstring''' lowerCamelCase__: Dict =["on_init_end", "on_train_begin"] lowerCamelCase__: List[str] =0 lowerCamelCase__: List[Any] =len(trainer.get_eval_dataloader()) lowerCamelCase__: Dict =["on_prediction_step"] * len(trainer.get_eval_dataloader()) + ["on_log", "on_evaluate"] for _ in range(trainer.state.num_train_epochs): expected_events.append("on_epoch_begin") for _ in range(UpperCAmelCase_): step += 1 expected_events += ["on_step_begin", "on_step_end"] if step % trainer.args.logging_steps == 0: expected_events.append("on_log") if trainer.args.evaluation_strategy == IntervalStrategy.STEPS and step % trainer.args.eval_steps == 0: expected_events += evaluation_events.copy() if step % trainer.args.save_steps == 0: expected_events.append("on_save") expected_events.append("on_epoch_end") if trainer.args.evaluation_strategy == IntervalStrategy.EPOCH: expected_events += evaluation_events.copy() expected_events += ["on_log", "on_train_end"] return expected_events def SCREAMING_SNAKE_CASE_ (self : int) ->Optional[int]: '''simple docstring''' lowerCamelCase__: Optional[Any] =self.get_trainer() lowerCamelCase__: Any =DEFAULT_CALLBACKS.copy() + [ProgressCallback] self.check_callbacks_equality(trainer.callback_handler.callbacks , UpperCAmelCase_) # Callbacks passed at init are added to the default callbacks lowerCamelCase__: Dict =self.get_trainer(callbacks=[MyTestTrainerCallback]) expected_callbacks.append(UpperCAmelCase_) self.check_callbacks_equality(trainer.callback_handler.callbacks , UpperCAmelCase_) # TrainingArguments.disable_tqdm controls if use ProgressCallback or PrinterCallback lowerCamelCase__: int =self.get_trainer(disable_tqdm=UpperCAmelCase_) lowerCamelCase__: Tuple =DEFAULT_CALLBACKS.copy() + [PrinterCallback] self.check_callbacks_equality(trainer.callback_handler.callbacks , UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : List[str]) ->Union[str, Any]: '''simple docstring''' lowerCamelCase__: Union[str, Any] =DEFAULT_CALLBACKS.copy() + [ProgressCallback] lowerCamelCase__: Optional[int] =self.get_trainer() # We can add, pop, or remove by class name trainer.remove_callback(UpperCAmelCase_) expected_callbacks.remove(UpperCAmelCase_) self.check_callbacks_equality(trainer.callback_handler.callbacks , UpperCAmelCase_) lowerCamelCase__: Dict =self.get_trainer() lowerCamelCase__: str =trainer.pop_callback(UpperCAmelCase_) self.assertEqual(cb.__class__ , UpperCAmelCase_) self.check_callbacks_equality(trainer.callback_handler.callbacks , UpperCAmelCase_) trainer.add_callback(UpperCAmelCase_) expected_callbacks.insert(0 , UpperCAmelCase_) self.check_callbacks_equality(trainer.callback_handler.callbacks , UpperCAmelCase_) # We can also add, pop, or remove by instance lowerCamelCase__: List[str] =self.get_trainer() lowerCamelCase__: List[str] =trainer.callback_handler.callbacks[0] trainer.remove_callback(UpperCAmelCase_) expected_callbacks.remove(UpperCAmelCase_) self.check_callbacks_equality(trainer.callback_handler.callbacks , UpperCAmelCase_) lowerCamelCase__: str =self.get_trainer() lowerCamelCase__: List[Any] =trainer.callback_handler.callbacks[0] lowerCamelCase__: Dict =trainer.pop_callback(UpperCAmelCase_) self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_) self.check_callbacks_equality(trainer.callback_handler.callbacks , UpperCAmelCase_) trainer.add_callback(UpperCAmelCase_) expected_callbacks.insert(0 , UpperCAmelCase_) self.check_callbacks_equality(trainer.callback_handler.callbacks , UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : List[Any]) ->int: '''simple docstring''' import warnings # XXX: for now ignore scatter_gather warnings in this test since it's not relevant to what's being tested warnings.simplefilter(action="ignore" , category=UpperCAmelCase_) lowerCamelCase__: Union[str, Any] =self.get_trainer(callbacks=[MyTestTrainerCallback]) trainer.train() lowerCamelCase__: int =trainer.callback_handler.callbacks[-2].events self.assertEqual(UpperCAmelCase_ , self.get_expected_events(UpperCAmelCase_)) # Independent log/save/eval lowerCamelCase__: Dict =self.get_trainer(callbacks=[MyTestTrainerCallback] , logging_steps=5) trainer.train() lowerCamelCase__: Optional[int] =trainer.callback_handler.callbacks[-2].events self.assertEqual(UpperCAmelCase_ , self.get_expected_events(UpperCAmelCase_)) lowerCamelCase__: Any =self.get_trainer(callbacks=[MyTestTrainerCallback] , save_steps=5) trainer.train() lowerCamelCase__: List[Any] =trainer.callback_handler.callbacks[-2].events self.assertEqual(UpperCAmelCase_ , self.get_expected_events(UpperCAmelCase_)) lowerCamelCase__: int =self.get_trainer(callbacks=[MyTestTrainerCallback] , eval_steps=5 , evaluation_strategy="steps") trainer.train() lowerCamelCase__: str =trainer.callback_handler.callbacks[-2].events self.assertEqual(UpperCAmelCase_ , self.get_expected_events(UpperCAmelCase_)) lowerCamelCase__: Dict =self.get_trainer(callbacks=[MyTestTrainerCallback] , evaluation_strategy="epoch") trainer.train() lowerCamelCase__: Tuple =trainer.callback_handler.callbacks[-2].events self.assertEqual(UpperCAmelCase_ , self.get_expected_events(UpperCAmelCase_)) # A bit of everything lowerCamelCase__: Tuple =self.get_trainer( callbacks=[MyTestTrainerCallback] , logging_steps=3 , save_steps=10 , eval_steps=5 , evaluation_strategy="steps" , ) trainer.train() lowerCamelCase__: int =trainer.callback_handler.callbacks[-2].events self.assertEqual(UpperCAmelCase_ , self.get_expected_events(UpperCAmelCase_)) # warning should be emitted for duplicated callbacks with patch("transformers.trainer_callback.logger.warning") as warn_mock: lowerCamelCase__: Optional[int] =self.get_trainer( callbacks=[MyTestTrainerCallback, MyTestTrainerCallback] , ) assert str(UpperCAmelCase_) in warn_mock.call_args[0][0]
273
1