code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
class A_ : def __init__( self : List[str] , UpperCAmelCase : Optional[int] ) -> Union[str, Any]: __lowerCAmelCase: Tuple = size __lowerCAmelCase: Union[str, Any] = [0] * size __lowerCAmelCase: int = [0] * size @staticmethod def UpperCAmelCase ( UpperCAmelCase : Any ) -> List[str]: return index | (index + 1) @staticmethod def UpperCAmelCase ( UpperCAmelCase : Optional[Any] ) -> Optional[int]: return (index & (index + 1)) - 1 def UpperCAmelCase ( self : Optional[Any] , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : str ) -> Tuple: __lowerCAmelCase: str = value while index < self.size: __lowerCAmelCase: Optional[int] = self.get_prev(_a ) + 1 if current_left_border == index: __lowerCAmelCase: str = value else: __lowerCAmelCase: Optional[Any] = max(_a , _a , _a ) __lowerCAmelCase: int = self.get_next(_a ) def UpperCAmelCase ( self : Tuple , UpperCAmelCase : Tuple , UpperCAmelCase : List[Any] ) -> List[Any]: right -= 1 # Because of right is exclusive __lowerCAmelCase: Any = 0 while left <= right: __lowerCAmelCase: List[Any] = self.get_prev(_a ) if left <= current_left: __lowerCAmelCase: Tuple = max(_a , self.tree[right] ) __lowerCAmelCase: int = current_left else: __lowerCAmelCase: Optional[Any] = max(_a , self.arr[right] ) right -= 1 return result if __name__ == "__main__": import doctest doctest.testmod()
322
import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipaConfig, BlipaForConditionalGeneration, BlipaProcessor, BlipaVisionConfig, BlipImageProcessor, OPTConfig, TaConfig, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def lowerCAmelCase_ ( ) -> str: '''simple docstring''' __magic_name__ : int = "https://storage.googleapis.com/sfr-vision-language-research/LAVIS/assets/merlion.png" __magic_name__ : Union[str, Any] = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ).convert("RGB" ) return image def lowerCAmelCase_ ( _snake_case : str ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : List[str] = [] # fmt: off # vision encoder rename_keys.append(("visual_encoder.cls_token", "vision_model.embeddings.class_embedding") ) rename_keys.append(("visual_encoder.pos_embed", "vision_model.embeddings.position_embedding") ) rename_keys.append(("visual_encoder.patch_embed.proj.weight", "vision_model.embeddings.patch_embedding.weight") ) rename_keys.append(("visual_encoder.patch_embed.proj.bias", "vision_model.embeddings.patch_embedding.bias") ) rename_keys.append(("ln_vision.weight", "vision_model.post_layernorm.weight") ) rename_keys.append(("ln_vision.bias", "vision_model.post_layernorm.bias") ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((F'''visual_encoder.blocks.{i}.norm1.weight''', F'''vision_model.encoder.layers.{i}.layer_norm1.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm1.bias''', F'''vision_model.encoder.layers.{i}.layer_norm1.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm2.weight''', F'''vision_model.encoder.layers.{i}.layer_norm2.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm2.bias''', F'''vision_model.encoder.layers.{i}.layer_norm2.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.qkv.weight''', F'''vision_model.encoder.layers.{i}.self_attn.qkv.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.proj.weight''', F'''vision_model.encoder.layers.{i}.self_attn.projection.weight''',) ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.proj.bias''', F'''vision_model.encoder.layers.{i}.self_attn.projection.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc1.weight''', F'''vision_model.encoder.layers.{i}.mlp.fc1.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc1.bias''', F'''vision_model.encoder.layers.{i}.mlp.fc1.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc2.weight''', F'''vision_model.encoder.layers.{i}.mlp.fc2.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc2.bias''', F'''vision_model.encoder.layers.{i}.mlp.fc2.bias''') ) # QFormer rename_keys.append(("Qformer.bert.embeddings.LayerNorm.weight", "qformer.layernorm.weight") ) rename_keys.append(("Qformer.bert.embeddings.LayerNorm.bias", "qformer.layernorm.bias") ) # fmt: on return rename_keys def lowerCAmelCase_ ( _snake_case : Any , _snake_case : Union[str, Any] , _snake_case : Optional[Any] ) -> int: '''simple docstring''' __magic_name__ : Tuple = dct.pop(_snake_case ) __magic_name__ : int = val def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : Optional[Any] ) -> Dict: '''simple docstring''' for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases __magic_name__ : List[Any] = state_dict.pop(F'''visual_encoder.blocks.{i}.attn.q_bias''' ) __magic_name__ : Optional[Any] = state_dict.pop(F'''visual_encoder.blocks.{i}.attn.v_bias''' ) # next, set bias in the state dict __magic_name__ : Optional[int] = torch.cat((q_bias, torch.zeros_like(_snake_case , requires_grad=_snake_case ), v_bias) ) __magic_name__ : Union[str, Any] = qkv_bias def lowerCAmelCase_ ( _snake_case : Dict , _snake_case : str ) -> int: '''simple docstring''' __magic_name__ : List[Any] = 364 if "coco" in model_name else 224 __magic_name__ : Union[str, Any] = BlipaVisionConfig(image_size=_snake_case ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "opt-2.7b" in model_name: __magic_name__ : List[str] = OPTConfig.from_pretrained("facebook/opt-2.7b" , eos_token_id=_snake_case ).to_dict() elif "opt-6.7b" in model_name: __magic_name__ : Any = OPTConfig.from_pretrained("facebook/opt-6.7b" , eos_token_id=_snake_case ).to_dict() elif "t5-xl" in model_name: __magic_name__ : Dict = TaConfig.from_pretrained("google/flan-t5-xl" , dense_act_fn="gelu" , bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: __magic_name__ : int = TaConfig.from_pretrained("google/flan-t5-xxl" , dense_act_fn="gelu" , bos_token_id=1 ).to_dict() __magic_name__ : List[Any] = BlipaConfig(vision_config=_snake_case , text_config=_snake_case ) return config, image_size @torch.no_grad() def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : str=None , _snake_case : Dict=False ) -> List[Any]: '''simple docstring''' __magic_name__ : Optional[int] = ( AutoTokenizer.from_pretrained("facebook/opt-2.7b" ) if "opt" in model_name else AutoTokenizer.from_pretrained("google/flan-t5-xl" ) ) __magic_name__ : List[Any] = tokenizer("\n" , add_special_tokens=_snake_case ).input_ids[0] __magic_name__ , __magic_name__ : Tuple = get_blipa_config(_snake_case , eos_token_id=_snake_case ) __magic_name__ : Union[str, Any] = BlipaForConditionalGeneration(_snake_case ).eval() __magic_name__ : Any = { "blip2-opt-2.7b": ("blip2_opt", "pretrain_opt2.7b"), "blip2-opt-6.7b": ("blip2_opt", "pretrain_opt6.7b"), "blip2-opt-2.7b-coco": ("blip2_opt", "caption_coco_opt2.7b"), "blip2-opt-6.7b-coco": ("blip2_opt", "caption_coco_opt6.7b"), "blip2-flan-t5-xl": ("blip2_t5", "pretrain_flant5xl"), "blip2-flan-t5-xl-coco": ("blip2_t5", "caption_coco_flant5xl"), "blip2-flan-t5-xxl": ("blip2_t5", "pretrain_flant5xxl"), } __magic_name__ , __magic_name__ : Union[str, Any] = model_name_to_original[model_name] # load original model print("Loading original model..." ) __magic_name__ : Union[str, Any] = "cuda" if torch.cuda.is_available() else "cpu" __magic_name__ , __magic_name__ , __magic_name__ : Optional[Any] = load_model_and_preprocess( name=_snake_case , model_type=_snake_case , is_eval=_snake_case , device=_snake_case ) original_model.eval() print("Done!" ) # update state dict keys __magic_name__ : Dict = original_model.state_dict() __magic_name__ : str = create_rename_keys(_snake_case ) for src, dest in rename_keys: rename_key(_snake_case , _snake_case , _snake_case ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): __magic_name__ : Any = state_dict.pop(_snake_case ) if key.startswith("Qformer.bert" ): __magic_name__ : Optional[int] = key.replace("Qformer.bert" , "qformer" ) if "attention.self" in key: __magic_name__ : Any = key.replace("self" , "attention" ) if "opt_proj" in key: __magic_name__ : Union[str, Any] = key.replace("opt_proj" , "language_projection" ) if "t5_proj" in key: __magic_name__ : Optional[int] = key.replace("t5_proj" , "language_projection" ) if key.startswith("opt" ): __magic_name__ : List[str] = key.replace("opt" , "language" ) if key.startswith("t5" ): __magic_name__ : Tuple = key.replace("t5" , "language" ) __magic_name__ : Dict = val # read in qv biases read_in_q_v_bias(_snake_case , _snake_case ) __magic_name__ , __magic_name__ : Tuple = hf_model.load_state_dict(_snake_case , strict=_snake_case ) assert len(_snake_case ) == 0 assert unexpected_keys == ["qformer.embeddings.position_ids"] __magic_name__ : List[Any] = load_demo_image() __magic_name__ : Tuple = vis_processors["eval"](_snake_case ).unsqueeze(0 ).to(_snake_case ) __magic_name__ : Dict = tokenizer(["\n"] , return_tensors="pt" ).input_ids.to(_snake_case ) # create processor __magic_name__ : Optional[Any] = BlipImageProcessor( size={"height": image_size, "width": image_size} , image_mean=_snake_case , image_std=_snake_case ) __magic_name__ : Dict = BlipaProcessor(image_processor=_snake_case , tokenizer=_snake_case ) __magic_name__ : Union[str, Any] = processor(images=_snake_case , return_tensors="pt" ).pixel_values.to(_snake_case ) # make sure processor creates exact same pixel values assert torch.allclose(_snake_case , _snake_case ) original_model.to(_snake_case ) hf_model.to(_snake_case ) with torch.no_grad(): if "opt" in model_name: __magic_name__ : List[Any] = original_model({"image": original_pixel_values, "text_input": [""]} ).logits __magic_name__ : Optional[int] = hf_model(_snake_case , _snake_case ).logits else: __magic_name__ : int = original_model( {"image": original_pixel_values, "text_input": ["\n"], "text_output": ["\n"]} ).logits __magic_name__ : Tuple = input_ids.masked_fill(input_ids == tokenizer.pad_token_id , -100 ) __magic_name__ : List[str] = hf_model(_snake_case , _snake_case , labels=_snake_case ).logits assert original_logits.shape == logits.shape print("First values of original logits:" , original_logits[0, :3, :3] ) print("First values of HF logits:" , logits[0, :3, :3] ) # assert values if model_name == "blip2-flan-t5-xl": __magic_name__ : List[str] = torch.tensor( [[-41.5_850, -4.4_440, -8.9_922], [-47.4_322, -5.9_143, -1.7_340]] , device=_snake_case ) assert torch.allclose(logits[0, :3, :3] , _snake_case , atol=1E-4 ) elif model_name == "blip2-flan-t5-xl-coco": __magic_name__ : Tuple = torch.tensor( [[-57.0_109, -9.8_967, -12.6_280], [-68.6_578, -12.7_191, -10.5_065]] , device=_snake_case ) else: # cast to same type __magic_name__ : str = logits.dtype assert torch.allclose(original_logits.to(_snake_case ) , _snake_case , atol=1E-2 ) print("Looks ok!" ) print("Generating a caption..." ) __magic_name__ : Optional[int] = "" __magic_name__ : Dict = tokenizer(_snake_case , return_tensors="pt" ).input_ids.to(_snake_case ) __magic_name__ : int = original_model.generate({"image": original_pixel_values} ) __magic_name__ : Optional[Any] = hf_model.generate( _snake_case , _snake_case , do_sample=_snake_case , num_beams=5 , max_length=30 , min_length=1 , top_p=0.9 , repetition_penalty=1.0 , length_penalty=1.0 , temperature=1 , ) print("Original generation:" , _snake_case ) __magic_name__ : Tuple = input_ids.shape[1] __magic_name__ : int = processor.batch_decode(outputs[:, prompt_length:] , skip_special_tokens=_snake_case ) __magic_name__ : Union[str, Any] = [text.strip() for text in output_text] print("HF generation:" , _snake_case ) if pytorch_dump_folder_path is not None: processor.save_pretrained(_snake_case ) hf_model.save_pretrained(_snake_case ) if push_to_hub: processor.push_to_hub(F'''nielsr/{model_name}''' ) hf_model.push_to_hub(F'''nielsr/{model_name}''' ) if __name__ == "__main__": snake_case : Any = argparse.ArgumentParser() snake_case : Union[str, Any] = [ "blip2-opt-2.7b", "blip2-opt-6.7b", "blip2-opt-2.7b-coco", "blip2-opt-6.7b-coco", "blip2-flan-t5-xl", "blip2-flan-t5-xl-coco", "blip2-flan-t5-xxl", ] parser.add_argument( "--model_name", default="blip2-opt-2.7b", choices=choices, type=str, help="Path to hf config.json of model to convert", ) parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model and processor to the hub after converting", ) snake_case : int = parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
281
0
from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class _snake_case ( _lowercase ): lowerCamelCase__: Any = "EncodecFeatureExtractor" lowerCamelCase__: Optional[int] = ("T5Tokenizer", "T5TokenizerFast") def __init__( self: Optional[int] , __lowerCamelCase: Optional[int] , __lowerCamelCase: str ) -> Optional[int]: super().__init__(_a , _a ) __UpperCAmelCase : int = self.feature_extractor __UpperCAmelCase : List[str] = False def _lowerCamelCase ( self: Any , __lowerCamelCase: Optional[Any]=None , __lowerCamelCase: List[str]=None , __lowerCamelCase: List[Any]=True ) -> List[str]: return self.tokenizer.get_decoder_prompt_ids(task=_a , language=_a , no_timestamps=_a ) def __call__( self: Dict , *__lowerCamelCase: Tuple , **__lowerCamelCase: Tuple ) -> Optional[Any]: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*_a , **_a ) __UpperCAmelCase : Union[str, Any] = kwargs.pop("audio" , _a ) __UpperCAmelCase : List[str] = kwargs.pop("sampling_rate" , _a ) __UpperCAmelCase : Tuple = kwargs.pop("text" , _a ) if len(_a ) > 0: __UpperCAmelCase : Union[str, Any] = args[0] __UpperCAmelCase : str = args[1:] if audio is None and text is None: raise ValueError("You need to specify either an `audio` or `text` input to process." ) if text is not None: __UpperCAmelCase : Dict = self.tokenizer(_a , **_a ) if audio is not None: __UpperCAmelCase : Any = self.feature_extractor(_a , *_a , sampling_rate=_a , **_a ) if audio is None: return inputs elif text is None: return audio_inputs else: __UpperCAmelCase : Union[str, Any] = audio_inputs["input_values"] if "padding_mask" in audio_inputs: __UpperCAmelCase : Optional[Any] = audio_inputs["padding_mask"] return inputs def _lowerCamelCase ( self: Any , *__lowerCamelCase: List[str] , **__lowerCamelCase: Union[str, Any] ) -> Dict: __UpperCAmelCase : Union[str, Any] = kwargs.pop("audio" , _a ) __UpperCAmelCase : List[Any] = kwargs.pop("padding_mask" , _a ) if len(_a ) > 0: __UpperCAmelCase : List[str] = args[0] __UpperCAmelCase : List[Any] = args[1:] if audio_values is not None: return self._decode_audio(_a , padding_mask=_a ) else: return self.tokenizer.batch_decode(*_a , **_a ) def _lowerCamelCase ( self: Optional[int] , *__lowerCamelCase: Tuple , **__lowerCamelCase: Optional[Any] ) -> Union[str, Any]: return self.tokenizer.decode(*_a , **_a ) def _lowerCamelCase ( self: List[Any] , __lowerCamelCase: List[str] , __lowerCamelCase: Union[str, Any] = None ) -> Any: __UpperCAmelCase : Optional[Any] = to_numpy(_a ) __UpperCAmelCase : Union[str, Any] = audio_values.shape if padding_mask is None: return list(_a ) __UpperCAmelCase : Tuple = to_numpy(_a ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) __UpperCAmelCase : str = seq_len - padding_mask.shape[-1] __UpperCAmelCase : str = 1 - self.feature_extractor.padding_value __UpperCAmelCase : List[Any] = np.pad(_a , ((0, 0), (0, difference)) , "constant" , constant_values=_a ) __UpperCAmelCase : Tuple = audio_values.tolist() for i in range(_a ): __UpperCAmelCase : Dict = np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] __UpperCAmelCase : List[str] = sliced_audio.reshape(_a , -1 ) return audio_values
157
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging snake_case : Dict = logging.get_logger(__name__) snake_case : Union[str, Any] = { "vocab_file": "vocab.txt", "merges_file": "bpe.codes", } snake_case : Dict = { "vocab_file": { "vinai/phobert-base": "https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt", "vinai/phobert-large": "https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt", }, "merges_file": { "vinai/phobert-base": "https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes", "vinai/phobert-large": "https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes", }, } snake_case : Union[str, Any] = { "vinai/phobert-base": 256, "vinai/phobert-large": 256, } def lowerCAmelCase_ ( _snake_case : str ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : List[str] = set() __magic_name__ : Any = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __magic_name__ : int = char __magic_name__ : List[str] = set(_snake_case ) return pairs class _snake_case ( snake_case ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , _a , _a , _a="<s>" , _a="</s>" , _a="</s>" , _a="<s>" , _a="<unk>" , _a="<pad>" , _a="<mask>" , **_a , ): super().__init__( bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , cls_token=_a , pad_token=_a , mask_token=_a , **_a , ) __magic_name__ : Dict = vocab_file __magic_name__ : Tuple = merges_file __magic_name__ : List[Any] = {} __magic_name__ : List[Any] = 0 __magic_name__ : Tuple = 1 __magic_name__ : int = 2 __magic_name__ : Union[str, Any] = 3 self.add_from_file(_a ) __magic_name__ : Optional[int] = {v: k for k, v in self.encoder.items()} with open(_a , encoding="utf-8" ) as merges_handle: __magic_name__ : List[str] = merges_handle.read().split("\n" )[:-1] __magic_name__ : Union[str, Any] = [tuple(merge.split()[:-1] ) for merge in merges] __magic_name__ : Union[str, Any] = dict(zip(_a , range(len(_a ) ) ) ) __magic_name__ : Optional[int] = {} def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __magic_name__ : Optional[Any] = [self.cls_token_id] __magic_name__ : Union[str, Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self , _a , _a = None , _a = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a , token_ids_a=_a , already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): __magic_name__ : Optional[Any] = [self.sep_token_id] __magic_name__ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def SCREAMING_SNAKE_CASE ( self ): return len(self.encoder ) def SCREAMING_SNAKE_CASE ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def SCREAMING_SNAKE_CASE ( self , _a ): if token in self.cache: return self.cache[token] __magic_name__ : List[Any] = tuple(_a ) __magic_name__ : List[Any] = tuple(list(word[:-1] ) + [word[-1] + "</w>"] ) __magic_name__ : Any = get_pairs(_a ) if not pairs: return token while True: __magic_name__ : str = min(_a , key=lambda _a : self.bpe_ranks.get(_a , float("inf" ) ) ) if bigram not in self.bpe_ranks: break __magic_name__ , __magic_name__ : List[str] = bigram __magic_name__ : List[str] = [] __magic_name__ : List[str] = 0 while i < len(_a ): try: __magic_name__ : Any = word.index(_a , _a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) __magic_name__ : Tuple = j if word[i] == first and i < len(_a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __magic_name__ : Union[str, Any] = tuple(_a ) __magic_name__ : Optional[int] = new_word if len(_a ) == 1: break else: __magic_name__ : List[Any] = get_pairs(_a ) __magic_name__ : Optional[int] = "@@ ".join(_a ) __magic_name__ : Tuple = word[:-4] __magic_name__ : str = word return word def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Optional[Any] = [] __magic_name__ : Dict = re.findall(r"\S+\n?" , _a ) for token in words: split_tokens.extend(list(self.bpe(_a ).split(" " ) ) ) return split_tokens def SCREAMING_SNAKE_CASE ( self , _a ): return self.encoder.get(_a , self.encoder.get(self.unk_token ) ) def SCREAMING_SNAKE_CASE ( self , _a ): return self.decoder.get(_a , self.unk_token ) def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Tuple = " ".join(_a ).replace("@@ " , "" ).strip() return out_string def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): if not os.path.isdir(_a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __magic_name__ : Optional[int] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) __magic_name__ : Union[str, Any] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ): copyfile(self.vocab_file , _a ) if os.path.abspath(self.merges_file ) != os.path.abspath(_a ): copyfile(self.merges_file , _a ) return out_vocab_file, out_merge_file def SCREAMING_SNAKE_CASE ( self , _a ): if isinstance(_a , _a ): try: with open(_a , "r" , encoding="utf-8" ) as fd: self.add_from_file(_a ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(f'''Incorrect encoding detected in {f}, please rebuild the dataset''' ) return __magic_name__ : List[Any] = f.readlines() for lineTmp in lines: __magic_name__ : Optional[Any] = lineTmp.strip() __magic_name__ : Union[str, Any] = line.rfind(" " ) if idx == -1: raise ValueError("Incorrect dictionary format, expected '<token> <cnt>'" ) __magic_name__ : Optional[int] = line[:idx] __magic_name__ : Dict = len(self.encoder )
281
0
"""simple docstring""" from pathlib import Path from typing import List from transformers import is_torch_available, is_vision_available from transformers.testing_utils import get_tests_dir, is_tool_test from transformers.tools.agent_types import AGENT_TYPE_MAPPING, AgentAudio, AgentImage, AgentText if is_torch_available(): import torch if is_vision_available(): from PIL import Image lowerCamelCase__ = ["text", "image", "audio"] def __lowerCAmelCase (_UpperCamelCase ): __lowerCAmelCase : Union[str, Any] = [] for input_type in input_types: if input_type == "text": inputs.append('Text input' ) elif input_type == "image": inputs.append( Image.open(Path(get_tests_dir('fixtures/tests_samples/COCO' ) ) / '000000039769.png' ).resize((512, 512) ) ) elif input_type == "audio": inputs.append(torch.ones(3000 ) ) elif isinstance(_snake_case , _snake_case ): inputs.append(create_inputs(_snake_case ) ) else: raise ValueError(F"Invalid type requested: {input_type}" ) return inputs def __lowerCAmelCase (_UpperCamelCase ): __lowerCAmelCase : Union[str, Any] = [] for output in outputs: if isinstance(_snake_case , (str, AgentText) ): output_types.append('text' ) elif isinstance(_snake_case , (Image.Image, AgentImage) ): output_types.append('image' ) elif isinstance(_snake_case , (torch.Tensor, AgentAudio) ): output_types.append('audio' ) else: raise ValueError(F"Invalid output: {output}" ) return output_types @is_tool_test class A__ : def __lowerCamelCase ( self ): self.assertTrue(hasattr(self.tool , 'inputs' ) ) self.assertTrue(hasattr(self.tool , 'outputs' ) ) __lowerCAmelCase : Union[str, Any] = self.tool.inputs for _input in inputs: if isinstance(_input , _a ): for __input in _input: self.assertTrue(__input in authorized_types ) else: self.assertTrue(_input in authorized_types ) __lowerCAmelCase : List[Any] = self.tool.outputs for _output in outputs: self.assertTrue(_output in authorized_types ) def __lowerCamelCase ( self ): __lowerCAmelCase : Union[str, Any] = create_inputs(self.tool.inputs ) __lowerCAmelCase : List[str] = self.tool(*_a ) # There is a single output if len(self.tool.outputs ) == 1: __lowerCAmelCase : Any = [outputs] self.assertListEqual(output_types(_a ) , self.tool.outputs ) def __lowerCamelCase ( self ): self.assertTrue(hasattr(self.tool , 'description' ) ) self.assertTrue(hasattr(self.tool , 'default_checkpoint' ) ) self.assertTrue(self.tool.description.startswith('This is a tool that' ) ) def __lowerCamelCase ( self ): __lowerCAmelCase : Optional[Any] = create_inputs(self.tool.inputs ) __lowerCAmelCase : Dict = self.tool(*_a ) if not isinstance(_a , _a ): __lowerCAmelCase : List[Any] = [outputs] self.assertEqual(len(_a ) , len(self.tool.outputs ) ) for output, output_type in zip(_a , self.tool.outputs ): __lowerCAmelCase : str = AGENT_TYPE_MAPPING[output_type] self.assertTrue(isinstance(_a , _a ) ) def __lowerCamelCase ( self ): __lowerCAmelCase : Tuple = create_inputs(self.tool.inputs ) __lowerCAmelCase : Optional[Any] = [] for _input, input_type in zip(_a , self.tool.inputs ): if isinstance(_a , _a ): _inputs.append([AGENT_TYPE_MAPPING[_input_type](_input ) for _input_type in input_type] ) else: _inputs.append(AGENT_TYPE_MAPPING[input_type](_input ) ) # Should not raise an error __lowerCAmelCase : List[Any] = self.tool(*_a ) if not isinstance(_a , _a ): __lowerCAmelCase : str = [outputs] self.assertEqual(len(_a ) , len(self.tool.outputs ) )
86
from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_ ( _snake_case : str = "laptop" ) -> DataFrame: '''simple docstring''' __magic_name__ : Tuple = F'''https://www.amazon.in/laptop/s?k={product}''' __magic_name__ : Dict = { "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36", "Accept-Language": "en-US, en;q=0.5", } __magic_name__ : Tuple = BeautifulSoup(requests.get(_snake_case , headers=_snake_case ).text ) # Initialize a Pandas dataframe with the column titles __magic_name__ : int = DataFrame( columns=[ "Product Title", "Product Link", "Current Price of the product", "Product Rating", "MRP of the product", "Discount", ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( "div" , attrs={"class": "s-result-item", "data-component-type": "s-search-result"} , ) , soup.find_all("div" , attrs={"class": "a-row a-size-base a-color-base"} ) , ): try: __magic_name__ : Dict = item.ha.text __magic_name__ : Optional[int] = "https://www.amazon.in/" + item.ha.a["href"] __magic_name__ : Optional[Any] = item.find("span" , attrs={"class": "a-offscreen"} ).text try: __magic_name__ : Union[str, Any] = item.find("span" , attrs={"class": "a-icon-alt"} ).text except AttributeError: __magic_name__ : Dict = "Not available" try: __magic_name__ : Optional[int] = ( "₹" + item.find( "span" , attrs={"class": "a-price a-text-price"} ).text.split("₹" )[1] ) except AttributeError: __magic_name__ : List[str] = "" try: __magic_name__ : int = float( ( ( float(product_mrp.strip("₹" ).replace("," , "" ) ) - float(product_price.strip("₹" ).replace("," , "" ) ) ) / float(product_mrp.strip("₹" ).replace("," , "" ) ) ) * 100 ) except ValueError: __magic_name__ : str = float("nan" ) except AttributeError: pass __magic_name__ : Optional[int] = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] __magic_name__ : Optional[Any] = " " __magic_name__ : str = " " data_frame.index += 1 return data_frame if __name__ == "__main__": snake_case : Any = "headphones" get_amazon_product_data(product).to_csv(F"Amazon Product Data for {product}.csv")
281
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class __snake_case ( unittest.TestCase ): def __init__( self : int , _lowercase : str , _lowercase : Optional[Any]=7 , _lowercase : Tuple=3 , _lowercase : Union[str, Any]=18 , _lowercase : Optional[int]=30 , _lowercase : Tuple=4_00 , _lowercase : Optional[Any]=True , _lowercase : Optional[Any]=None , _lowercase : List[Any]=True , _lowercase : Optional[int]=None , _lowercase : str=True , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = size if size is not None else {"shortest_edge": 20} SCREAMING_SNAKE_CASE__ = crop_size if crop_size is not None else {"height": 18, "width": 18} SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = image_size SCREAMING_SNAKE_CASE__ = min_resolution SCREAMING_SNAKE_CASE__ = max_resolution SCREAMING_SNAKE_CASE__ = do_resize SCREAMING_SNAKE_CASE__ = size SCREAMING_SNAKE_CASE__ = do_center_crop SCREAMING_SNAKE_CASE__ = crop_size SCREAMING_SNAKE_CASE__ = do_flip_channel_order def __a ( self : Optional[Any] ): """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class __snake_case ( lowerCamelCase_ , unittest.TestCase ): lowerCAmelCase_ = MobileViTImageProcessor if is_vision_available() else None def __a ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE__ = MobileViTImageProcessingTester(self ) @property def __a ( self : Optional[int] ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def __a ( self : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_a , """do_resize""" ) ) self.assertTrue(hasattr(_a , """size""" ) ) self.assertTrue(hasattr(_a , """do_center_crop""" ) ) self.assertTrue(hasattr(_a , """center_crop""" ) ) self.assertTrue(hasattr(_a , """do_flip_channel_order""" ) ) def __a ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 20} ) self.assertEqual(image_processor.crop_size , {"""height""": 18, """width""": 18} ) SCREAMING_SNAKE_CASE__ = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) def __a ( self : str ): """simple docstring""" pass def __a ( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_a ) for image in image_inputs: self.assertIsInstance(_a , Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE__ = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched SCREAMING_SNAKE_CASE__ = image_processing(_a , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def __a ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_a , numpify=_a ) for image in image_inputs: self.assertIsInstance(_a , np.ndarray ) # Test not batched input SCREAMING_SNAKE_CASE__ = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched SCREAMING_SNAKE_CASE__ = image_processing(_a , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def __a ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_a , torchify=_a ) for image in image_inputs: self.assertIsInstance(_a , torch.Tensor ) # Test not batched input SCREAMING_SNAKE_CASE__ = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched SCREAMING_SNAKE_CASE__ = image_processing(_a , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , )
219
from __future__ import annotations class _snake_case : def __init__( self , _a ): __magic_name__ : Optional[Any] = data __magic_name__ : Node | None = None __magic_name__ : Node | None = None def lowerCAmelCase_ ( _snake_case : Node | None ) -> None: # In Order traversal of the tree '''simple docstring''' if tree: display(tree.left ) print(tree.data ) display(tree.right ) def lowerCAmelCase_ ( _snake_case : Node | None ) -> int: '''simple docstring''' return 1 + max(depth_of_tree(tree.left ) , depth_of_tree(tree.right ) ) if tree else 0 def lowerCAmelCase_ ( _snake_case : Node ) -> bool: '''simple docstring''' if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def lowerCAmelCase_ ( ) -> None: # Main function for testing. '''simple docstring''' __magic_name__ : int = Node(1 ) __magic_name__ : Union[str, Any] = Node(2 ) __magic_name__ : Tuple = Node(3 ) __magic_name__ : Optional[Any] = Node(4 ) __magic_name__ : Union[str, Any] = Node(5 ) __magic_name__ : Any = Node(6 ) __magic_name__ : int = Node(7 ) __magic_name__ : List[str] = Node(8 ) __magic_name__ : Union[str, Any] = Node(9 ) print(is_full_binary_tree(_snake_case ) ) print(depth_of_tree(_snake_case ) ) print("Tree is: " ) display(_snake_case ) if __name__ == "__main__": main()
281
0
'''simple docstring''' import warnings from diffusers import StableDiffusionImgaImgPipeline # noqa F401 warnings.warn( '''The `image_to_image.py` script is outdated. Please use directly `from diffusers import''' ''' StableDiffusionImg2ImgPipeline` instead.''' )
97
def lowerCAmelCase_ ( _snake_case : str , _snake_case : str ) -> bool: '''simple docstring''' __magic_name__ : Union[str, Any] = len(_snake_case ) + 1 __magic_name__ : List[str] = len(_snake_case ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length i of input_string matches with prefix string of length j of # given pattern. # "dp" stands for dynamic programming. __magic_name__ : str = [[0 for i in range(_snake_case )] for j in range(_snake_case )] # since string of zero length match pattern of zero length __magic_name__ : Optional[int] = 1 # since pattern of zero length will never match with string of non-zero length for i in range(1 , _snake_case ): __magic_name__ : Optional[int] = 0 # since string of zero length will match with pattern where there # is at least one * alternatively for j in range(1 , _snake_case ): __magic_name__ : Union[str, Any] = dp[0][j - 2] if pattern[j - 1] == "*" else 0 # now using bottom-up approach to find for all remaining lengths for i in range(1 , _snake_case ): for j in range(1 , _snake_case ): if input_string[i - 1] == pattern[j - 1] or pattern[j - 1] == ".": __magic_name__ : Optional[int] = dp[i - 1][j - 1] elif pattern[j - 1] == "*": if dp[i][j - 2] == 1: __magic_name__ : Optional[Any] = 1 elif pattern[j - 2] in (input_string[i - 1], "."): __magic_name__ : List[Any] = dp[i - 1][j] else: __magic_name__ : Union[str, Any] = 0 else: __magic_name__ : Dict = 0 return bool(dp[-1][-1] ) if __name__ == "__main__": import doctest doctest.testmod() # inputing the strings # input_string = input("input a string :") # pattern = input("input a pattern :") snake_case : Optional[Any] = "aab" snake_case : List[str] = "c*a*b" # using function to check whether given string matches the given pattern if match_pattern(input_string, pattern): print(F"{input_string} matches the given pattern {pattern}") else: print(F"{input_string} does not match with the given pattern {pattern}")
281
0
"""simple docstring""" import argparse import os import shutil from pathlib import Path import onnx import torch from packaging import version from torch.onnx import export from diffusers import OnnxRuntimeModel, OnnxStableDiffusionPipeline, StableDiffusionPipeline A: Tuple = version.parse(version.parse(torch.__version__).base_version) < version.parse("1.11") def _snake_case ( UpperCamelCase : Union[str, Any] , UpperCamelCase : tuple , UpperCamelCase : Path , UpperCamelCase : Optional[Any] , UpperCamelCase : Any , UpperCamelCase : Dict , UpperCamelCase : Optional[Any] , UpperCamelCase : List[str]=False , ): output_path.parent.mkdir(parents=_snake_case , exist_ok=_snake_case ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( _snake_case , _snake_case , f=output_path.as_posix() , input_names=_snake_case , output_names=_snake_case , dynamic_axes=_snake_case , do_constant_folding=_snake_case , use_external_data_format=_snake_case , enable_onnx_checker=_snake_case , opset_version=_snake_case , ) else: export( _snake_case , _snake_case , f=output_path.as_posix() , input_names=_snake_case , output_names=_snake_case , dynamic_axes=_snake_case , do_constant_folding=_snake_case , opset_version=_snake_case , ) @torch.no_grad() def _snake_case ( UpperCamelCase : str , UpperCamelCase : str , UpperCamelCase : int , UpperCamelCase : bool = False ): UpperCAmelCase : Tuple = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): UpperCAmelCase : List[str] = "cuda" elif fpaa and not torch.cuda.is_available(): raise ValueError("""`float16` model export is only supported on GPUs with CUDA""" ) else: UpperCAmelCase : Tuple = "cpu" UpperCAmelCase : Optional[Any] = StableDiffusionPipeline.from_pretrained(_snake_case , torch_dtype=_snake_case ).to(_snake_case ) UpperCAmelCase : List[str] = Path(_snake_case ) # TEXT ENCODER UpperCAmelCase : Tuple = pipeline.text_encoder.config.max_position_embeddings UpperCAmelCase : List[str] = pipeline.text_encoder.config.hidden_size UpperCAmelCase : List[Any] = pipeline.tokenizer( """A sample prompt""" , padding="""max_length""" , max_length=pipeline.tokenizer.model_max_length , truncation=_snake_case , return_tensors="""pt""" , ) onnx_export( pipeline.text_encoder , model_args=(text_input.input_ids.to(device=_snake_case , dtype=torch.intaa )) , output_path=output_path / """text_encoder""" / """model.onnx""" , ordered_input_names=["""input_ids"""] , output_names=["""last_hidden_state""", """pooler_output"""] , dynamic_axes={ """input_ids""": {0: """batch""", 1: """sequence"""}, } , opset=_snake_case , ) del pipeline.text_encoder # UNET UpperCAmelCase : Any = pipeline.unet.config.in_channels UpperCAmelCase : Tuple = pipeline.unet.config.sample_size UpperCAmelCase : int = output_path / "unet" / "model.onnx" onnx_export( pipeline.unet , model_args=( torch.randn(2 , _snake_case , _snake_case , _snake_case ).to(device=_snake_case , dtype=_snake_case ), torch.randn(2 ).to(device=_snake_case , dtype=_snake_case ), torch.randn(2 , _snake_case , _snake_case ).to(device=_snake_case , dtype=_snake_case ), False, ) , output_path=_snake_case , ordered_input_names=["""sample""", """timestep""", """encoder_hidden_states""", """return_dict"""] , output_names=["""out_sample"""] , dynamic_axes={ """sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, """timestep""": {0: """batch"""}, """encoder_hidden_states""": {0: """batch""", 1: """sequence"""}, } , opset=_snake_case , use_external_data_format=_snake_case , ) UpperCAmelCase : Any = str(unet_path.absolute().as_posix() ) UpperCAmelCase : int = os.path.dirname(_snake_case ) UpperCAmelCase : Tuple = onnx.load(_snake_case ) # clean up existing tensor files shutil.rmtree(_snake_case ) os.mkdir(_snake_case ) # collate external tensor files into one onnx.save_model( _snake_case , _snake_case , save_as_external_data=_snake_case , all_tensors_to_one_file=_snake_case , location="""weights.pb""" , convert_attribute=_snake_case , ) del pipeline.unet # VAE ENCODER UpperCAmelCase : int = pipeline.vae UpperCAmelCase : Tuple = vae_encoder.config.in_channels UpperCAmelCase : List[Any] = vae_encoder.config.sample_size # need to get the raw tensor output (sample) from the encoder UpperCAmelCase : Union[str, Any] = lambda UpperCamelCase , UpperCamelCase : vae_encoder.encode(_snake_case , _snake_case )[0].sample() onnx_export( _snake_case , model_args=( torch.randn(1 , _snake_case , _snake_case , _snake_case ).to(device=_snake_case , dtype=_snake_case ), False, ) , output_path=output_path / """vae_encoder""" / """model.onnx""" , ordered_input_names=["""sample""", """return_dict"""] , output_names=["""latent_sample"""] , dynamic_axes={ """sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, } , opset=_snake_case , ) # VAE DECODER UpperCAmelCase : Dict = pipeline.vae UpperCAmelCase : Tuple = vae_decoder.config.latent_channels UpperCAmelCase : Optional[Any] = vae_decoder.config.out_channels # forward only through the decoder part UpperCAmelCase : str = vae_encoder.decode onnx_export( _snake_case , model_args=( torch.randn(1 , _snake_case , _snake_case , _snake_case ).to(device=_snake_case , dtype=_snake_case ), False, ) , output_path=output_path / """vae_decoder""" / """model.onnx""" , ordered_input_names=["""latent_sample""", """return_dict"""] , output_names=["""sample"""] , dynamic_axes={ """latent_sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, } , opset=_snake_case , ) del pipeline.vae # SAFETY CHECKER if pipeline.safety_checker is not None: UpperCAmelCase : List[Any] = pipeline.safety_checker UpperCAmelCase : Optional[int] = safety_checker.config.vision_config.num_channels UpperCAmelCase : str = safety_checker.config.vision_config.image_size UpperCAmelCase : List[Any] = safety_checker.forward_onnx onnx_export( pipeline.safety_checker , model_args=( torch.randn( 1 , _snake_case , _snake_case , _snake_case , ).to(device=_snake_case , dtype=_snake_case ), torch.randn(1 , _snake_case , _snake_case , _snake_case ).to(device=_snake_case , dtype=_snake_case ), ) , output_path=output_path / """safety_checker""" / """model.onnx""" , ordered_input_names=["""clip_input""", """images"""] , output_names=["""out_images""", """has_nsfw_concepts"""] , dynamic_axes={ """clip_input""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, """images""": {0: """batch""", 1: """height""", 2: """width""", 3: """channels"""}, } , opset=_snake_case , ) del pipeline.safety_checker UpperCAmelCase : Optional[int] = OnnxRuntimeModel.from_pretrained(output_path / """safety_checker""" ) UpperCAmelCase : Tuple = pipeline.feature_extractor else: UpperCAmelCase : Optional[Any] = None UpperCAmelCase : Dict = None UpperCAmelCase : Optional[int] = OnnxStableDiffusionPipeline( vae_encoder=OnnxRuntimeModel.from_pretrained(output_path / """vae_encoder""" ) , vae_decoder=OnnxRuntimeModel.from_pretrained(output_path / """vae_decoder""" ) , text_encoder=OnnxRuntimeModel.from_pretrained(output_path / """text_encoder""" ) , tokenizer=pipeline.tokenizer , unet=OnnxRuntimeModel.from_pretrained(output_path / """unet""" ) , scheduler=pipeline.scheduler , safety_checker=_snake_case , feature_extractor=_snake_case , requires_safety_checker=safety_checker is not None , ) onnx_pipeline.save_pretrained(_snake_case ) print("""ONNX pipeline saved to""" , _snake_case ) del pipeline del onnx_pipeline UpperCAmelCase : int = OnnxStableDiffusionPipeline.from_pretrained(_snake_case , provider="""CPUExecutionProvider""" ) print("""ONNX pipeline is loadable""" ) if __name__ == "__main__": A: List[str] = argparse.ArgumentParser() parser.add_argument( "--model_path", type=str, required=True, help="Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).", ) parser.add_argument("--output_path", type=str, required=True, help="Path to the output model.") parser.add_argument( "--opset", default=1_4, type=int, help="The version of the ONNX operator set to use.", ) parser.add_argument("--fp16", action="store_true", default=False, help="Export the models in `float16` mode") A: List[Any] = parser.parse_args() convert_models(args.model_path, args.output_path, args.opset, args.fpaa)
109
import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class _snake_case : @staticmethod def SCREAMING_SNAKE_CASE ( *_a , **_a ): pass def lowerCAmelCase_ ( _snake_case : Image ) -> str: '''simple docstring''' __magic_name__ : Optional[int] = hashlib.mda(image.tobytes() ) return m.hexdigest()[:10] def lowerCAmelCase_ ( _snake_case : Image ) -> Dict: '''simple docstring''' __magic_name__ : List[Any] = np.array(_snake_case ) __magic_name__ : Optional[int] = npimg.shape return {"hash": hashimage(_snake_case ), "shape": shape} @is_pipeline_test @require_vision @require_torch class _snake_case ( unittest.TestCase ): UpperCamelCase__ = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) UpperCamelCase__ = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def SCREAMING_SNAKE_CASE ( self , _a , _a , _a ): __magic_name__ : Dict = MaskGenerationPipeline(model=_a , image_processor=_a ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self , _a , _a ): pass @require_tf @unittest.skip("Image segmentation not implemented in TF" ) def SCREAMING_SNAKE_CASE ( self ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = pipeline("mask-generation" , model="facebook/sam-vit-huge" ) __magic_name__ : str = image_segmenter("http://images.cocodataset.org/val2017/000000039769.jpg" , points_per_batch=256 ) # Shortening by hashing __magic_name__ : Dict = [] for i, o in enumerate(outputs["masks"] ): new_outupt += [{"mask": mask_to_test_readable(_a ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(_a , decimals=4 ) , [ {"mask": {"hash": "115ad19f5f", "shape": (480, 640)}, "scores": 1.04_44}, {"mask": {"hash": "6affa964c6", "shape": (480, 640)}, "scores": 1.0_21}, {"mask": {"hash": "dfe28a0388", "shape": (480, 640)}, "scores": 1.01_67}, {"mask": {"hash": "c0a5f4a318", "shape": (480, 640)}, "scores": 1.01_32}, {"mask": {"hash": "fe8065c197", "shape": (480, 640)}, "scores": 1.00_53}, {"mask": {"hash": "e2d0b7a0b7", "shape": (480, 640)}, "scores": 0.99_67}, {"mask": {"hash": "453c7844bd", "shape": (480, 640)}, "scores": 0.9_93}, {"mask": {"hash": "3d44f2926d", "shape": (480, 640)}, "scores": 0.99_09}, {"mask": {"hash": "64033ddc3f", "shape": (480, 640)}, "scores": 0.98_79}, {"mask": {"hash": "801064ff79", "shape": (480, 640)}, "scores": 0.98_34}, {"mask": {"hash": "6172f276ef", "shape": (480, 640)}, "scores": 0.97_16}, {"mask": {"hash": "b49e60e084", "shape": (480, 640)}, "scores": 0.96_12}, {"mask": {"hash": "a811e775fd", "shape": (480, 640)}, "scores": 0.95_99}, {"mask": {"hash": "a6a8ebcf4b", "shape": (480, 640)}, "scores": 0.95_52}, {"mask": {"hash": "9d8257e080", "shape": (480, 640)}, "scores": 0.95_32}, {"mask": {"hash": "32de6454a8", "shape": (480, 640)}, "scores": 0.95_16}, {"mask": {"hash": "af3d4af2c8", "shape": (480, 640)}, "scores": 0.94_99}, {"mask": {"hash": "3c6db475fb", "shape": (480, 640)}, "scores": 0.94_83}, {"mask": {"hash": "c290813fb9", "shape": (480, 640)}, "scores": 0.94_64}, {"mask": {"hash": "b6f0b8f606", "shape": (480, 640)}, "scores": 0.9_43}, {"mask": {"hash": "92ce16bfdf", "shape": (480, 640)}, "scores": 0.9_43}, {"mask": {"hash": "c749b25868", "shape": (480, 640)}, "scores": 0.94_08}, {"mask": {"hash": "efb6cab859", "shape": (480, 640)}, "scores": 0.93_35}, {"mask": {"hash": "1ff2eafb30", "shape": (480, 640)}, "scores": 0.93_26}, {"mask": {"hash": "788b798e24", "shape": (480, 640)}, "scores": 0.92_62}, {"mask": {"hash": "abea804f0e", "shape": (480, 640)}, "scores": 0.89_99}, {"mask": {"hash": "7b9e8ddb73", "shape": (480, 640)}, "scores": 0.89_86}, {"mask": {"hash": "cd24047c8a", "shape": (480, 640)}, "scores": 0.89_84}, {"mask": {"hash": "6943e6bcbd", "shape": (480, 640)}, "scores": 0.88_73}, {"mask": {"hash": "b5f47c9191", "shape": (480, 640)}, "scores": 0.88_71} ] , ) # fmt: on @require_torch @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : str = "facebook/sam-vit-huge" __magic_name__ : str = pipeline("mask-generation" , model=_a ) __magic_name__ : Tuple = image_segmenter( "http://images.cocodataset.org/val2017/000000039769.jpg" , pred_iou_thresh=1 , points_per_batch=256 ) # Shortening by hashing __magic_name__ : Any = [] for i, o in enumerate(outputs["masks"] ): new_outupt += [{"mask": mask_to_test_readable(_a ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(_a , decimals=4 ) , [ {"mask": {"hash": "115ad19f5f", "shape": (480, 640)}, "scores": 1.04_44}, {"mask": {"hash": "6affa964c6", "shape": (480, 640)}, "scores": 1.02_10}, {"mask": {"hash": "dfe28a0388", "shape": (480, 640)}, "scores": 1.01_67}, {"mask": {"hash": "c0a5f4a318", "shape": (480, 640)}, "scores": 1.01_32}, {"mask": {"hash": "fe8065c197", "shape": (480, 640)}, "scores": 1.00_53}, ] , )
281
0
import webbrowser from sys import argv from urllib.parse import parse_qs, quote import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": _A = "%20".join(argv[1:]) if len(argv) > 1 else quote(str(input('''Search: '''))) print('''Googling.....''') _A = f"https://www.google.com/search?q={query}&num=100" _A = requests.get( url, headers={'''User-Agent''': str(UserAgent().random)}, ) try: _A = ( BeautifulSoup(res.text, '''html.parser''') .find('''div''', attrs={'''class''': '''yuRUbf'''}) .find('''a''') .get('''href''') ) except AttributeError: _A = parse_qs( BeautifulSoup(res.text, '''html.parser''') .find('''div''', attrs={'''class''': '''kCrYT'''}) .find('''a''') .get('''href''') )["url"][0] webbrowser.open(link)
278
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets snake_case : List[Any] = "\\n@inproceedings{lin-2004-rouge,\n title = \"{ROUGE}: A Package for Automatic Evaluation of Summaries\",\n author = \"Lin, Chin-Yew\",\n booktitle = \"Text Summarization Branches Out\",\n month = jul,\n year = \"2004\",\n address = \"Barcelona, Spain\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W04-1013\",\n pages = \"74--81\",\n}\n" snake_case : Any = "\\nROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for\nevaluating automatic summarization and machine translation software in natural language processing.\nThe metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation.\n\nNote that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters.\n\nThis metrics is a wrapper around Google Research reimplementation of ROUGE:\nhttps://github.com/google-research/google-research/tree/master/rouge\n" snake_case : str = "\nCalculates average rouge scores for a list of hypotheses and references\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n rouge_types: A list of rouge types to calculate.\n Valid names:\n `\"rouge{n}\"` (e.g. `\"rouge1\"`, `\"rouge2\"`) where: {n} is the n-gram based scoring,\n `\"rougeL\"`: Longest common subsequence based scoring.\n `\"rougeLSum\"`: rougeLsum splits text using `\"\n\"`.\n See details in https://github.com/huggingface/datasets/issues/617\n use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes.\n use_aggregator: Return aggregates if this is set to True\nReturns:\n rouge1: rouge_1 (precision, recall, f1),\n rouge2: rouge_2 (precision, recall, f1),\n rougeL: rouge_l (precision, recall, f1),\n rougeLsum: rouge_lsum (precision, recall, f1)\nExamples:\n\n >>> rouge = datasets.load_metric('rouge')\n >>> predictions = [\"hello there\", \"general kenobi\"]\n >>> references = [\"hello there\", \"general kenobi\"]\n >>> results = rouge.compute(predictions=predictions, references=references)\n >>> print(list(results.keys()))\n ['rouge1', 'rouge2', 'rougeL', 'rougeLsum']\n >>> print(results[\"rouge1\"])\n AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0))\n >>> print(results[\"rouge1\"].mid.fmeasure)\n 1.0\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _snake_case ( datasets.Metric ): def SCREAMING_SNAKE_CASE ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/google-research/google-research/tree/master/rouge"] , reference_urls=[ "https://en.wikipedia.org/wiki/ROUGE_(metric)", "https://github.com/google-research/google-research/tree/master/rouge", ] , ) def SCREAMING_SNAKE_CASE ( self , _a , _a , _a=None , _a=True , _a=False ): if rouge_types is None: __magic_name__ : str = ["rouge1", "rouge2", "rougeL", "rougeLsum"] __magic_name__ : List[str] = rouge_scorer.RougeScorer(rouge_types=_a , use_stemmer=_a ) if use_aggregator: __magic_name__ : Dict = scoring.BootstrapAggregator() else: __magic_name__ : str = [] for ref, pred in zip(_a , _a ): __magic_name__ : Union[str, Any] = scorer.score(_a , _a ) if use_aggregator: aggregator.add_scores(_a ) else: scores.append(_a ) if use_aggregator: __magic_name__ : Any = aggregator.aggregate() else: __magic_name__ : List[Any] = {} for key in scores[0]: __magic_name__ : str = [score[key] for score in scores] return result
281
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __UpperCamelCase = { "configuration_mega": ["MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP", "MegaConfig", "MegaOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ "MEGA_PRETRAINED_MODEL_ARCHIVE_LIST", "MegaForCausalLM", "MegaForMaskedLM", "MegaForMultipleChoice", "MegaForQuestionAnswering", "MegaForSequenceClassification", "MegaForTokenClassification", "MegaModel", "MegaPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mega import MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP, MegaConfig, MegaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mega import ( MEGA_PRETRAINED_MODEL_ARCHIVE_LIST, MegaForCausalLM, MegaForMaskedLM, MegaForMultipleChoice, MegaForQuestionAnswering, MegaForSequenceClassification, MegaForTokenClassification, MegaModel, MegaPreTrainedModel, ) else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
113
snake_case : Optional[int] = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/" def lowerCAmelCase_ ( _snake_case : bytes ) -> bytes: '''simple docstring''' if not isinstance(_snake_case , _snake_case ): __magic_name__ : Tuple = F'''a bytes-like object is required, not \'{data.__class__.__name__}\'''' raise TypeError(_snake_case ) __magic_name__ : Optional[int] = "".join(bin(_snake_case )[2:].zfill(8 ) for byte in data ) __magic_name__ : List[Any] = len(_snake_case ) % 6 != 0 if padding_needed: # The padding that will be added later __magic_name__ : List[str] = B"=" * ((6 - len(_snake_case ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(_snake_case ) % 6) else: __magic_name__ : List[str] = B"" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(_snake_case ) , 6 ) ).encode() + padding ) def lowerCAmelCase_ ( _snake_case : str ) -> bytes: '''simple docstring''' if not isinstance(_snake_case , _snake_case ) and not isinstance(_snake_case , _snake_case ): __magic_name__ : List[str] = ( "argument should be a bytes-like object or ASCII string, " F'''not \'{encoded_data.__class__.__name__}\'''' ) raise TypeError(_snake_case ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(_snake_case , _snake_case ): try: __magic_name__ : List[Any] = encoded_data.decode("utf-8" ) except UnicodeDecodeError: raise ValueError("base64 encoded data should only contain ASCII characters" ) __magic_name__ : List[str] = encoded_data.count("=" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(_snake_case ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one __magic_name__ : Optional[int] = encoded_data[:-padding] __magic_name__ : Dict = "".join( bin(B64_CHARSET.index(_snake_case ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: __magic_name__ : Union[str, Any] = "".join( bin(B64_CHARSET.index(_snake_case ) )[2:].zfill(6 ) for char in encoded_data ) __magic_name__ : List[Any] = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(_snake_case ) , 8 ) ] return bytes(_snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
281
0
"""simple docstring""" def _snake_case ( lowercase__ = 1000 ): _lowerCamelCase : Dict = 2**power _lowerCamelCase : Optional[Any] = str(_snake_case ) _lowerCamelCase : List[Any] = list(_snake_case ) _lowerCamelCase : Optional[int] = 0 for i in list_num: sum_of_num += int(_snake_case ) return sum_of_num if __name__ == "__main__": lowercase__ = int(input("""Enter the power of 2: """).strip()) print("""2 ^ """, power, """ = """, 2**power) lowercase__ = solution(power) print("""Sum of the digits is: """, result)
96
import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class _snake_case ( unittest.TestCase ): def __init__( self , _a , _a=13 , _a=7 , _a=True , _a=True , _a=True , _a=True , _a=99 , _a=32 , _a=5 , _a=4 , _a=37 , _a="gelu" , _a=0.1 , _a=0.1 , _a=512 , _a=16 , _a=2 , _a=0.02 , _a=4 , ): __magic_name__ : List[Any] = parent __magic_name__ : Optional[Any] = batch_size __magic_name__ : Dict = seq_length __magic_name__ : Union[str, Any] = is_training __magic_name__ : Optional[Any] = use_attention_mask __magic_name__ : Optional[Any] = use_token_type_ids __magic_name__ : int = use_labels __magic_name__ : List[Any] = vocab_size __magic_name__ : Union[str, Any] = hidden_size __magic_name__ : Optional[Any] = num_hidden_layers __magic_name__ : int = num_attention_heads __magic_name__ : Any = intermediate_size __magic_name__ : List[Any] = hidden_act __magic_name__ : List[Any] = hidden_dropout_prob __magic_name__ : Optional[int] = attention_probs_dropout_prob __magic_name__ : List[Any] = max_position_embeddings __magic_name__ : Tuple = type_vocab_size __magic_name__ : List[str] = type_sequence_label_size __magic_name__ : Dict = initializer_range __magic_name__ : List[Any] = num_choices def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __magic_name__ : List[Any] = None if self.use_attention_mask: __magic_name__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) __magic_name__ : str = None if self.use_token_type_ids: __magic_name__ : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __magic_name__ : List[str] = RobertaPreLayerNormConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_a , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : int = self.prepare_config_and_inputs() __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : List[Any] = config_and_inputs __magic_name__ : List[str] = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = self.prepare_config_and_inputs() __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Union[str, Any] = config_and_inputs __magic_name__ : Tuple = True __magic_name__ : int = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __magic_name__ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class _snake_case ( snake_case , unittest.TestCase ): UpperCamelCase__ = True UpperCamelCase__ = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[Any] = FlaxRobertaPreLayerNormModelTester(self ) @slow def SCREAMING_SNAKE_CASE ( self ): for model_class_name in self.all_model_classes: __magic_name__ : Optional[Any] = model_class_name.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=_a ) __magic_name__ : Dict = model(np.ones((1, 1) ) ) self.assertIsNotNone(_a ) @require_flax class _snake_case ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = FlaxRobertaPreLayerNormForMaskedLM.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=_a ) __magic_name__ : Union[str, Any] = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) __magic_name__ : List[str] = model(_a )[0] __magic_name__ : str = [1, 11, 50_265] self.assertEqual(list(output.shape ) , _a ) # compare the actual values for a slice. __magic_name__ : List[str] = np.array( [[[40.48_80, 18.01_99, -5.23_67], [-1.88_77, -4.08_85, 10.70_85], [-2.26_13, -5.61_10, 7.26_65]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , _a , atol=1e-4 ) ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[str] = FlaxRobertaPreLayerNormModel.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=_a ) __magic_name__ : Tuple = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) __magic_name__ : Tuple = model(_a )[0] # compare the actual values for a slice. __magic_name__ : Dict = np.array( [[[0.02_08, -0.03_56, 0.02_37], [-0.15_69, -0.04_11, -0.26_26], [0.18_79, 0.01_25, -0.00_89]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , _a , atol=1e-4 ) )
281
0
"""simple docstring""" import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) class lowerCAmelCase_ ( lowerCAmelCase ): """simple docstring""" def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): """simple docstring""" warnings.warn( 'The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use ChineseCLIPImageProcessor instead.' , _a , ) super().__init__(*_a , **_a )
150
def lowerCAmelCase_ ( _snake_case : list[list[int | float]] ) -> int: '''simple docstring''' __magic_name__ : Any = len(_snake_case ) __magic_name__ : Optional[Any] = len(matrix[0] ) __magic_name__ : Union[str, Any] = min(_snake_case , _snake_case ) for row in range(_snake_case ): # Check if diagonal element is not zero if matrix[row][row] != 0: # Eliminate all the elements below the diagonal for col in range(row + 1 , _snake_case ): __magic_name__ : Optional[Any] = matrix[col][row] / matrix[row][row] for i in range(_snake_case , _snake_case ): matrix[col][i] -= multiplier * matrix[row][i] else: # Find a non-zero diagonal element to swap rows __magic_name__ : str = True for i in range(row + 1 , _snake_case ): if matrix[i][row] != 0: __magic_name__ , __magic_name__ : List[str] = matrix[i], matrix[row] __magic_name__ : Union[str, Any] = False break if reduce: rank -= 1 for i in range(_snake_case ): __magic_name__ : Any = matrix[i][rank] # Reduce the row pointer by one to stay on the same row row -= 1 return rank if __name__ == "__main__": import doctest doctest.testmod()
281
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase : int = logging.get_logger(__name__) UpperCAmelCase : Union[str, Any] = { "funnel-transformer/small": "https://huggingface.co/funnel-transformer/small/resolve/main/config.json", "funnel-transformer/small-base": "https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json", "funnel-transformer/medium": "https://huggingface.co/funnel-transformer/medium/resolve/main/config.json", "funnel-transformer/medium-base": "https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json", "funnel-transformer/intermediate": ( "https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json" ), "funnel-transformer/intermediate-base": ( "https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json" ), "funnel-transformer/large": "https://huggingface.co/funnel-transformer/large/resolve/main/config.json", "funnel-transformer/large-base": "https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json", "funnel-transformer/xlarge": "https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json", "funnel-transformer/xlarge-base": "https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json", } class lowerCamelCase__ ( A ): """simple docstring""" __a = """funnel""" __a = { """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", } def __init__( self : Optional[int] , UpperCamelCase : Union[str, Any]=30_522 , UpperCamelCase : Tuple=[4, 4, 4] , UpperCamelCase : Dict=None , UpperCamelCase : List[Any]=2 , UpperCamelCase : Optional[Any]=768 , UpperCamelCase : Optional[Any]=12 , UpperCamelCase : Dict=64 , UpperCamelCase : int=3_072 , UpperCamelCase : Optional[int]="gelu_new" , UpperCamelCase : Optional[int]=0.1 , UpperCamelCase : Any=0.1 , UpperCamelCase : List[str]=0.0 , UpperCamelCase : Dict=0.1 , UpperCamelCase : Tuple=None , UpperCamelCase : Any=1e-9 , UpperCamelCase : Union[str, Any]="mean" , UpperCamelCase : Tuple="relative_shift" , UpperCamelCase : List[Any]=True , UpperCamelCase : Tuple=True , UpperCamelCase : int=True , **UpperCamelCase : Dict , ): '''simple docstring''' __UpperCAmelCase : List[Any] = vocab_size __UpperCAmelCase : Tuple = block_sizes __UpperCAmelCase : Tuple = [1] * len(_a ) if block_repeats is None else block_repeats assert len(_a ) == len( self.block_repeats ), "`block_sizes` and `block_repeats` should have the same length." __UpperCAmelCase : str = num_decoder_layers __UpperCAmelCase : str = d_model __UpperCAmelCase : Dict = n_head __UpperCAmelCase : Tuple = d_head __UpperCAmelCase : str = d_inner __UpperCAmelCase : Optional[Any] = hidden_act __UpperCAmelCase : Any = hidden_dropout __UpperCAmelCase : Dict = attention_dropout __UpperCAmelCase : Any = activation_dropout __UpperCAmelCase : Tuple = initializer_range __UpperCAmelCase : Optional[int] = initializer_std __UpperCAmelCase : List[str] = layer_norm_eps assert pooling_type in [ "mean", "max", ], f'''Got {pooling_type} for `pooling_type` but only \'mean\' and \'max\' are supported.''' __UpperCAmelCase : Any = pooling_type assert attention_type in [ "relative_shift", "factorized", ], f'''Got {attention_type} for `attention_type` but only \'relative_shift\' and \'factorized\' are supported.''' __UpperCAmelCase : Any = attention_type __UpperCAmelCase : int = separate_cls __UpperCAmelCase : Dict = truncate_seq __UpperCAmelCase : List[Any] = pool_q_only super().__init__(**_a ) @property def lowerCamelCase__ ( self : List[str] ): '''simple docstring''' return sum(self.block_sizes ) @num_hidden_layers.setter def lowerCamelCase__ ( self : Optional[Any] , UpperCamelCase : Dict ): '''simple docstring''' raise NotImplementedError( """This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`.""" ) @property def lowerCamelCase__ ( self : List[Any] ): '''simple docstring''' return len(self.block_sizes ) @num_blocks.setter def lowerCamelCase__ ( self : Tuple , UpperCamelCase : List[Any] ): '''simple docstring''' raise NotImplementedError("""This model does not support the setting of `num_blocks`. Please set `block_sizes`.""" )
115
import argparse import collections import json import os import re import string import sys import numpy as np snake_case : Dict = re.compile(R"\b(a|an|the)\b", re.UNICODE) snake_case : Optional[int] = None def lowerCAmelCase_ ( ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : Any = argparse.ArgumentParser("Official evaluation script for SQuAD version 2.0." ) parser.add_argument("data_file" , metavar="data.json" , help="Input data JSON file." ) parser.add_argument("pred_file" , metavar="pred.json" , help="Model predictions." ) parser.add_argument( "--out-file" , "-o" , metavar="eval.json" , help="Write accuracy metrics to file (default is stdout)." ) parser.add_argument( "--na-prob-file" , "-n" , metavar="na_prob.json" , help="Model estimates of probability of no answer." ) parser.add_argument( "--na-prob-thresh" , "-t" , type=_snake_case , default=1.0 , help="Predict \"\" if no-answer probability exceeds this (default = 1.0)." , ) parser.add_argument( "--out-image-dir" , "-p" , metavar="out_images" , default=_snake_case , help="Save precision-recall curves to directory." ) parser.add_argument("--verbose" , "-v" , action="store_true" ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def lowerCAmelCase_ ( _snake_case : Optional[Any] ) -> Tuple: '''simple docstring''' __magic_name__ : Optional[int] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: __magic_name__ : str = bool(qa["answers"]["text"] ) return qid_to_has_ans def lowerCAmelCase_ ( _snake_case : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' def remove_articles(_snake_case : List[str] ): return ARTICLES_REGEX.sub(" " , _snake_case ) def white_space_fix(_snake_case : Optional[int] ): return " ".join(text.split() ) def remove_punc(_snake_case : Optional[int] ): __magic_name__ : Dict = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_snake_case : str ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_snake_case ) ) ) ) def lowerCAmelCase_ ( _snake_case : Any ) -> Optional[Any]: '''simple docstring''' if not s: return [] return normalize_answer(_snake_case ).split() def lowerCAmelCase_ ( _snake_case : str , _snake_case : Dict ) -> Tuple: '''simple docstring''' return int(normalize_answer(_snake_case ) == normalize_answer(_snake_case ) ) def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : int ) -> str: '''simple docstring''' __magic_name__ : Any = get_tokens(_snake_case ) __magic_name__ : Optional[int] = get_tokens(_snake_case ) __magic_name__ : Tuple = collections.Counter(_snake_case ) & collections.Counter(_snake_case ) __magic_name__ : Tuple = sum(common.values() ) if len(_snake_case ) == 0 or len(_snake_case ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 __magic_name__ : Dict = 1.0 * num_same / len(_snake_case ) __magic_name__ : Optional[Any] = 1.0 * num_same / len(_snake_case ) __magic_name__ : List[Any] = (2 * precision * recall) / (precision + recall) return fa def lowerCAmelCase_ ( _snake_case : Optional[Any] , _snake_case : List[Any] ) -> List[Any]: '''simple docstring''' __magic_name__ : Union[str, Any] = {} __magic_name__ : int = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: __magic_name__ : Union[str, Any] = qa["id"] __magic_name__ : Any = [t for t in qa["answers"]["text"] if normalize_answer(_snake_case )] if not gold_answers: # For unanswerable questions, only correct answer is empty string __magic_name__ : Tuple = [""] if qid not in preds: print(F'''Missing prediction for {qid}''' ) continue __magic_name__ : Any = preds[qid] # Take max over all gold answers __magic_name__ : List[Any] = max(compute_exact(_snake_case , _snake_case ) for a in gold_answers ) __magic_name__ : int = max(compute_fa(_snake_case , _snake_case ) for a in gold_answers ) return exact_scores, fa_scores def lowerCAmelCase_ ( _snake_case : Optional[Any] , _snake_case : List[Any] , _snake_case : Optional[int] , _snake_case : Dict ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : str = {} for qid, s in scores.items(): __magic_name__ : Dict = na_probs[qid] > na_prob_thresh if pred_na: __magic_name__ : str = float(not qid_to_has_ans[qid] ) else: __magic_name__ : Optional[int] = s return new_scores def lowerCAmelCase_ ( _snake_case : List[Any] , _snake_case : List[str] , _snake_case : Tuple=None ) -> Tuple: '''simple docstring''' if not qid_list: __magic_name__ : Any = len(_snake_case ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores.values() ) / total), ("f1", 100.0 * sum(fa_scores.values() ) / total), ("total", total), ] ) else: __magic_name__ : Tuple = len(_snake_case ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores[k] for k in qid_list ) / total), ("f1", 100.0 * sum(fa_scores[k] for k in qid_list ) / total), ("total", total), ] ) def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : str , _snake_case : str ) -> Dict: '''simple docstring''' for k in new_eval: __magic_name__ : int = new_eval[k] def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Dict , _snake_case : Optional[Any] , _snake_case : Union[str, Any] ) -> str: '''simple docstring''' plt.step(_snake_case , _snake_case , color="b" , alpha=0.2 , where="post" ) plt.fill_between(_snake_case , _snake_case , step="post" , alpha=0.2 , color="b" ) plt.xlabel("Recall" ) plt.ylabel("Precision" ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(_snake_case ) plt.savefig(_snake_case ) plt.clf() def lowerCAmelCase_ ( _snake_case : Dict , _snake_case : Any , _snake_case : Optional[int] , _snake_case : List[Any] , _snake_case : Optional[int]=None , _snake_case : int=None ) -> str: '''simple docstring''' __magic_name__ : Union[str, Any] = sorted(_snake_case , key=lambda _snake_case : na_probs[k] ) __magic_name__ : Optional[int] = 0.0 __magic_name__ : str = 1.0 __magic_name__ : str = 0.0 __magic_name__ : List[str] = [1.0] __magic_name__ : str = [0.0] __magic_name__ : Optional[Any] = 0.0 for i, qid in enumerate(_snake_case ): if qid_to_has_ans[qid]: true_pos += scores[qid] __magic_name__ : List[str] = true_pos / float(i + 1 ) __magic_name__ : Any = true_pos / float(_snake_case ) if i == len(_snake_case ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(_snake_case ) recalls.append(_snake_case ) if out_image: plot_pr_curve(_snake_case , _snake_case , _snake_case , _snake_case ) return {"ap": 100.0 * avg_prec} def lowerCAmelCase_ ( _snake_case : Tuple , _snake_case : Optional[Any] , _snake_case : Optional[int] , _snake_case : Optional[Any] , _snake_case : Any , _snake_case : List[Any] ) -> Union[str, Any]: '''simple docstring''' if out_image_dir and not os.path.exists(_snake_case ): os.makedirs(_snake_case ) __magic_name__ : Any = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return __magic_name__ : str = make_precision_recall_eval( _snake_case , _snake_case , _snake_case , _snake_case , out_image=os.path.join(_snake_case , "pr_exact.png" ) , title="Precision-Recall curve for Exact Match score" , ) __magic_name__ : Union[str, Any] = make_precision_recall_eval( _snake_case , _snake_case , _snake_case , _snake_case , out_image=os.path.join(_snake_case , "pr_f1.png" ) , title="Precision-Recall curve for F1 score" , ) __magic_name__ : str = {k: float(_snake_case ) for k, v in qid_to_has_ans.items()} __magic_name__ : str = make_precision_recall_eval( _snake_case , _snake_case , _snake_case , _snake_case , out_image=os.path.join(_snake_case , "pr_oracle.png" ) , title="Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)" , ) merge_eval(_snake_case , _snake_case , "pr_exact" ) merge_eval(_snake_case , _snake_case , "pr_f1" ) merge_eval(_snake_case , _snake_case , "pr_oracle" ) def lowerCAmelCase_ ( _snake_case : int , _snake_case : Optional[Any] , _snake_case : List[str] , _snake_case : Optional[Any] ) -> Dict: '''simple docstring''' if not qid_list: return __magic_name__ : Dict = [na_probs[k] for k in qid_list] __magic_name__ : str = np.ones_like(_snake_case ) / float(len(_snake_case ) ) plt.hist(_snake_case , weights=_snake_case , bins=20 , range=(0.0, 1.0) ) plt.xlabel("Model probability of no-answer" ) plt.ylabel("Proportion of dataset" ) plt.title(F'''Histogram of no-answer probability: {name}''' ) plt.savefig(os.path.join(_snake_case , F'''na_prob_hist_{name}.png''' ) ) plt.clf() def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Tuple , _snake_case : List[str] , _snake_case : Dict ) -> List[Any]: '''simple docstring''' __magic_name__ : Union[str, Any] = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) __magic_name__ : List[str] = num_no_ans __magic_name__ : Dict = cur_score __magic_name__ : Dict = 0.0 __magic_name__ : Any = sorted(_snake_case , key=lambda _snake_case : na_probs[k] ) for i, qid in enumerate(_snake_case ): if qid not in scores: continue if qid_to_has_ans[qid]: __magic_name__ : Union[str, Any] = scores[qid] else: if preds[qid]: __magic_name__ : List[Any] = -1 else: __magic_name__ : Optional[int] = 0 cur_score += diff if cur_score > best_score: __magic_name__ : Optional[int] = cur_score __magic_name__ : List[Any] = na_probs[qid] return 100.0 * best_score / len(_snake_case ), best_thresh def lowerCAmelCase_ ( _snake_case : int , _snake_case : str , _snake_case : List[str] , _snake_case : Tuple , _snake_case : List[Any] , _snake_case : Dict ) -> Optional[Any]: '''simple docstring''' __magic_name__ , __magic_name__ : List[str] = find_best_thresh(_snake_case , _snake_case , _snake_case , _snake_case ) __magic_name__ , __magic_name__ : int = find_best_thresh(_snake_case , _snake_case , _snake_case , _snake_case ) __magic_name__ : Optional[int] = best_exact __magic_name__ : List[Any] = exact_thresh __magic_name__ : Dict = best_fa __magic_name__ : Any = fa_thresh def lowerCAmelCase_ ( ) -> int: '''simple docstring''' with open(OPTS.data_file ) as f: __magic_name__ : Optional[Any] = json.load(_snake_case ) __magic_name__ : List[Any] = dataset_json["data"] with open(OPTS.pred_file ) as f: __magic_name__ : Optional[Any] = json.load(_snake_case ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: __magic_name__ : Any = json.load(_snake_case ) else: __magic_name__ : Any = {k: 0.0 for k in preds} __magic_name__ : str = make_qid_to_has_ans(_snake_case ) # maps qid to True/False __magic_name__ : Tuple = [k for k, v in qid_to_has_ans.items() if v] __magic_name__ : Optional[Any] = [k for k, v in qid_to_has_ans.items() if not v] __magic_name__ , __magic_name__ : Union[str, Any] = get_raw_scores(_snake_case , _snake_case ) __magic_name__ : Optional[Any] = apply_no_ans_threshold(_snake_case , _snake_case , _snake_case , OPTS.na_prob_thresh ) __magic_name__ : Optional[Any] = apply_no_ans_threshold(_snake_case , _snake_case , _snake_case , OPTS.na_prob_thresh ) __magic_name__ : List[Any] = make_eval_dict(_snake_case , _snake_case ) if has_ans_qids: __magic_name__ : int = make_eval_dict(_snake_case , _snake_case , qid_list=_snake_case ) merge_eval(_snake_case , _snake_case , "HasAns" ) if no_ans_qids: __magic_name__ : List[Any] = make_eval_dict(_snake_case , _snake_case , qid_list=_snake_case ) merge_eval(_snake_case , _snake_case , "NoAns" ) if OPTS.na_prob_file: find_all_best_thresh(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case , OPTS.out_image_dir ) histogram_na_prob(_snake_case , _snake_case , OPTS.out_image_dir , "hasAns" ) histogram_na_prob(_snake_case , _snake_case , OPTS.out_image_dir , "noAns" ) if OPTS.out_file: with open(OPTS.out_file , "w" ) as f: json.dump(_snake_case , _snake_case ) else: print(json.dumps(_snake_case , indent=2 ) ) if __name__ == "__main__": snake_case : int = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("Agg") import matplotlib.pyplot as plt main()
281
0
import math def _a ( SCREAMING_SNAKE_CASE : float , SCREAMING_SNAKE_CASE : float ) -> float: """simple docstring""" return math.pow(_snake_case , 2 ) - a def _a ( SCREAMING_SNAKE_CASE : float ) -> float: """simple docstring""" return 2 * x def _a ( SCREAMING_SNAKE_CASE : float ) -> float: """simple docstring""" __lowerCAmelCase: Optional[int] = 2.0 while start <= a: __lowerCAmelCase: str = math.pow(_snake_case , 2 ) return start def _a ( SCREAMING_SNAKE_CASE : float , SCREAMING_SNAKE_CASE : int = 99_99 , SCREAMING_SNAKE_CASE : float = 0.0_0_0_0_0_0_0_0_0_0_0_0_0_1 ) -> float: """simple docstring""" if a < 0: raise ValueError('math domain error' ) __lowerCAmelCase: Optional[int] = get_initial_point(_snake_case ) for _ in range(_snake_case ): __lowerCAmelCase: int = value __lowerCAmelCase: str = value - fx(_snake_case , _snake_case ) / fx_derivative(_snake_case ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
322
import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin snake_case : str = "▁" snake_case : List[Any] = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class _snake_case ( snake_case , unittest.TestCase ): UpperCamelCase__ = BigBirdTokenizer UpperCamelCase__ = BigBirdTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True def SCREAMING_SNAKE_CASE ( self ): super().setUp() __magic_name__ : Optional[Any] = self.tokenizer_class(_a , keep_accents=_a ) tokenizer.save_pretrained(self.tmpdirname ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Union[str, Any] = "<s>" __magic_name__ : Dict = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_a ) , _a ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_a ) , _a ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "[MASK]" ) self.assertEqual(len(_a ) , 1_004 ) def SCREAMING_SNAKE_CASE ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def SCREAMING_SNAKE_CASE ( self ): if not self.test_rust_tokenizer: return __magic_name__ : Dict = self.get_tokenizer() __magic_name__ : str = self.get_rust_tokenizer() __magic_name__ : Any = "I was born in 92000, and this is falsé." __magic_name__ : Dict = tokenizer.tokenize(_a ) __magic_name__ : Any = rust_tokenizer.tokenize(_a ) self.assertListEqual(_a , _a ) __magic_name__ : List[Any] = tokenizer.encode(_a , add_special_tokens=_a ) __magic_name__ : List[str] = rust_tokenizer.encode(_a , add_special_tokens=_a ) self.assertListEqual(_a , _a ) __magic_name__ : str = self.get_rust_tokenizer() __magic_name__ : Dict = tokenizer.encode(_a ) __magic_name__ : Optional[int] = rust_tokenizer.encode(_a ) self.assertListEqual(_a , _a ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = BigBirdTokenizer(_a , keep_accents=_a ) __magic_name__ : str = tokenizer.tokenize("This is a test" ) self.assertListEqual(_a , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_a ) , [285, 46, 10, 170, 382] , ) __magic_name__ : Dict = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( _a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) __magic_name__ : Union[str, Any] = tokenizer.convert_tokens_to_ids(_a ) self.assertListEqual( _a , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) __magic_name__ : int = tokenizer.convert_ids_to_tokens(_a ) self.assertListEqual( _a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def SCREAMING_SNAKE_CASE ( self ): return BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Any = "Hello World!" __magic_name__ : Dict = [65, 18_536, 2_260, 101, 66] self.assertListEqual(_a , self.big_tokenizer.encode(_a ) ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) # fmt: off __magic_name__ : List[str] = [65, 871, 419, 358, 946, 991, 2_521, 452, 358, 1_357, 387, 7_751, 3_536, 112, 985, 456, 126, 865, 938, 5_400, 5_734, 458, 1_368, 467, 786, 2_462, 5_246, 1_159, 633, 865, 4_519, 457, 582, 852, 2_557, 427, 916, 508, 405, 34_324, 497, 391, 408, 11_342, 1_244, 385, 100, 938, 985, 456, 574, 362, 12_597, 3_200, 3_129, 1_172, 66] # noqa: E231 # fmt: on self.assertListEqual(_a , self.big_tokenizer.encode(_a ) ) @require_torch @slow def SCREAMING_SNAKE_CASE ( self ): import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence __magic_name__ : Optional[Any] = list(self.big_tokenizer.get_vocab().keys() )[:10] __magic_name__ : List[Any] = " ".join(_a ) __magic_name__ : Any = self.big_tokenizer.encode_plus(_a , return_tensors="pt" , return_token_type_ids=_a ) __magic_name__ : Union[str, Any] = self.big_tokenizer.batch_encode_plus( [sequence + " " + sequence] , return_tensors="pt" , return_token_type_ids=_a ) __magic_name__ : List[str] = BigBirdConfig(attention_type="original_full" ) __magic_name__ : Optional[int] = BigBirdModel(_a ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_a ) model(**_a ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : int = BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) __magic_name__ : int = tokenizer.decode(tokenizer("Paris is the [MASK]." ).input_ids ) self.assertTrue(decoded_text == "[CLS] Paris is the[MASK].[SEP]" ) @slow def SCREAMING_SNAKE_CASE ( self ): # fmt: off __magic_name__ : Optional[Any] = {"input_ids": [[65, 39_286, 458, 36_335, 2_001, 456, 13_073, 13_266, 455, 113, 7_746, 1_741, 11_157, 391, 13_073, 13_266, 455, 113, 3_967, 35_412, 113, 4_936, 109, 3_870, 2_377, 113, 30_084, 45_720, 458, 134, 17_496, 112, 503, 11_672, 113, 118, 112, 5_665, 13_347, 38_687, 112, 1_496, 31_389, 112, 3_268, 47_264, 134, 962, 112, 16_377, 8_035, 23_130, 430, 12_169, 15_518, 28_592, 458, 146, 41_697, 109, 391, 12_169, 15_518, 16_689, 458, 146, 41_358, 109, 452, 726, 4_034, 111, 763, 35_412, 5_082, 388, 1_903, 111, 9_051, 391, 2_870, 48_918, 1_900, 1_123, 550, 998, 112, 9_586, 15_985, 455, 391, 410, 22_955, 37_636, 114, 66], [65, 448, 17_496, 419, 3_663, 385, 763, 113, 27_533, 2_870, 3_283, 13_043, 1_639, 24_713, 523, 656, 24_013, 18_550, 2_521, 517, 27_014, 21_244, 420, 1_212, 1_465, 391, 927, 4_833, 388, 578, 11_786, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 484, 2_169, 7_687, 21_932, 18_146, 726, 363, 17_032, 3_391, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_a , model_name="google/bigbird-roberta-base" , revision="215c99f1600e06f83acce68422f2035b2b5c3510" , )
281
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case = { "configuration_biogpt": ["BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BioGptConfig"], "tokenization_biogpt": ["BioGptTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ "BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST", "BioGptForCausalLM", "BioGptForTokenClassification", "BioGptForSequenceClassification", "BioGptModel", "BioGptPreTrainedModel", ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
157
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging snake_case : int = logging.get_logger(__name__) snake_case : List[str] = {"vocab_file": "spiece.model"} snake_case : List[str] = { "vocab_file": { "albert-base-v1": "https://huggingface.co/albert-base-v1/resolve/main/spiece.model", "albert-large-v1": "https://huggingface.co/albert-large-v1/resolve/main/spiece.model", "albert-xlarge-v1": "https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model", "albert-xxlarge-v1": "https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model", "albert-base-v2": "https://huggingface.co/albert-base-v2/resolve/main/spiece.model", "albert-large-v2": "https://huggingface.co/albert-large-v2/resolve/main/spiece.model", "albert-xlarge-v2": "https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model", "albert-xxlarge-v2": "https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model", } } snake_case : Tuple = { "albert-base-v1": 512, "albert-large-v1": 512, "albert-xlarge-v1": 512, "albert-xxlarge-v1": 512, "albert-base-v2": 512, "albert-large-v2": 512, "albert-xlarge-v2": 512, "albert-xxlarge-v2": 512, } snake_case : List[str] = "▁" class _snake_case ( snake_case ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , _a , _a=True , _a=True , _a=False , _a="[CLS]" , _a="[SEP]" , _a="<unk>" , _a="[SEP]" , _a="<pad>" , _a="[CLS]" , _a="[MASK]" , _a = None , **_a , ): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. __magic_name__ : str = ( AddedToken(_a , lstrip=_a , rstrip=_a , normalized=_a ) if isinstance(_a , _a ) else mask_token ) __magic_name__ : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=_a , remove_space=_a , keep_accents=_a , bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , pad_token=_a , cls_token=_a , mask_token=_a , sp_model_kwargs=self.sp_model_kwargs , **_a , ) __magic_name__ : Dict = do_lower_case __magic_name__ : Tuple = remove_space __magic_name__ : Union[str, Any] = keep_accents __magic_name__ : Tuple = vocab_file __magic_name__ : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_a ) @property def SCREAMING_SNAKE_CASE ( self ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[str] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): __magic_name__ : List[str] = self.__dict__.copy() __magic_name__ : Any = None return state def __setstate__( self , _a ): __magic_name__ : Union[str, Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): __magic_name__ : str = {} __magic_name__ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self , _a ): if self.remove_space: __magic_name__ : List[Any] = " ".join(inputs.strip().split() ) else: __magic_name__ : str = inputs __magic_name__ : int = outputs.replace("``" , "\"" ).replace("''" , "\"" ) if not self.keep_accents: __magic_name__ : str = unicodedata.normalize("NFKD" , _a ) __magic_name__ : Tuple = "".join([c for c in outputs if not unicodedata.combining(_a )] ) if self.do_lower_case: __magic_name__ : int = outputs.lower() return outputs def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Optional[Any] = self.preprocess_text(_a ) __magic_name__ : Dict = self.sp_model.encode(_a , out_type=_a ) __magic_name__ : Any = [] for piece in pieces: if len(_a ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): __magic_name__ : Optional[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(_a , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __magic_name__ : List[str] = cur_pieces[1:] else: __magic_name__ : Optional[int] = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(_a ) else: new_pieces.append(_a ) return new_pieces def SCREAMING_SNAKE_CASE ( self , _a ): return self.sp_model.PieceToId(_a ) def SCREAMING_SNAKE_CASE ( self , _a ): return self.sp_model.IdToPiece(_a ) def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Any = [] __magic_name__ : Union[str, Any] = "" __magic_name__ : int = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(_a ) + token __magic_name__ : List[Any] = True __magic_name__ : Optional[int] = [] else: current_sub_tokens.append(_a ) __magic_name__ : Optional[Any] = False out_string += self.sp_model.decode(_a ) return out_string.strip() def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): __magic_name__ : List[str] = [self.sep_token_id] __magic_name__ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self , _a , _a = None , _a = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a , token_ids_a=_a , already_has_special_tokens=_a ) if token_ids_a is not None: return [1] + ([0] * len(_a )) + [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1] def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): __magic_name__ : Optional[int] = [self.sep_token_id] __magic_name__ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): if not os.path.isdir(_a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __magic_name__ : List[str] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _a ) elif not os.path.isfile(self.vocab_file ): with open(_a , "wb" ) as fi: __magic_name__ : Union[str, Any] = self.sp_model.serialized_model_proto() fi.write(_a ) return (out_vocab_file,)
281
0
"""simple docstring""" import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { "vocab_file": "vocab.txt", "merges_file": "bpe.codes", } lowerCamelCase__ = { "vocab_file": { "vinai/phobert-base": "https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt", "vinai/phobert-large": "https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt", }, "merges_file": { "vinai/phobert-base": "https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes", "vinai/phobert-large": "https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes", }, } lowerCamelCase__ = { "vinai/phobert-base": 256, "vinai/phobert-large": 256, } def __lowerCAmelCase (_UpperCamelCase ): __lowerCAmelCase : List[str] = set() __lowerCAmelCase : Any = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowerCAmelCase : int = char __lowerCAmelCase : List[str] = set(_snake_case ) return pairs class A__ ( _lowerCamelCase): A_ : Union[str, Any] = VOCAB_FILES_NAMES A_ : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP A_ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="<s>" , _SCREAMING_SNAKE_CASE="</s>" , _SCREAMING_SNAKE_CASE="</s>" , _SCREAMING_SNAKE_CASE="<s>" , _SCREAMING_SNAKE_CASE="<unk>" , _SCREAMING_SNAKE_CASE="<pad>" , _SCREAMING_SNAKE_CASE="<mask>" , **_SCREAMING_SNAKE_CASE , ): super().__init__( bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , cls_token=_a , pad_token=_a , mask_token=_a , **_a , ) __lowerCAmelCase : Dict = vocab_file __lowerCAmelCase : Tuple = merges_file __lowerCAmelCase : List[Any] = {} __lowerCAmelCase : List[Any] = 0 __lowerCAmelCase : Tuple = 1 __lowerCAmelCase : int = 2 __lowerCAmelCase : Union[str, Any] = 3 self.add_from_file(_a ) __lowerCAmelCase : Optional[int] = {v: k for k, v in self.encoder.items()} with open(_a , encoding='utf-8' ) as merges_handle: __lowerCAmelCase : List[str] = merges_handle.read().split('\n' )[:-1] __lowerCAmelCase : Union[str, Any] = [tuple(merge.split()[:-1] ) for merge in merges] __lowerCAmelCase : Union[str, Any] = dict(zip(_a , range(len(_a ) ) ) ) __lowerCAmelCase : Optional[int] = {} def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __lowerCAmelCase : Optional[Any] = [self.cls_token_id] __lowerCAmelCase : Union[str, Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a , token_ids_a=_a , already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ): __lowerCAmelCase : Optional[Any] = [self.sep_token_id] __lowerCAmelCase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __lowerCamelCase ( self ): return len(self.encoder ) def __lowerCamelCase ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE ): if token in self.cache: return self.cache[token] __lowerCAmelCase : List[Any] = tuple(_a ) __lowerCAmelCase : List[Any] = tuple(list(word[:-1] ) + [word[-1] + '</w>'] ) __lowerCAmelCase : Any = get_pairs(_a ) if not pairs: return token while True: __lowerCAmelCase : str = min(_a , key=lambda _SCREAMING_SNAKE_CASE : self.bpe_ranks.get(_a , float('inf' ) ) ) if bigram not in self.bpe_ranks: break __lowerCAmelCase : List[str] = bigram __lowerCAmelCase : List[str] = [] __lowerCAmelCase : List[str] = 0 while i < len(_a ): try: __lowerCAmelCase : Any = word.index(_a , _a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) __lowerCAmelCase : Tuple = j if word[i] == first and i < len(_a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowerCAmelCase : Union[str, Any] = tuple(_a ) __lowerCAmelCase : Optional[int] = new_word if len(_a ) == 1: break else: __lowerCAmelCase : List[Any] = get_pairs(_a ) __lowerCAmelCase : Optional[int] = "@@ ".join(_a ) __lowerCAmelCase : Tuple = word[:-4] __lowerCAmelCase : str = word return word def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE ): __lowerCAmelCase : Optional[Any] = [] __lowerCAmelCase : Dict = re.findall(R'\S+\n?' , _a ) for token in words: split_tokens.extend(list(self.bpe(_a ).split(' ' ) ) ) return split_tokens def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE ): return self.encoder.get(_a , self.encoder.get(self.unk_token ) ) def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE ): return self.decoder.get(_a , self.unk_token ) def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE ): __lowerCAmelCase : Tuple = " ".join(_a ).replace('@@ ' , '' ).strip() return out_string def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ): if not os.path.isdir(_a ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return __lowerCAmelCase : Optional[int] = os.path.join( _a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) __lowerCAmelCase : Union[str, Any] = os.path.join( _a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ): copyfile(self.vocab_file , _a ) if os.path.abspath(self.merges_file ) != os.path.abspath(_a ): copyfile(self.merges_file , _a ) return out_vocab_file, out_merge_file def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE ): if isinstance(_a , _a ): try: with open(_a , 'r' , encoding='utf-8' ) as fd: self.add_from_file(_a ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(f"Incorrect encoding detected in {f}, please rebuild the dataset" ) return __lowerCAmelCase : List[Any] = f.readlines() for lineTmp in lines: __lowerCAmelCase : Optional[Any] = lineTmp.strip() __lowerCAmelCase : Union[str, Any] = line.rfind(' ' ) if idx == -1: raise ValueError('Incorrect dictionary format, expected \'<token> <cnt>\'' ) __lowerCAmelCase : Optional[int] = line[:idx] __lowerCAmelCase : Dict = len(self.encoder )
86
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : Tuple , _snake_case : List[Any] , _snake_case : Optional[Any] ) -> Optional[Any]: '''simple docstring''' if isinstance(_snake_case , _snake_case ): __magic_name__ : Union[str, Any] = np.full((len(_snake_case ), sequence_length, 2) , _snake_case ) else: __magic_name__ : List[Any] = np.full((len(_snake_case ), sequence_length) , _snake_case ) for i, tensor in enumerate(_snake_case ): if padding_side == "right": if isinstance(_snake_case , _snake_case ): __magic_name__ : Optional[Any] = tensor[:sequence_length] else: __magic_name__ : Union[str, Any] = tensor[:sequence_length] else: if isinstance(_snake_case , _snake_case ): __magic_name__ : List[Any] = tensor[:sequence_length] else: __magic_name__ : Optional[Any] = tensor[:sequence_length] return out_tensor.tolist() def lowerCAmelCase_ ( _snake_case : Optional[int] ) -> Tuple: '''simple docstring''' __magic_name__ : Union[str, Any] = ord(_snake_case ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 123 and cp <= 126): return True __magic_name__ : Any = unicodedata.category(_snake_case ) if cat.startswith("P" ): return True return False @dataclass class _snake_case ( snake_case ): UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = -100 UpperCamelCase__ = "pt" def SCREAMING_SNAKE_CASE ( self , _a ): import torch __magic_name__ : List[str] = "label" if "label" in features[0].keys() else "labels" __magic_name__ : Union[str, Any] = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __magic_name__ : Optional[int] = self.tokenizer.pad( _a , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="pt" if labels is None else None , ) if labels is None: return batch __magic_name__ : Dict = torch.tensor(batch["entity_ids"] ).shape[1] __magic_name__ : List[Any] = self.tokenizer.padding_side if padding_side == "right": __magic_name__ : str = [ list(_a ) + [self.label_pad_token_id] * (sequence_length - len(_a )) for label in labels ] else: __magic_name__ : int = [ [self.label_pad_token_id] * (sequence_length - len(_a )) + list(_a ) for label in labels ] __magic_name__ : Dict = [feature["ner_tags"] for feature in features] __magic_name__ : List[Any] = padding_tensor(_a , -1 , _a , _a ) __magic_name__ : Any = [feature["original_entity_spans"] for feature in features] __magic_name__ : Any = padding_tensor(_a , (-1, -1) , _a , _a ) __magic_name__ : List[Any] = {k: torch.tensor(_a , dtype=torch.intaa ) for k, v in batch.items()} return batch
281
0
import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) __lowerCamelCase : int = pytest.mark.integration @pytest.mark.parametrize("""path""" , ["""paws""", """csv"""] ) def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Dict , __UpperCamelCase : str ) -> Dict: """simple docstring""" inspect_dataset(_snake_case , _snake_case ) SCREAMING_SNAKE_CASE__ = path + ".py" assert script_name in os.listdir(_snake_case ) assert "__pycache__" not in os.listdir(_snake_case ) @pytest.mark.filterwarnings("""ignore:inspect_metric is deprecated:FutureWarning""" ) @pytest.mark.filterwarnings("""ignore:metric_module_factory is deprecated:FutureWarning""" ) @pytest.mark.parametrize("""path""" , ["""accuracy"""] ) def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Dict , __UpperCamelCase : Any ) -> Optional[int]: """simple docstring""" inspect_metric(_snake_case , _snake_case ) SCREAMING_SNAKE_CASE__ = path + ".py" assert script_name in os.listdir(_snake_case ) assert "__pycache__" not in os.listdir(_snake_case ) @pytest.mark.parametrize( """path, config_name, expected_splits""" , [ ("""squad""", """plain_text""", ["""train""", """validation"""]), ("""dalle-mini/wit""", """dalle-mini--wit""", ["""train"""]), ("""paws""", """labeled_final""", ["""train""", """test""", """validation"""]), ] , ) def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Dict , __UpperCamelCase : int , __UpperCamelCase : str ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = get_dataset_config_info(_snake_case , config_name=_snake_case ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( """path, config_name, expected_exception""" , [ ("""paws""", None, ValueError), ] , ) def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : Union[str, Any] , __UpperCamelCase : int ) -> int: """simple docstring""" with pytest.raises(_snake_case ): get_dataset_config_info(_snake_case , config_name=_snake_case ) @pytest.mark.parametrize( """path, expected""" , [ ("""squad""", """plain_text"""), ("""acronym_identification""", """default"""), ("""lhoestq/squad""", """plain_text"""), ("""lhoestq/test""", """default"""), ("""lhoestq/demo1""", """lhoestq--demo1"""), ("""dalle-mini/wit""", """dalle-mini--wit"""), ] , ) def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Dict , __UpperCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = get_dataset_config_names(_snake_case ) assert expected in config_names @pytest.mark.parametrize( """path, expected_configs, expected_splits_in_first_config""" , [ ("""squad""", ["""plain_text"""], ["""train""", """validation"""]), ("""dalle-mini/wit""", ["""dalle-mini--wit"""], ["""train"""]), ("""paws""", ["""labeled_final""", """labeled_swap""", """unlabeled_final"""], ["""train""", """test""", """validation"""]), ] , ) def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : int , __UpperCamelCase : List[str] , __UpperCamelCase : int ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = get_dataset_infos(_snake_case ) assert list(infos.keys() ) == expected_configs SCREAMING_SNAKE_CASE__ = expected_configs[0] assert expected_config in infos SCREAMING_SNAKE_CASE__ = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( """path, expected_config, expected_splits""" , [ ("""squad""", """plain_text""", ["""train""", """validation"""]), ("""dalle-mini/wit""", """dalle-mini--wit""", ["""train"""]), ("""paws""", """labeled_final""", ["""train""", """test""", """validation"""]), ] , ) def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : int , __UpperCamelCase : Dict , __UpperCamelCase : int ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = get_dataset_infos(_snake_case ) assert expected_config in infos SCREAMING_SNAKE_CASE__ = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( """path, config_name, expected_exception""" , [ ("""paws""", None, ValueError), ] , ) def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : List[Any] , __UpperCamelCase : Tuple , __UpperCamelCase : Any ) -> Tuple: """simple docstring""" with pytest.raises(_snake_case ): get_dataset_split_names(_snake_case , config_name=_snake_case )
219
import math def lowerCAmelCase_ ( _snake_case : float , _snake_case : float ) -> float: '''simple docstring''' return math.pow(_snake_case , 2 ) - a def lowerCAmelCase_ ( _snake_case : float ) -> float: '''simple docstring''' return 2 * x def lowerCAmelCase_ ( _snake_case : float ) -> float: '''simple docstring''' __magic_name__ : Optional[int] = 2.0 while start <= a: __magic_name__ : str = math.pow(_snake_case , 2 ) return start def lowerCAmelCase_ ( _snake_case : float , _snake_case : int = 9999 , _snake_case : float = 0.00_000_000_000_001 ) -> float: '''simple docstring''' if a < 0: raise ValueError("math domain error" ) __magic_name__ : Optional[int] = get_initial_point(_snake_case ) for _ in range(_snake_case ): __magic_name__ : int = value __magic_name__ : str = value - fx(_snake_case , _snake_case ) / fx_derivative(_snake_case ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
281
0
'''simple docstring''' def a ( __a ) -> list: '''simple docstring''' if n_term == "": return [] UpperCamelCase__ :list = [] for temp in range(int(_snake_case ) ): series.append(f'''1/{temp + 1}''' if series else '''1''' ) return series if __name__ == "__main__": __snake_case = input('''Enter the last number (nth term) of the Harmonic Series''') print('''Formula of Harmonic Series => 1+1/2+1/3 ..... 1/n''') print(harmonic_series(nth_term))
97
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class _snake_case : UpperCamelCase__ = LEDConfig UpperCamelCase__ = {} UpperCamelCase__ = 'gelu' def __init__( self , _a , _a=13 , _a=7 , _a=True , _a=False , _a=99 , _a=32 , _a=2 , _a=4 , _a=37 , _a=0.1 , _a=0.1 , _a=20 , _a=2 , _a=1 , _a=0 , _a=4 , ): __magic_name__ : int = parent __magic_name__ : Optional[int] = batch_size __magic_name__ : Tuple = seq_length __magic_name__ : List[Any] = is_training __magic_name__ : Dict = use_labels __magic_name__ : Optional[Any] = vocab_size __magic_name__ : int = hidden_size __magic_name__ : Optional[int] = num_hidden_layers __magic_name__ : Optional[int] = num_attention_heads __magic_name__ : Tuple = intermediate_size __magic_name__ : Any = hidden_dropout_prob __magic_name__ : Optional[int] = attention_probs_dropout_prob __magic_name__ : List[str] = max_position_embeddings __magic_name__ : Any = eos_token_id __magic_name__ : str = pad_token_id __magic_name__ : int = bos_token_id __magic_name__ : Optional[int] = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after __magic_name__ : Tuple = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests __magic_name__ : Tuple = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __magic_name__ : int = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __magic_name__ : int = tf.concat([input_ids, eos_tensor] , axis=1 ) __magic_name__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __magic_name__ : Dict = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) __magic_name__ : List[str] = prepare_led_inputs_dict(_a , _a , _a ) __magic_name__ : Union[str, Any] = tf.concat( [tf.zeros_like(_a )[:, :-1], tf.ones_like(_a )[:, -1:]] , axis=-1 , ) __magic_name__ : List[Any] = global_attention_mask return config, inputs_dict def SCREAMING_SNAKE_CASE ( self , _a , _a ): __magic_name__ : Dict = TFLEDModel(config=_a ).get_decoder() __magic_name__ : Optional[int] = inputs_dict["input_ids"] __magic_name__ : Union[str, Any] = input_ids[:1, :] __magic_name__ : str = inputs_dict["attention_mask"][:1, :] __magic_name__ : int = 1 # first forward pass __magic_name__ : Tuple = model(_a , attention_mask=_a , use_cache=_a ) __magic_name__ , __magic_name__ : str = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __magic_name__ : Optional[int] = ids_tensor((self.batch_size, 3) , config.vocab_size ) __magic_name__ : Any = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __magic_name__ : Optional[Any] = tf.concat([input_ids, next_tokens] , axis=-1 ) __magic_name__ : List[Any] = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __magic_name__ : List[str] = model(_a , attention_mask=_a )[0] __magic_name__ : Dict = model(_a , attention_mask=_a , past_key_values=_a )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __magic_name__ : List[Any] = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __magic_name__ : Union[str, Any] = output_from_no_past[:, -3:, random_slice_idx] __magic_name__ : List[str] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_a , _a , rtol=1e-3 ) def lowerCAmelCase_ ( _snake_case : Any , _snake_case : List[Any] , _snake_case : Any , _snake_case : str=None , _snake_case : List[str]=None , _snake_case : int=None , _snake_case : Any=None , ) -> int: '''simple docstring''' if attention_mask is None: __magic_name__ : str = tf.cast(tf.math.not_equal(_snake_case , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: __magic_name__ : List[Any] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: __magic_name__ : Dict = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: __magic_name__ : Any = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class _snake_case ( snake_case , snake_case , unittest.TestCase ): UpperCamelCase__ = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () UpperCamelCase__ = (TFLEDForConditionalGeneration,) if is_tf_available() else () UpperCamelCase__ = ( { 'conversational': TFLEDForConditionalGeneration, 'feature-extraction': TFLEDModel, 'summarization': TFLEDForConditionalGeneration, 'text2text-generation': TFLEDForConditionalGeneration, 'translation': TFLEDForConditionalGeneration, } if is_tf_available() else {} ) UpperCamelCase__ = True UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = TFLEDModelTester(self ) __magic_name__ : List[Any] = ConfigTester(self , config_class=_a ) def SCREAMING_SNAKE_CASE ( self ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_a ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ , __magic_name__ : int = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ : List[str] = tf.zeros_like(inputs_dict["attention_mask"] ) __magic_name__ : Optional[Any] = 2 __magic_name__ : Tuple = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict["global_attention_mask"] , ) __magic_name__ : Any = True __magic_name__ : str = self.model_tester.seq_length __magic_name__ : Dict = self.model_tester.encoder_seq_length def check_decoder_attentions_output(_a ): __magic_name__ : str = outputs.decoder_attentions self.assertEqual(len(_a ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(_a ): __magic_name__ : Any = [t.numpy() for t in outputs.encoder_attentions] __magic_name__ : Tuple = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(_a ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(_a ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: __magic_name__ : Union[str, Any] = True __magic_name__ : List[str] = False __magic_name__ : Tuple = False __magic_name__ : Optional[int] = model_class(_a ) __magic_name__ : str = model(self._prepare_for_class(_a , _a ) ) __magic_name__ : Any = len(_a ) self.assertEqual(config.output_hidden_states , _a ) check_encoder_attentions_output(_a ) if self.is_encoder_decoder: __magic_name__ : Tuple = model_class(_a ) __magic_name__ : Optional[Any] = model(self._prepare_for_class(_a , _a ) ) self.assertEqual(config.output_hidden_states , _a ) check_decoder_attentions_output(_a ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] __magic_name__ : Dict = True __magic_name__ : str = model_class(_a ) __magic_name__ : Any = model(self._prepare_for_class(_a , _a ) ) self.assertEqual(config.output_hidden_states , _a ) check_encoder_attentions_output(_a ) # Check attention is always last and order is fine __magic_name__ : Union[str, Any] = True __magic_name__ : Union[str, Any] = True __magic_name__ : List[str] = model_class(_a ) __magic_name__ : Any = model(self._prepare_for_class(_a , _a ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(_a ) ) self.assertEqual(model.config.output_hidden_states , _a ) check_encoder_attentions_output(_a ) @unittest.skip("LED keeps using potentially symbolic tensors in conditionals and breaks tracing." ) def SCREAMING_SNAKE_CASE ( self ): pass def SCREAMING_SNAKE_CASE ( self ): # TODO: Head-masking not yet implement pass def lowerCAmelCase_ ( _snake_case : int ) -> Optional[int]: '''simple docstring''' return tf.constant(_snake_case , dtype=tf.intaa ) snake_case : Optional[int] = 1E-4 @slow @require_tf class _snake_case ( unittest.TestCase ): def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[Any] = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ).led # change to intended input here __magic_name__ : Optional[int] = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) __magic_name__ : str = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) __magic_name__ : Any = prepare_led_inputs_dict(model.config , _a , _a ) __magic_name__ : List[Any] = model(**_a )[0] __magic_name__ : List[str] = (1, 1_024, 768) self.assertEqual(output.shape , _a ) # change to expected output here __magic_name__ : int = tf.convert_to_tensor( [[2.30_50, 2.82_79, 0.65_31], [-1.84_57, -0.14_55, -3.56_61], [-1.01_86, 0.45_86, -2.20_43]] , ) tf.debugging.assert_near(output[:, :3, :3] , _a , atol=1e-3 ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Tuple = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ) # change to intended input here __magic_name__ : int = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) __magic_name__ : Tuple = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) __magic_name__ : Optional[Any] = prepare_led_inputs_dict(model.config , _a , _a ) __magic_name__ : Union[str, Any] = model(**_a )[0] __magic_name__ : Optional[int] = (1, 1_024, model.config.vocab_size) self.assertEqual(output.shape , _a ) # change to expected output here __magic_name__ : str = tf.convert_to_tensor( [[33.65_07, 6.45_72, 16.80_89], [5.87_39, -2.42_38, 11.29_02], [-3.21_39, -4.31_49, 4.27_83]] , ) tf.debugging.assert_near(output[:, :3, :3] , _a , atol=1e-3 , rtol=1e-3 )
281
0
"""simple docstring""" def _snake_case ( UpperCamelCase : int , UpperCamelCase : int ): UpperCAmelCase : Dict = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): UpperCAmelCase : Any = n - k # Calculate C(n,k) for i in range(_snake_case ): result *= n - i result //= i + 1 return result def _snake_case ( UpperCamelCase : int ): return binomial_coefficient(2 * node_count , _snake_case ) // (node_count + 1) def _snake_case ( UpperCamelCase : int ): if n < 0: raise ValueError("""factorial() not defined for negative values""" ) UpperCAmelCase : Optional[Any] = 1 for i in range(1 , n + 1 ): result *= i return result def _snake_case ( UpperCamelCase : int ): return catalan_number(_snake_case ) * factorial(_snake_case ) if __name__ == "__main__": A: Dict = int(input("Enter the number of nodes: ").strip() or 0) if node_count <= 0: raise ValueError("We need some nodes to work with.") print( f"""Given {node_count} nodes, there are {binary_tree_count(node_count)} """ f"""binary trees and {catalan_number(node_count)} binary search trees.""" )
109
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() snake_case : Optional[Any] = logging.get_logger(__name__) def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Union[str, Any]=False ) -> List[str]: '''simple docstring''' __magic_name__ : Union[str, Any] = [] # fmt: off # stem: rename_keys.append(("cls_token", "vit.embeddings.cls_token") ) rename_keys.append(("pos_embed", "vit.embeddings.position_embeddings") ) rename_keys.append(("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight") ) rename_keys.append(("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias") ) # backbone rename_keys.append(("patch_embed.backbone.stem.conv.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.bias", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias") ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias''') ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''vit.encoder.layer.{i}.output.dense.bias''') ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" __magic_name__ : int = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) # fmt: on return rename_keys def lowerCAmelCase_ ( _snake_case : Any , _snake_case : Any , _snake_case : Dict=False ) -> int: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: __magic_name__ : int = "" else: __magic_name__ : Union[str, Any] = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __magic_name__ : Optional[Any] = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) __magic_name__ : int = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict __magic_name__ : Dict = in_proj_weight[ : config.hidden_size, : ] __magic_name__ : List[str] = in_proj_bias[: config.hidden_size] __magic_name__ : List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __magic_name__ : Optional[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __magic_name__ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] __magic_name__ : int = in_proj_bias[-config.hidden_size :] def lowerCAmelCase_ ( _snake_case : List[str] ) -> List[str]: '''simple docstring''' __magic_name__ : List[str] = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(_snake_case , _snake_case ) def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : int , _snake_case : Union[str, Any] ) -> Optional[int]: '''simple docstring''' __magic_name__ : int = dct.pop(_snake_case ) __magic_name__ : List[Any] = val def lowerCAmelCase_ ( ) -> Dict: '''simple docstring''' __magic_name__ : List[str] = "http://images.cocodataset.org/val2017/000000039769.jpg" __magic_name__ : List[str] = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return im @torch.no_grad() def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : Any , _snake_case : int=False ) -> Dict: '''simple docstring''' __magic_name__ : List[str] = BitConfig( global_padding="same" , layer_type="bottleneck" , depths=(3, 4, 9) , out_features=["stage3"] , embedding_dynamic_padding=_snake_case , ) __magic_name__ : List[str] = ViTHybridConfig(backbone_config=_snake_case , image_size=384 , num_labels=1000 ) __magic_name__ : str = False # load original model from timm __magic_name__ : Union[str, Any] = timm.create_model(_snake_case , pretrained=_snake_case ) timm_model.eval() # load state_dict of original model, remove and rename some keys __magic_name__ : List[Any] = timm_model.state_dict() if base_model: remove_classification_head_(_snake_case ) __magic_name__ : Tuple = create_rename_keys(_snake_case , _snake_case ) for src, dest in rename_keys: rename_key(_snake_case , _snake_case , _snake_case ) read_in_q_k_v(_snake_case , _snake_case , _snake_case ) __magic_name__ : List[str] = "huggingface/label-files" __magic_name__ : int = "imagenet-1k-id2label.json" __magic_name__ : Optional[int] = json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type="dataset" ) , "r" ) ) __magic_name__ : int = {int(_snake_case ): v for k, v in idalabel.items()} __magic_name__ : List[str] = idalabel __magic_name__ : List[str] = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": __magic_name__ : List[str] = ViTHybridModel(_snake_case ).eval() else: __magic_name__ : str = ViTHybridForImageClassification(_snake_case ).eval() model.load_state_dict(_snake_case ) # create image processor __magic_name__ : List[Any] = create_transform(**resolve_data_config({} , model=_snake_case ) ) __magic_name__ : int = transform.transforms __magic_name__ : List[str] = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } __magic_name__ : int = ViTHybridImageProcessor( do_resize=_snake_case , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_snake_case , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=_snake_case , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) __magic_name__ : List[Any] = prepare_img() __magic_name__ : Any = transform(_snake_case ).unsqueeze(0 ) __magic_name__ : Tuple = processor(_snake_case , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(_snake_case , _snake_case ) # verify logits with torch.no_grad(): __magic_name__ : Optional[int] = model(_snake_case ) __magic_name__ : List[str] = outputs.logits print("Predicted class:" , logits.argmax(-1 ).item() ) if base_model: __magic_name__ : List[str] = timm_model.forward_features(_snake_case ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_snake_case , outputs.pooler_output , atol=1E-3 ) else: __magic_name__ : Any = timm_model(_snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_snake_case , outputs.logits , atol=1E-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(_snake_case ).mkdir(exist_ok=_snake_case ) print(F'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(_snake_case ) print(F'''Saving processor to {pytorch_dump_folder_path}''' ) processor.save_pretrained(_snake_case ) if push_to_hub: print(F'''Pushing model and processor to the hub {vit_name}''' ) model.push_to_hub(F'''ybelkada/{vit_name}''' ) processor.push_to_hub(F'''ybelkada/{vit_name}''' ) if __name__ == "__main__": snake_case : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--vit_name", default="vit_base_r50_s16_384", type=str, help="Name of the hybrid ViT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to upload the model to the HuggingFace hub." ) snake_case : List[Any] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
281
0
import numpy as np def __UpperCamelCase ( _A , _A ): return np.where(vector > 0 , _snake_case , (alpha * (np.exp(_snake_case ) - 1)) ) if __name__ == "__main__": import doctest doctest.testmod()
278
# This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny model through reduction of a normal pre-trained model, but keeping the # full vocab, merges file, and thus also resulting in a larger model due to a large vocab size. # This gives ~3MB in total for all files. # # If you want a 50 times smaller than this see `fsmt-make-super-tiny-model.py`, which is slightly more complicated # # # It will be used then as "stas/tiny-wmt19-en-de" # Build from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration snake_case : List[str] = "facebook/wmt19-en-de" snake_case : Dict = FSMTTokenizer.from_pretrained(mname) # get the correct vocab sizes, etc. from the master model snake_case : List[str] = FSMTConfig.from_pretrained(mname) config.update( dict( d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) ) snake_case : int = FSMTForConditionalGeneration(config) print(F"num of params {tiny_model.num_parameters()}") # Test snake_case : Optional[Any] = tokenizer(["Making tiny model"], return_tensors="pt") snake_case : List[str] = tiny_model(**batch) print("test output:", len(outputs.logits[0])) # Save snake_case : Dict = "tiny-wmt19-en-de" tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F"Generated {mname_tiny}") # Upload # transformers-cli upload tiny-wmt19-en-de
281
0
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. __UpperCamelCase = abspath(join(dirname(__file__), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def lowercase (SCREAMING_SNAKE_CASE_ : Tuple ) -> Any: config.addinivalue_line( 'markers' , 'is_pt_tf_cross_test: mark test to run only when PT and TF interactions are tested' ) config.addinivalue_line( 'markers' , 'is_pt_flax_cross_test: mark test to run only when PT and FLAX interactions are tested' ) config.addinivalue_line('markers' , 'is_pipeline_test: mark test to run only when pipelines are tested' ) config.addinivalue_line('markers' , 'is_staging_test: mark test to run only in the staging environment' ) config.addinivalue_line('markers' , 'accelerate_tests: mark test that require accelerate' ) config.addinivalue_line('markers' , 'tool_tests: mark the tool tests that are run on their specific schedule' ) def lowercase (SCREAMING_SNAKE_CASE_ : List[str] ) -> Dict: from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(_snake_case ) def lowercase (SCREAMING_SNAKE_CASE_ : Any ) -> str: from transformers.testing_utils import pytest_terminal_summary_main SCREAMING_SNAKE_CASE = terminalreporter.config.getoption('--make-reports' ) if make_reports: pytest_terminal_summary_main(_snake_case , id=_snake_case ) def lowercase (SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : Tuple ) -> Optional[int]: if exitstatus == 5: SCREAMING_SNAKE_CASE = 0 # Doctest custom flag to ignore output. __UpperCamelCase = doctest.register_optionflag('''IGNORE_RESULT''') __UpperCamelCase = doctest.OutputChecker class lowerCAmelCase ( lowerCamelCase_ ): '''simple docstring''' def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: if IGNORE_RESULT & optionflags: return True return OutputChecker.check_output(self , _a , _a , _a ) __UpperCamelCase = CustomOutputChecker __UpperCamelCase = HfDoctestModule __UpperCamelCase = HfDocTestParser
113
import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s", datefmt="%m/%d/%Y %H:%M:%S", level=logging.INFO ) snake_case : Optional[int] = logging.getLogger(__name__) def lowerCAmelCase_ ( _snake_case : Dict , _snake_case : Union[str, Any] ) -> Tuple: '''simple docstring''' __magic_name__ : List[str] = np.argmax(_snake_case , axis=1 ) return np.sum(outputs == labels ) def lowerCAmelCase_ ( _snake_case : Optional[Any] ) -> Dict: '''simple docstring''' with open(_snake_case , encoding="utf_8" ) as f: __magic_name__ : List[str] = csv.reader(_snake_case ) __magic_name__ : List[Any] = [] next(_snake_case ) # skip the first line for line in tqdm(_snake_case ): output.append((" ".join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def lowerCAmelCase_ ( _snake_case : str , _snake_case : Tuple , _snake_case : Union[str, Any] , _snake_case : List[Any] , _snake_case : Tuple , _snake_case : Optional[int] ) -> int: '''simple docstring''' __magic_name__ : Optional[int] = [] for dataset in encoded_datasets: __magic_name__ : Union[str, Any] = len(_snake_case ) __magic_name__ : Dict = np.zeros((n_batch, 2, input_len) , dtype=np.intaa ) __magic_name__ : List[str] = np.zeros((n_batch, 2) , dtype=np.intaa ) __magic_name__ : Optional[int] = np.full((n_batch, 2, input_len) , fill_value=-100 , dtype=np.intaa ) __magic_name__ : int = np.zeros((n_batch,) , dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(_snake_case ): __magic_name__ : Dict = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __magic_name__ : Dict = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __magic_name__ : str = with_conta __magic_name__ : Tuple = with_conta __magic_name__ : Union[str, Any] = len(_snake_case ) - 1 __magic_name__ : int = len(_snake_case ) - 1 __magic_name__ : Optional[Any] = with_conta __magic_name__ : Optional[Any] = with_conta __magic_name__ : Optional[int] = mc_label __magic_name__ : str = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(_snake_case ) for t in all_inputs ) ) return tensor_datasets def lowerCAmelCase_ ( ) -> List[Any]: '''simple docstring''' __magic_name__ : Any = argparse.ArgumentParser() parser.add_argument("--model_name" , type=_snake_case , default="openai-gpt" , help="pretrained model name" ) parser.add_argument("--do_train" , action="store_true" , help="Whether to run training." ) parser.add_argument("--do_eval" , action="store_true" , help="Whether to run eval on the dev set." ) parser.add_argument( "--output_dir" , default=_snake_case , type=_snake_case , required=_snake_case , help="The output directory where the model predictions and checkpoints will be written." , ) parser.add_argument("--train_dataset" , type=_snake_case , default="" ) parser.add_argument("--eval_dataset" , type=_snake_case , default="" ) parser.add_argument("--seed" , type=_snake_case , default=42 ) parser.add_argument("--num_train_epochs" , type=_snake_case , default=3 ) parser.add_argument("--train_batch_size" , type=_snake_case , default=8 ) parser.add_argument("--eval_batch_size" , type=_snake_case , default=16 ) parser.add_argument("--adam_epsilon" , default=1E-8 , type=_snake_case , help="Epsilon for Adam optimizer." ) parser.add_argument("--max_grad_norm" , type=_snake_case , default=1 ) parser.add_argument( "--max_steps" , default=-1 , type=_snake_case , help=( "If > 0: set total number of training steps to perform. Override num_train_epochs." ) , ) parser.add_argument( "--gradient_accumulation_steps" , type=_snake_case , default=1 , help="Number of updates steps to accumulate before performing a backward/update pass." , ) parser.add_argument("--learning_rate" , type=_snake_case , default=6.25E-5 ) parser.add_argument("--warmup_steps" , default=0 , type=_snake_case , help="Linear warmup over warmup_steps." ) parser.add_argument("--lr_schedule" , type=_snake_case , default="warmup_linear" ) parser.add_argument("--weight_decay" , type=_snake_case , default=0.01 ) parser.add_argument("--lm_coef" , type=_snake_case , default=0.9 ) parser.add_argument("--n_valid" , type=_snake_case , default=374 ) parser.add_argument("--server_ip" , type=_snake_case , default="" , help="Can be used for distant debugging." ) parser.add_argument("--server_port" , type=_snake_case , default="" , help="Can be used for distant debugging." ) __magic_name__ : List[Any] = parser.parse_args() print(_snake_case ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print("Waiting for debugger attach" ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=_snake_case ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) __magic_name__ : Dict = torch.device("cuda" if torch.cuda.is_available() else "cpu" ) __magic_name__ : Optional[int] = torch.cuda.device_count() logger.info("device: {}, n_gpu {}".format(_snake_case , _snake_case ) ) if not args.do_train and not args.do_eval: raise ValueError("At least one of `do_train` or `do_eval` must be True." ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset __magic_name__ : List[Any] = ["_start_", "_delimiter_", "_classify_"] __magic_name__ : Optional[int] = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(_snake_case ) __magic_name__ : Optional[Any] = tokenizer.convert_tokens_to_ids(_snake_case ) __magic_name__ : List[str] = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(_snake_case ) ) model.to(_snake_case ) # Load and encode the datasets def tokenize_and_encode(_snake_case : str ): if isinstance(_snake_case , _snake_case ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(_snake_case ) ) elif isinstance(_snake_case , _snake_case ): return obj return [tokenize_and_encode(_snake_case ) for o in obj] logger.info("Encoding dataset..." ) __magic_name__ : Optional[int] = load_rocstories_dataset(args.train_dataset ) __magic_name__ : str = load_rocstories_dataset(args.eval_dataset ) __magic_name__ : int = (train_dataset, eval_dataset) __magic_name__ : List[str] = tokenize_and_encode(_snake_case ) # Compute the max input length for the Transformer __magic_name__ : Optional[Any] = model.config.n_positions // 2 - 2 __magic_name__ : Optional[int] = max( len(story[:max_length] ) + max(len(conta[:max_length] ) , len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) __magic_name__ : List[str] = min(_snake_case , model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders __magic_name__ : List[Any] = pre_process_datasets(_snake_case , _snake_case , _snake_case , *_snake_case ) __magic_name__ , __magic_name__ : Optional[int] = tensor_datasets[0], tensor_datasets[1] __magic_name__ : Tuple = TensorDataset(*_snake_case ) __magic_name__ : Union[str, Any] = RandomSampler(_snake_case ) __magic_name__ : Dict = DataLoader(_snake_case , sampler=_snake_case , batch_size=args.train_batch_size ) __magic_name__ : Any = TensorDataset(*_snake_case ) __magic_name__ : Optional[Any] = SequentialSampler(_snake_case ) __magic_name__ : int = DataLoader(_snake_case , sampler=_snake_case , batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: __magic_name__ : Tuple = args.max_steps __magic_name__ : List[str] = args.max_steps // (len(_snake_case ) // args.gradient_accumulation_steps) + 1 else: __magic_name__ : List[str] = len(_snake_case ) // args.gradient_accumulation_steps * args.num_train_epochs __magic_name__ : str = list(model.named_parameters() ) __magic_name__ : Dict = ["bias", "LayerNorm.bias", "LayerNorm.weight"] __magic_name__ : str = [ { "params": [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], "weight_decay": args.weight_decay, }, {"params": [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], "weight_decay": 0.0}, ] __magic_name__ : str = AdamW(_snake_case , lr=args.learning_rate , eps=args.adam_epsilon ) __magic_name__ : List[str] = get_linear_schedule_with_warmup( _snake_case , num_warmup_steps=args.warmup_steps , num_training_steps=_snake_case ) if args.do_train: __magic_name__ , __magic_name__ , __magic_name__ : Union[str, Any] = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) , desc="Epoch" ): __magic_name__ : List[str] = 0 __magic_name__ : Tuple = 0 __magic_name__ : Dict = tqdm(_snake_case , desc="Training" ) for step, batch in enumerate(_snake_case ): __magic_name__ : Optional[Any] = tuple(t.to(_snake_case ) for t in batch ) __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Dict = batch __magic_name__ : Optional[Any] = model(_snake_case , mc_token_ids=_snake_case , lm_labels=_snake_case , mc_labels=_snake_case ) __magic_name__ : Optional[Any] = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() __magic_name__ : List[str] = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 __magic_name__ : int = "Training loss: {:.2e} lr: {:.2e}".format(_snake_case , scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer __magic_name__ : Dict = model.module if hasattr(_snake_case , "module" ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` __magic_name__ : List[Any] = os.path.join(args.output_dir , _snake_case ) __magic_name__ : Dict = os.path.join(args.output_dir , _snake_case ) torch.save(model_to_save.state_dict() , _snake_case ) model_to_save.config.to_json_file(_snake_case ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned __magic_name__ : Dict = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) __magic_name__ : Optional[int] = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(_snake_case ) if args.do_eval: model.eval() __magic_name__ , __magic_name__ : Any = 0, 0 __magic_name__ , __magic_name__ : Union[str, Any] = 0, 0 for batch in tqdm(_snake_case , desc="Evaluating" ): __magic_name__ : int = tuple(t.to(_snake_case ) for t in batch ) __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Union[str, Any] = batch with torch.no_grad(): __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Dict = model( _snake_case , mc_token_ids=_snake_case , lm_labels=_snake_case , mc_labels=_snake_case ) __magic_name__ : Tuple = mc_logits.detach().cpu().numpy() __magic_name__ : Any = mc_labels.to("cpu" ).numpy() __magic_name__ : str = accuracy(_snake_case , _snake_case ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 __magic_name__ : Tuple = eval_loss / nb_eval_steps __magic_name__ : List[Any] = eval_accuracy / nb_eval_examples __magic_name__ : int = tr_loss / nb_tr_steps if args.do_train else None __magic_name__ : Any = {"eval_loss": eval_loss, "eval_accuracy": eval_accuracy, "train_loss": train_loss} __magic_name__ : int = os.path.join(args.output_dir , "eval_results.txt" ) with open(_snake_case , "w" ) as writer: logger.info("***** Eval results *****" ) for key in sorted(result.keys() ): logger.info(" %s = %s" , _snake_case , str(result[key] ) ) writer.write("%s = %s\n" % (key, str(result[key] )) ) if __name__ == "__main__": main()
281
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowercase__ = {"configuration_fnet": ["FNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "FNetConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["FNetTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["FNetTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ "FNET_PRETRAINED_MODEL_ARCHIVE_LIST", "FNetForMaskedLM", "FNetForMultipleChoice", "FNetForNextSentencePrediction", "FNetForPreTraining", "FNetForQuestionAnswering", "FNetForSequenceClassification", "FNetForTokenClassification", "FNetLayer", "FNetModel", "FNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_fnet import FNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet import FNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet_fast import FNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_fnet import ( FNET_PRETRAINED_MODEL_ARCHIVE_LIST, FNetForMaskedLM, FNetForMultipleChoice, FNetForNextSentencePrediction, FNetForPreTraining, FNetForQuestionAnswering, FNetForSequenceClassification, FNetForTokenClassification, FNetLayer, FNetModel, FNetPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
281
0
"""simple docstring""" import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class lowerCAmelCase_ ( lowerCAmelCase ): """simple docstring""" def snake_case ( self ): """simple docstring""" snake_case = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_a , 'hidden_sizes' ) ) self.parent.assertTrue(hasattr(_a , 'num_attention_heads' ) ) self.parent.assertTrue(hasattr(_a , 'num_encoder_blocks' ) ) class lowerCAmelCase_ : """simple docstring""" def __init__( self , lowerCAmelCase , lowerCAmelCase=13 , lowerCAmelCase=64 , lowerCAmelCase=3 , lowerCAmelCase=4 , lowerCAmelCase=[2, 2, 2, 2] , lowerCAmelCase=[8, 4, 2, 1] , lowerCAmelCase=[16, 32, 64, 1_28] , lowerCAmelCase=[1, 4, 8, 16] , lowerCAmelCase=[1, 2, 4, 8] , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=0.02 , lowerCAmelCase=3 , lowerCAmelCase=None , ): """simple docstring""" snake_case = parent snake_case = batch_size snake_case = image_size snake_case = num_channels snake_case = num_encoder_blocks snake_case = sr_ratios snake_case = depths snake_case = hidden_sizes snake_case = downsampling_rates snake_case = num_attention_heads snake_case = is_training snake_case = use_labels snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = initializer_range snake_case = num_labels snake_case = scope def snake_case ( self ): """simple docstring""" snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case = None if self.use_labels: snake_case = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) snake_case = self.get_config() return config, pixel_values, labels def snake_case ( self ): """simple docstring""" return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def snake_case ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" snake_case = SegformerModel(config=_a ) model.to(_a ) model.eval() snake_case = model(_a ) snake_case = self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def snake_case ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" snake_case = self.num_labels snake_case = SegformerForSemanticSegmentation(_a ) model.to(_a ) model.eval() snake_case = model(_a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) snake_case = model(_a , labels=_a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss , 0.0 ) def snake_case ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" snake_case = 1 snake_case = SegformerForSemanticSegmentation(config=_a ) model.to(_a ) model.eval() snake_case = torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size) ).to(_a ) snake_case = model(_a , labels=_a ) self.parent.assertGreater(result.loss , 0.0 ) def snake_case ( self ): """simple docstring""" snake_case = self.prepare_config_and_inputs() snake_case = config_and_inputs snake_case = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase_ ( lowerCAmelCase , lowerCAmelCase , unittest.TestCase ): """simple docstring""" _lowerCAmelCase : int = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) _lowerCAmelCase : Dict = ( { """feature-extraction""": SegformerModel, """image-classification""": SegformerForImageClassification, """image-segmentation""": SegformerForSemanticSegmentation, } if is_torch_available() else {} ) _lowerCAmelCase : List[Any] = True _lowerCAmelCase : Tuple = False _lowerCAmelCase : str = False _lowerCAmelCase : List[str] = False def snake_case ( self ): """simple docstring""" snake_case = SegformerModelTester(self ) snake_case = SegformerConfigTester(self , config_class=_a ) def snake_case ( self ): """simple docstring""" self.config_tester.run_common_tests() def snake_case ( self ): """simple docstring""" snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def snake_case ( self ): """simple docstring""" snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*_a ) def snake_case ( self ): """simple docstring""" snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*_a ) @unittest.skip('SegFormer does not use inputs_embeds' ) def snake_case ( self ): """simple docstring""" pass @unittest.skip('SegFormer does not have get_input_embeddings method and get_output_embeddings methods' ) def snake_case ( self ): """simple docstring""" pass def snake_case ( self ): """simple docstring""" snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(_a ) snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case = [*signature.parameters.keys()] snake_case = ["pixel_values"] self.assertListEqual(arg_names[:1] , _a ) def snake_case ( self ): """simple docstring""" snake_case = self.model_tester.prepare_config_and_inputs_for_common() snake_case = True for model_class in self.all_model_classes: snake_case = True snake_case = False snake_case = True snake_case = model_class(_a ) model.to(_a ) model.eval() with torch.no_grad(): snake_case = model(**self._prepare_for_class(_a , _a ) ) snake_case = outputs.attentions snake_case = sum(self.model_tester.depths ) self.assertEqual(len(_a ) , _a ) # check that output_attentions also work using config del inputs_dict["output_attentions"] snake_case = True snake_case = model_class(_a ) model.to(_a ) model.eval() with torch.no_grad(): snake_case = model(**self._prepare_for_class(_a , _a ) ) snake_case = outputs.attentions self.assertEqual(len(_a ) , _a ) # verify the first attentions (first block, first layer) snake_case = (self.model_tester.image_size // 4) ** 2 snake_case = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) snake_case = (self.model_tester.image_size // 32) ** 2 snake_case = (self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) snake_case = len(_a ) # Check attention is always last and order is fine snake_case = True snake_case = True snake_case = model_class(_a ) model.to(_a ) model.eval() with torch.no_grad(): snake_case = model(**self._prepare_for_class(_a , _a ) ) self.assertEqual(out_len + 1 , len(_a ) ) snake_case = outputs.attentions self.assertEqual(len(_a ) , _a ) # verify the first attentions (first block, first layer) snake_case = (self.model_tester.image_size // 4) ** 2 snake_case = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def snake_case ( self ): """simple docstring""" def check_hidden_states_output(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): snake_case = model_class(_a ) model.to(_a ) model.eval() with torch.no_grad(): snake_case = model(**self._prepare_for_class(_a , _a ) ) snake_case = outputs.hidden_states snake_case = self.model_tester.num_encoder_blocks self.assertEqual(len(_a ) , _a ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = True check_hidden_states_output(_a , _a , _a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case = True check_hidden_states_output(_a , _a , _a ) def snake_case ( self ): """simple docstring""" if not self.model_tester.is_training: return snake_case = self.model_tester.prepare_config_and_inputs_for_common() snake_case = True for model_class in self.all_model_classes: if model_class in get_values(_a ): continue snake_case = model_class(_a ) model.to(_a ) model.train() snake_case = self._prepare_for_class(_a , _a , return_labels=_a ) snake_case = model(**_a ).loss loss.backward() @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def snake_case ( self ): """simple docstring""" pass @slow def snake_case ( self ): """simple docstring""" for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case = SegformerModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def lowerCAmelCase__ ( ) -> Tuple: """simple docstring""" snake_case = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch class lowerCAmelCase_ ( unittest.TestCase ): """simple docstring""" @slow def snake_case ( self ): """simple docstring""" snake_case = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=_a , align=_a , do_random_crop=_a ) snake_case = SegformerForSemanticSegmentation.from_pretrained('nvidia/segformer-b0-finetuned-ade-512-512' ).to( _a ) snake_case = prepare_img() snake_case = image_processor(images=_a , return_tensors='pt' ) snake_case = encoded_inputs.pixel_values.to(_a ) with torch.no_grad(): snake_case = model(_a ) snake_case = torch.Size((1, model.config.num_labels, 1_28, 1_28) ) self.assertEqual(outputs.logits.shape , _a ) snake_case = torch.tensor( [ [[-4.63_10, -5.52_32, -6.23_56], [-5.19_21, -6.14_44, -6.59_96], [-5.44_24, -6.27_90, -6.75_74]], [[-12.13_91, -13.31_22, -13.95_54], [-12.87_32, -13.93_52, -14.35_63], [-12.94_38, -13.82_26, -14.25_13]], [[-12.51_34, -13.46_86, -14.49_15], [-12.86_69, -14.43_43, -14.77_58], [-13.25_23, -14.58_19, -15.06_94]], ] ).to(_a ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , _a , atol=1E-4 ) ) @slow def snake_case ( self ): """simple docstring""" snake_case = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=_a , align=_a , do_random_crop=_a ) snake_case = SegformerForSemanticSegmentation.from_pretrained( 'nvidia/segformer-b1-finetuned-cityscapes-1024-1024' ).to(_a ) snake_case = prepare_img() snake_case = image_processor(images=_a , return_tensors='pt' ) snake_case = encoded_inputs.pixel_values.to(_a ) with torch.no_grad(): snake_case = model(_a ) snake_case = torch.Size((1, model.config.num_labels, 1_28, 1_28) ) self.assertEqual(outputs.logits.shape , _a ) snake_case = torch.tensor( [ [[-13.57_48, -13.91_11, -12.65_00], [-14.35_00, -15.36_83, -14.23_28], [-14.75_32, -16.04_24, -15.60_87]], [[-17.16_51, -15.87_25, -12.96_53], [-17.25_80, -17.37_18, -14.82_23], [-16.60_58, -16.87_83, -16.74_52]], [[-3.64_56, -3.02_09, -1.42_03], [-3.07_97, -3.19_59, -2.00_00], [-1.87_57, -1.92_17, -1.69_97]], ] ).to(_a ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , _a , atol=1E-1 ) ) @slow def snake_case ( self ): """simple docstring""" snake_case = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=_a , align=_a , do_random_crop=_a ) snake_case = SegformerForSemanticSegmentation.from_pretrained('nvidia/segformer-b0-finetuned-ade-512-512' ).to( _a ) snake_case = prepare_img() snake_case = image_processor(images=_a , return_tensors='pt' ) snake_case = encoded_inputs.pixel_values.to(_a ) with torch.no_grad(): snake_case = model(_a ) snake_case = outputs.logits.detach().cpu() snake_case = image_processor.post_process_semantic_segmentation(outputs=_a , target_sizes=[(5_00, 3_00)] ) snake_case = torch.Size((5_00, 3_00) ) self.assertEqual(segmentation[0].shape , _a ) snake_case = image_processor.post_process_semantic_segmentation(outputs=_a ) snake_case = torch.Size((1_28, 1_28) ) self.assertEqual(segmentation[0].shape , _a )
150
import importlib import os import fsspec import pytest from fsspec import register_implementation from fsspec.registry import _registry as _fsspec_registry from datasets.filesystems import COMPRESSION_FILESYSTEMS, HfFileSystem, extract_path_from_uri, is_remote_filesystem from .utils import require_lza, require_zstandard def lowerCAmelCase_ ( _snake_case : List[Any] ) -> List[Any]: '''simple docstring''' assert "mock" in _fsspec_registry assert "bz2" in _fsspec_registry def lowerCAmelCase_ ( ) -> Tuple: '''simple docstring''' assert "mock" not in _fsspec_registry assert "bz2" in _fsspec_registry def lowerCAmelCase_ ( ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : Dict = "mock-s3-bucket" __magic_name__ : Any = F'''s3://{mock_bucket}''' __magic_name__ : str = extract_path_from_uri(_snake_case ) assert dataset_path.startswith("s3://" ) is False __magic_name__ : Tuple = "./local/path" __magic_name__ : Optional[Any] = extract_path_from_uri(_snake_case ) assert dataset_path == new_dataset_path def lowerCAmelCase_ ( _snake_case : List[str] ) -> Optional[Any]: '''simple docstring''' __magic_name__ : str = is_remote_filesystem(_snake_case ) assert is_remote is True __magic_name__ : Optional[int] = fsspec.filesystem("file" ) __magic_name__ : int = is_remote_filesystem(_snake_case ) assert is_remote is False @pytest.mark.parametrize("compression_fs_class" , _snake_case ) def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : Optional[Any] , _snake_case : int , _snake_case : Tuple , _snake_case : Any , _snake_case : Union[str, Any] , _snake_case : Any ) -> int: '''simple docstring''' __magic_name__ : Any = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_file, "bz2": bza_file, "lz4": lza_file} __magic_name__ : str = input_paths[compression_fs_class.protocol] if input_path is None: __magic_name__ : Dict = F'''for \'{compression_fs_class.protocol}\' compression protocol, ''' if compression_fs_class.protocol == "lz4": reason += require_lza.kwargs["reason"] elif compression_fs_class.protocol == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(_snake_case ) __magic_name__ : str = fsspec.filesystem(compression_fs_class.protocol , fo=_snake_case ) assert isinstance(_snake_case , _snake_case ) __magic_name__ : int = os.path.basename(_snake_case ) __magic_name__ : Optional[int] = expected_filename[: expected_filename.rindex("." )] assert fs.glob("*" ) == [expected_filename] with fs.open(_snake_case , "r" , encoding="utf-8" ) as f, open(_snake_case , encoding="utf-8" ) as expected_file: assert f.read() == expected_file.read() @pytest.mark.parametrize("protocol" , ["zip", "gzip"] ) def lowerCAmelCase_ ( _snake_case : List[Any] , _snake_case : Optional[Any] , _snake_case : Optional[Any] ) -> str: '''simple docstring''' __magic_name__ : int = {"zip": zip_jsonl_path, "gzip": jsonl_gz_path} __magic_name__ : int = compressed_file_paths[protocol] __magic_name__ : Tuple = "dataset.jsonl" __magic_name__ : List[str] = F'''{protocol}://{member_file_path}::{compressed_file_path}''' __magic_name__ , *__magic_name__ : Optional[Any] = fsspec.get_fs_token_paths(_snake_case ) assert fs.isfile(_snake_case ) assert not fs.isfile("non_existing_" + member_file_path ) @pytest.mark.integration def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Dict , _snake_case : List[str] , _snake_case : Tuple ) -> str: '''simple docstring''' __magic_name__ : int = hf_api.dataset_info(_snake_case , token=_snake_case ) __magic_name__ : Optional[Any] = HfFileSystem(repo_info=_snake_case , token=_snake_case ) assert sorted(hffs.glob("*" ) ) == [".gitattributes", "data"] assert hffs.isdir("data" ) assert hffs.isfile(".gitattributes" ) and hffs.isfile("data/text_data.txt" ) with open(_snake_case ) as f: assert hffs.open("data/text_data.txt" , "r" ).read() == f.read() def lowerCAmelCase_ ( ) -> Optional[int]: '''simple docstring''' __magic_name__ : Optional[Any] = "bz2" # Import module import datasets.filesystems # Overwrite protocol and reload register_implementation(_snake_case , _snake_case , clobber=_snake_case ) with pytest.warns(_snake_case ) as warning_info: importlib.reload(datasets.filesystems ) assert len(_snake_case ) == 1 assert ( str(warning_info[0].message ) == F'''A filesystem protocol was already set for {protocol} and will be overwritten.''' )
281
0
"""simple docstring""" import gc import threading import time import psutil import torch class lowerCamelCase__ : """simple docstring""" def __init__( self : Union[str, Any] ): '''simple docstring''' __UpperCAmelCase : List[Any] = psutil.Process() __UpperCAmelCase : Union[str, Any] = False def lowerCamelCase__ ( self : int ): '''simple docstring''' __UpperCAmelCase : List[Any] = -1 while True: __UpperCAmelCase : Tuple = max(self.process.memory_info().rss , self.cpu_memory_peak ) # can't sleep or will not catch the peak right (this comment is here on purpose) if not self.peak_monitoring: break def lowerCamelCase__ ( self : Union[str, Any] ): '''simple docstring''' __UpperCAmelCase : Dict = True __UpperCAmelCase : Optional[int] = threading.Thread(target=self.peak_monitor ) __UpperCAmelCase : List[Any] = True self.thread.start() def lowerCamelCase__ ( self : List[Any] ): '''simple docstring''' __UpperCAmelCase : int = False self.thread.join() return self.cpu_memory_peak UpperCAmelCase : Union[str, Any] = PeakCPUMemory() def lowerCamelCase ( ) -> Union[str, Any]: '''simple docstring''' __UpperCAmelCase : Tuple = {"time": time.time()} gc.collect() torch.cuda.empty_cache() # CPU mem __UpperCAmelCase : Optional[int] = psutil.Process().memory_info().rss cpu_peak_tracker.start() # GPU mem for i in range(torch.cuda.device_count() ): __UpperCAmelCase : Dict = torch.cuda.memory_allocated(_snake_case ) torch.cuda.reset_peak_memory_stats() return measures def lowerCamelCase ( _UpperCamelCase : str ) -> Optional[Any]: '''simple docstring''' __UpperCAmelCase : int = {"time": time.time() - start_measures["time"]} gc.collect() torch.cuda.empty_cache() # CPU mem __UpperCAmelCase : Optional[Any] = (psutil.Process().memory_info().rss - start_measures["cpu"]) / 2**2_0 __UpperCAmelCase : str = (cpu_peak_tracker.stop() - start_measures["cpu"]) / 2**2_0 # GPU mem for i in range(torch.cuda.device_count() ): __UpperCAmelCase : Optional[int] = (torch.cuda.memory_allocated(_snake_case ) - start_measures[str(_snake_case )]) / 2**2_0 __UpperCAmelCase : Optional[Any] = (torch.cuda.max_memory_allocated(_snake_case ) - start_measures[str(_snake_case )]) / 2**2_0 return measures def lowerCamelCase ( _UpperCamelCase : str , _UpperCamelCase : Dict ) -> str: '''simple docstring''' print(f'''{description}:''' ) print(f'''- Time: {measures["time"]:.2f}s''' ) for i in range(torch.cuda.device_count() ): print(f'''- GPU {i} allocated: {measures[str(_snake_case )]:.2f}MiB''' ) __UpperCAmelCase : List[Any] = measures[f'''{i}-peak'''] print(f'''- GPU {i} peak: {peak:.2f}MiB''' ) print(f'''- CPU RAM allocated: {measures["cpu"]:.2f}MiB''' ) print(f'''- CPU RAM peak: {measures["cpu-peak"]:.2f}MiB''' )
115
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case : Dict = logging.get_logger(__name__) snake_case : List[Any] = { "YituTech/conv-bert-base": "https://huggingface.co/YituTech/conv-bert-base/resolve/main/config.json", "YituTech/conv-bert-medium-small": ( "https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/config.json" ), "YituTech/conv-bert-small": "https://huggingface.co/YituTech/conv-bert-small/resolve/main/config.json", # See all ConvBERT models at https://huggingface.co/models?filter=convbert } class _snake_case ( snake_case ): UpperCamelCase__ = 'convbert' def __init__( self , _a=30_522 , _a=768 , _a=12 , _a=12 , _a=3_072 , _a="gelu" , _a=0.1 , _a=0.1 , _a=512 , _a=2 , _a=0.02 , _a=1e-12 , _a=1 , _a=0 , _a=2 , _a=768 , _a=2 , _a=9 , _a=1 , _a=None , **_a , ): super().__init__( pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , **_a , ) __magic_name__ : Tuple = vocab_size __magic_name__ : List[Any] = hidden_size __magic_name__ : Union[str, Any] = num_hidden_layers __magic_name__ : List[Any] = num_attention_heads __magic_name__ : str = intermediate_size __magic_name__ : Any = hidden_act __magic_name__ : List[Any] = hidden_dropout_prob __magic_name__ : Optional[int] = attention_probs_dropout_prob __magic_name__ : Tuple = max_position_embeddings __magic_name__ : str = type_vocab_size __magic_name__ : List[str] = initializer_range __magic_name__ : Tuple = layer_norm_eps __magic_name__ : List[Any] = embedding_size __magic_name__ : List[Any] = head_ratio __magic_name__ : str = conv_kernel_size __magic_name__ : Dict = num_groups __magic_name__ : str = classifier_dropout class _snake_case ( snake_case ): @property def SCREAMING_SNAKE_CASE ( self ): if self.task == "multiple-choice": __magic_name__ : Dict = {0: "batch", 1: "choice", 2: "sequence"} else: __magic_name__ : Dict = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis), ] )
281
0
def _a ( SCREAMING_SNAKE_CASE : Optional[int]=2_81_23 ) -> Optional[Any]: """simple docstring""" __lowerCAmelCase: Optional[Any] = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i __lowerCAmelCase: int = set() __lowerCAmelCase: List[str] = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(_snake_case ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
322
import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipaConfig, BlipaForConditionalGeneration, BlipaProcessor, BlipaVisionConfig, BlipImageProcessor, OPTConfig, TaConfig, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def lowerCAmelCase_ ( ) -> str: '''simple docstring''' __magic_name__ : int = "https://storage.googleapis.com/sfr-vision-language-research/LAVIS/assets/merlion.png" __magic_name__ : Union[str, Any] = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ).convert("RGB" ) return image def lowerCAmelCase_ ( _snake_case : str ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : List[str] = [] # fmt: off # vision encoder rename_keys.append(("visual_encoder.cls_token", "vision_model.embeddings.class_embedding") ) rename_keys.append(("visual_encoder.pos_embed", "vision_model.embeddings.position_embedding") ) rename_keys.append(("visual_encoder.patch_embed.proj.weight", "vision_model.embeddings.patch_embedding.weight") ) rename_keys.append(("visual_encoder.patch_embed.proj.bias", "vision_model.embeddings.patch_embedding.bias") ) rename_keys.append(("ln_vision.weight", "vision_model.post_layernorm.weight") ) rename_keys.append(("ln_vision.bias", "vision_model.post_layernorm.bias") ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((F'''visual_encoder.blocks.{i}.norm1.weight''', F'''vision_model.encoder.layers.{i}.layer_norm1.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm1.bias''', F'''vision_model.encoder.layers.{i}.layer_norm1.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm2.weight''', F'''vision_model.encoder.layers.{i}.layer_norm2.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm2.bias''', F'''vision_model.encoder.layers.{i}.layer_norm2.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.qkv.weight''', F'''vision_model.encoder.layers.{i}.self_attn.qkv.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.proj.weight''', F'''vision_model.encoder.layers.{i}.self_attn.projection.weight''',) ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.proj.bias''', F'''vision_model.encoder.layers.{i}.self_attn.projection.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc1.weight''', F'''vision_model.encoder.layers.{i}.mlp.fc1.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc1.bias''', F'''vision_model.encoder.layers.{i}.mlp.fc1.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc2.weight''', F'''vision_model.encoder.layers.{i}.mlp.fc2.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc2.bias''', F'''vision_model.encoder.layers.{i}.mlp.fc2.bias''') ) # QFormer rename_keys.append(("Qformer.bert.embeddings.LayerNorm.weight", "qformer.layernorm.weight") ) rename_keys.append(("Qformer.bert.embeddings.LayerNorm.bias", "qformer.layernorm.bias") ) # fmt: on return rename_keys def lowerCAmelCase_ ( _snake_case : Any , _snake_case : Union[str, Any] , _snake_case : Optional[Any] ) -> int: '''simple docstring''' __magic_name__ : Tuple = dct.pop(_snake_case ) __magic_name__ : int = val def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : Optional[Any] ) -> Dict: '''simple docstring''' for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases __magic_name__ : List[Any] = state_dict.pop(F'''visual_encoder.blocks.{i}.attn.q_bias''' ) __magic_name__ : Optional[Any] = state_dict.pop(F'''visual_encoder.blocks.{i}.attn.v_bias''' ) # next, set bias in the state dict __magic_name__ : Optional[int] = torch.cat((q_bias, torch.zeros_like(_snake_case , requires_grad=_snake_case ), v_bias) ) __magic_name__ : Union[str, Any] = qkv_bias def lowerCAmelCase_ ( _snake_case : Dict , _snake_case : str ) -> int: '''simple docstring''' __magic_name__ : List[Any] = 364 if "coco" in model_name else 224 __magic_name__ : Union[str, Any] = BlipaVisionConfig(image_size=_snake_case ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "opt-2.7b" in model_name: __magic_name__ : List[str] = OPTConfig.from_pretrained("facebook/opt-2.7b" , eos_token_id=_snake_case ).to_dict() elif "opt-6.7b" in model_name: __magic_name__ : Any = OPTConfig.from_pretrained("facebook/opt-6.7b" , eos_token_id=_snake_case ).to_dict() elif "t5-xl" in model_name: __magic_name__ : Dict = TaConfig.from_pretrained("google/flan-t5-xl" , dense_act_fn="gelu" , bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: __magic_name__ : int = TaConfig.from_pretrained("google/flan-t5-xxl" , dense_act_fn="gelu" , bos_token_id=1 ).to_dict() __magic_name__ : List[Any] = BlipaConfig(vision_config=_snake_case , text_config=_snake_case ) return config, image_size @torch.no_grad() def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : str=None , _snake_case : Dict=False ) -> List[Any]: '''simple docstring''' __magic_name__ : Optional[int] = ( AutoTokenizer.from_pretrained("facebook/opt-2.7b" ) if "opt" in model_name else AutoTokenizer.from_pretrained("google/flan-t5-xl" ) ) __magic_name__ : List[Any] = tokenizer("\n" , add_special_tokens=_snake_case ).input_ids[0] __magic_name__ , __magic_name__ : Tuple = get_blipa_config(_snake_case , eos_token_id=_snake_case ) __magic_name__ : Union[str, Any] = BlipaForConditionalGeneration(_snake_case ).eval() __magic_name__ : Any = { "blip2-opt-2.7b": ("blip2_opt", "pretrain_opt2.7b"), "blip2-opt-6.7b": ("blip2_opt", "pretrain_opt6.7b"), "blip2-opt-2.7b-coco": ("blip2_opt", "caption_coco_opt2.7b"), "blip2-opt-6.7b-coco": ("blip2_opt", "caption_coco_opt6.7b"), "blip2-flan-t5-xl": ("blip2_t5", "pretrain_flant5xl"), "blip2-flan-t5-xl-coco": ("blip2_t5", "caption_coco_flant5xl"), "blip2-flan-t5-xxl": ("blip2_t5", "pretrain_flant5xxl"), } __magic_name__ , __magic_name__ : Union[str, Any] = model_name_to_original[model_name] # load original model print("Loading original model..." ) __magic_name__ : Union[str, Any] = "cuda" if torch.cuda.is_available() else "cpu" __magic_name__ , __magic_name__ , __magic_name__ : Optional[Any] = load_model_and_preprocess( name=_snake_case , model_type=_snake_case , is_eval=_snake_case , device=_snake_case ) original_model.eval() print("Done!" ) # update state dict keys __magic_name__ : Dict = original_model.state_dict() __magic_name__ : str = create_rename_keys(_snake_case ) for src, dest in rename_keys: rename_key(_snake_case , _snake_case , _snake_case ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): __magic_name__ : Any = state_dict.pop(_snake_case ) if key.startswith("Qformer.bert" ): __magic_name__ : Optional[int] = key.replace("Qformer.bert" , "qformer" ) if "attention.self" in key: __magic_name__ : Any = key.replace("self" , "attention" ) if "opt_proj" in key: __magic_name__ : Union[str, Any] = key.replace("opt_proj" , "language_projection" ) if "t5_proj" in key: __magic_name__ : Optional[int] = key.replace("t5_proj" , "language_projection" ) if key.startswith("opt" ): __magic_name__ : List[str] = key.replace("opt" , "language" ) if key.startswith("t5" ): __magic_name__ : Tuple = key.replace("t5" , "language" ) __magic_name__ : Dict = val # read in qv biases read_in_q_v_bias(_snake_case , _snake_case ) __magic_name__ , __magic_name__ : Tuple = hf_model.load_state_dict(_snake_case , strict=_snake_case ) assert len(_snake_case ) == 0 assert unexpected_keys == ["qformer.embeddings.position_ids"] __magic_name__ : List[Any] = load_demo_image() __magic_name__ : Tuple = vis_processors["eval"](_snake_case ).unsqueeze(0 ).to(_snake_case ) __magic_name__ : Dict = tokenizer(["\n"] , return_tensors="pt" ).input_ids.to(_snake_case ) # create processor __magic_name__ : Optional[Any] = BlipImageProcessor( size={"height": image_size, "width": image_size} , image_mean=_snake_case , image_std=_snake_case ) __magic_name__ : Dict = BlipaProcessor(image_processor=_snake_case , tokenizer=_snake_case ) __magic_name__ : Union[str, Any] = processor(images=_snake_case , return_tensors="pt" ).pixel_values.to(_snake_case ) # make sure processor creates exact same pixel values assert torch.allclose(_snake_case , _snake_case ) original_model.to(_snake_case ) hf_model.to(_snake_case ) with torch.no_grad(): if "opt" in model_name: __magic_name__ : List[Any] = original_model({"image": original_pixel_values, "text_input": [""]} ).logits __magic_name__ : Optional[int] = hf_model(_snake_case , _snake_case ).logits else: __magic_name__ : int = original_model( {"image": original_pixel_values, "text_input": ["\n"], "text_output": ["\n"]} ).logits __magic_name__ : Tuple = input_ids.masked_fill(input_ids == tokenizer.pad_token_id , -100 ) __magic_name__ : List[str] = hf_model(_snake_case , _snake_case , labels=_snake_case ).logits assert original_logits.shape == logits.shape print("First values of original logits:" , original_logits[0, :3, :3] ) print("First values of HF logits:" , logits[0, :3, :3] ) # assert values if model_name == "blip2-flan-t5-xl": __magic_name__ : List[str] = torch.tensor( [[-41.5_850, -4.4_440, -8.9_922], [-47.4_322, -5.9_143, -1.7_340]] , device=_snake_case ) assert torch.allclose(logits[0, :3, :3] , _snake_case , atol=1E-4 ) elif model_name == "blip2-flan-t5-xl-coco": __magic_name__ : Tuple = torch.tensor( [[-57.0_109, -9.8_967, -12.6_280], [-68.6_578, -12.7_191, -10.5_065]] , device=_snake_case ) else: # cast to same type __magic_name__ : str = logits.dtype assert torch.allclose(original_logits.to(_snake_case ) , _snake_case , atol=1E-2 ) print("Looks ok!" ) print("Generating a caption..." ) __magic_name__ : Optional[int] = "" __magic_name__ : Dict = tokenizer(_snake_case , return_tensors="pt" ).input_ids.to(_snake_case ) __magic_name__ : int = original_model.generate({"image": original_pixel_values} ) __magic_name__ : Optional[Any] = hf_model.generate( _snake_case , _snake_case , do_sample=_snake_case , num_beams=5 , max_length=30 , min_length=1 , top_p=0.9 , repetition_penalty=1.0 , length_penalty=1.0 , temperature=1 , ) print("Original generation:" , _snake_case ) __magic_name__ : Tuple = input_ids.shape[1] __magic_name__ : int = processor.batch_decode(outputs[:, prompt_length:] , skip_special_tokens=_snake_case ) __magic_name__ : Union[str, Any] = [text.strip() for text in output_text] print("HF generation:" , _snake_case ) if pytorch_dump_folder_path is not None: processor.save_pretrained(_snake_case ) hf_model.save_pretrained(_snake_case ) if push_to_hub: processor.push_to_hub(F'''nielsr/{model_name}''' ) hf_model.push_to_hub(F'''nielsr/{model_name}''' ) if __name__ == "__main__": snake_case : Any = argparse.ArgumentParser() snake_case : Union[str, Any] = [ "blip2-opt-2.7b", "blip2-opt-6.7b", "blip2-opt-2.7b-coco", "blip2-opt-6.7b-coco", "blip2-flan-t5-xl", "blip2-flan-t5-xl-coco", "blip2-flan-t5-xxl", ] parser.add_argument( "--model_name", default="blip2-opt-2.7b", choices=choices, type=str, help="Path to hf config.json of model to convert", ) parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model and processor to the hub after converting", ) snake_case : int = parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
281
0
def _UpperCamelCase ( snake_case__, snake_case__ ) -> bool: __UpperCAmelCase : Union[str, Any] = len(_snake_case ) + 1 __UpperCAmelCase : List[str] = len(_snake_case ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length i of input_string matches with prefix string of length j of # given pattern. # "dp" stands for dynamic programming. __UpperCAmelCase : str = [[0 for i in range(_snake_case )] for j in range(_snake_case )] # since string of zero length match pattern of zero length __UpperCAmelCase : Optional[int] = 1 # since pattern of zero length will never match with string of non-zero length for i in range(1, _snake_case ): __UpperCAmelCase : Optional[int] = 0 # since string of zero length will match with pattern where there # is at least one * alternatively for j in range(1, _snake_case ): __UpperCAmelCase : Union[str, Any] = dp[0][j - 2] if pattern[j - 1] == "*" else 0 # now using bottom-up approach to find for all remaining lengths for i in range(1, _snake_case ): for j in range(1, _snake_case ): if input_string[i - 1] == pattern[j - 1] or pattern[j - 1] == ".": __UpperCAmelCase : Optional[int] = dp[i - 1][j - 1] elif pattern[j - 1] == "*": if dp[i][j - 2] == 1: __UpperCAmelCase : Optional[Any] = 1 elif pattern[j - 2] in (input_string[i - 1], "."): __UpperCAmelCase : List[Any] = dp[i - 1][j] else: __UpperCAmelCase : Union[str, Any] = 0 else: __UpperCAmelCase : Dict = 0 return bool(dp[-1][-1] ) if __name__ == "__main__": import doctest doctest.testmod() # inputing the strings # input_string = input("input a string :") # pattern = input("input a pattern :") _snake_case = "aab" _snake_case = "c*a*b" # using function to check whether given string matches the given pattern if match_pattern(input_string, pattern): print(F'{input_string} matches the given pattern {pattern}') else: print(F'{input_string} does not match with the given pattern {pattern}')
157
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging snake_case : Dict = logging.get_logger(__name__) snake_case : Union[str, Any] = { "vocab_file": "vocab.txt", "merges_file": "bpe.codes", } snake_case : Dict = { "vocab_file": { "vinai/phobert-base": "https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt", "vinai/phobert-large": "https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt", }, "merges_file": { "vinai/phobert-base": "https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes", "vinai/phobert-large": "https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes", }, } snake_case : Union[str, Any] = { "vinai/phobert-base": 256, "vinai/phobert-large": 256, } def lowerCAmelCase_ ( _snake_case : str ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : List[str] = set() __magic_name__ : Any = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __magic_name__ : int = char __magic_name__ : List[str] = set(_snake_case ) return pairs class _snake_case ( snake_case ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , _a , _a , _a="<s>" , _a="</s>" , _a="</s>" , _a="<s>" , _a="<unk>" , _a="<pad>" , _a="<mask>" , **_a , ): super().__init__( bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , cls_token=_a , pad_token=_a , mask_token=_a , **_a , ) __magic_name__ : Dict = vocab_file __magic_name__ : Tuple = merges_file __magic_name__ : List[Any] = {} __magic_name__ : List[Any] = 0 __magic_name__ : Tuple = 1 __magic_name__ : int = 2 __magic_name__ : Union[str, Any] = 3 self.add_from_file(_a ) __magic_name__ : Optional[int] = {v: k for k, v in self.encoder.items()} with open(_a , encoding="utf-8" ) as merges_handle: __magic_name__ : List[str] = merges_handle.read().split("\n" )[:-1] __magic_name__ : Union[str, Any] = [tuple(merge.split()[:-1] ) for merge in merges] __magic_name__ : Union[str, Any] = dict(zip(_a , range(len(_a ) ) ) ) __magic_name__ : Optional[int] = {} def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __magic_name__ : Optional[Any] = [self.cls_token_id] __magic_name__ : Union[str, Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self , _a , _a = None , _a = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a , token_ids_a=_a , already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): __magic_name__ : Optional[Any] = [self.sep_token_id] __magic_name__ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def SCREAMING_SNAKE_CASE ( self ): return len(self.encoder ) def SCREAMING_SNAKE_CASE ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def SCREAMING_SNAKE_CASE ( self , _a ): if token in self.cache: return self.cache[token] __magic_name__ : List[Any] = tuple(_a ) __magic_name__ : List[Any] = tuple(list(word[:-1] ) + [word[-1] + "</w>"] ) __magic_name__ : Any = get_pairs(_a ) if not pairs: return token while True: __magic_name__ : str = min(_a , key=lambda _a : self.bpe_ranks.get(_a , float("inf" ) ) ) if bigram not in self.bpe_ranks: break __magic_name__ , __magic_name__ : List[str] = bigram __magic_name__ : List[str] = [] __magic_name__ : List[str] = 0 while i < len(_a ): try: __magic_name__ : Any = word.index(_a , _a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) __magic_name__ : Tuple = j if word[i] == first and i < len(_a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __magic_name__ : Union[str, Any] = tuple(_a ) __magic_name__ : Optional[int] = new_word if len(_a ) == 1: break else: __magic_name__ : List[Any] = get_pairs(_a ) __magic_name__ : Optional[int] = "@@ ".join(_a ) __magic_name__ : Tuple = word[:-4] __magic_name__ : str = word return word def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Optional[Any] = [] __magic_name__ : Dict = re.findall(r"\S+\n?" , _a ) for token in words: split_tokens.extend(list(self.bpe(_a ).split(" " ) ) ) return split_tokens def SCREAMING_SNAKE_CASE ( self , _a ): return self.encoder.get(_a , self.encoder.get(self.unk_token ) ) def SCREAMING_SNAKE_CASE ( self , _a ): return self.decoder.get(_a , self.unk_token ) def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Tuple = " ".join(_a ).replace("@@ " , "" ).strip() return out_string def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): if not os.path.isdir(_a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __magic_name__ : Optional[int] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) __magic_name__ : Union[str, Any] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ): copyfile(self.vocab_file , _a ) if os.path.abspath(self.merges_file ) != os.path.abspath(_a ): copyfile(self.merges_file , _a ) return out_vocab_file, out_merge_file def SCREAMING_SNAKE_CASE ( self , _a ): if isinstance(_a , _a ): try: with open(_a , "r" , encoding="utf-8" ) as fd: self.add_from_file(_a ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(f'''Incorrect encoding detected in {f}, please rebuild the dataset''' ) return __magic_name__ : List[Any] = f.readlines() for lineTmp in lines: __magic_name__ : Optional[Any] = lineTmp.strip() __magic_name__ : Union[str, Any] = line.rfind(" " ) if idx == -1: raise ValueError("Incorrect dictionary format, expected '<token> <cnt>'" ) __magic_name__ : Optional[int] = line[:idx] __magic_name__ : Dict = len(self.encoder )
281
0
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, PNDMScheduler, StableDiffusionLDMaDPipeline, UNetaDConditionModel, ) from diffusers.utils import nightly, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS enable_full_determinism() class A__ ( unittest.TestCase): A_ : Tuple = StableDiffusionLDMaDPipeline A_ : Tuple = TEXT_TO_IMAGE_PARAMS A_ : List[Any] = TEXT_TO_IMAGE_BATCH_PARAMS A_ : List[Any] = TEXT_TO_IMAGE_IMAGE_PARAMS def __lowerCamelCase ( self ): torch.manual_seed(0 ) __lowerCAmelCase : Tuple = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) __lowerCAmelCase : int = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=_a , set_alpha_to_one=_a , ) torch.manual_seed(0 ) __lowerCAmelCase : int = AutoencoderKL( block_out_channels=[32, 64] , in_channels=6 , out_channels=6 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCAmelCase : List[str] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) __lowerCAmelCase : Optional[Any] = CLIPTextModel(_a ) __lowerCAmelCase : Union[str, Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __lowerCAmelCase : List[Any] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=0 ): if str(_a ).startswith('mps' ): __lowerCAmelCase : Tuple = torch.manual_seed(_a ) else: __lowerCAmelCase : Optional[int] = torch.Generator(device=_a ).manual_seed(_a ) __lowerCAmelCase : Optional[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def __lowerCamelCase ( self ): __lowerCAmelCase : str = "cpu" # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Optional[int] = self.get_dummy_components() __lowerCAmelCase : List[str] = StableDiffusionLDMaDPipeline(**_a ) __lowerCAmelCase : List[Any] = ldmad_pipe.to(_a ) ldmad_pipe.set_progress_bar_config(disable=_a ) __lowerCAmelCase : Optional[Any] = self.get_dummy_inputs(_a ) __lowerCAmelCase : List[str] = ldmad_pipe(**_a ) __lowerCAmelCase : Tuple = output.rgb, output.depth __lowerCAmelCase : Union[str, Any] = rgb[0, -3:, -3:, -1] __lowerCAmelCase : List[str] = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) __lowerCAmelCase : str = np.array( [0.3733_8176, 0.7_0247, 0.7420_3193, 0.5164_3604, 0.5825_6793, 0.6093_2136, 0.418_1095, 0.4835_5877, 0.4653_5262] ) __lowerCAmelCase : Optional[int] = np.array([1_03.4_67_27, 85.81_2004, 87.84_9236] ) assert np.abs(image_slice_rgb.flatten() - expected_slice_rgb ).max() < 1E-2 assert np.abs(image_slice_depth.flatten() - expected_slice_depth ).max() < 1E-2 def __lowerCamelCase ( self ): __lowerCAmelCase : Optional[Any] = self.get_dummy_components() __lowerCAmelCase : Tuple = StableDiffusionLDMaDPipeline(**_a ) __lowerCAmelCase : Optional[Any] = ldmad_pipe.to(_a ) ldmad_pipe.set_progress_bar_config(disable=_a ) __lowerCAmelCase : Any = self.get_dummy_inputs(_a ) __lowerCAmelCase : List[str] = 3 * [inputs["prompt"]] # forward __lowerCAmelCase : Union[str, Any] = ldmad_pipe(**_a ) __lowerCAmelCase : int = output.rgb, output.depth __lowerCAmelCase : Union[str, Any] = rgb_slice_a[0, -3:, -3:, -1] __lowerCAmelCase : List[str] = depth_slice_a[0, -3:, -1] __lowerCAmelCase : List[str] = self.get_dummy_inputs(_a ) __lowerCAmelCase : List[str] = 3 * [inputs.pop('prompt' )] __lowerCAmelCase : List[Any] = ldmad_pipe.tokenizer( _a , padding='max_length' , max_length=ldmad_pipe.tokenizer.model_max_length , truncation=_a , return_tensors='pt' , ) __lowerCAmelCase : Any = text_inputs["input_ids"].to(_a ) __lowerCAmelCase : Optional[int] = ldmad_pipe.text_encoder(_a )[0] __lowerCAmelCase : List[Any] = prompt_embeds # forward __lowerCAmelCase : Dict = ldmad_pipe(**_a ) __lowerCAmelCase : Optional[int] = output.rgb, output.depth __lowerCAmelCase : str = rgb_slice_a[0, -3:, -3:, -1] __lowerCAmelCase : Union[str, Any] = depth_slice_a[0, -3:, -1] assert np.abs(rgb_slice_a.flatten() - rgb_slice_a.flatten() ).max() < 1E-4 assert np.abs(depth_slice_a.flatten() - depth_slice_a.flatten() ).max() < 1E-4 def __lowerCamelCase ( self ): __lowerCAmelCase : List[Any] = "cpu" # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : List[Any] = self.get_dummy_components() __lowerCAmelCase : List[str] = PNDMScheduler(skip_prk_steps=_a ) __lowerCAmelCase : Optional[Any] = StableDiffusionLDMaDPipeline(**_a ) __lowerCAmelCase : Union[str, Any] = ldmad_pipe.to(_a ) ldmad_pipe.set_progress_bar_config(disable=_a ) __lowerCAmelCase : str = self.get_dummy_inputs(_a ) __lowerCAmelCase : Tuple = "french fries" __lowerCAmelCase : List[Any] = ldmad_pipe(**_a , negative_prompt=_a ) __lowerCAmelCase : Union[str, Any] = output.rgb, output.depth __lowerCAmelCase : Dict = rgb[0, -3:, -3:, -1] __lowerCAmelCase : List[str] = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) __lowerCAmelCase : Dict = np.array( [0.3_7044, 0.7181_1503, 0.722_3251, 0.4860_3675, 0.563_8391, 0.636_4948, 0.4283_3704, 0.490_1315, 0.4792_6217] ) __lowerCAmelCase : Dict = np.array([1_07.8_47_38, 84.6_2802, 89.96_2135] ) assert np.abs(rgb_slice.flatten() - expected_slice_rgb ).max() < 1E-2 assert np.abs(depth_slice.flatten() - expected_slice_depth ).max() < 1E-2 @slow @require_torch_gpu class A__ ( unittest.TestCase): def __lowerCamelCase ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="cpu" , _SCREAMING_SNAKE_CASE=torch.floataa , _SCREAMING_SNAKE_CASE=0 ): __lowerCAmelCase : Tuple = torch.Generator(device=_a ).manual_seed(_a ) __lowerCAmelCase : Union[str, Any] = np.random.RandomState(_a ).standard_normal((1, 4, 64, 64) ) __lowerCAmelCase : Tuple = torch.from_numpy(_a ).to(device=_a , dtype=_a ) __lowerCAmelCase : Union[str, Any] = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def __lowerCamelCase ( self ): __lowerCAmelCase : Tuple = StableDiffusionLDMaDPipeline.from_pretrained('Intel/ldm3d' ) __lowerCAmelCase : Optional[Any] = ldmad_pipe.to(_a ) ldmad_pipe.set_progress_bar_config(disable=_a ) __lowerCAmelCase : List[Any] = self.get_inputs(_a ) __lowerCAmelCase : List[str] = ldmad_pipe(**_a ) __lowerCAmelCase : int = output.rgb, output.depth __lowerCAmelCase : Dict = rgb[0, -3:, -3:, -1].flatten() __lowerCAmelCase : str = rgb[0, -3:, -1].flatten() assert rgb.shape == (1, 5_12, 5_12, 3) assert depth.shape == (1, 5_12, 5_12) __lowerCAmelCase : Union[str, Any] = np.array( [0.5380_5465, 0.5670_7305, 0.548_6515, 0.5701_2236, 0.581_4511, 0.5625_3487, 0.5484_3014, 0.5509_2263, 0.645_9706] ) __lowerCAmelCase : Any = np.array( [0.926_3781, 0.667_8672, 0.548_6515, 0.9220_2145, 0.6783_1135, 0.5625_3487, 0.924_1694, 0.755_1478, 0.645_9706] ) assert np.abs(rgb_slice - expected_slice_rgb ).max() < 3E-3 assert np.abs(depth_slice - expected_slice_depth ).max() < 3E-3 @nightly @require_torch_gpu class A__ ( unittest.TestCase): def __lowerCamelCase ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="cpu" , _SCREAMING_SNAKE_CASE=torch.floataa , _SCREAMING_SNAKE_CASE=0 ): __lowerCAmelCase : Union[str, Any] = torch.Generator(device=_a ).manual_seed(_a ) __lowerCAmelCase : Tuple = np.random.RandomState(_a ).standard_normal((1, 4, 64, 64) ) __lowerCAmelCase : List[str] = torch.from_numpy(_a ).to(device=_a , dtype=_a ) __lowerCAmelCase : Optional[Any] = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 50, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def __lowerCamelCase ( self ): __lowerCAmelCase : int = StableDiffusionLDMaDPipeline.from_pretrained('Intel/ldm3d' ).to(_a ) ldmad_pipe.set_progress_bar_config(disable=_a ) __lowerCAmelCase : Tuple = self.get_inputs(_a ) __lowerCAmelCase : Dict = ldmad_pipe(**_a ) __lowerCAmelCase : Tuple = output.rgb, output.depth __lowerCAmelCase : List[Any] = 0.49_5586 __lowerCAmelCase : Any = 0.3379_5515 __lowerCAmelCase : str = 1_12.4_85_18 __lowerCAmelCase : Tuple = 98.48_9746 assert np.abs(expected_rgb_mean - rgb.mean() ) < 1E-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1E-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1E-3 assert np.abs(expected_depth_std - depth.std() ) < 1E-3 def __lowerCamelCase ( self ): __lowerCAmelCase : str = StableDiffusionLDMaDPipeline.from_pretrained('Intel/ldm3d-4c' ).to(_a ) ldmad_pipe.set_progress_bar_config(disable=_a ) __lowerCAmelCase : str = self.get_inputs(_a ) __lowerCAmelCase : int = ldmad_pipe(**_a ) __lowerCAmelCase : Any = output.rgb, output.depth __lowerCAmelCase : Optional[int] = 0.419_4127 __lowerCAmelCase : Any = 0.3537_5586 __lowerCAmelCase : Tuple = 0.563_8502 __lowerCAmelCase : Dict = 0.3468_6103 assert rgb.shape == (1, 5_12, 5_12, 3) assert depth.shape == (1, 5_12, 5_12, 1) assert np.abs(expected_rgb_mean - rgb.mean() ) < 1E-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1E-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1E-3 assert np.abs(expected_depth_std - depth.std() ) < 1E-3
86
from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_ ( _snake_case : str = "laptop" ) -> DataFrame: '''simple docstring''' __magic_name__ : Tuple = F'''https://www.amazon.in/laptop/s?k={product}''' __magic_name__ : Dict = { "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36", "Accept-Language": "en-US, en;q=0.5", } __magic_name__ : Tuple = BeautifulSoup(requests.get(_snake_case , headers=_snake_case ).text ) # Initialize a Pandas dataframe with the column titles __magic_name__ : int = DataFrame( columns=[ "Product Title", "Product Link", "Current Price of the product", "Product Rating", "MRP of the product", "Discount", ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( "div" , attrs={"class": "s-result-item", "data-component-type": "s-search-result"} , ) , soup.find_all("div" , attrs={"class": "a-row a-size-base a-color-base"} ) , ): try: __magic_name__ : Dict = item.ha.text __magic_name__ : Optional[int] = "https://www.amazon.in/" + item.ha.a["href"] __magic_name__ : Optional[Any] = item.find("span" , attrs={"class": "a-offscreen"} ).text try: __magic_name__ : Union[str, Any] = item.find("span" , attrs={"class": "a-icon-alt"} ).text except AttributeError: __magic_name__ : Dict = "Not available" try: __magic_name__ : Optional[int] = ( "₹" + item.find( "span" , attrs={"class": "a-price a-text-price"} ).text.split("₹" )[1] ) except AttributeError: __magic_name__ : List[str] = "" try: __magic_name__ : int = float( ( ( float(product_mrp.strip("₹" ).replace("," , "" ) ) - float(product_price.strip("₹" ).replace("," , "" ) ) ) / float(product_mrp.strip("₹" ).replace("," , "" ) ) ) * 100 ) except ValueError: __magic_name__ : str = float("nan" ) except AttributeError: pass __magic_name__ : Optional[int] = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] __magic_name__ : Optional[Any] = " " __magic_name__ : str = " " data_frame.index += 1 return data_frame if __name__ == "__main__": snake_case : Any = "headphones" get_amazon_product_data(product).to_csv(F"Amazon Product Data for {product}.csv")
281
0
def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : int , __UpperCamelCase : int ) -> int: """simple docstring""" return int(input_a == input_a == 0 ) def __SCREAMING_SNAKE_CASE ( ) -> None: """simple docstring""" print("""Truth Table of NOR Gate:""" ) print("""| Input 1 | Input 2 | Output |""" ) print(f"""| 0 | 0 | {nor_gate(0 , 0 )} |""" ) print(f"""| 0 | 1 | {nor_gate(0 , 1 )} |""" ) print(f"""| 1 | 0 | {nor_gate(1 , 0 )} |""" ) print(f"""| 1 | 1 | {nor_gate(1 , 1 )} |""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
219
from __future__ import annotations class _snake_case : def __init__( self , _a ): __magic_name__ : Optional[Any] = data __magic_name__ : Node | None = None __magic_name__ : Node | None = None def lowerCAmelCase_ ( _snake_case : Node | None ) -> None: # In Order traversal of the tree '''simple docstring''' if tree: display(tree.left ) print(tree.data ) display(tree.right ) def lowerCAmelCase_ ( _snake_case : Node | None ) -> int: '''simple docstring''' return 1 + max(depth_of_tree(tree.left ) , depth_of_tree(tree.right ) ) if tree else 0 def lowerCAmelCase_ ( _snake_case : Node ) -> bool: '''simple docstring''' if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def lowerCAmelCase_ ( ) -> None: # Main function for testing. '''simple docstring''' __magic_name__ : int = Node(1 ) __magic_name__ : Union[str, Any] = Node(2 ) __magic_name__ : Tuple = Node(3 ) __magic_name__ : Optional[Any] = Node(4 ) __magic_name__ : Union[str, Any] = Node(5 ) __magic_name__ : Any = Node(6 ) __magic_name__ : int = Node(7 ) __magic_name__ : List[str] = Node(8 ) __magic_name__ : Union[str, Any] = Node(9 ) print(is_full_binary_tree(_snake_case ) ) print(depth_of_tree(_snake_case ) ) print("Tree is: " ) display(_snake_case ) if __name__ == "__main__": main()
281
0
'''simple docstring''' import unittest import numpy as np from transformers import is_flax_available from transformers.testing_utils import require_flax from ..test_modeling_flax_common import ids_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.generation import ( FlaxForcedBOSTokenLogitsProcessor, FlaxForcedEOSTokenLogitsProcessor, FlaxLogitsProcessorList, FlaxMinLengthLogitsProcessor, FlaxTemperatureLogitsWarper, FlaxTopKLogitsWarper, FlaxTopPLogitsWarper, ) @require_flax class lowercase ( unittest.TestCase ): """simple docstring""" def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :List[str] = jnp.ones((batch_size, length) ) / length return scores def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = None UpperCamelCase__ :Tuple = 20 UpperCamelCase__ :Optional[int] = self._get_uniform_logits(batch_size=2 , length=_a ) # tweak scores to not be uniform anymore UpperCamelCase__ :Optional[int] = scores.at[1, 5].set((1 / length) + 0.1 ) # peak, 1st batch UpperCamelCase__ :Any = scores.at[1, 10].set((1 / length) - 0.4 ) # valley, 1st batch # compute softmax UpperCamelCase__ :Optional[int] = jax.nn.softmax(_a , axis=-1 ) UpperCamelCase__ :List[str] = FlaxTemperatureLogitsWarper(temperature=0.5 ) UpperCamelCase__ :List[Any] = FlaxTemperatureLogitsWarper(temperature=1.3 ) UpperCamelCase__ :List[str] = jax.nn.softmax(temp_dist_warper_sharper(_a , scores.copy() , cur_len=_a ) , axis=-1 ) UpperCamelCase__ :Optional[Any] = jax.nn.softmax(temp_dist_warper_smoother(_a , scores.copy() , cur_len=_a ) , axis=-1 ) # uniform distribution stays uniform self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_sharp[0, :] , atol=1e-3 ) ) self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_smooth[0, :] , atol=1e-3 ) ) # sharp peaks get higher, valleys get lower self.assertLess(probs[1, :].max() , warped_prob_sharp[1, :].max() ) self.assertGreater(probs[1, :].min() , warped_prob_sharp[1, :].min() ) # smooth peaks get lower, valleys get higher self.assertGreater(probs[1, :].max() , warped_prob_smooth[1, :].max() ) self.assertLess(probs[1, :].min() , warped_prob_smooth[1, :].min() ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Union[str, Any] = None UpperCamelCase__ :Optional[Any] = 10 UpperCamelCase__ :Union[str, Any] = 2 # create ramp distribution UpperCamelCase__ :List[Any] = np.broadcast_to(np.arange(_a )[None, :] , (batch_size, vocab_size) ).copy() UpperCamelCase__ :Optional[Any] = ramp_logits[1:, : vocab_size // 2] + vocab_size UpperCamelCase__ :List[Any] = FlaxTopKLogitsWarper(3 ) UpperCamelCase__ :Optional[Any] = top_k_warp(_a , _a , cur_len=_a ) # check that correct tokens are filtered self.assertListEqual(jnp.isinf(scores[0] ).tolist() , 7 * [True] + 3 * [False] ) self.assertListEqual(jnp.isinf(scores[1] ).tolist() , 2 * [True] + 3 * [False] + 5 * [True] ) # check special case UpperCamelCase__ :Dict = 5 UpperCamelCase__ :Optional[Any] = FlaxTopKLogitsWarper(top_k=1 , filter_value=0.0 , min_tokens_to_keep=3 ) UpperCamelCase__ :Dict = np.broadcast_to(np.arange(_a )[None, :] , (batch_size, length) ).copy() UpperCamelCase__ :str = top_k_warp_safety_check(_a , _a , cur_len=_a ) # min_tokens overwrites k: 3 tokens are kept => 2 tokens are nullified self.assertListEqual((scores == 0.0).sum(axis=-1 ).tolist() , [2, 2] ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :List[Any] = None UpperCamelCase__ :Optional[Any] = 10 UpperCamelCase__ :Optional[Any] = 2 # create distribution and take log (inverse to Softmax as taken in TopPLogitsWarper) UpperCamelCase__ :List[str] = np.log(np.array([[0.3, 0.1, 0.1, 0.5], [0.15, 0.3, 0.3, 0.25]] ) ) UpperCamelCase__ :Optional[Any] = FlaxTopPLogitsWarper(0.8 ) UpperCamelCase__ :Dict = np.exp(top_p_warp(_a , _a , cur_len=_a ) ) # dist should be filtered to keep min num values so that sum is >= top_p # exp (-inf) => 0 UpperCamelCase__ :int = np.array([[0.3, 0.0, 0.0, 0.5], [0.0, 0.3, 0.3, 0.25]] ) self.assertTrue(np.allclose(_a , _a , atol=1e-3 ) ) # check edge cases with negative and extreme logits UpperCamelCase__ :Optional[Any] = np.broadcast_to(np.arange(_a )[None, :] , (batch_size, vocab_size) ).copy() - ( vocab_size // 2 ) # make ramp_logits more extreme UpperCamelCase__ :Optional[Any] = ramp_logits[1] * 100.0 # make sure at least 2 tokens are kept UpperCamelCase__ :Union[str, Any] = FlaxTopPLogitsWarper(0.9 , min_tokens_to_keep=2 , filter_value=0.0 ) UpperCamelCase__ :str = top_p_warp(_a , _a , cur_len=_a ) # first batch should keep three tokens, second batch would keep only 1, but due to `min_tokens_to_keep=2` keeps 2. self.assertListEqual((filtered_dist != 0.0).sum(axis=-1 ).tolist() , [3, 2] ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Any = 20 UpperCamelCase__ :List[Any] = 4 UpperCamelCase__ :Dict = 0 UpperCamelCase__ :Dict = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_a ) # check that min length is applied at length 5 UpperCamelCase__ :Union[str, Any] = ids_tensor((batch_size, 20) , vocab_size=20 ) UpperCamelCase__ :Any = 5 UpperCamelCase__ :Tuple = self._get_uniform_logits(_a , _a ) UpperCamelCase__ :Union[str, Any] = min_dist_processor(_a , _a , cur_len=_a ) self.assertListEqual(scores_before_min_length[:, eos_token_id].tolist() , 4 * [-float('''inf''' )] ) # check that min length is not applied anymore at length 15 UpperCamelCase__ :Dict = self._get_uniform_logits(_a , _a ) UpperCamelCase__ :List[str] = 15 UpperCamelCase__ :str = min_dist_processor(_a , _a , cur_len=_a ) self.assertFalse(jnp.isinf(_a ).any() ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Union[str, Any] = 20 UpperCamelCase__ :Any = 4 UpperCamelCase__ :Union[str, Any] = 0 UpperCamelCase__ :Tuple = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_a ) # check that all scores are -inf except the bos_token_id score UpperCamelCase__ :Optional[Any] = ids_tensor((batch_size, 1) , vocab_size=20 ) UpperCamelCase__ :Optional[Any] = 1 UpperCamelCase__ :List[str] = self._get_uniform_logits(_a , _a ) UpperCamelCase__ :Optional[int] = logits_processor(_a , _a , cur_len=_a ) self.assertTrue(jnp.isneginf(scores[:, bos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, bos_token_id].tolist() , 4 * [0] ) # score for bos_token_id shold be zero # check that bos_token_id is not forced if current length is greater than 1 UpperCamelCase__ :Optional[int] = 3 UpperCamelCase__ :List[Any] = self._get_uniform_logits(_a , _a ) UpperCamelCase__ :List[Any] = logits_processor(_a , _a , cur_len=_a ) self.assertFalse(jnp.isinf(_a ).any() ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :List[str] = 20 UpperCamelCase__ :Optional[Any] = 4 UpperCamelCase__ :Any = 0 UpperCamelCase__ :Optional[Any] = 5 UpperCamelCase__ :Dict = FlaxForcedEOSTokenLogitsProcessor(max_length=_a , eos_token_id=_a ) # check that all scores are -inf except the eos_token_id when max_length is reached UpperCamelCase__ :Optional[int] = ids_tensor((batch_size, 4) , vocab_size=20 ) UpperCamelCase__ :List[Any] = 4 UpperCamelCase__ :Optional[int] = self._get_uniform_logits(_a , _a ) UpperCamelCase__ :List[Any] = logits_processor(_a , _a , cur_len=_a ) self.assertTrue(jnp.isneginf(scores[:, eos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, eos_token_id].tolist() , 4 * [0] ) # score for eos_token_id should be zero # check that eos_token_id is not forced if max_length is not reached UpperCamelCase__ :List[str] = 3 UpperCamelCase__ :Union[str, Any] = self._get_uniform_logits(_a , _a ) UpperCamelCase__ :List[Any] = logits_processor(_a , _a , cur_len=_a ) self.assertFalse(jnp.isinf(_a ).any() ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[int] = 4 UpperCamelCase__ :Optional[int] = 10 UpperCamelCase__ :str = 15 UpperCamelCase__ :Tuple = 2 UpperCamelCase__ :int = 1 UpperCamelCase__ :Dict = 15 # dummy input_ids and scores UpperCamelCase__ :Optional[Any] = ids_tensor((batch_size, sequence_length) , _a ) UpperCamelCase__ :Optional[Any] = input_ids.copy() UpperCamelCase__ :Tuple = self._get_uniform_logits(_a , _a ) UpperCamelCase__ :Dict = scores.copy() # instantiate all dist processors UpperCamelCase__ :Union[str, Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) UpperCamelCase__ :int = FlaxTopKLogitsWarper(3 ) UpperCamelCase__ :List[Any] = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors UpperCamelCase__ :str = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_a ) UpperCamelCase__ :Any = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_a ) UpperCamelCase__ :Optional[int] = FlaxForcedEOSTokenLogitsProcessor(max_length=_a , eos_token_id=_a ) UpperCamelCase__ :Optional[int] = 10 # no processor list UpperCamelCase__ :List[str] = temp_dist_warp(_a , _a , cur_len=_a ) UpperCamelCase__ :str = top_k_warp(_a , _a , cur_len=_a ) UpperCamelCase__ :Tuple = top_p_warp(_a , _a , cur_len=_a ) UpperCamelCase__ :str = min_dist_proc(_a , _a , cur_len=_a ) UpperCamelCase__ :Tuple = bos_dist_proc(_a , _a , cur_len=_a ) UpperCamelCase__ :str = eos_dist_proc(_a , _a , cur_len=_a ) # with processor list UpperCamelCase__ :List[str] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) UpperCamelCase__ :Tuple = processor(_a , _a , cur_len=_a ) # scores should be equal self.assertTrue(jnp.allclose(_a , _a , atol=1e-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :int = 4 UpperCamelCase__ :Dict = 10 UpperCamelCase__ :str = 15 UpperCamelCase__ :List[Any] = 2 UpperCamelCase__ :List[Any] = 1 UpperCamelCase__ :Dict = 15 # dummy input_ids and scores UpperCamelCase__ :List[str] = ids_tensor((batch_size, sequence_length) , _a ) UpperCamelCase__ :int = input_ids.copy() UpperCamelCase__ :Optional[int] = self._get_uniform_logits(_a , _a ) UpperCamelCase__ :Optional[int] = scores.copy() # instantiate all dist processors UpperCamelCase__ :Dict = FlaxTemperatureLogitsWarper(temperature=0.5 ) UpperCamelCase__ :Optional[int] = FlaxTopKLogitsWarper(3 ) UpperCamelCase__ :Dict = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors UpperCamelCase__ :Optional[Any] = FlaxMinLengthLogitsProcessor(min_length=10 , eos_token_id=_a ) UpperCamelCase__ :Optional[int] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=_a ) UpperCamelCase__ :Union[str, Any] = FlaxForcedEOSTokenLogitsProcessor(max_length=_a , eos_token_id=_a ) UpperCamelCase__ :List[Any] = 10 # no processor list def run_no_processor_list(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): UpperCamelCase__ :List[str] = temp_dist_warp(_a , _a , cur_len=_a ) UpperCamelCase__ :int = top_k_warp(_a , _a , cur_len=_a ) UpperCamelCase__ :Any = top_p_warp(_a , _a , cur_len=_a ) UpperCamelCase__ :str = min_dist_proc(_a , _a , cur_len=_a ) UpperCamelCase__ :List[str] = bos_dist_proc(_a , _a , cur_len=_a ) UpperCamelCase__ :Optional[int] = eos_dist_proc(_a , _a , cur_len=_a ) return scores # with processor list def run_processor_list(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): UpperCamelCase__ :Optional[int] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) UpperCamelCase__ :int = processor(_a , _a , cur_len=_a ) return scores UpperCamelCase__ :Optional[int] = jax.jit(_a ) UpperCamelCase__ :List[str] = jax.jit(_a ) UpperCamelCase__ :Optional[Any] = jitted_run_no_processor_list(_a , _a , _a ) UpperCamelCase__ :Tuple = jitted_run_processor_list(_a , _a , _a ) # scores should be equal self.assertTrue(jnp.allclose(_a , _a , atol=1e-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() )
97
def lowerCAmelCase_ ( _snake_case : str , _snake_case : str ) -> bool: '''simple docstring''' __magic_name__ : Union[str, Any] = len(_snake_case ) + 1 __magic_name__ : List[str] = len(_snake_case ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length i of input_string matches with prefix string of length j of # given pattern. # "dp" stands for dynamic programming. __magic_name__ : str = [[0 for i in range(_snake_case )] for j in range(_snake_case )] # since string of zero length match pattern of zero length __magic_name__ : Optional[int] = 1 # since pattern of zero length will never match with string of non-zero length for i in range(1 , _snake_case ): __magic_name__ : Optional[int] = 0 # since string of zero length will match with pattern where there # is at least one * alternatively for j in range(1 , _snake_case ): __magic_name__ : Union[str, Any] = dp[0][j - 2] if pattern[j - 1] == "*" else 0 # now using bottom-up approach to find for all remaining lengths for i in range(1 , _snake_case ): for j in range(1 , _snake_case ): if input_string[i - 1] == pattern[j - 1] or pattern[j - 1] == ".": __magic_name__ : Optional[int] = dp[i - 1][j - 1] elif pattern[j - 1] == "*": if dp[i][j - 2] == 1: __magic_name__ : Optional[Any] = 1 elif pattern[j - 2] in (input_string[i - 1], "."): __magic_name__ : List[Any] = dp[i - 1][j] else: __magic_name__ : Union[str, Any] = 0 else: __magic_name__ : Dict = 0 return bool(dp[-1][-1] ) if __name__ == "__main__": import doctest doctest.testmod() # inputing the strings # input_string = input("input a string :") # pattern = input("input a pattern :") snake_case : Optional[Any] = "aab" snake_case : List[str] = "c*a*b" # using function to check whether given string matches the given pattern if match_pattern(input_string, pattern): print(F"{input_string} matches the given pattern {pattern}") else: print(F"{input_string} does not match with the given pattern {pattern}")
281
0
"""simple docstring""" from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def _snake_case ( ): UpperCAmelCase : List[str] = { "repo_name": ["test_repo1", "test_repo2", "test_repo3"], "path": ["test_1.py", "test_2.py", "unit_test.py"], "content": ["a " * 20, "a " * 30, "b " * 7], } UpperCAmelCase : List[str] = Dataset.from_dict(_snake_case ) return dataset class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ ): def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' UpperCAmelCase : str = get_dataset() UpperCAmelCase : Optional[Any] = make_duplicate_clusters(_a , 0.85 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : Optional[int] = get_dataset() UpperCAmelCase : List[Any] = deduplicate_dataset(_a ) self.assertEqual(len(_a ) , 2 ) print(_a ) self.assertEqual(duplicate_clusters[0][0]["""copies"""] , 2 ) self.assertEqual(duplicate_clusters[0][0]["""is_extreme"""] , _a )
109
import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class _snake_case : @staticmethod def SCREAMING_SNAKE_CASE ( *_a , **_a ): pass def lowerCAmelCase_ ( _snake_case : Image ) -> str: '''simple docstring''' __magic_name__ : Optional[int] = hashlib.mda(image.tobytes() ) return m.hexdigest()[:10] def lowerCAmelCase_ ( _snake_case : Image ) -> Dict: '''simple docstring''' __magic_name__ : List[Any] = np.array(_snake_case ) __magic_name__ : Optional[int] = npimg.shape return {"hash": hashimage(_snake_case ), "shape": shape} @is_pipeline_test @require_vision @require_torch class _snake_case ( unittest.TestCase ): UpperCamelCase__ = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) UpperCamelCase__ = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def SCREAMING_SNAKE_CASE ( self , _a , _a , _a ): __magic_name__ : Dict = MaskGenerationPipeline(model=_a , image_processor=_a ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self , _a , _a ): pass @require_tf @unittest.skip("Image segmentation not implemented in TF" ) def SCREAMING_SNAKE_CASE ( self ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = pipeline("mask-generation" , model="facebook/sam-vit-huge" ) __magic_name__ : str = image_segmenter("http://images.cocodataset.org/val2017/000000039769.jpg" , points_per_batch=256 ) # Shortening by hashing __magic_name__ : Dict = [] for i, o in enumerate(outputs["masks"] ): new_outupt += [{"mask": mask_to_test_readable(_a ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(_a , decimals=4 ) , [ {"mask": {"hash": "115ad19f5f", "shape": (480, 640)}, "scores": 1.04_44}, {"mask": {"hash": "6affa964c6", "shape": (480, 640)}, "scores": 1.0_21}, {"mask": {"hash": "dfe28a0388", "shape": (480, 640)}, "scores": 1.01_67}, {"mask": {"hash": "c0a5f4a318", "shape": (480, 640)}, "scores": 1.01_32}, {"mask": {"hash": "fe8065c197", "shape": (480, 640)}, "scores": 1.00_53}, {"mask": {"hash": "e2d0b7a0b7", "shape": (480, 640)}, "scores": 0.99_67}, {"mask": {"hash": "453c7844bd", "shape": (480, 640)}, "scores": 0.9_93}, {"mask": {"hash": "3d44f2926d", "shape": (480, 640)}, "scores": 0.99_09}, {"mask": {"hash": "64033ddc3f", "shape": (480, 640)}, "scores": 0.98_79}, {"mask": {"hash": "801064ff79", "shape": (480, 640)}, "scores": 0.98_34}, {"mask": {"hash": "6172f276ef", "shape": (480, 640)}, "scores": 0.97_16}, {"mask": {"hash": "b49e60e084", "shape": (480, 640)}, "scores": 0.96_12}, {"mask": {"hash": "a811e775fd", "shape": (480, 640)}, "scores": 0.95_99}, {"mask": {"hash": "a6a8ebcf4b", "shape": (480, 640)}, "scores": 0.95_52}, {"mask": {"hash": "9d8257e080", "shape": (480, 640)}, "scores": 0.95_32}, {"mask": {"hash": "32de6454a8", "shape": (480, 640)}, "scores": 0.95_16}, {"mask": {"hash": "af3d4af2c8", "shape": (480, 640)}, "scores": 0.94_99}, {"mask": {"hash": "3c6db475fb", "shape": (480, 640)}, "scores": 0.94_83}, {"mask": {"hash": "c290813fb9", "shape": (480, 640)}, "scores": 0.94_64}, {"mask": {"hash": "b6f0b8f606", "shape": (480, 640)}, "scores": 0.9_43}, {"mask": {"hash": "92ce16bfdf", "shape": (480, 640)}, "scores": 0.9_43}, {"mask": {"hash": "c749b25868", "shape": (480, 640)}, "scores": 0.94_08}, {"mask": {"hash": "efb6cab859", "shape": (480, 640)}, "scores": 0.93_35}, {"mask": {"hash": "1ff2eafb30", "shape": (480, 640)}, "scores": 0.93_26}, {"mask": {"hash": "788b798e24", "shape": (480, 640)}, "scores": 0.92_62}, {"mask": {"hash": "abea804f0e", "shape": (480, 640)}, "scores": 0.89_99}, {"mask": {"hash": "7b9e8ddb73", "shape": (480, 640)}, "scores": 0.89_86}, {"mask": {"hash": "cd24047c8a", "shape": (480, 640)}, "scores": 0.89_84}, {"mask": {"hash": "6943e6bcbd", "shape": (480, 640)}, "scores": 0.88_73}, {"mask": {"hash": "b5f47c9191", "shape": (480, 640)}, "scores": 0.88_71} ] , ) # fmt: on @require_torch @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : str = "facebook/sam-vit-huge" __magic_name__ : str = pipeline("mask-generation" , model=_a ) __magic_name__ : Tuple = image_segmenter( "http://images.cocodataset.org/val2017/000000039769.jpg" , pred_iou_thresh=1 , points_per_batch=256 ) # Shortening by hashing __magic_name__ : Any = [] for i, o in enumerate(outputs["masks"] ): new_outupt += [{"mask": mask_to_test_readable(_a ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(_a , decimals=4 ) , [ {"mask": {"hash": "115ad19f5f", "shape": (480, 640)}, "scores": 1.04_44}, {"mask": {"hash": "6affa964c6", "shape": (480, 640)}, "scores": 1.02_10}, {"mask": {"hash": "dfe28a0388", "shape": (480, 640)}, "scores": 1.01_67}, {"mask": {"hash": "c0a5f4a318", "shape": (480, 640)}, "scores": 1.01_32}, {"mask": {"hash": "fe8065c197", "shape": (480, 640)}, "scores": 1.00_53}, ] , )
281
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _A = logging.get_logger(__name__) _A = {} class A ( __UpperCAmelCase ): __snake_case = 'llama' __snake_case = ['past_key_values'] def __init__( self, UpperCamelCase__=3_2000, UpperCamelCase__=4096, UpperCamelCase__=1_1008, UpperCamelCase__=32, UpperCamelCase__=32, UpperCamelCase__=None, UpperCamelCase__="silu", UpperCamelCase__=2048, UpperCamelCase__=0.02, UpperCamelCase__=1E-6, UpperCamelCase__=True, UpperCamelCase__=0, UpperCamelCase__=1, UpperCamelCase__=2, UpperCamelCase__=1, UpperCamelCase__=False, UpperCamelCase__=None, **UpperCamelCase__, ): """simple docstring""" lowerCAmelCase_ = vocab_size lowerCAmelCase_ = max_position_embeddings lowerCAmelCase_ = hidden_size lowerCAmelCase_ = intermediate_size lowerCAmelCase_ = num_hidden_layers lowerCAmelCase_ = num_attention_heads # for backward compatibility if num_key_value_heads is None: lowerCAmelCase_ = num_attention_heads lowerCAmelCase_ = num_key_value_heads lowerCAmelCase_ = hidden_act lowerCAmelCase_ = initializer_range lowerCAmelCase_ = rms_norm_eps lowerCAmelCase_ = pretraining_tp lowerCAmelCase_ = use_cache lowerCAmelCase_ = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=_a, bos_token_id=_a, eos_token_id=_a, tie_word_embeddings=_a, **_a, ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" if self.rope_scaling is None: return if not isinstance(self.rope_scaling, _a ) or len(self.rope_scaling ) != 2: raise ValueError( '''`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ''' f"got {self.rope_scaling}" ) lowerCAmelCase_ = self.rope_scaling.get('''type''', _a ) lowerCAmelCase_ = self.rope_scaling.get('''factor''', _a ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}" ) if rope_scaling_factor is None or not isinstance(_a, _a ) or rope_scaling_factor <= 1.0: raise ValueError(f"`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}" )
278
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets snake_case : List[Any] = "\\n@inproceedings{lin-2004-rouge,\n title = \"{ROUGE}: A Package for Automatic Evaluation of Summaries\",\n author = \"Lin, Chin-Yew\",\n booktitle = \"Text Summarization Branches Out\",\n month = jul,\n year = \"2004\",\n address = \"Barcelona, Spain\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W04-1013\",\n pages = \"74--81\",\n}\n" snake_case : Any = "\\nROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for\nevaluating automatic summarization and machine translation software in natural language processing.\nThe metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation.\n\nNote that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters.\n\nThis metrics is a wrapper around Google Research reimplementation of ROUGE:\nhttps://github.com/google-research/google-research/tree/master/rouge\n" snake_case : str = "\nCalculates average rouge scores for a list of hypotheses and references\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n rouge_types: A list of rouge types to calculate.\n Valid names:\n `\"rouge{n}\"` (e.g. `\"rouge1\"`, `\"rouge2\"`) where: {n} is the n-gram based scoring,\n `\"rougeL\"`: Longest common subsequence based scoring.\n `\"rougeLSum\"`: rougeLsum splits text using `\"\n\"`.\n See details in https://github.com/huggingface/datasets/issues/617\n use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes.\n use_aggregator: Return aggregates if this is set to True\nReturns:\n rouge1: rouge_1 (precision, recall, f1),\n rouge2: rouge_2 (precision, recall, f1),\n rougeL: rouge_l (precision, recall, f1),\n rougeLsum: rouge_lsum (precision, recall, f1)\nExamples:\n\n >>> rouge = datasets.load_metric('rouge')\n >>> predictions = [\"hello there\", \"general kenobi\"]\n >>> references = [\"hello there\", \"general kenobi\"]\n >>> results = rouge.compute(predictions=predictions, references=references)\n >>> print(list(results.keys()))\n ['rouge1', 'rouge2', 'rougeL', 'rougeLsum']\n >>> print(results[\"rouge1\"])\n AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0))\n >>> print(results[\"rouge1\"].mid.fmeasure)\n 1.0\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _snake_case ( datasets.Metric ): def SCREAMING_SNAKE_CASE ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/google-research/google-research/tree/master/rouge"] , reference_urls=[ "https://en.wikipedia.org/wiki/ROUGE_(metric)", "https://github.com/google-research/google-research/tree/master/rouge", ] , ) def SCREAMING_SNAKE_CASE ( self , _a , _a , _a=None , _a=True , _a=False ): if rouge_types is None: __magic_name__ : str = ["rouge1", "rouge2", "rougeL", "rougeLsum"] __magic_name__ : List[str] = rouge_scorer.RougeScorer(rouge_types=_a , use_stemmer=_a ) if use_aggregator: __magic_name__ : Dict = scoring.BootstrapAggregator() else: __magic_name__ : str = [] for ref, pred in zip(_a , _a ): __magic_name__ : Union[str, Any] = scorer.score(_a , _a ) if use_aggregator: aggregator.add_scores(_a ) else: scores.append(_a ) if use_aggregator: __magic_name__ : Any = aggregator.aggregate() else: __magic_name__ : List[Any] = {} for key in scores[0]: __magic_name__ : str = [score[key] for score in scores] return result
281
0
"""simple docstring""" import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging __UpperCamelCase = "\\n\n" __UpperCamelCase = "\nPerplexity (PPL) is one of the most common metrics for evaluating language models.\nIt is defined as the exponentiated average negative log-likelihood of a sequence.\n\nFor more information, see https://huggingface.co/docs/transformers/perplexity\n" __UpperCamelCase = "\nArgs:\n model_id (str): model used for calculating Perplexity\n NOTE: Perplexity can only be calculated for causal language models.\n This includes models such as gpt2, causal variations of bert,\n causal versions of t5, and more (the full list can be found\n in the AutoModelForCausalLM documentation here:\n https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM )\n\n input_texts (list of str): input text, each separate text snippet\n is one list entry.\n batch_size (int): the batch size to run texts through the model. Defaults to 16.\n add_start_token (bool): whether to add the start token to the texts,\n so the perplexity can include the probability of the first word. Defaults to True.\n device (str): device to run on, defaults to 'cuda' when available\nReturns:\n perplexity: dictionary containing the perplexity scores for the texts\n in the input list, as well as the mean perplexity. If one of the input texts is\n longer than the max input length of the model, then it is truncated to the\n max length for the perplexity computation.\nExamples:\n Example 1:\n >>> perplexity = datasets.load_metric(\"perplexity\")\n >>> input_texts = [\"lorem ipsum\", \"Happy Birthday!\", \"Bienvenue\"]\n >>> results = perplexity.compute(model_id='gpt2',\n ... add_start_token=False,\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n ['perplexities', 'mean_perplexity']\n >>> print(round(results[\"mean_perplexity\"], 2))\n 78.22\n >>> print(round(results[\"perplexities\"][0], 2))\n 11.11\n\n Example 2:\n >>> perplexity = datasets.load_metric(\"perplexity\")\n >>> input_texts = datasets.load_dataset(\"wikitext\",\n ... \"wikitext-2-raw-v1\",\n ... split=\"test\")[\"text\"][:50] # doctest:+ELLIPSIS\n [...]\n >>> input_texts = [s for s in input_texts if s!='']\n >>> results = perplexity.compute(model_id='gpt2',\n ... input_texts=input_texts) # doctest:+ELLIPSIS\n >>> print(list(results.keys()))\n ['perplexities', 'mean_perplexity']\n >>> print(round(results[\"mean_perplexity\"], 2))\n 60.35\n >>> print(round(results[\"perplexities\"][0], 2))\n 81.12\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase ( datasets.Metric ): '''simple docstring''' def __A ( self ) -> Dict: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'input_texts': datasets.Value('string' ), } ) , reference_urls=['https://huggingface.co/docs/transformers/perplexity'] , ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 16 , lowerCAmelCase__ = True , lowerCAmelCase__=None ) -> Tuple: if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": SCREAMING_SNAKE_CASE = "cuda" else: SCREAMING_SNAKE_CASE = "cuda" if torch.cuda.is_available() else "cpu" SCREAMING_SNAKE_CASE = AutoModelForCausalLM.from_pretrained(_a ) SCREAMING_SNAKE_CASE = model.to(_a ) SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained(_a ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: SCREAMING_SNAKE_CASE = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(_a ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({'pad_token': existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" SCREAMING_SNAKE_CASE = model.config.max_length - 1 else: SCREAMING_SNAKE_CASE = model.config.max_length SCREAMING_SNAKE_CASE = tokenizer( _a , add_special_tokens=_a , padding=_a , truncation=_a , max_length=_a , return_tensors='pt' , return_attention_mask=_a , ).to(_a ) SCREAMING_SNAKE_CASE = encodings["input_ids"] SCREAMING_SNAKE_CASE = encodings["attention_mask"] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = CrossEntropyLoss(reduction='none' ) for start_index in logging.tqdm(range(0 , len(_a ) , _a ) ): SCREAMING_SNAKE_CASE = min(start_index + batch_size , len(_a ) ) SCREAMING_SNAKE_CASE = encoded_texts[start_index:end_index] SCREAMING_SNAKE_CASE = attn_masks[start_index:end_index] if add_start_token: SCREAMING_SNAKE_CASE = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(_a ) SCREAMING_SNAKE_CASE = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) SCREAMING_SNAKE_CASE = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(_a ), attn_mask] , dim=1 ) SCREAMING_SNAKE_CASE = encoded_batch with torch.no_grad(): SCREAMING_SNAKE_CASE = model(_a , attention_mask=_a ).logits SCREAMING_SNAKE_CASE = out_logits[..., :-1, :].contiguous() SCREAMING_SNAKE_CASE = labels[..., 1:].contiguous() SCREAMING_SNAKE_CASE = attn_mask[..., 1:].contiguous() SCREAMING_SNAKE_CASE = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , _a ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(_a )}
113
snake_case : Optional[int] = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/" def lowerCAmelCase_ ( _snake_case : bytes ) -> bytes: '''simple docstring''' if not isinstance(_snake_case , _snake_case ): __magic_name__ : Tuple = F'''a bytes-like object is required, not \'{data.__class__.__name__}\'''' raise TypeError(_snake_case ) __magic_name__ : Optional[int] = "".join(bin(_snake_case )[2:].zfill(8 ) for byte in data ) __magic_name__ : List[Any] = len(_snake_case ) % 6 != 0 if padding_needed: # The padding that will be added later __magic_name__ : List[str] = B"=" * ((6 - len(_snake_case ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(_snake_case ) % 6) else: __magic_name__ : List[str] = B"" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(_snake_case ) , 6 ) ).encode() + padding ) def lowerCAmelCase_ ( _snake_case : str ) -> bytes: '''simple docstring''' if not isinstance(_snake_case , _snake_case ) and not isinstance(_snake_case , _snake_case ): __magic_name__ : List[str] = ( "argument should be a bytes-like object or ASCII string, " F'''not \'{encoded_data.__class__.__name__}\'''' ) raise TypeError(_snake_case ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(_snake_case , _snake_case ): try: __magic_name__ : List[Any] = encoded_data.decode("utf-8" ) except UnicodeDecodeError: raise ValueError("base64 encoded data should only contain ASCII characters" ) __magic_name__ : List[str] = encoded_data.count("=" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(_snake_case ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one __magic_name__ : Optional[int] = encoded_data[:-padding] __magic_name__ : Dict = "".join( bin(B64_CHARSET.index(_snake_case ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: __magic_name__ : Union[str, Any] = "".join( bin(B64_CHARSET.index(_snake_case ) )[2:].zfill(6 ) for char in encoded_data ) __magic_name__ : List[Any] = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(_snake_case ) , 8 ) ] return bytes(_snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
281
0
"""simple docstring""" import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder lowercase__ = "__DUMMY_TRANSFORMERS_USER__" lowercase__ = "Dummy User" lowercase__ = "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" lowercase__ = "https://hub-ci.huggingface.co" lowercase__ = CI_HUB_ENDPOINT + "/datasets/{repo_id}/resolve/{revision}/{path}" lowercase__ = CI_HUB_ENDPOINT + "/{repo_id}/resolve/{revision}/{filename}" lowercase__ = Path("""~/.huggingface/hub_ci_token""").expanduser() @pytest.fixture def _snake_case ( lowercase__ ): monkeypatch.setattr( 'huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE' , _snake_case ) @pytest.fixture def _snake_case ( lowercase__ ): monkeypatch.setattr('datasets.config.HF_ENDPOINT' , _snake_case ) monkeypatch.setattr('datasets.config.HUB_DATASETS_URL' , _snake_case ) @pytest.fixture def _snake_case ( lowercase__ ): monkeypatch.setattr('huggingface_hub.hf_api.HfFolder.path_token' , _snake_case ) @pytest.fixture def _snake_case ( lowercase__ , lowercase__ ): HfFolder.save_token(_snake_case ) yield HfFolder.delete_token() @pytest.fixture(scope='session' ) def _snake_case ( ): return HfApi(endpoint=_snake_case ) @pytest.fixture(scope='session' ) def _snake_case ( lowercase__ ): _lowerCamelCase : List[Any] = HfFolder.get_token() HfFolder.save_token(_snake_case ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(_snake_case ) @pytest.fixture def _snake_case ( lowercase__ ): def _cleanup_repo(lowercase__ ): hf_api.delete_repo(_snake_case , token=_snake_case , repo_type='dataset' ) return _cleanup_repo @pytest.fixture def _snake_case ( lowercase__ ): @contextmanager def _temporary_repo(lowercase__ ): try: yield repo_id finally: cleanup_repo(_snake_case ) return _temporary_repo @pytest.fixture(scope='session' ) def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : Union[str, Any] = f'''repo_txt_data-{int(time.time() * 10E3 )}''' _lowerCamelCase : List[Any] = f'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_snake_case , token=_snake_case , repo_type='dataset' , private=_snake_case ) hf_api.upload_file( token=_snake_case , path_or_fileobj=str(_snake_case ) , path_in_repo='data/text_data.txt' , repo_id=_snake_case , repo_type='dataset' , ) yield repo_id try: hf_api.delete_repo(_snake_case , token=_snake_case , repo_type='dataset' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope='session' ) def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : Union[str, Any] = f'''repo_zipped_txt_data-{int(time.time() * 10E3 )}''' _lowerCamelCase : List[str] = f'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_snake_case , token=_snake_case , repo_type='dataset' , private=_snake_case ) hf_api.upload_file( token=_snake_case , path_or_fileobj=str(_snake_case ) , path_in_repo='data.zip' , repo_id=_snake_case , repo_type='dataset' , ) yield repo_id try: hf_api.delete_repo(_snake_case , token=_snake_case , repo_type='dataset' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope='session' ) def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : int = f'''repo_zipped_img_data-{int(time.time() * 10E3 )}''' _lowerCamelCase : int = f'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_snake_case , token=_snake_case , repo_type='dataset' , private=_snake_case ) hf_api.upload_file( token=_snake_case , path_or_fileobj=str(_snake_case ) , path_in_repo='data.zip' , repo_id=_snake_case , repo_type='dataset' , ) yield repo_id try: hf_api.delete_repo(_snake_case , token=_snake_case , repo_type='dataset' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): return hf_private_dataset_repo_zipped_img_data_
96
import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class _snake_case ( unittest.TestCase ): def __init__( self , _a , _a=13 , _a=7 , _a=True , _a=True , _a=True , _a=True , _a=99 , _a=32 , _a=5 , _a=4 , _a=37 , _a="gelu" , _a=0.1 , _a=0.1 , _a=512 , _a=16 , _a=2 , _a=0.02 , _a=4 , ): __magic_name__ : List[Any] = parent __magic_name__ : Optional[Any] = batch_size __magic_name__ : Dict = seq_length __magic_name__ : Union[str, Any] = is_training __magic_name__ : Optional[Any] = use_attention_mask __magic_name__ : Optional[Any] = use_token_type_ids __magic_name__ : int = use_labels __magic_name__ : List[Any] = vocab_size __magic_name__ : Union[str, Any] = hidden_size __magic_name__ : Optional[Any] = num_hidden_layers __magic_name__ : int = num_attention_heads __magic_name__ : Any = intermediate_size __magic_name__ : List[Any] = hidden_act __magic_name__ : List[Any] = hidden_dropout_prob __magic_name__ : Optional[int] = attention_probs_dropout_prob __magic_name__ : List[Any] = max_position_embeddings __magic_name__ : Tuple = type_vocab_size __magic_name__ : List[str] = type_sequence_label_size __magic_name__ : Dict = initializer_range __magic_name__ : List[Any] = num_choices def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __magic_name__ : List[Any] = None if self.use_attention_mask: __magic_name__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) __magic_name__ : str = None if self.use_token_type_ids: __magic_name__ : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __magic_name__ : List[str] = RobertaPreLayerNormConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_a , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : int = self.prepare_config_and_inputs() __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : List[Any] = config_and_inputs __magic_name__ : List[str] = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = self.prepare_config_and_inputs() __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Union[str, Any] = config_and_inputs __magic_name__ : Tuple = True __magic_name__ : int = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __magic_name__ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class _snake_case ( snake_case , unittest.TestCase ): UpperCamelCase__ = True UpperCamelCase__ = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[Any] = FlaxRobertaPreLayerNormModelTester(self ) @slow def SCREAMING_SNAKE_CASE ( self ): for model_class_name in self.all_model_classes: __magic_name__ : Optional[Any] = model_class_name.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=_a ) __magic_name__ : Dict = model(np.ones((1, 1) ) ) self.assertIsNotNone(_a ) @require_flax class _snake_case ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = FlaxRobertaPreLayerNormForMaskedLM.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=_a ) __magic_name__ : Union[str, Any] = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) __magic_name__ : List[str] = model(_a )[0] __magic_name__ : str = [1, 11, 50_265] self.assertEqual(list(output.shape ) , _a ) # compare the actual values for a slice. __magic_name__ : List[str] = np.array( [[[40.48_80, 18.01_99, -5.23_67], [-1.88_77, -4.08_85, 10.70_85], [-2.26_13, -5.61_10, 7.26_65]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , _a , atol=1e-4 ) ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[str] = FlaxRobertaPreLayerNormModel.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=_a ) __magic_name__ : Tuple = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) __magic_name__ : Tuple = model(_a )[0] # compare the actual values for a slice. __magic_name__ : Dict = np.array( [[[0.02_08, -0.03_56, 0.02_37], [-0.15_69, -0.04_11, -0.26_26], [0.18_79, 0.01_25, -0.00_89]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , _a , atol=1e-4 ) )
281
0
"""simple docstring""" SCREAMING_SNAKE_CASE__ = 256 # Modulus to hash a string SCREAMING_SNAKE_CASE__ = 1_000_003 def lowerCAmelCase__ ( _UpperCamelCase : str , _UpperCamelCase : str ) -> bool: """simple docstring""" snake_case = len(_snake_case ) snake_case = len(_snake_case ) if p_len > t_len: return False snake_case = 0 snake_case = 0 snake_case = 1 # Calculating the hash of pattern and substring of text for i in range(_snake_case ): snake_case = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus snake_case = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue snake_case = (modulus_power * alphabet_size) % modulus for i in range(0 , t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash snake_case = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def lowerCAmelCase__ ( ) -> None: """simple docstring""" snake_case = "abc1abc12" snake_case = "alskfjaldsabc1abc1abc12k23adsfabcabc" snake_case = "alskfjaldsk23adsfabcabc" assert rabin_karp(_snake_case , _snake_case ) and not rabin_karp(_snake_case , _snake_case ) # Test 2) snake_case = "ABABX" snake_case = "ABABZABABYABABX" assert rabin_karp(_snake_case , _snake_case ) # Test 3) snake_case = "AAAB" snake_case = "ABAAAAAB" assert rabin_karp(_snake_case , _snake_case ) # Test 4) snake_case = "abcdabcy" snake_case = "abcxabcdabxabcdabcdabcy" assert rabin_karp(_snake_case , _snake_case ) # Test 5) snake_case = "Lü" snake_case = "Lüsai" assert rabin_karp(_snake_case , _snake_case ) snake_case = "Lue" assert not rabin_karp(_snake_case , _snake_case ) print('Success.' ) if __name__ == "__main__": test_rabin_karp()
150
def lowerCAmelCase_ ( _snake_case : list[list[int | float]] ) -> int: '''simple docstring''' __magic_name__ : Any = len(_snake_case ) __magic_name__ : Optional[Any] = len(matrix[0] ) __magic_name__ : Union[str, Any] = min(_snake_case , _snake_case ) for row in range(_snake_case ): # Check if diagonal element is not zero if matrix[row][row] != 0: # Eliminate all the elements below the diagonal for col in range(row + 1 , _snake_case ): __magic_name__ : Optional[Any] = matrix[col][row] / matrix[row][row] for i in range(_snake_case , _snake_case ): matrix[col][i] -= multiplier * matrix[row][i] else: # Find a non-zero diagonal element to swap rows __magic_name__ : str = True for i in range(row + 1 , _snake_case ): if matrix[i][row] != 0: __magic_name__ , __magic_name__ : List[str] = matrix[i], matrix[row] __magic_name__ : Union[str, Any] = False break if reduce: rank -= 1 for i in range(_snake_case ): __magic_name__ : Any = matrix[i][rank] # Reduce the row pointer by one to stay on the same row row -= 1 return rank if __name__ == "__main__": import doctest doctest.testmod()
281
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase : Any = logging.get_logger(__name__) UpperCAmelCase : Optional[int] = { "EleutherAI/gpt-neox-20b": "https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json", # See all GPTNeoX models at https://huggingface.co/models?filter=gpt_neox } class lowerCamelCase__ ( A ): """simple docstring""" __a = """gpt_neox""" def __init__( self : str , UpperCamelCase : Dict=50_432 , UpperCamelCase : Optional[int]=6_144 , UpperCamelCase : Union[str, Any]=44 , UpperCamelCase : List[Any]=64 , UpperCamelCase : Dict=24_576 , UpperCamelCase : Optional[int]="gelu" , UpperCamelCase : Optional[int]=0.25 , UpperCamelCase : int=10_000 , UpperCamelCase : List[Any]=0.0 , UpperCamelCase : str=0.0 , UpperCamelCase : List[str]=0.1 , UpperCamelCase : Optional[int]=2_048 , UpperCamelCase : Union[str, Any]=0.02 , UpperCamelCase : Tuple=1e-5 , UpperCamelCase : List[str]=True , UpperCamelCase : Tuple=0 , UpperCamelCase : List[Any]=2 , UpperCamelCase : Optional[Any]=False , UpperCamelCase : str=True , UpperCamelCase : Union[str, Any]=None , **UpperCamelCase : str , ): '''simple docstring''' super().__init__(bos_token_id=_a , eos_token_id=_a , **_a ) __UpperCAmelCase : List[Any] = vocab_size __UpperCAmelCase : List[Any] = max_position_embeddings __UpperCAmelCase : Optional[Any] = hidden_size __UpperCAmelCase : Union[str, Any] = num_hidden_layers __UpperCAmelCase : Optional[Any] = num_attention_heads __UpperCAmelCase : List[Any] = intermediate_size __UpperCAmelCase : List[str] = hidden_act __UpperCAmelCase : Union[str, Any] = rotary_pct __UpperCAmelCase : Any = rotary_emb_base __UpperCAmelCase : int = attention_dropout __UpperCAmelCase : Optional[int] = hidden_dropout __UpperCAmelCase : List[str] = classifier_dropout __UpperCAmelCase : List[Any] = initializer_range __UpperCAmelCase : str = layer_norm_eps __UpperCAmelCase : Optional[int] = use_cache __UpperCAmelCase : Any = tie_word_embeddings __UpperCAmelCase : Optional[Any] = use_parallel_residual __UpperCAmelCase : Dict = rope_scaling self._rope_scaling_validation() if self.hidden_size % self.num_attention_heads != 0: raise ValueError( """The hidden size is not divisble by the number of attention heads! Make sure to update them!""" ) def lowerCamelCase__ ( self : Tuple ): '''simple docstring''' if self.rope_scaling is None: return if not isinstance(self.rope_scaling , _a ) or len(self.rope_scaling ) != 2: raise ValueError( """`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, """ f'''got {self.rope_scaling}''' ) __UpperCAmelCase : Dict = self.rope_scaling.get("""type""" , _a ) __UpperCAmelCase : int = self.rope_scaling.get("""factor""" , _a ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f'''`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}''' ) if rope_scaling_factor is None or not isinstance(_a , _a ) or rope_scaling_factor <= 1.0: raise ValueError(f'''`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}''' )
115
import argparse import collections import json import os import re import string import sys import numpy as np snake_case : Dict = re.compile(R"\b(a|an|the)\b", re.UNICODE) snake_case : Optional[int] = None def lowerCAmelCase_ ( ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : Any = argparse.ArgumentParser("Official evaluation script for SQuAD version 2.0." ) parser.add_argument("data_file" , metavar="data.json" , help="Input data JSON file." ) parser.add_argument("pred_file" , metavar="pred.json" , help="Model predictions." ) parser.add_argument( "--out-file" , "-o" , metavar="eval.json" , help="Write accuracy metrics to file (default is stdout)." ) parser.add_argument( "--na-prob-file" , "-n" , metavar="na_prob.json" , help="Model estimates of probability of no answer." ) parser.add_argument( "--na-prob-thresh" , "-t" , type=_snake_case , default=1.0 , help="Predict \"\" if no-answer probability exceeds this (default = 1.0)." , ) parser.add_argument( "--out-image-dir" , "-p" , metavar="out_images" , default=_snake_case , help="Save precision-recall curves to directory." ) parser.add_argument("--verbose" , "-v" , action="store_true" ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def lowerCAmelCase_ ( _snake_case : Optional[Any] ) -> Tuple: '''simple docstring''' __magic_name__ : Optional[int] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: __magic_name__ : str = bool(qa["answers"]["text"] ) return qid_to_has_ans def lowerCAmelCase_ ( _snake_case : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' def remove_articles(_snake_case : List[str] ): return ARTICLES_REGEX.sub(" " , _snake_case ) def white_space_fix(_snake_case : Optional[int] ): return " ".join(text.split() ) def remove_punc(_snake_case : Optional[int] ): __magic_name__ : Dict = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_snake_case : str ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_snake_case ) ) ) ) def lowerCAmelCase_ ( _snake_case : Any ) -> Optional[Any]: '''simple docstring''' if not s: return [] return normalize_answer(_snake_case ).split() def lowerCAmelCase_ ( _snake_case : str , _snake_case : Dict ) -> Tuple: '''simple docstring''' return int(normalize_answer(_snake_case ) == normalize_answer(_snake_case ) ) def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : int ) -> str: '''simple docstring''' __magic_name__ : Any = get_tokens(_snake_case ) __magic_name__ : Optional[int] = get_tokens(_snake_case ) __magic_name__ : Tuple = collections.Counter(_snake_case ) & collections.Counter(_snake_case ) __magic_name__ : Tuple = sum(common.values() ) if len(_snake_case ) == 0 or len(_snake_case ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 __magic_name__ : Dict = 1.0 * num_same / len(_snake_case ) __magic_name__ : Optional[Any] = 1.0 * num_same / len(_snake_case ) __magic_name__ : List[Any] = (2 * precision * recall) / (precision + recall) return fa def lowerCAmelCase_ ( _snake_case : Optional[Any] , _snake_case : List[Any] ) -> List[Any]: '''simple docstring''' __magic_name__ : Union[str, Any] = {} __magic_name__ : int = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: __magic_name__ : Union[str, Any] = qa["id"] __magic_name__ : Any = [t for t in qa["answers"]["text"] if normalize_answer(_snake_case )] if not gold_answers: # For unanswerable questions, only correct answer is empty string __magic_name__ : Tuple = [""] if qid not in preds: print(F'''Missing prediction for {qid}''' ) continue __magic_name__ : Any = preds[qid] # Take max over all gold answers __magic_name__ : List[Any] = max(compute_exact(_snake_case , _snake_case ) for a in gold_answers ) __magic_name__ : int = max(compute_fa(_snake_case , _snake_case ) for a in gold_answers ) return exact_scores, fa_scores def lowerCAmelCase_ ( _snake_case : Optional[Any] , _snake_case : List[Any] , _snake_case : Optional[int] , _snake_case : Dict ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : str = {} for qid, s in scores.items(): __magic_name__ : Dict = na_probs[qid] > na_prob_thresh if pred_na: __magic_name__ : str = float(not qid_to_has_ans[qid] ) else: __magic_name__ : Optional[int] = s return new_scores def lowerCAmelCase_ ( _snake_case : List[Any] , _snake_case : List[str] , _snake_case : Tuple=None ) -> Tuple: '''simple docstring''' if not qid_list: __magic_name__ : Any = len(_snake_case ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores.values() ) / total), ("f1", 100.0 * sum(fa_scores.values() ) / total), ("total", total), ] ) else: __magic_name__ : Tuple = len(_snake_case ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores[k] for k in qid_list ) / total), ("f1", 100.0 * sum(fa_scores[k] for k in qid_list ) / total), ("total", total), ] ) def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : str , _snake_case : str ) -> Dict: '''simple docstring''' for k in new_eval: __magic_name__ : int = new_eval[k] def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Dict , _snake_case : Optional[Any] , _snake_case : Union[str, Any] ) -> str: '''simple docstring''' plt.step(_snake_case , _snake_case , color="b" , alpha=0.2 , where="post" ) plt.fill_between(_snake_case , _snake_case , step="post" , alpha=0.2 , color="b" ) plt.xlabel("Recall" ) plt.ylabel("Precision" ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(_snake_case ) plt.savefig(_snake_case ) plt.clf() def lowerCAmelCase_ ( _snake_case : Dict , _snake_case : Any , _snake_case : Optional[int] , _snake_case : List[Any] , _snake_case : Optional[int]=None , _snake_case : int=None ) -> str: '''simple docstring''' __magic_name__ : Union[str, Any] = sorted(_snake_case , key=lambda _snake_case : na_probs[k] ) __magic_name__ : Optional[int] = 0.0 __magic_name__ : str = 1.0 __magic_name__ : str = 0.0 __magic_name__ : List[str] = [1.0] __magic_name__ : str = [0.0] __magic_name__ : Optional[Any] = 0.0 for i, qid in enumerate(_snake_case ): if qid_to_has_ans[qid]: true_pos += scores[qid] __magic_name__ : List[str] = true_pos / float(i + 1 ) __magic_name__ : Any = true_pos / float(_snake_case ) if i == len(_snake_case ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(_snake_case ) recalls.append(_snake_case ) if out_image: plot_pr_curve(_snake_case , _snake_case , _snake_case , _snake_case ) return {"ap": 100.0 * avg_prec} def lowerCAmelCase_ ( _snake_case : Tuple , _snake_case : Optional[Any] , _snake_case : Optional[int] , _snake_case : Optional[Any] , _snake_case : Any , _snake_case : List[Any] ) -> Union[str, Any]: '''simple docstring''' if out_image_dir and not os.path.exists(_snake_case ): os.makedirs(_snake_case ) __magic_name__ : Any = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return __magic_name__ : str = make_precision_recall_eval( _snake_case , _snake_case , _snake_case , _snake_case , out_image=os.path.join(_snake_case , "pr_exact.png" ) , title="Precision-Recall curve for Exact Match score" , ) __magic_name__ : Union[str, Any] = make_precision_recall_eval( _snake_case , _snake_case , _snake_case , _snake_case , out_image=os.path.join(_snake_case , "pr_f1.png" ) , title="Precision-Recall curve for F1 score" , ) __magic_name__ : str = {k: float(_snake_case ) for k, v in qid_to_has_ans.items()} __magic_name__ : str = make_precision_recall_eval( _snake_case , _snake_case , _snake_case , _snake_case , out_image=os.path.join(_snake_case , "pr_oracle.png" ) , title="Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)" , ) merge_eval(_snake_case , _snake_case , "pr_exact" ) merge_eval(_snake_case , _snake_case , "pr_f1" ) merge_eval(_snake_case , _snake_case , "pr_oracle" ) def lowerCAmelCase_ ( _snake_case : int , _snake_case : Optional[Any] , _snake_case : List[str] , _snake_case : Optional[Any] ) -> Dict: '''simple docstring''' if not qid_list: return __magic_name__ : Dict = [na_probs[k] for k in qid_list] __magic_name__ : str = np.ones_like(_snake_case ) / float(len(_snake_case ) ) plt.hist(_snake_case , weights=_snake_case , bins=20 , range=(0.0, 1.0) ) plt.xlabel("Model probability of no-answer" ) plt.ylabel("Proportion of dataset" ) plt.title(F'''Histogram of no-answer probability: {name}''' ) plt.savefig(os.path.join(_snake_case , F'''na_prob_hist_{name}.png''' ) ) plt.clf() def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Tuple , _snake_case : List[str] , _snake_case : Dict ) -> List[Any]: '''simple docstring''' __magic_name__ : Union[str, Any] = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) __magic_name__ : List[str] = num_no_ans __magic_name__ : Dict = cur_score __magic_name__ : Dict = 0.0 __magic_name__ : Any = sorted(_snake_case , key=lambda _snake_case : na_probs[k] ) for i, qid in enumerate(_snake_case ): if qid not in scores: continue if qid_to_has_ans[qid]: __magic_name__ : Union[str, Any] = scores[qid] else: if preds[qid]: __magic_name__ : List[Any] = -1 else: __magic_name__ : Optional[int] = 0 cur_score += diff if cur_score > best_score: __magic_name__ : Optional[int] = cur_score __magic_name__ : List[Any] = na_probs[qid] return 100.0 * best_score / len(_snake_case ), best_thresh def lowerCAmelCase_ ( _snake_case : int , _snake_case : str , _snake_case : List[str] , _snake_case : Tuple , _snake_case : List[Any] , _snake_case : Dict ) -> Optional[Any]: '''simple docstring''' __magic_name__ , __magic_name__ : List[str] = find_best_thresh(_snake_case , _snake_case , _snake_case , _snake_case ) __magic_name__ , __magic_name__ : int = find_best_thresh(_snake_case , _snake_case , _snake_case , _snake_case ) __magic_name__ : Optional[int] = best_exact __magic_name__ : List[Any] = exact_thresh __magic_name__ : Dict = best_fa __magic_name__ : Any = fa_thresh def lowerCAmelCase_ ( ) -> int: '''simple docstring''' with open(OPTS.data_file ) as f: __magic_name__ : Optional[Any] = json.load(_snake_case ) __magic_name__ : List[Any] = dataset_json["data"] with open(OPTS.pred_file ) as f: __magic_name__ : Optional[Any] = json.load(_snake_case ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: __magic_name__ : Any = json.load(_snake_case ) else: __magic_name__ : Any = {k: 0.0 for k in preds} __magic_name__ : str = make_qid_to_has_ans(_snake_case ) # maps qid to True/False __magic_name__ : Tuple = [k for k, v in qid_to_has_ans.items() if v] __magic_name__ : Optional[Any] = [k for k, v in qid_to_has_ans.items() if not v] __magic_name__ , __magic_name__ : Union[str, Any] = get_raw_scores(_snake_case , _snake_case ) __magic_name__ : Optional[Any] = apply_no_ans_threshold(_snake_case , _snake_case , _snake_case , OPTS.na_prob_thresh ) __magic_name__ : Optional[Any] = apply_no_ans_threshold(_snake_case , _snake_case , _snake_case , OPTS.na_prob_thresh ) __magic_name__ : List[Any] = make_eval_dict(_snake_case , _snake_case ) if has_ans_qids: __magic_name__ : int = make_eval_dict(_snake_case , _snake_case , qid_list=_snake_case ) merge_eval(_snake_case , _snake_case , "HasAns" ) if no_ans_qids: __magic_name__ : List[Any] = make_eval_dict(_snake_case , _snake_case , qid_list=_snake_case ) merge_eval(_snake_case , _snake_case , "NoAns" ) if OPTS.na_prob_file: find_all_best_thresh(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case , OPTS.out_image_dir ) histogram_na_prob(_snake_case , _snake_case , OPTS.out_image_dir , "hasAns" ) histogram_na_prob(_snake_case , _snake_case , OPTS.out_image_dir , "noAns" ) if OPTS.out_file: with open(OPTS.out_file , "w" ) as f: json.dump(_snake_case , _snake_case ) else: print(json.dumps(_snake_case , indent=2 ) ) if __name__ == "__main__": snake_case : int = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("Agg") import matplotlib.pyplot as plt main()
281
0
from .imports import is_tqdm_available if is_tqdm_available(): from tqdm.auto import tqdm as _tqdm from ..state import PartialState def _a ( SCREAMING_SNAKE_CASE : bool = True , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : int ) -> Dict: """simple docstring""" if not is_tqdm_available(): raise ImportError('Accelerate\'s `tqdm` module requires `tqdm` to be installed. Please run `pip install tqdm`.' ) __lowerCAmelCase: str = False if main_process_only: __lowerCAmelCase: Tuple = PartialState().local_process_index == 0 return _tqdm(*_snake_case , **_snake_case , disable=_snake_case )
322
import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin snake_case : str = "▁" snake_case : List[Any] = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class _snake_case ( snake_case , unittest.TestCase ): UpperCamelCase__ = BigBirdTokenizer UpperCamelCase__ = BigBirdTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True def SCREAMING_SNAKE_CASE ( self ): super().setUp() __magic_name__ : Optional[Any] = self.tokenizer_class(_a , keep_accents=_a ) tokenizer.save_pretrained(self.tmpdirname ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Union[str, Any] = "<s>" __magic_name__ : Dict = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_a ) , _a ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_a ) , _a ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "[MASK]" ) self.assertEqual(len(_a ) , 1_004 ) def SCREAMING_SNAKE_CASE ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def SCREAMING_SNAKE_CASE ( self ): if not self.test_rust_tokenizer: return __magic_name__ : Dict = self.get_tokenizer() __magic_name__ : str = self.get_rust_tokenizer() __magic_name__ : Any = "I was born in 92000, and this is falsé." __magic_name__ : Dict = tokenizer.tokenize(_a ) __magic_name__ : Any = rust_tokenizer.tokenize(_a ) self.assertListEqual(_a , _a ) __magic_name__ : List[Any] = tokenizer.encode(_a , add_special_tokens=_a ) __magic_name__ : List[str] = rust_tokenizer.encode(_a , add_special_tokens=_a ) self.assertListEqual(_a , _a ) __magic_name__ : str = self.get_rust_tokenizer() __magic_name__ : Dict = tokenizer.encode(_a ) __magic_name__ : Optional[int] = rust_tokenizer.encode(_a ) self.assertListEqual(_a , _a ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = BigBirdTokenizer(_a , keep_accents=_a ) __magic_name__ : str = tokenizer.tokenize("This is a test" ) self.assertListEqual(_a , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_a ) , [285, 46, 10, 170, 382] , ) __magic_name__ : Dict = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( _a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) __magic_name__ : Union[str, Any] = tokenizer.convert_tokens_to_ids(_a ) self.assertListEqual( _a , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) __magic_name__ : int = tokenizer.convert_ids_to_tokens(_a ) self.assertListEqual( _a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def SCREAMING_SNAKE_CASE ( self ): return BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Any = "Hello World!" __magic_name__ : Dict = [65, 18_536, 2_260, 101, 66] self.assertListEqual(_a , self.big_tokenizer.encode(_a ) ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) # fmt: off __magic_name__ : List[str] = [65, 871, 419, 358, 946, 991, 2_521, 452, 358, 1_357, 387, 7_751, 3_536, 112, 985, 456, 126, 865, 938, 5_400, 5_734, 458, 1_368, 467, 786, 2_462, 5_246, 1_159, 633, 865, 4_519, 457, 582, 852, 2_557, 427, 916, 508, 405, 34_324, 497, 391, 408, 11_342, 1_244, 385, 100, 938, 985, 456, 574, 362, 12_597, 3_200, 3_129, 1_172, 66] # noqa: E231 # fmt: on self.assertListEqual(_a , self.big_tokenizer.encode(_a ) ) @require_torch @slow def SCREAMING_SNAKE_CASE ( self ): import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence __magic_name__ : Optional[Any] = list(self.big_tokenizer.get_vocab().keys() )[:10] __magic_name__ : List[Any] = " ".join(_a ) __magic_name__ : Any = self.big_tokenizer.encode_plus(_a , return_tensors="pt" , return_token_type_ids=_a ) __magic_name__ : Union[str, Any] = self.big_tokenizer.batch_encode_plus( [sequence + " " + sequence] , return_tensors="pt" , return_token_type_ids=_a ) __magic_name__ : List[str] = BigBirdConfig(attention_type="original_full" ) __magic_name__ : Optional[int] = BigBirdModel(_a ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_a ) model(**_a ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : int = BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) __magic_name__ : int = tokenizer.decode(tokenizer("Paris is the [MASK]." ).input_ids ) self.assertTrue(decoded_text == "[CLS] Paris is the[MASK].[SEP]" ) @slow def SCREAMING_SNAKE_CASE ( self ): # fmt: off __magic_name__ : Optional[Any] = {"input_ids": [[65, 39_286, 458, 36_335, 2_001, 456, 13_073, 13_266, 455, 113, 7_746, 1_741, 11_157, 391, 13_073, 13_266, 455, 113, 3_967, 35_412, 113, 4_936, 109, 3_870, 2_377, 113, 30_084, 45_720, 458, 134, 17_496, 112, 503, 11_672, 113, 118, 112, 5_665, 13_347, 38_687, 112, 1_496, 31_389, 112, 3_268, 47_264, 134, 962, 112, 16_377, 8_035, 23_130, 430, 12_169, 15_518, 28_592, 458, 146, 41_697, 109, 391, 12_169, 15_518, 16_689, 458, 146, 41_358, 109, 452, 726, 4_034, 111, 763, 35_412, 5_082, 388, 1_903, 111, 9_051, 391, 2_870, 48_918, 1_900, 1_123, 550, 998, 112, 9_586, 15_985, 455, 391, 410, 22_955, 37_636, 114, 66], [65, 448, 17_496, 419, 3_663, 385, 763, 113, 27_533, 2_870, 3_283, 13_043, 1_639, 24_713, 523, 656, 24_013, 18_550, 2_521, 517, 27_014, 21_244, 420, 1_212, 1_465, 391, 927, 4_833, 388, 578, 11_786, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 484, 2_169, 7_687, 21_932, 18_146, 726, 363, 17_032, 3_391, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_a , model_name="google/bigbird-roberta-base" , revision="215c99f1600e06f83acce68422f2035b2b5c3510" , )
281
0
import importlib import os import fsspec import pytest from fsspec import register_implementation from fsspec.registry import _registry as _fsspec_registry from datasets.filesystems import COMPRESSION_FILESYSTEMS, HfFileSystem, extract_path_from_uri, is_remote_filesystem from .utils import require_lza, require_zstandard def _UpperCamelCase ( snake_case__ ) -> List[Any]: assert "mock" in _fsspec_registry assert "bz2" in _fsspec_registry def _UpperCamelCase ( ) -> Tuple: assert "mock" not in _fsspec_registry assert "bz2" in _fsspec_registry def _UpperCamelCase ( ) -> Union[str, Any]: __UpperCAmelCase : Dict = "mock-s3-bucket" __UpperCAmelCase : Any = f'''s3://{mock_bucket}''' __UpperCAmelCase : str = extract_path_from_uri(_snake_case ) assert dataset_path.startswith("s3://" ) is False __UpperCAmelCase : Tuple = "./local/path" __UpperCAmelCase : Optional[Any] = extract_path_from_uri(_snake_case ) assert dataset_path == new_dataset_path def _UpperCamelCase ( snake_case__ ) -> Optional[Any]: __UpperCAmelCase : str = is_remote_filesystem(_snake_case ) assert is_remote is True __UpperCAmelCase : Optional[int] = fsspec.filesystem("file" ) __UpperCAmelCase : int = is_remote_filesystem(_snake_case ) assert is_remote is False @pytest.mark.parametrize("compression_fs_class", _snake_case ) def _UpperCamelCase ( snake_case__, snake_case__, snake_case__, snake_case__, snake_case__, snake_case__, snake_case__ ) -> int: __UpperCAmelCase : Any = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_file, "bz2": bza_file, "lz4": lza_file} __UpperCAmelCase : str = input_paths[compression_fs_class.protocol] if input_path is None: __UpperCAmelCase : Dict = f'''for \'{compression_fs_class.protocol}\' compression protocol, ''' if compression_fs_class.protocol == "lz4": reason += require_lza.kwargs["reason"] elif compression_fs_class.protocol == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(_snake_case ) __UpperCAmelCase : str = fsspec.filesystem(compression_fs_class.protocol, fo=_snake_case ) assert isinstance(_snake_case, _snake_case ) __UpperCAmelCase : int = os.path.basename(_snake_case ) __UpperCAmelCase : Optional[int] = expected_filename[: expected_filename.rindex("." )] assert fs.glob("*" ) == [expected_filename] with fs.open(_snake_case, "r", encoding="utf-8" ) as f, open(_snake_case, encoding="utf-8" ) as expected_file: assert f.read() == expected_file.read() @pytest.mark.parametrize("protocol", ["zip", "gzip"] ) def _UpperCamelCase ( snake_case__, snake_case__, snake_case__ ) -> str: __UpperCAmelCase : int = {"zip": zip_jsonl_path, "gzip": jsonl_gz_path} __UpperCAmelCase : int = compressed_file_paths[protocol] __UpperCAmelCase : Tuple = "dataset.jsonl" __UpperCAmelCase : List[str] = f'''{protocol}://{member_file_path}::{compressed_file_path}''' __UpperCAmelCase : Optional[Any] = fsspec.get_fs_token_paths(_snake_case ) assert fs.isfile(_snake_case ) assert not fs.isfile("non_existing_" + member_file_path ) @pytest.mark.integration def _UpperCamelCase ( snake_case__, snake_case__, snake_case__, snake_case__ ) -> str: __UpperCAmelCase : int = hf_api.dataset_info(_snake_case, token=_snake_case ) __UpperCAmelCase : Optional[Any] = HfFileSystem(repo_info=_snake_case, token=_snake_case ) assert sorted(hffs.glob("*" ) ) == [".gitattributes", "data"] assert hffs.isdir("data" ) assert hffs.isfile(".gitattributes" ) and hffs.isfile("data/text_data.txt" ) with open(_snake_case ) as f: assert hffs.open("data/text_data.txt", "r" ).read() == f.read() def _UpperCamelCase ( ) -> Optional[int]: __UpperCAmelCase : Optional[Any] = "bz2" # Import module import datasets.filesystems # Overwrite protocol and reload register_implementation(_snake_case, _snake_case, clobber=_snake_case ) with pytest.warns(_snake_case ) as warning_info: importlib.reload(datasets.filesystems ) assert len(_snake_case ) == 1 assert ( str(warning_info[0].message ) == f'''A filesystem protocol was already set for {protocol} and will be overwritten.''' )
157
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging snake_case : int = logging.get_logger(__name__) snake_case : List[str] = {"vocab_file": "spiece.model"} snake_case : List[str] = { "vocab_file": { "albert-base-v1": "https://huggingface.co/albert-base-v1/resolve/main/spiece.model", "albert-large-v1": "https://huggingface.co/albert-large-v1/resolve/main/spiece.model", "albert-xlarge-v1": "https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model", "albert-xxlarge-v1": "https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model", "albert-base-v2": "https://huggingface.co/albert-base-v2/resolve/main/spiece.model", "albert-large-v2": "https://huggingface.co/albert-large-v2/resolve/main/spiece.model", "albert-xlarge-v2": "https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model", "albert-xxlarge-v2": "https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model", } } snake_case : Tuple = { "albert-base-v1": 512, "albert-large-v1": 512, "albert-xlarge-v1": 512, "albert-xxlarge-v1": 512, "albert-base-v2": 512, "albert-large-v2": 512, "albert-xlarge-v2": 512, "albert-xxlarge-v2": 512, } snake_case : List[str] = "▁" class _snake_case ( snake_case ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , _a , _a=True , _a=True , _a=False , _a="[CLS]" , _a="[SEP]" , _a="<unk>" , _a="[SEP]" , _a="<pad>" , _a="[CLS]" , _a="[MASK]" , _a = None , **_a , ): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. __magic_name__ : str = ( AddedToken(_a , lstrip=_a , rstrip=_a , normalized=_a ) if isinstance(_a , _a ) else mask_token ) __magic_name__ : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=_a , remove_space=_a , keep_accents=_a , bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , pad_token=_a , cls_token=_a , mask_token=_a , sp_model_kwargs=self.sp_model_kwargs , **_a , ) __magic_name__ : Dict = do_lower_case __magic_name__ : Tuple = remove_space __magic_name__ : Union[str, Any] = keep_accents __magic_name__ : Tuple = vocab_file __magic_name__ : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_a ) @property def SCREAMING_SNAKE_CASE ( self ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[str] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): __magic_name__ : List[str] = self.__dict__.copy() __magic_name__ : Any = None return state def __setstate__( self , _a ): __magic_name__ : Union[str, Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): __magic_name__ : str = {} __magic_name__ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self , _a ): if self.remove_space: __magic_name__ : List[Any] = " ".join(inputs.strip().split() ) else: __magic_name__ : str = inputs __magic_name__ : int = outputs.replace("``" , "\"" ).replace("''" , "\"" ) if not self.keep_accents: __magic_name__ : str = unicodedata.normalize("NFKD" , _a ) __magic_name__ : Tuple = "".join([c for c in outputs if not unicodedata.combining(_a )] ) if self.do_lower_case: __magic_name__ : int = outputs.lower() return outputs def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Optional[Any] = self.preprocess_text(_a ) __magic_name__ : Dict = self.sp_model.encode(_a , out_type=_a ) __magic_name__ : Any = [] for piece in pieces: if len(_a ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): __magic_name__ : Optional[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(_a , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __magic_name__ : List[str] = cur_pieces[1:] else: __magic_name__ : Optional[int] = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(_a ) else: new_pieces.append(_a ) return new_pieces def SCREAMING_SNAKE_CASE ( self , _a ): return self.sp_model.PieceToId(_a ) def SCREAMING_SNAKE_CASE ( self , _a ): return self.sp_model.IdToPiece(_a ) def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Any = [] __magic_name__ : Union[str, Any] = "" __magic_name__ : int = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(_a ) + token __magic_name__ : List[Any] = True __magic_name__ : Optional[int] = [] else: current_sub_tokens.append(_a ) __magic_name__ : Optional[Any] = False out_string += self.sp_model.decode(_a ) return out_string.strip() def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): __magic_name__ : List[str] = [self.sep_token_id] __magic_name__ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self , _a , _a = None , _a = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a , token_ids_a=_a , already_has_special_tokens=_a ) if token_ids_a is not None: return [1] + ([0] * len(_a )) + [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1] def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): __magic_name__ : Optional[int] = [self.sep_token_id] __magic_name__ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): if not os.path.isdir(_a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __magic_name__ : List[str] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _a ) elif not os.path.isfile(self.vocab_file ): with open(_a , "wb" ) as fi: __magic_name__ : Union[str, Any] = self.sp_model.serialized_model_proto() fi.write(_a ) return (out_vocab_file,)
281
0
"""simple docstring""" import os from argparse import ArgumentParser from typing import List import torch.utils.data from datasets import Dataset, IterableDataset from datasets.distributed import split_dataset_by_node lowerCamelCase__ = 4 lowerCamelCase__ = 3 class A__ ( _lowerCamelCase): pass def __lowerCAmelCase (_UpperCamelCase ): for shard in shards: for i in range(_snake_case ): yield {"i": i, "shard": shard} def __lowerCAmelCase (): __lowerCAmelCase : Optional[int] = int(os.environ['RANK'] ) __lowerCAmelCase : Union[str, Any] = int(os.environ['WORLD_SIZE'] ) __lowerCAmelCase : Optional[int] = ArgumentParser() parser.add_argument('--streaming' , type=_snake_case ) parser.add_argument('--local_rank' , type=_snake_case ) parser.add_argument('--num_workers' , type=_snake_case , default=0 ) __lowerCAmelCase : Tuple = parser.parse_args() __lowerCAmelCase : Optional[Any] = args.streaming __lowerCAmelCase : Any = args.num_workers __lowerCAmelCase : Optional[int] = {"shards": [F"shard_{shard_idx}" for shard_idx in range(_snake_case )]} __lowerCAmelCase : List[Any] = IterableDataset.from_generator(_snake_case , gen_kwargs=_snake_case ) if not streaming: __lowerCAmelCase : Dict = Dataset.from_list(list(_snake_case ) ) __lowerCAmelCase : List[str] = split_dataset_by_node(_snake_case , rank=_snake_case , world_size=_snake_case ) __lowerCAmelCase : Optional[Any] = torch.utils.data.DataLoader(_snake_case , num_workers=_snake_case ) __lowerCAmelCase : Optional[int] = NUM_SHARDS * NUM_ITEMS_PER_SHARD __lowerCAmelCase : Dict = full_size // world_size expected_local_size += int(rank < (full_size % world_size) ) __lowerCAmelCase : Optional[int] = sum(1 for _ in dataloader ) if local_size != expected_local_size: raise FailedTestError(F"local_size {local_size} != expected_local_size {expected_local_size}" ) if __name__ == "__main__": main()
86
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : Tuple , _snake_case : List[Any] , _snake_case : Optional[Any] ) -> Optional[Any]: '''simple docstring''' if isinstance(_snake_case , _snake_case ): __magic_name__ : Union[str, Any] = np.full((len(_snake_case ), sequence_length, 2) , _snake_case ) else: __magic_name__ : List[Any] = np.full((len(_snake_case ), sequence_length) , _snake_case ) for i, tensor in enumerate(_snake_case ): if padding_side == "right": if isinstance(_snake_case , _snake_case ): __magic_name__ : Optional[Any] = tensor[:sequence_length] else: __magic_name__ : Union[str, Any] = tensor[:sequence_length] else: if isinstance(_snake_case , _snake_case ): __magic_name__ : List[Any] = tensor[:sequence_length] else: __magic_name__ : Optional[Any] = tensor[:sequence_length] return out_tensor.tolist() def lowerCAmelCase_ ( _snake_case : Optional[int] ) -> Tuple: '''simple docstring''' __magic_name__ : Union[str, Any] = ord(_snake_case ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 123 and cp <= 126): return True __magic_name__ : Any = unicodedata.category(_snake_case ) if cat.startswith("P" ): return True return False @dataclass class _snake_case ( snake_case ): UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = -100 UpperCamelCase__ = "pt" def SCREAMING_SNAKE_CASE ( self , _a ): import torch __magic_name__ : List[str] = "label" if "label" in features[0].keys() else "labels" __magic_name__ : Union[str, Any] = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __magic_name__ : Optional[int] = self.tokenizer.pad( _a , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="pt" if labels is None else None , ) if labels is None: return batch __magic_name__ : Dict = torch.tensor(batch["entity_ids"] ).shape[1] __magic_name__ : List[Any] = self.tokenizer.padding_side if padding_side == "right": __magic_name__ : str = [ list(_a ) + [self.label_pad_token_id] * (sequence_length - len(_a )) for label in labels ] else: __magic_name__ : int = [ [self.label_pad_token_id] * (sequence_length - len(_a )) + list(_a ) for label in labels ] __magic_name__ : Dict = [feature["ner_tags"] for feature in features] __magic_name__ : List[Any] = padding_tensor(_a , -1 , _a , _a ) __magic_name__ : Any = [feature["original_entity_spans"] for feature in features] __magic_name__ : Any = padding_tensor(_a , (-1, -1) , _a , _a ) __magic_name__ : List[Any] = {k: torch.tensor(_a , dtype=torch.intaa ) for k, v in batch.items()} return batch
281
0
import math import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from .attention_processor import Attention from .embeddings import get_timestep_embedding from .modeling_utils import ModelMixin class __snake_case ( lowerCamelCase_ , lowerCamelCase_ ): @register_to_config def __init__( self : List[str] , _lowercase : Dict = 1_28 , _lowercase : List[Any] = 2_56 , _lowercase : str = 20_00.0 , _lowercase : List[Any] = 7_68 , _lowercase : Union[str, Any] = 12 , _lowercase : Dict = 12 , _lowercase : Any = 64 , _lowercase : Union[str, Any] = 20_48 , _lowercase : List[str] = 0.1 , ): """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE__ = nn.Sequential( nn.Linear(_a , d_model * 4 , bias=_a ) , nn.SiLU() , nn.Linear(d_model * 4 , d_model * 4 , bias=_a ) , nn.SiLU() , ) SCREAMING_SNAKE_CASE__ = nn.Embedding(_a , _a ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = nn.Linear(_a , _a , bias=_a ) SCREAMING_SNAKE_CASE__ = nn.Dropout(p=_a ) SCREAMING_SNAKE_CASE__ = nn.ModuleList() for lyr_num in range(_a ): # FiLM conditional T5 decoder SCREAMING_SNAKE_CASE__ = DecoderLayer(d_model=_a , d_kv=_a , num_heads=_a , d_ff=_a , dropout_rate=_a ) self.decoders.append(_a ) SCREAMING_SNAKE_CASE__ = TaLayerNorm(_a ) SCREAMING_SNAKE_CASE__ = nn.Dropout(p=_a ) SCREAMING_SNAKE_CASE__ = nn.Linear(_a , _a , bias=_a ) def __a ( self : List[Any] , _lowercase : Dict , _lowercase : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ = torch.mul(query_input.unsqueeze(-1 ) , key_input.unsqueeze(-2 ) ) return mask.unsqueeze(-3 ) def __a ( self : Optional[int] , _lowercase : str , _lowercase : Optional[int] , _lowercase : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ = decoder_input_tokens.shape assert decoder_noise_time.shape == (batch,) # decoder_noise_time is in [0, 1), so rescale to expected timing range. SCREAMING_SNAKE_CASE__ = get_timestep_embedding( decoder_noise_time * self.config.max_decoder_noise_time , embedding_dim=self.config.d_model , max_period=self.config.max_decoder_noise_time , ).to(dtype=self.dtype ) SCREAMING_SNAKE_CASE__ = self.conditioning_emb(_a ).unsqueeze(1 ) assert conditioning_emb.shape == (batch, 1, self.config.d_model * 4) SCREAMING_SNAKE_CASE__ = decoder_input_tokens.shape[1] # If we want to use relative positions for audio context, we can just offset # this sequence by the length of encodings_and_masks. SCREAMING_SNAKE_CASE__ = torch.broadcast_to( torch.arange(_a , device=decoder_input_tokens.device ) , (batch, seq_length) , ) SCREAMING_SNAKE_CASE__ = self.position_encoding(_a ) SCREAMING_SNAKE_CASE__ = self.continuous_inputs_projection(_a ) inputs += position_encodings SCREAMING_SNAKE_CASE__ = self.dropout(_a ) # decoder: No padding present. SCREAMING_SNAKE_CASE__ = torch.ones( decoder_input_tokens.shape[:2] , device=decoder_input_tokens.device , dtype=inputs.dtype ) # Translate encoding masks to encoder-decoder masks. SCREAMING_SNAKE_CASE__ = [(x, self.encoder_decoder_mask(_a , _a )) for x, y in encodings_and_masks] # cross attend style: concat encodings SCREAMING_SNAKE_CASE__ = torch.cat([x[0] for x in encodings_and_encdec_masks] , dim=1 ) SCREAMING_SNAKE_CASE__ = torch.cat([x[1] for x in encodings_and_encdec_masks] , dim=-1 ) for lyr in self.decoders: SCREAMING_SNAKE_CASE__ = lyr( _a , conditioning_emb=_a , encoder_hidden_states=_a , encoder_attention_mask=_a , )[0] SCREAMING_SNAKE_CASE__ = self.decoder_norm(_a ) SCREAMING_SNAKE_CASE__ = self.post_dropout(_a ) SCREAMING_SNAKE_CASE__ = self.spec_out(_a ) return spec_out class __snake_case ( nn.Module ): def __init__( self : Tuple , _lowercase : Any , _lowercase : List[str] , _lowercase : Tuple , _lowercase : List[Any] , _lowercase : str , _lowercase : List[str]=1E-6 ): """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE__ = nn.ModuleList() # cond self attention: layer 0 self.layer.append( TaLayerSelfAttentionCond(d_model=_a , d_kv=_a , num_heads=_a , dropout_rate=_a ) ) # cross attention: layer 1 self.layer.append( TaLayerCrossAttention( d_model=_a , d_kv=_a , num_heads=_a , dropout_rate=_a , layer_norm_epsilon=_a , ) ) # Film Cond MLP + dropout: last layer self.layer.append( TaLayerFFCond(d_model=_a , d_ff=_a , dropout_rate=_a , layer_norm_epsilon=_a ) ) def __a ( self : int , _lowercase : Optional[Any] , _lowercase : Union[str, Any]=None , _lowercase : Optional[int]=None , _lowercase : Dict=None , _lowercase : List[Any]=None , _lowercase : int=None , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.layer[0]( _a , conditioning_emb=_a , attention_mask=_a , ) if encoder_hidden_states is not None: SCREAMING_SNAKE_CASE__ = torch.where(encoder_attention_mask > 0 , 0 , -1E10 ).to( encoder_hidden_states.dtype ) SCREAMING_SNAKE_CASE__ = self.layer[1]( _a , key_value_states=_a , attention_mask=_a , ) # Apply Film Conditional Feed Forward layer SCREAMING_SNAKE_CASE__ = self.layer[-1](_a , _a ) return (hidden_states,) class __snake_case ( nn.Module ): def __init__( self : List[Any] , _lowercase : Any , _lowercase : Tuple , _lowercase : Dict , _lowercase : int ): """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE__ = TaLayerNorm(_a ) SCREAMING_SNAKE_CASE__ = TaFiLMLayer(in_features=d_model * 4 , out_features=_a ) SCREAMING_SNAKE_CASE__ = Attention(query_dim=_a , heads=_a , dim_head=_a , out_bias=_a , scale_qk=_a ) SCREAMING_SNAKE_CASE__ = nn.Dropout(_a ) def __a ( self : Dict , _lowercase : Any , _lowercase : Optional[int]=None , _lowercase : Optional[int]=None , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.layer_norm(_a ) if conditioning_emb is not None: SCREAMING_SNAKE_CASE__ = self.FiLMLayer(_a , _a ) # Self-attention block SCREAMING_SNAKE_CASE__ = self.attention(_a ) SCREAMING_SNAKE_CASE__ = hidden_states + self.dropout(_a ) return hidden_states class __snake_case ( nn.Module ): def __init__( self : Any , _lowercase : Optional[int] , _lowercase : Tuple , _lowercase : Union[str, Any] , _lowercase : Union[str, Any] , _lowercase : Dict ): """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE__ = Attention(query_dim=_a , heads=_a , dim_head=_a , out_bias=_a , scale_qk=_a ) SCREAMING_SNAKE_CASE__ = TaLayerNorm(_a , eps=_a ) SCREAMING_SNAKE_CASE__ = nn.Dropout(_a ) def __a ( self : Tuple , _lowercase : Dict , _lowercase : List[Any]=None , _lowercase : Any=None , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.layer_norm(_a ) SCREAMING_SNAKE_CASE__ = self.attention( _a , encoder_hidden_states=_a , attention_mask=attention_mask.squeeze(1 ) , ) SCREAMING_SNAKE_CASE__ = hidden_states + self.dropout(_a ) return layer_output class __snake_case ( nn.Module ): def __init__( self : List[str] , _lowercase : int , _lowercase : int , _lowercase : str , _lowercase : List[Any] ): """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE__ = TaDenseGatedActDense(d_model=_a , d_ff=_a , dropout_rate=_a ) SCREAMING_SNAKE_CASE__ = TaFiLMLayer(in_features=d_model * 4 , out_features=_a ) SCREAMING_SNAKE_CASE__ = TaLayerNorm(_a , eps=_a ) SCREAMING_SNAKE_CASE__ = nn.Dropout(_a ) def __a ( self : Tuple , _lowercase : Union[str, Any] , _lowercase : int=None ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.layer_norm(_a ) if conditioning_emb is not None: SCREAMING_SNAKE_CASE__ = self.film(_a , _a ) SCREAMING_SNAKE_CASE__ = self.DenseReluDense(_a ) SCREAMING_SNAKE_CASE__ = hidden_states + self.dropout(_a ) return hidden_states class __snake_case ( nn.Module ): def __init__( self : Union[str, Any] , _lowercase : Optional[int] , _lowercase : Optional[int] , _lowercase : Tuple ): """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE__ = nn.Linear(_a , _a , bias=_a ) SCREAMING_SNAKE_CASE__ = nn.Linear(_a , _a , bias=_a ) SCREAMING_SNAKE_CASE__ = nn.Linear(_a , _a , bias=_a ) SCREAMING_SNAKE_CASE__ = nn.Dropout(_a ) SCREAMING_SNAKE_CASE__ = NewGELUActivation() def __a ( self : str , _lowercase : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.act(self.wi_a(_a ) ) SCREAMING_SNAKE_CASE__ = self.wi_a(_a ) SCREAMING_SNAKE_CASE__ = hidden_gelu * hidden_linear SCREAMING_SNAKE_CASE__ = self.dropout(_a ) SCREAMING_SNAKE_CASE__ = self.wo(_a ) return hidden_states class __snake_case ( nn.Module ): def __init__( self : List[Any] , _lowercase : Any , _lowercase : List[str]=1E-6 ): """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE__ = nn.Parameter(torch.ones(_a ) ) SCREAMING_SNAKE_CASE__ = eps def __a ( self : Optional[int] , _lowercase : Any ): """simple docstring""" SCREAMING_SNAKE_CASE__ = hidden_states.to(torch.floataa ).pow(2 ).mean(-1 , keepdim=_a ) SCREAMING_SNAKE_CASE__ = hidden_states * torch.rsqrt(variance + self.variance_epsilon ) # convert into half-precision if necessary if self.weight.dtype in [torch.floataa, torch.bfloataa]: SCREAMING_SNAKE_CASE__ = hidden_states.to(self.weight.dtype ) return self.weight * hidden_states class __snake_case ( nn.Module ): def __a ( self : Any , _lowercase : str ): """simple docstring""" return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi ) * (input + 0.04_47_15 * torch.pow(_a , 3.0 )) )) class __snake_case ( nn.Module ): def __init__( self : List[Any] , _lowercase : Any , _lowercase : List[str] ): """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE__ = nn.Linear(_a , out_features * 2 , bias=_a ) def __a ( self : Union[str, Any] , _lowercase : Any , _lowercase : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.scale_bias(_a ) SCREAMING_SNAKE_CASE__ = torch.chunk(_a , 2 , -1 ) SCREAMING_SNAKE_CASE__ = x * (1 + scale) + shift return x
219
import math def lowerCAmelCase_ ( _snake_case : float , _snake_case : float ) -> float: '''simple docstring''' return math.pow(_snake_case , 2 ) - a def lowerCAmelCase_ ( _snake_case : float ) -> float: '''simple docstring''' return 2 * x def lowerCAmelCase_ ( _snake_case : float ) -> float: '''simple docstring''' __magic_name__ : Optional[int] = 2.0 while start <= a: __magic_name__ : str = math.pow(_snake_case , 2 ) return start def lowerCAmelCase_ ( _snake_case : float , _snake_case : int = 9999 , _snake_case : float = 0.00_000_000_000_001 ) -> float: '''simple docstring''' if a < 0: raise ValueError("math domain error" ) __magic_name__ : Optional[int] = get_initial_point(_snake_case ) for _ in range(_snake_case ): __magic_name__ : int = value __magic_name__ : str = value - fx(_snake_case , _snake_case ) / fx_derivative(_snake_case ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
281
0
'''simple docstring''' import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowercase ( A__ , A__ , unittest.TestCase ): """simple docstring""" _a = IFImgaImgSuperResolutionPipeline _a = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'width', 'height'} _a = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'original_image'} ) _a = PipelineTesterMixin.required_optional_params - {'latents'} def lowerCAmelCase__ ( self ): '''simple docstring''' return self._get_superresolution_dummy_components() def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_=0 ): '''simple docstring''' if str(_a ).startswith('''mps''' ): UpperCamelCase__ :Dict = torch.manual_seed(_a ) else: UpperCamelCase__ :Any = torch.Generator(device=_a ).manual_seed(_a ) UpperCamelCase__ :Union[str, Any] = floats_tensor((1, 3, 32, 32) , rng=random.Random(_a ) ).to(_a ) UpperCamelCase__ :Dict = floats_tensor((1, 3, 16, 16) , rng=random.Random(_a ) ).to(_a ) UpperCamelCase__ :List[str] = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def lowerCAmelCase__ ( self ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def lowerCAmelCase__ ( self ): '''simple docstring''' self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def lowerCAmelCase__ ( self ): '''simple docstring''' super().test_save_load_floataa(expected_max_diff=1e-1 ) def lowerCAmelCase__ ( self ): '''simple docstring''' self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def lowerCAmelCase__ ( self ): '''simple docstring''' self._test_save_load_local() def lowerCAmelCase__ ( self ): '''simple docstring''' self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
97
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class _snake_case : UpperCamelCase__ = LEDConfig UpperCamelCase__ = {} UpperCamelCase__ = 'gelu' def __init__( self , _a , _a=13 , _a=7 , _a=True , _a=False , _a=99 , _a=32 , _a=2 , _a=4 , _a=37 , _a=0.1 , _a=0.1 , _a=20 , _a=2 , _a=1 , _a=0 , _a=4 , ): __magic_name__ : int = parent __magic_name__ : Optional[int] = batch_size __magic_name__ : Tuple = seq_length __magic_name__ : List[Any] = is_training __magic_name__ : Dict = use_labels __magic_name__ : Optional[Any] = vocab_size __magic_name__ : int = hidden_size __magic_name__ : Optional[int] = num_hidden_layers __magic_name__ : Optional[int] = num_attention_heads __magic_name__ : Tuple = intermediate_size __magic_name__ : Any = hidden_dropout_prob __magic_name__ : Optional[int] = attention_probs_dropout_prob __magic_name__ : List[str] = max_position_embeddings __magic_name__ : Any = eos_token_id __magic_name__ : str = pad_token_id __magic_name__ : int = bos_token_id __magic_name__ : Optional[int] = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after __magic_name__ : Tuple = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests __magic_name__ : Tuple = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __magic_name__ : int = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __magic_name__ : int = tf.concat([input_ids, eos_tensor] , axis=1 ) __magic_name__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __magic_name__ : Dict = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) __magic_name__ : List[str] = prepare_led_inputs_dict(_a , _a , _a ) __magic_name__ : Union[str, Any] = tf.concat( [tf.zeros_like(_a )[:, :-1], tf.ones_like(_a )[:, -1:]] , axis=-1 , ) __magic_name__ : List[Any] = global_attention_mask return config, inputs_dict def SCREAMING_SNAKE_CASE ( self , _a , _a ): __magic_name__ : Dict = TFLEDModel(config=_a ).get_decoder() __magic_name__ : Optional[int] = inputs_dict["input_ids"] __magic_name__ : Union[str, Any] = input_ids[:1, :] __magic_name__ : str = inputs_dict["attention_mask"][:1, :] __magic_name__ : int = 1 # first forward pass __magic_name__ : Tuple = model(_a , attention_mask=_a , use_cache=_a ) __magic_name__ , __magic_name__ : str = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __magic_name__ : Optional[int] = ids_tensor((self.batch_size, 3) , config.vocab_size ) __magic_name__ : Any = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __magic_name__ : Optional[Any] = tf.concat([input_ids, next_tokens] , axis=-1 ) __magic_name__ : List[Any] = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __magic_name__ : List[str] = model(_a , attention_mask=_a )[0] __magic_name__ : Dict = model(_a , attention_mask=_a , past_key_values=_a )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __magic_name__ : List[Any] = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __magic_name__ : Union[str, Any] = output_from_no_past[:, -3:, random_slice_idx] __magic_name__ : List[str] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_a , _a , rtol=1e-3 ) def lowerCAmelCase_ ( _snake_case : Any , _snake_case : List[Any] , _snake_case : Any , _snake_case : str=None , _snake_case : List[str]=None , _snake_case : int=None , _snake_case : Any=None , ) -> int: '''simple docstring''' if attention_mask is None: __magic_name__ : str = tf.cast(tf.math.not_equal(_snake_case , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: __magic_name__ : List[Any] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: __magic_name__ : Dict = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: __magic_name__ : Any = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class _snake_case ( snake_case , snake_case , unittest.TestCase ): UpperCamelCase__ = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () UpperCamelCase__ = (TFLEDForConditionalGeneration,) if is_tf_available() else () UpperCamelCase__ = ( { 'conversational': TFLEDForConditionalGeneration, 'feature-extraction': TFLEDModel, 'summarization': TFLEDForConditionalGeneration, 'text2text-generation': TFLEDForConditionalGeneration, 'translation': TFLEDForConditionalGeneration, } if is_tf_available() else {} ) UpperCamelCase__ = True UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = TFLEDModelTester(self ) __magic_name__ : List[Any] = ConfigTester(self , config_class=_a ) def SCREAMING_SNAKE_CASE ( self ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_a ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ , __magic_name__ : int = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ : List[str] = tf.zeros_like(inputs_dict["attention_mask"] ) __magic_name__ : Optional[Any] = 2 __magic_name__ : Tuple = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict["global_attention_mask"] , ) __magic_name__ : Any = True __magic_name__ : str = self.model_tester.seq_length __magic_name__ : Dict = self.model_tester.encoder_seq_length def check_decoder_attentions_output(_a ): __magic_name__ : str = outputs.decoder_attentions self.assertEqual(len(_a ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(_a ): __magic_name__ : Any = [t.numpy() for t in outputs.encoder_attentions] __magic_name__ : Tuple = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(_a ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(_a ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: __magic_name__ : Union[str, Any] = True __magic_name__ : List[str] = False __magic_name__ : Tuple = False __magic_name__ : Optional[int] = model_class(_a ) __magic_name__ : str = model(self._prepare_for_class(_a , _a ) ) __magic_name__ : Any = len(_a ) self.assertEqual(config.output_hidden_states , _a ) check_encoder_attentions_output(_a ) if self.is_encoder_decoder: __magic_name__ : Tuple = model_class(_a ) __magic_name__ : Optional[Any] = model(self._prepare_for_class(_a , _a ) ) self.assertEqual(config.output_hidden_states , _a ) check_decoder_attentions_output(_a ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] __magic_name__ : Dict = True __magic_name__ : str = model_class(_a ) __magic_name__ : Any = model(self._prepare_for_class(_a , _a ) ) self.assertEqual(config.output_hidden_states , _a ) check_encoder_attentions_output(_a ) # Check attention is always last and order is fine __magic_name__ : Union[str, Any] = True __magic_name__ : Union[str, Any] = True __magic_name__ : List[str] = model_class(_a ) __magic_name__ : Any = model(self._prepare_for_class(_a , _a ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(_a ) ) self.assertEqual(model.config.output_hidden_states , _a ) check_encoder_attentions_output(_a ) @unittest.skip("LED keeps using potentially symbolic tensors in conditionals and breaks tracing." ) def SCREAMING_SNAKE_CASE ( self ): pass def SCREAMING_SNAKE_CASE ( self ): # TODO: Head-masking not yet implement pass def lowerCAmelCase_ ( _snake_case : int ) -> Optional[int]: '''simple docstring''' return tf.constant(_snake_case , dtype=tf.intaa ) snake_case : Optional[int] = 1E-4 @slow @require_tf class _snake_case ( unittest.TestCase ): def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[Any] = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ).led # change to intended input here __magic_name__ : Optional[int] = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) __magic_name__ : str = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) __magic_name__ : Any = prepare_led_inputs_dict(model.config , _a , _a ) __magic_name__ : List[Any] = model(**_a )[0] __magic_name__ : List[str] = (1, 1_024, 768) self.assertEqual(output.shape , _a ) # change to expected output here __magic_name__ : int = tf.convert_to_tensor( [[2.30_50, 2.82_79, 0.65_31], [-1.84_57, -0.14_55, -3.56_61], [-1.01_86, 0.45_86, -2.20_43]] , ) tf.debugging.assert_near(output[:, :3, :3] , _a , atol=1e-3 ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Tuple = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ) # change to intended input here __magic_name__ : int = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) __magic_name__ : Tuple = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) __magic_name__ : Optional[Any] = prepare_led_inputs_dict(model.config , _a , _a ) __magic_name__ : Union[str, Any] = model(**_a )[0] __magic_name__ : Optional[int] = (1, 1_024, model.config.vocab_size) self.assertEqual(output.shape , _a ) # change to expected output here __magic_name__ : str = tf.convert_to_tensor( [[33.65_07, 6.45_72, 16.80_89], [5.87_39, -2.42_38, 11.29_02], [-3.21_39, -4.31_49, 4.27_83]] , ) tf.debugging.assert_near(output[:, :3, :3] , _a , atol=1e-3 , rtol=1e-3 )
281
0
"""simple docstring""" import unittest import numpy as np from transformers.file_utils import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=7 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=18 , _SCREAMING_SNAKE_CASE=30 , _SCREAMING_SNAKE_CASE=400 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=[0.5, 0.5, 0.5] , _SCREAMING_SNAKE_CASE=[0.5, 0.5, 0.5] , ) -> Tuple: '''simple docstring''' UpperCAmelCase : int = size if size is not None else {"height": 18, "width": 18} UpperCAmelCase : Union[str, Any] = parent UpperCAmelCase : int = batch_size UpperCAmelCase : List[Any] = num_channels UpperCAmelCase : str = image_size UpperCAmelCase : Optional[int] = min_resolution UpperCAmelCase : List[str] = max_resolution UpperCAmelCase : str = do_resize UpperCAmelCase : Optional[int] = size UpperCAmelCase : Tuple = do_normalize UpperCAmelCase : Optional[Any] = image_mean UpperCAmelCase : int = image_std def SCREAMING_SNAKE_CASE ( self ) -> Dict: '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ , unittest.TestCase ): __lowerCAmelCase : str = DPTImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self ) -> Tuple: '''simple docstring''' UpperCAmelCase : List[Any] = DPTImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self ) -> Dict: '''simple docstring''' UpperCAmelCase : List[str] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_a , """image_mean""" ) ) self.assertTrue(hasattr(_a , """image_std""" ) ) self.assertTrue(hasattr(_a , """do_normalize""" ) ) self.assertTrue(hasattr(_a , """do_resize""" ) ) self.assertTrue(hasattr(_a , """size""" ) ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase : int = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 18} ) UpperCAmelCase : List[str] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_a ) for image in image_inputs: self.assertIsInstance(_a , Image.Image ) # Test not batched input UpperCAmelCase : Optional[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched UpperCAmelCase : str = image_processing(_a , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: '''simple docstring''' UpperCAmelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_a , numpify=_a ) for image in image_inputs: self.assertIsInstance(_a , np.ndarray ) # Test not batched input UpperCAmelCase : Optional[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched UpperCAmelCase : List[Any] = image_processing(_a , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase : int = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_a , torchify=_a ) for image in image_inputs: self.assertIsInstance(_a , torch.Tensor ) # Test not batched input UpperCAmelCase : str = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched UpperCAmelCase : Any = image_processing(_a , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , )
109
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() snake_case : Optional[Any] = logging.get_logger(__name__) def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Union[str, Any]=False ) -> List[str]: '''simple docstring''' __magic_name__ : Union[str, Any] = [] # fmt: off # stem: rename_keys.append(("cls_token", "vit.embeddings.cls_token") ) rename_keys.append(("pos_embed", "vit.embeddings.position_embeddings") ) rename_keys.append(("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight") ) rename_keys.append(("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias") ) # backbone rename_keys.append(("patch_embed.backbone.stem.conv.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.bias", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias") ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias''') ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''vit.encoder.layer.{i}.output.dense.bias''') ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" __magic_name__ : int = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) # fmt: on return rename_keys def lowerCAmelCase_ ( _snake_case : Any , _snake_case : Any , _snake_case : Dict=False ) -> int: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: __magic_name__ : int = "" else: __magic_name__ : Union[str, Any] = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __magic_name__ : Optional[Any] = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) __magic_name__ : int = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict __magic_name__ : Dict = in_proj_weight[ : config.hidden_size, : ] __magic_name__ : List[str] = in_proj_bias[: config.hidden_size] __magic_name__ : List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __magic_name__ : Optional[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __magic_name__ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] __magic_name__ : int = in_proj_bias[-config.hidden_size :] def lowerCAmelCase_ ( _snake_case : List[str] ) -> List[str]: '''simple docstring''' __magic_name__ : List[str] = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(_snake_case , _snake_case ) def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : int , _snake_case : Union[str, Any] ) -> Optional[int]: '''simple docstring''' __magic_name__ : int = dct.pop(_snake_case ) __magic_name__ : List[Any] = val def lowerCAmelCase_ ( ) -> Dict: '''simple docstring''' __magic_name__ : List[str] = "http://images.cocodataset.org/val2017/000000039769.jpg" __magic_name__ : List[str] = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return im @torch.no_grad() def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : Any , _snake_case : int=False ) -> Dict: '''simple docstring''' __magic_name__ : List[str] = BitConfig( global_padding="same" , layer_type="bottleneck" , depths=(3, 4, 9) , out_features=["stage3"] , embedding_dynamic_padding=_snake_case , ) __magic_name__ : List[str] = ViTHybridConfig(backbone_config=_snake_case , image_size=384 , num_labels=1000 ) __magic_name__ : str = False # load original model from timm __magic_name__ : Union[str, Any] = timm.create_model(_snake_case , pretrained=_snake_case ) timm_model.eval() # load state_dict of original model, remove and rename some keys __magic_name__ : List[Any] = timm_model.state_dict() if base_model: remove_classification_head_(_snake_case ) __magic_name__ : Tuple = create_rename_keys(_snake_case , _snake_case ) for src, dest in rename_keys: rename_key(_snake_case , _snake_case , _snake_case ) read_in_q_k_v(_snake_case , _snake_case , _snake_case ) __magic_name__ : List[str] = "huggingface/label-files" __magic_name__ : int = "imagenet-1k-id2label.json" __magic_name__ : Optional[int] = json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type="dataset" ) , "r" ) ) __magic_name__ : int = {int(_snake_case ): v for k, v in idalabel.items()} __magic_name__ : List[str] = idalabel __magic_name__ : List[str] = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": __magic_name__ : List[str] = ViTHybridModel(_snake_case ).eval() else: __magic_name__ : str = ViTHybridForImageClassification(_snake_case ).eval() model.load_state_dict(_snake_case ) # create image processor __magic_name__ : List[Any] = create_transform(**resolve_data_config({} , model=_snake_case ) ) __magic_name__ : int = transform.transforms __magic_name__ : List[str] = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } __magic_name__ : int = ViTHybridImageProcessor( do_resize=_snake_case , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_snake_case , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=_snake_case , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) __magic_name__ : List[Any] = prepare_img() __magic_name__ : Any = transform(_snake_case ).unsqueeze(0 ) __magic_name__ : Tuple = processor(_snake_case , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(_snake_case , _snake_case ) # verify logits with torch.no_grad(): __magic_name__ : Optional[int] = model(_snake_case ) __magic_name__ : List[str] = outputs.logits print("Predicted class:" , logits.argmax(-1 ).item() ) if base_model: __magic_name__ : List[str] = timm_model.forward_features(_snake_case ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_snake_case , outputs.pooler_output , atol=1E-3 ) else: __magic_name__ : Any = timm_model(_snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_snake_case , outputs.logits , atol=1E-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(_snake_case ).mkdir(exist_ok=_snake_case ) print(F'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(_snake_case ) print(F'''Saving processor to {pytorch_dump_folder_path}''' ) processor.save_pretrained(_snake_case ) if push_to_hub: print(F'''Pushing model and processor to the hub {vit_name}''' ) model.push_to_hub(F'''ybelkada/{vit_name}''' ) processor.push_to_hub(F'''ybelkada/{vit_name}''' ) if __name__ == "__main__": snake_case : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--vit_name", default="vit_base_r50_s16_384", type=str, help="Name of the hybrid ViT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to upload the model to the HuggingFace hub." ) snake_case : List[Any] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
281
0
import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class A ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): __snake_case = StableDiffusionDiffEditPipeline __snake_case = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'height', 'width', 'image'} | {'image_latents'} __snake_case = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {'image'} | {'image_latents'} __snake_case = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __snake_case = frozenset([] ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowerCAmelCase_ = UNetaDConditionModel( block_out_channels=(32, 64), layers_per_block=2, sample_size=32, in_channels=4, out_channels=4, down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D'''), up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D'''), cross_attention_dim=32, attention_head_dim=(2, 4), use_linear_projection=_a, ) lowerCAmelCase_ = DDIMScheduler( beta_start=0.00_085, beta_end=0.012, beta_schedule='''scaled_linear''', clip_sample=_a, set_alpha_to_one=_a, ) lowerCAmelCase_ = DDIMInverseScheduler( beta_start=0.00_085, beta_end=0.012, beta_schedule='''scaled_linear''', clip_sample=_a, set_alpha_to_zero=_a, ) torch.manual_seed(0 ) lowerCAmelCase_ = AutoencoderKL( block_out_channels=[32, 64], in_channels=3, out_channels=3, down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''], up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''], latent_channels=4, sample_size=128, ) torch.manual_seed(0 ) lowerCAmelCase_ = CLIPTextConfig( bos_token_id=0, eos_token_id=2, hidden_size=32, intermediate_size=37, layer_norm_eps=1E-05, num_attention_heads=4, num_hidden_layers=5, pad_token_id=1, vocab_size=1000, hidden_act='''gelu''', projection_dim=512, ) lowerCAmelCase_ = CLIPTextModel(_a ) lowerCAmelCase_ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) lowerCAmelCase_ = { "unet": unet, "scheduler": scheduler, "inverse_scheduler": inverse_scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__, UpperCamelCase__=0 ): """simple docstring""" lowerCAmelCase_ = floats_tensor((1, 16, 16), rng=random.Random(_a ) ).to(_a ) lowerCAmelCase_ = floats_tensor((1, 2, 4, 16, 16), rng=random.Random(_a ) ).to(_a ) if str(_a ).startswith('''mps''' ): lowerCAmelCase_ = torch.manual_seed(_a ) else: lowerCAmelCase_ = torch.Generator(device=_a ).manual_seed(_a ) lowerCAmelCase_ = { "prompt": "a dog and a newt", "mask_image": mask, "image_latents": latents, "generator": generator, "num_inference_steps": 2, "inpaint_strength": 1.0, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__, UpperCamelCase__=0 ): """simple docstring""" lowerCAmelCase_ = floats_tensor((1, 3, 32, 32), rng=random.Random(_a ) ).to(_a ) lowerCAmelCase_ = image.cpu().permute(0, 2, 3, 1 )[0] lowerCAmelCase_ = Image.fromarray(np.uinta(_a ) ).convert('''RGB''' ) if str(_a ).startswith('''mps''' ): lowerCAmelCase_ = torch.manual_seed(_a ) else: lowerCAmelCase_ = torch.Generator(device=_a ).manual_seed(_a ) lowerCAmelCase_ = { "image": image, "source_prompt": "a cat and a frog", "target_prompt": "a dog and a newt", "generator": generator, "num_inference_steps": 2, "num_maps_per_mask": 2, "mask_encode_strength": 1.0, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__, UpperCamelCase__=0 ): """simple docstring""" lowerCAmelCase_ = floats_tensor((1, 3, 32, 32), rng=random.Random(_a ) ).to(_a ) lowerCAmelCase_ = image.cpu().permute(0, 2, 3, 1 )[0] lowerCAmelCase_ = Image.fromarray(np.uinta(_a ) ).convert('''RGB''' ) if str(_a ).startswith('''mps''' ): lowerCAmelCase_ = torch.manual_seed(_a ) else: lowerCAmelCase_ = torch.Generator(device=_a ).manual_seed(_a ) lowerCAmelCase_ = { "image": image, "prompt": "a cat and a frog", "generator": generator, "num_inference_steps": 2, "inpaint_strength": 1.0, "guidance_scale": 6.0, "decode_latents": True, "output_type": "numpy", } return inputs def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" if not hasattr(self.pipeline_class, '''_optional_components''' ): return lowerCAmelCase_ = self.get_dummy_components() lowerCAmelCase_ = self.pipeline_class(**_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(_a, _a, _a ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) lowerCAmelCase_ = self.get_dummy_inputs(_a ) lowerCAmelCase_ = pipe(**_a )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_a ) lowerCAmelCase_ = self.pipeline_class.from_pretrained(_a ) pipe_loaded.to(_a ) pipe_loaded.set_progress_bar_config(disable=_a ) for optional_component in pipe._optional_components: self.assertTrue( getattr(_a, _a ) is None, f"`{optional_component}` did not stay set to None after loading.", ) lowerCAmelCase_ = self.get_dummy_inputs(_a ) lowerCAmelCase_ = pipe_loaded(**_a )[0] lowerCAmelCase_ = np.abs(output - output_loaded ).max() self.assertLess(_a, 1E-4 ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = "cpu" lowerCAmelCase_ = self.get_dummy_components() lowerCAmelCase_ = self.pipeline_class(**_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) lowerCAmelCase_ = self.get_dummy_mask_inputs(_a ) lowerCAmelCase_ = pipe.generate_mask(**_a ) lowerCAmelCase_ = mask[0, -3:, -3:] self.assertEqual(mask.shape, (1, 16, 16) ) lowerCAmelCase_ = np.array([0] * 9 ) lowerCAmelCase_ = np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(_a, 1E-3 ) self.assertEqual(mask[0, -3, -4], 0 ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = "cpu" lowerCAmelCase_ = self.get_dummy_components() lowerCAmelCase_ = self.pipeline_class(**_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) lowerCAmelCase_ = self.get_dummy_inversion_inputs(_a ) lowerCAmelCase_ = pipe.invert(**_a ).images lowerCAmelCase_ = image[0, -1, -3:, -3:] self.assertEqual(image.shape, (2, 32, 32, 3) ) lowerCAmelCase_ = np.array( [0.5_150, 0.5_134, 0.5_043, 0.5_376, 0.4_694, 0.51_050, 0.5_015, 0.4_407, 0.4_799], ) lowerCAmelCase_ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_a, 1E-3 ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=5E-3 ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = "cpu" lowerCAmelCase_ = self.get_dummy_components() lowerCAmelCase_ = {"beta_start": 0.00_085, "beta_end": 0.012, "beta_schedule": "scaled_linear"} lowerCAmelCase_ = DPMSolverMultistepScheduler(**_a ) lowerCAmelCase_ = DPMSolverMultistepInverseScheduler(**_a ) lowerCAmelCase_ = self.pipeline_class(**_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) lowerCAmelCase_ = self.get_dummy_inversion_inputs(_a ) lowerCAmelCase_ = pipe.invert(**_a ).images lowerCAmelCase_ = image[0, -1, -3:, -3:] self.assertEqual(image.shape, (2, 32, 32, 3) ) lowerCAmelCase_ = np.array( [0.5_150, 0.5_134, 0.5_043, 0.5_376, 0.4_694, 0.51_050, 0.5_015, 0.4_407, 0.4_799], ) lowerCAmelCase_ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_a, 1E-3 ) @require_torch_gpu @slow class A ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def SCREAMING_SNAKE_CASE__ ( cls ): """simple docstring""" lowerCAmelCase_ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png''' ) lowerCAmelCase_ = raw_image.convert('''RGB''' ).resize((768, 768) ) lowerCAmelCase_ = raw_image def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = torch.manual_seed(0 ) lowerCAmelCase_ = StableDiffusionDiffEditPipeline.from_pretrained( '''stabilityai/stable-diffusion-2-1''', safety_checker=_a, torch_dtype=torch.floataa ) lowerCAmelCase_ = DDIMScheduler.from_config(pipe.scheduler.config ) lowerCAmelCase_ = DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=_a ) lowerCAmelCase_ = "a bowl of fruit" lowerCAmelCase_ = "a bowl of pears" lowerCAmelCase_ = pipe.generate_mask( image=self.raw_image, source_prompt=_a, target_prompt=_a, generator=_a, ) lowerCAmelCase_ = pipe.invert( prompt=_a, image=self.raw_image, inpaint_strength=0.7, generator=_a ).latents lowerCAmelCase_ = pipe( prompt=_a, mask_image=_a, image_latents=_a, generator=_a, negative_prompt=_a, inpaint_strength=0.7, output_type='''numpy''', ).images[0] lowerCAmelCase_ = ( np.array( load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/diffedit/pears.png''' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5E-1 def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = torch.manual_seed(0 ) lowerCAmelCase_ = StableDiffusionDiffEditPipeline.from_pretrained( '''stabilityai/stable-diffusion-2-1''', safety_checker=_a, torch_dtype=torch.floataa ) lowerCAmelCase_ = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) lowerCAmelCase_ = DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=_a ) lowerCAmelCase_ = "a bowl of fruit" lowerCAmelCase_ = "a bowl of pears" lowerCAmelCase_ = pipe.generate_mask( image=self.raw_image, source_prompt=_a, target_prompt=_a, generator=_a, ) lowerCAmelCase_ = pipe.invert( prompt=_a, image=self.raw_image, inpaint_strength=0.7, generator=_a, num_inference_steps=25, ).latents lowerCAmelCase_ = pipe( prompt=_a, mask_image=_a, image_latents=_a, generator=_a, negative_prompt=_a, inpaint_strength=0.7, num_inference_steps=25, output_type='''numpy''', ).images[0] lowerCAmelCase_ = ( np.array( load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/diffedit/pears.png''' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5E-1
278
# This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny model through reduction of a normal pre-trained model, but keeping the # full vocab, merges file, and thus also resulting in a larger model due to a large vocab size. # This gives ~3MB in total for all files. # # If you want a 50 times smaller than this see `fsmt-make-super-tiny-model.py`, which is slightly more complicated # # # It will be used then as "stas/tiny-wmt19-en-de" # Build from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration snake_case : List[str] = "facebook/wmt19-en-de" snake_case : Dict = FSMTTokenizer.from_pretrained(mname) # get the correct vocab sizes, etc. from the master model snake_case : List[str] = FSMTConfig.from_pretrained(mname) config.update( dict( d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) ) snake_case : int = FSMTForConditionalGeneration(config) print(F"num of params {tiny_model.num_parameters()}") # Test snake_case : Optional[Any] = tokenizer(["Making tiny model"], return_tensors="pt") snake_case : List[str] = tiny_model(**batch) print("test output:", len(outputs.logits[0])) # Save snake_case : Dict = "tiny-wmt19-en-de" tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F"Generated {mname_tiny}") # Upload # transformers-cli upload tiny-wmt19-en-de
281
0
"""simple docstring""" import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets __UpperCamelCase = "\\n@inproceedings{lin-2004-rouge,\n title = \"{ROUGE}: A Package for Automatic Evaluation of Summaries\",\n author = \"Lin, Chin-Yew\",\n booktitle = \"Text Summarization Branches Out\",\n month = jul,\n year = \"2004\",\n address = \"Barcelona, Spain\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W04-1013\",\n pages = \"74--81\",\n}\n" __UpperCamelCase = "\\nROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for\nevaluating automatic summarization and machine translation software in natural language processing.\nThe metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation.\n\nNote that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters.\n\nThis metrics is a wrapper around Google Research reimplementation of ROUGE:\nhttps://github.com/google-research/google-research/tree/master/rouge\n" __UpperCamelCase = "\nCalculates average rouge scores for a list of hypotheses and references\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n rouge_types: A list of rouge types to calculate.\n Valid names:\n `\"rouge{n}\"` (e.g. `\"rouge1\"`, `\"rouge2\"`) where: {n} is the n-gram based scoring,\n `\"rougeL\"`: Longest common subsequence based scoring.\n `\"rougeLSum\"`: rougeLsum splits text using `\"\n\"`.\n See details in https://github.com/huggingface/datasets/issues/617\n use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes.\n use_aggregator: Return aggregates if this is set to True\nReturns:\n rouge1: rouge_1 (precision, recall, f1),\n rouge2: rouge_2 (precision, recall, f1),\n rougeL: rouge_l (precision, recall, f1),\n rougeLsum: rouge_lsum (precision, recall, f1)\nExamples:\n\n >>> rouge = datasets.load_metric('rouge')\n >>> predictions = [\"hello there\", \"general kenobi\"]\n >>> references = [\"hello there\", \"general kenobi\"]\n >>> results = rouge.compute(predictions=predictions, references=references)\n >>> print(list(results.keys()))\n ['rouge1', 'rouge2', 'rougeL', 'rougeLsum']\n >>> print(results[\"rouge1\"])\n AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0))\n >>> print(results[\"rouge1\"].mid.fmeasure)\n 1.0\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase ( datasets.Metric ): '''simple docstring''' def __A ( self ) -> Optional[int]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/google-research/google-research/tree/master/rouge'] , reference_urls=[ 'https://en.wikipedia.org/wiki/ROUGE_(metric)', 'https://github.com/google-research/google-research/tree/master/rouge', ] , ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=False ) -> Union[str, Any]: if rouge_types is None: SCREAMING_SNAKE_CASE = ["rouge1", "rouge2", "rougeL", "rougeLsum"] SCREAMING_SNAKE_CASE = rouge_scorer.RougeScorer(rouge_types=_a , use_stemmer=_a ) if use_aggregator: SCREAMING_SNAKE_CASE = scoring.BootstrapAggregator() else: SCREAMING_SNAKE_CASE = [] for ref, pred in zip(_a , _a ): SCREAMING_SNAKE_CASE = scorer.score(_a , _a ) if use_aggregator: aggregator.add_scores(_a ) else: scores.append(_a ) if use_aggregator: SCREAMING_SNAKE_CASE = aggregator.aggregate() else: SCREAMING_SNAKE_CASE = {} for key in scores[0]: SCREAMING_SNAKE_CASE = [score[key] for score in scores] return result
113
import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s", datefmt="%m/%d/%Y %H:%M:%S", level=logging.INFO ) snake_case : Optional[int] = logging.getLogger(__name__) def lowerCAmelCase_ ( _snake_case : Dict , _snake_case : Union[str, Any] ) -> Tuple: '''simple docstring''' __magic_name__ : List[str] = np.argmax(_snake_case , axis=1 ) return np.sum(outputs == labels ) def lowerCAmelCase_ ( _snake_case : Optional[Any] ) -> Dict: '''simple docstring''' with open(_snake_case , encoding="utf_8" ) as f: __magic_name__ : List[str] = csv.reader(_snake_case ) __magic_name__ : List[Any] = [] next(_snake_case ) # skip the first line for line in tqdm(_snake_case ): output.append((" ".join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def lowerCAmelCase_ ( _snake_case : str , _snake_case : Tuple , _snake_case : Union[str, Any] , _snake_case : List[Any] , _snake_case : Tuple , _snake_case : Optional[int] ) -> int: '''simple docstring''' __magic_name__ : Optional[int] = [] for dataset in encoded_datasets: __magic_name__ : Union[str, Any] = len(_snake_case ) __magic_name__ : Dict = np.zeros((n_batch, 2, input_len) , dtype=np.intaa ) __magic_name__ : List[str] = np.zeros((n_batch, 2) , dtype=np.intaa ) __magic_name__ : Optional[int] = np.full((n_batch, 2, input_len) , fill_value=-100 , dtype=np.intaa ) __magic_name__ : int = np.zeros((n_batch,) , dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(_snake_case ): __magic_name__ : Dict = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __magic_name__ : Dict = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __magic_name__ : str = with_conta __magic_name__ : Tuple = with_conta __magic_name__ : Union[str, Any] = len(_snake_case ) - 1 __magic_name__ : int = len(_snake_case ) - 1 __magic_name__ : Optional[Any] = with_conta __magic_name__ : Optional[Any] = with_conta __magic_name__ : Optional[int] = mc_label __magic_name__ : str = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(_snake_case ) for t in all_inputs ) ) return tensor_datasets def lowerCAmelCase_ ( ) -> List[Any]: '''simple docstring''' __magic_name__ : Any = argparse.ArgumentParser() parser.add_argument("--model_name" , type=_snake_case , default="openai-gpt" , help="pretrained model name" ) parser.add_argument("--do_train" , action="store_true" , help="Whether to run training." ) parser.add_argument("--do_eval" , action="store_true" , help="Whether to run eval on the dev set." ) parser.add_argument( "--output_dir" , default=_snake_case , type=_snake_case , required=_snake_case , help="The output directory where the model predictions and checkpoints will be written." , ) parser.add_argument("--train_dataset" , type=_snake_case , default="" ) parser.add_argument("--eval_dataset" , type=_snake_case , default="" ) parser.add_argument("--seed" , type=_snake_case , default=42 ) parser.add_argument("--num_train_epochs" , type=_snake_case , default=3 ) parser.add_argument("--train_batch_size" , type=_snake_case , default=8 ) parser.add_argument("--eval_batch_size" , type=_snake_case , default=16 ) parser.add_argument("--adam_epsilon" , default=1E-8 , type=_snake_case , help="Epsilon for Adam optimizer." ) parser.add_argument("--max_grad_norm" , type=_snake_case , default=1 ) parser.add_argument( "--max_steps" , default=-1 , type=_snake_case , help=( "If > 0: set total number of training steps to perform. Override num_train_epochs." ) , ) parser.add_argument( "--gradient_accumulation_steps" , type=_snake_case , default=1 , help="Number of updates steps to accumulate before performing a backward/update pass." , ) parser.add_argument("--learning_rate" , type=_snake_case , default=6.25E-5 ) parser.add_argument("--warmup_steps" , default=0 , type=_snake_case , help="Linear warmup over warmup_steps." ) parser.add_argument("--lr_schedule" , type=_snake_case , default="warmup_linear" ) parser.add_argument("--weight_decay" , type=_snake_case , default=0.01 ) parser.add_argument("--lm_coef" , type=_snake_case , default=0.9 ) parser.add_argument("--n_valid" , type=_snake_case , default=374 ) parser.add_argument("--server_ip" , type=_snake_case , default="" , help="Can be used for distant debugging." ) parser.add_argument("--server_port" , type=_snake_case , default="" , help="Can be used for distant debugging." ) __magic_name__ : List[Any] = parser.parse_args() print(_snake_case ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print("Waiting for debugger attach" ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=_snake_case ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) __magic_name__ : Dict = torch.device("cuda" if torch.cuda.is_available() else "cpu" ) __magic_name__ : Optional[int] = torch.cuda.device_count() logger.info("device: {}, n_gpu {}".format(_snake_case , _snake_case ) ) if not args.do_train and not args.do_eval: raise ValueError("At least one of `do_train` or `do_eval` must be True." ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset __magic_name__ : List[Any] = ["_start_", "_delimiter_", "_classify_"] __magic_name__ : Optional[int] = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(_snake_case ) __magic_name__ : Optional[Any] = tokenizer.convert_tokens_to_ids(_snake_case ) __magic_name__ : List[str] = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(_snake_case ) ) model.to(_snake_case ) # Load and encode the datasets def tokenize_and_encode(_snake_case : str ): if isinstance(_snake_case , _snake_case ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(_snake_case ) ) elif isinstance(_snake_case , _snake_case ): return obj return [tokenize_and_encode(_snake_case ) for o in obj] logger.info("Encoding dataset..." ) __magic_name__ : Optional[int] = load_rocstories_dataset(args.train_dataset ) __magic_name__ : str = load_rocstories_dataset(args.eval_dataset ) __magic_name__ : int = (train_dataset, eval_dataset) __magic_name__ : List[str] = tokenize_and_encode(_snake_case ) # Compute the max input length for the Transformer __magic_name__ : Optional[Any] = model.config.n_positions // 2 - 2 __magic_name__ : Optional[int] = max( len(story[:max_length] ) + max(len(conta[:max_length] ) , len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) __magic_name__ : List[str] = min(_snake_case , model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders __magic_name__ : List[Any] = pre_process_datasets(_snake_case , _snake_case , _snake_case , *_snake_case ) __magic_name__ , __magic_name__ : Optional[int] = tensor_datasets[0], tensor_datasets[1] __magic_name__ : Tuple = TensorDataset(*_snake_case ) __magic_name__ : Union[str, Any] = RandomSampler(_snake_case ) __magic_name__ : Dict = DataLoader(_snake_case , sampler=_snake_case , batch_size=args.train_batch_size ) __magic_name__ : Any = TensorDataset(*_snake_case ) __magic_name__ : Optional[Any] = SequentialSampler(_snake_case ) __magic_name__ : int = DataLoader(_snake_case , sampler=_snake_case , batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: __magic_name__ : Tuple = args.max_steps __magic_name__ : List[str] = args.max_steps // (len(_snake_case ) // args.gradient_accumulation_steps) + 1 else: __magic_name__ : List[str] = len(_snake_case ) // args.gradient_accumulation_steps * args.num_train_epochs __magic_name__ : str = list(model.named_parameters() ) __magic_name__ : Dict = ["bias", "LayerNorm.bias", "LayerNorm.weight"] __magic_name__ : str = [ { "params": [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], "weight_decay": args.weight_decay, }, {"params": [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], "weight_decay": 0.0}, ] __magic_name__ : str = AdamW(_snake_case , lr=args.learning_rate , eps=args.adam_epsilon ) __magic_name__ : List[str] = get_linear_schedule_with_warmup( _snake_case , num_warmup_steps=args.warmup_steps , num_training_steps=_snake_case ) if args.do_train: __magic_name__ , __magic_name__ , __magic_name__ : Union[str, Any] = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) , desc="Epoch" ): __magic_name__ : List[str] = 0 __magic_name__ : Tuple = 0 __magic_name__ : Dict = tqdm(_snake_case , desc="Training" ) for step, batch in enumerate(_snake_case ): __magic_name__ : Optional[Any] = tuple(t.to(_snake_case ) for t in batch ) __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Dict = batch __magic_name__ : Optional[Any] = model(_snake_case , mc_token_ids=_snake_case , lm_labels=_snake_case , mc_labels=_snake_case ) __magic_name__ : Optional[Any] = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() __magic_name__ : List[str] = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 __magic_name__ : int = "Training loss: {:.2e} lr: {:.2e}".format(_snake_case , scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer __magic_name__ : Dict = model.module if hasattr(_snake_case , "module" ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` __magic_name__ : List[Any] = os.path.join(args.output_dir , _snake_case ) __magic_name__ : Dict = os.path.join(args.output_dir , _snake_case ) torch.save(model_to_save.state_dict() , _snake_case ) model_to_save.config.to_json_file(_snake_case ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned __magic_name__ : Dict = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) __magic_name__ : Optional[int] = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(_snake_case ) if args.do_eval: model.eval() __magic_name__ , __magic_name__ : Any = 0, 0 __magic_name__ , __magic_name__ : Union[str, Any] = 0, 0 for batch in tqdm(_snake_case , desc="Evaluating" ): __magic_name__ : int = tuple(t.to(_snake_case ) for t in batch ) __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Union[str, Any] = batch with torch.no_grad(): __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Dict = model( _snake_case , mc_token_ids=_snake_case , lm_labels=_snake_case , mc_labels=_snake_case ) __magic_name__ : Tuple = mc_logits.detach().cpu().numpy() __magic_name__ : Any = mc_labels.to("cpu" ).numpy() __magic_name__ : str = accuracy(_snake_case , _snake_case ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 __magic_name__ : Tuple = eval_loss / nb_eval_steps __magic_name__ : List[Any] = eval_accuracy / nb_eval_examples __magic_name__ : int = tr_loss / nb_tr_steps if args.do_train else None __magic_name__ : Any = {"eval_loss": eval_loss, "eval_accuracy": eval_accuracy, "train_loss": train_loss} __magic_name__ : int = os.path.join(args.output_dir , "eval_results.txt" ) with open(_snake_case , "w" ) as writer: logger.info("***** Eval results *****" ) for key in sorted(result.keys() ): logger.info(" %s = %s" , _snake_case , str(result[key] ) ) writer.write("%s = %s\n" % (key, str(result[key] )) ) if __name__ == "__main__": main()
281
0
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """Salesforce/blip-image-captioning-base""" lowerCamelCase__ = ( """This is a tool that generates a description of an image. It takes an input named `image` which should be the """ """image to caption, and returns a text that contains the description in English.""" ) lowerCamelCase__ = """image_captioner""" lowerCamelCase__ = AutoModelForVisionaSeq lowerCamelCase__ = ["""image"""] lowerCamelCase__ = ["""text"""] def __init__( self , *lowercase , **lowercase ): requires_backends(self , ['vision'] ) super().__init__(*_a , **_a ) def A_ ( self , lowercase ): return self.pre_processor(images=_a , return_tensors='pt' ) def A_ ( self , lowercase ): return self.model.generate(**_a ) def A_ ( self , lowercase ): return self.pre_processor.batch_decode(_a , skip_special_tokens=_a )[0].strip()
96
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
281
0
"""simple docstring""" import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaProcessor, is_apex_available, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse("1.6"): SCREAMING_SNAKE_CASE__ = True from torch.cuda.amp import autocast SCREAMING_SNAKE_CASE__ = logging.getLogger(__name__) def lowerCAmelCase__ ( _UpperCamelCase : List[str]=None , _UpperCamelCase : Union[str, Any]=None ) -> Dict: """simple docstring""" return field(default_factory=lambda: default , metadata=_snake_case ) @dataclass class lowerCAmelCase_ : """simple docstring""" _lowerCAmelCase : Dict = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) _lowerCAmelCase : Tuple = field( default=lowerCAmelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) _lowerCAmelCase : str = field( default=lowerCAmelCase , metadata={"""help""": """Whether to freeze the feature extractor layers of the model."""} ) _lowerCAmelCase : Optional[Any] = field( default=0.1 , metadata={"""help""": """The dropout ratio for the attention probabilities."""} ) _lowerCAmelCase : str = field( default=0.1 , metadata={"""help""": """The dropout ratio for activations inside the fully connected layer."""} ) _lowerCAmelCase : List[str] = field( default=0.1 , metadata={ """help""": """The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler.""" } , ) _lowerCAmelCase : Optional[Any] = field( default=0.1 , metadata={"""help""": """The dropout probabilitiy for all 1D convolutional layers in feature extractor."""} , ) _lowerCAmelCase : Optional[int] = field( default=0.05 , metadata={ """help""": ( """Propability of each feature vector along the time axis to be chosen as the start of the vector""" """span to be masked. Approximately ``mask_time_prob * sequence_length // mask_time_length`` feature""" """vectors will be masked along the time axis. This is only relevant if ``apply_spec_augment is True``.""" ) } , ) _lowerCAmelCase : str = field(default=0.0 , metadata={"""help""": """The LayerDrop probability."""} ) @dataclass class lowerCAmelCase_ : """simple docstring""" _lowerCAmelCase : Any = field( default=lowerCAmelCase , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) _lowerCAmelCase : List[Any] = field( default="""train+validation""" , metadata={ """help""": """The name of the training data set split to use (via the datasets library). Defaults to \'train\'""" } , ) _lowerCAmelCase : Optional[Any] = field( default=lowerCAmelCase , metadata={"""help""": """Overwrite the cached preprocessed datasets or not."""} ) _lowerCAmelCase : Dict = field( default=lowerCAmelCase , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) _lowerCAmelCase : Dict = field( default=lowerCAmelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) _lowerCAmelCase : Dict = field( default=lowerCAmelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of validation examples to this """ """value if set.""" ) } , ) _lowerCAmelCase : Optional[int] = list_field( default=[""",""", """?""", """.""", """!""", """-""", """;""", """:""", """\"\"""", """%""", """\'""", """\"""", """�"""] , metadata={"""help""": """A list of characters to remove from the transcripts."""} , ) @dataclass class lowerCAmelCase_ : """simple docstring""" _lowerCAmelCase : Union[str, Any] = 42 _lowerCAmelCase : Tuple = True _lowerCAmelCase : Optional[int] = None _lowerCAmelCase : Dict = None _lowerCAmelCase : Optional[int] = None _lowerCAmelCase : int = None def __call__( self , lowerCAmelCase ): """simple docstring""" snake_case = [{"input_values": feature["input_values"]} for feature in features] snake_case = [{"input_ids": feature["labels"]} for feature in features] snake_case = self.processor.pad( _a , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='pt' , ) snake_case = self.processor.pad( labels=_a , padding=self.padding , max_length=self.max_length_labels , pad_to_multiple_of=self.pad_to_multiple_of_labels , return_tensors='pt' , ) # replace padding with -100 to ignore loss correctly snake_case = labels_batch["input_ids"].masked_fill(labels_batch.attention_mask.ne(1 ) , -1_00 ) snake_case = labels return batch class lowerCAmelCase_ ( lowerCAmelCase ): """simple docstring""" def snake_case ( self , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" model.train() snake_case = self._prepare_inputs(_a ) if self.use_amp: with autocast(): snake_case = self.compute_loss(_a , _a ) else: snake_case = self.compute_loss(_a , _a ) if self.args.n_gpu > 1: if model.module.config.ctc_loss_reduction == "mean": snake_case = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": snake_case = loss.sum() / (inputs["labels"] >= 0).sum() else: raise ValueError(F"""{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']""" ) if self.args.gradient_accumulation_steps > 1: snake_case = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(_a ).backward() elif self.use_apex: with amp.scale_loss(_a , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(_a ) else: loss.backward() return loss.detach() def lowerCAmelCase__ ( ) -> Optional[Any]: """simple docstring""" snake_case = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. snake_case = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: snake_case = parser.parse_args_into_dataclasses() # Detecting last checkpoint. snake_case = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: snake_case = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. """ 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None: logger.info( f"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( f"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + f"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info('Training/evaluation parameters %s' , _snake_case ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: snake_case = datasets.load_dataset( 'common_voice' , data_args.dataset_config_name , split=data_args.train_split_name ) snake_case = datasets.load_dataset('common_voice' , data_args.dataset_config_name , split='test' ) # Create and save tokenizer snake_case = f"""[{"".join(data_args.chars_to_ignore )}]""" def remove_special_characters(_UpperCamelCase : Union[str, Any] ): snake_case = re.sub(_snake_case , '' , batch['sentence'] ).lower() + " " return batch snake_case = train_dataset.map(_snake_case , remove_columns=['sentence'] ) snake_case = eval_dataset.map(_snake_case , remove_columns=['sentence'] ) def extract_all_chars(_UpperCamelCase : Union[str, Any] ): snake_case = " ".join(batch['text'] ) snake_case = list(set(_snake_case ) ) return {"vocab": [vocab], "all_text": [all_text]} snake_case = train_dataset.map( _snake_case , batched=_snake_case , batch_size=-1 , keep_in_memory=_snake_case , remove_columns=train_dataset.column_names , ) snake_case = train_dataset.map( _snake_case , batched=_snake_case , batch_size=-1 , keep_in_memory=_snake_case , remove_columns=eval_dataset.column_names , ) snake_case = list(set(vocab_train['vocab'][0] ) | set(vocab_test['vocab'][0] ) ) snake_case = {v: k for k, v in enumerate(_snake_case )} snake_case = vocab_dict[" "] del vocab_dict[" "] snake_case = len(_snake_case ) snake_case = len(_snake_case ) with open('vocab.json' , 'w' ) as vocab_file: json.dump(_snake_case , _snake_case ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case = WavaVecaCTCTokenizer( 'vocab.json' , unk_token='[UNK]' , pad_token='[PAD]' , word_delimiter_token='|' , ) snake_case = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0.0 , do_normalize=_snake_case , return_attention_mask=_snake_case ) snake_case = WavaVecaProcessor(feature_extractor=_snake_case , tokenizer=_snake_case ) snake_case = WavaVecaForCTC.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , activation_dropout=model_args.activation_dropout , attention_dropout=model_args.attention_dropout , hidden_dropout=model_args.hidden_dropout , feat_proj_dropout=model_args.feat_proj_dropout , mask_time_prob=model_args.mask_time_prob , gradient_checkpointing=training_args.gradient_checkpointing , layerdrop=model_args.layerdrop , ctc_loss_reduction='mean' , pad_token_id=processor.tokenizer.pad_token_id , vocab_size=len(processor.tokenizer ) , ) if data_args.max_train_samples is not None: snake_case = min(len(_snake_case ) , data_args.max_train_samples ) snake_case = train_dataset.select(range(_snake_case ) ) if data_args.max_val_samples is not None: snake_case = eval_dataset.select(range(data_args.max_val_samples ) ) snake_case = torchaudio.transforms.Resample(4_8_0_0_0 , 1_6_0_0_0 ) # Preprocessing the datasets. # We need to read the aduio files as arrays and tokenize the targets. def speech_file_to_array_fn(_UpperCamelCase : List[str] ): snake_case = torchaudio.load(batch['path'] ) snake_case = resampler(_snake_case ).squeeze().numpy() snake_case = 1_6_0_0_0 snake_case = batch["text"] return batch snake_case = train_dataset.map( _snake_case , remove_columns=train_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) snake_case = eval_dataset.map( _snake_case , remove_columns=eval_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) def prepare_dataset(_UpperCamelCase : Any ): # check that all files have the correct sampling rate assert ( len(set(batch['sampling_rate'] ) ) == 1 ), f"""Make sure all inputs have the same sampling rate of {processor.feature_extractor.sampling_rate}.""" snake_case = processor( audio=batch['speech'] , text=batch['target_text'] , sampling_rate=batch['sampling_rate'][0] ) batch.update(_snake_case ) return batch snake_case = train_dataset.map( _snake_case , remove_columns=train_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=_snake_case , num_proc=data_args.preprocessing_num_workers , ) snake_case = eval_dataset.map( _snake_case , remove_columns=eval_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=_snake_case , num_proc=data_args.preprocessing_num_workers , ) # Metric snake_case = datasets.load_metric('wer' ) def compute_metrics(_UpperCamelCase : List[Any] ): snake_case = pred.predictions snake_case = np.argmax(_snake_case , axis=-1 ) snake_case = processor.tokenizer.pad_token_id snake_case = processor.batch_decode(_snake_case ) # we do not want to group tokens when computing the metrics snake_case = processor.batch_decode(pred.label_ids , group_tokens=_snake_case ) snake_case = wer_metric.compute(predictions=_snake_case , references=_snake_case ) return {"wer": wer} if model_args.freeze_feature_extractor: model.freeze_feature_extractor() # Data collator snake_case = DataCollatorCTCWithPadding(processor=_snake_case , padding=_snake_case ) # Initialize our Trainer snake_case = CTCTrainer( model=_snake_case , data_collator=_snake_case , args=_snake_case , compute_metrics=_snake_case , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=processor.feature_extractor , ) # Training if training_args.do_train: if last_checkpoint is not None: snake_case = last_checkpoint elif os.path.isdir(model_args.model_name_or_path ): snake_case = model_args.model_name_or_path else: snake_case = None # Save the feature_extractor and the tokenizer if is_main_process(training_args.local_rank ): processor.save_pretrained(training_args.output_dir ) snake_case = trainer.train(resume_from_checkpoint=_snake_case ) trainer.save_model() snake_case = train_result.metrics snake_case = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(_snake_case ) ) snake_case = min(_snake_case , len(_snake_case ) ) trainer.log_metrics('train' , _snake_case ) trainer.save_metrics('train' , _snake_case ) trainer.save_state() # Evaluation snake_case = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) snake_case = trainer.evaluate() snake_case = data_args.max_val_samples if data_args.max_val_samples is not None else len(_snake_case ) snake_case = min(_snake_case , len(_snake_case ) ) trainer.log_metrics('eval' , _snake_case ) trainer.save_metrics('eval' , _snake_case ) return results if __name__ == "__main__": main()
150
import importlib import os import fsspec import pytest from fsspec import register_implementation from fsspec.registry import _registry as _fsspec_registry from datasets.filesystems import COMPRESSION_FILESYSTEMS, HfFileSystem, extract_path_from_uri, is_remote_filesystem from .utils import require_lza, require_zstandard def lowerCAmelCase_ ( _snake_case : List[Any] ) -> List[Any]: '''simple docstring''' assert "mock" in _fsspec_registry assert "bz2" in _fsspec_registry def lowerCAmelCase_ ( ) -> Tuple: '''simple docstring''' assert "mock" not in _fsspec_registry assert "bz2" in _fsspec_registry def lowerCAmelCase_ ( ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : Dict = "mock-s3-bucket" __magic_name__ : Any = F'''s3://{mock_bucket}''' __magic_name__ : str = extract_path_from_uri(_snake_case ) assert dataset_path.startswith("s3://" ) is False __magic_name__ : Tuple = "./local/path" __magic_name__ : Optional[Any] = extract_path_from_uri(_snake_case ) assert dataset_path == new_dataset_path def lowerCAmelCase_ ( _snake_case : List[str] ) -> Optional[Any]: '''simple docstring''' __magic_name__ : str = is_remote_filesystem(_snake_case ) assert is_remote is True __magic_name__ : Optional[int] = fsspec.filesystem("file" ) __magic_name__ : int = is_remote_filesystem(_snake_case ) assert is_remote is False @pytest.mark.parametrize("compression_fs_class" , _snake_case ) def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : Optional[Any] , _snake_case : int , _snake_case : Tuple , _snake_case : Any , _snake_case : Union[str, Any] , _snake_case : Any ) -> int: '''simple docstring''' __magic_name__ : Any = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_file, "bz2": bza_file, "lz4": lza_file} __magic_name__ : str = input_paths[compression_fs_class.protocol] if input_path is None: __magic_name__ : Dict = F'''for \'{compression_fs_class.protocol}\' compression protocol, ''' if compression_fs_class.protocol == "lz4": reason += require_lza.kwargs["reason"] elif compression_fs_class.protocol == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(_snake_case ) __magic_name__ : str = fsspec.filesystem(compression_fs_class.protocol , fo=_snake_case ) assert isinstance(_snake_case , _snake_case ) __magic_name__ : int = os.path.basename(_snake_case ) __magic_name__ : Optional[int] = expected_filename[: expected_filename.rindex("." )] assert fs.glob("*" ) == [expected_filename] with fs.open(_snake_case , "r" , encoding="utf-8" ) as f, open(_snake_case , encoding="utf-8" ) as expected_file: assert f.read() == expected_file.read() @pytest.mark.parametrize("protocol" , ["zip", "gzip"] ) def lowerCAmelCase_ ( _snake_case : List[Any] , _snake_case : Optional[Any] , _snake_case : Optional[Any] ) -> str: '''simple docstring''' __magic_name__ : int = {"zip": zip_jsonl_path, "gzip": jsonl_gz_path} __magic_name__ : int = compressed_file_paths[protocol] __magic_name__ : Tuple = "dataset.jsonl" __magic_name__ : List[str] = F'''{protocol}://{member_file_path}::{compressed_file_path}''' __magic_name__ , *__magic_name__ : Optional[Any] = fsspec.get_fs_token_paths(_snake_case ) assert fs.isfile(_snake_case ) assert not fs.isfile("non_existing_" + member_file_path ) @pytest.mark.integration def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Dict , _snake_case : List[str] , _snake_case : Tuple ) -> str: '''simple docstring''' __magic_name__ : int = hf_api.dataset_info(_snake_case , token=_snake_case ) __magic_name__ : Optional[Any] = HfFileSystem(repo_info=_snake_case , token=_snake_case ) assert sorted(hffs.glob("*" ) ) == [".gitattributes", "data"] assert hffs.isdir("data" ) assert hffs.isfile(".gitattributes" ) and hffs.isfile("data/text_data.txt" ) with open(_snake_case ) as f: assert hffs.open("data/text_data.txt" , "r" ).read() == f.read() def lowerCAmelCase_ ( ) -> Optional[int]: '''simple docstring''' __magic_name__ : Optional[Any] = "bz2" # Import module import datasets.filesystems # Overwrite protocol and reload register_implementation(_snake_case , _snake_case , clobber=_snake_case ) with pytest.warns(_snake_case ) as warning_info: importlib.reload(datasets.filesystems ) assert len(_snake_case ) == 1 assert ( str(warning_info[0].message ) == F'''A filesystem protocol was already set for {protocol} and will be overwritten.''' )
281
0
"""simple docstring""" from __future__ import annotations def lowerCamelCase ( _UpperCamelCase : float , _UpperCamelCase : float , _UpperCamelCase : float , ) -> tuple: '''simple docstring''' if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError("""You cannot supply more or less than 2 values""" ) elif electron_conc < 0: raise ValueError("""Electron concentration cannot be negative in a semiconductor""" ) elif hole_conc < 0: raise ValueError("""Hole concentration cannot be negative in a semiconductor""" ) elif intrinsic_conc < 0: raise ValueError( """Intrinsic concentration cannot be negative in a semiconductor""" ) elif electron_conc == 0: return ( "electron_conc", intrinsic_conc**2 / hole_conc, ) elif hole_conc == 0: return ( "hole_conc", intrinsic_conc**2 / electron_conc, ) elif intrinsic_conc == 0: return ( "intrinsic_conc", (electron_conc * hole_conc) ** 0.5, ) else: return (-1, -1) if __name__ == "__main__": import doctest doctest.testmod()
115
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case : Dict = logging.get_logger(__name__) snake_case : List[Any] = { "YituTech/conv-bert-base": "https://huggingface.co/YituTech/conv-bert-base/resolve/main/config.json", "YituTech/conv-bert-medium-small": ( "https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/config.json" ), "YituTech/conv-bert-small": "https://huggingface.co/YituTech/conv-bert-small/resolve/main/config.json", # See all ConvBERT models at https://huggingface.co/models?filter=convbert } class _snake_case ( snake_case ): UpperCamelCase__ = 'convbert' def __init__( self , _a=30_522 , _a=768 , _a=12 , _a=12 , _a=3_072 , _a="gelu" , _a=0.1 , _a=0.1 , _a=512 , _a=2 , _a=0.02 , _a=1e-12 , _a=1 , _a=0 , _a=2 , _a=768 , _a=2 , _a=9 , _a=1 , _a=None , **_a , ): super().__init__( pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , **_a , ) __magic_name__ : Tuple = vocab_size __magic_name__ : List[Any] = hidden_size __magic_name__ : Union[str, Any] = num_hidden_layers __magic_name__ : List[Any] = num_attention_heads __magic_name__ : str = intermediate_size __magic_name__ : Any = hidden_act __magic_name__ : List[Any] = hidden_dropout_prob __magic_name__ : Optional[int] = attention_probs_dropout_prob __magic_name__ : Tuple = max_position_embeddings __magic_name__ : str = type_vocab_size __magic_name__ : List[str] = initializer_range __magic_name__ : Tuple = layer_norm_eps __magic_name__ : List[Any] = embedding_size __magic_name__ : List[Any] = head_ratio __magic_name__ : str = conv_kernel_size __magic_name__ : Dict = num_groups __magic_name__ : str = classifier_dropout class _snake_case ( snake_case ): @property def SCREAMING_SNAKE_CASE ( self ): if self.task == "multiple-choice": __magic_name__ : Dict = {0: "batch", 1: "choice", 2: "sequence"} else: __magic_name__ : Dict = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis), ] )
281
0
from __future__ import annotations from math import gcd def _a ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int = 2 , SCREAMING_SNAKE_CASE : int = 1 , SCREAMING_SNAKE_CASE : int = 3 , ) -> int | None: """simple docstring""" if num < 2: raise ValueError('The input value cannot be less than 2' ) # Because of the relationship between ``f(f(x))`` and ``f(x)``, this # algorithm struggles to find factors that are divisible by two. # As a workaround, we specifically check for two and even inputs. # See: https://math.stackexchange.com/a/2856214/165820 if num > 2 and num % 2 == 0: return 2 # Pollard's Rho algorithm requires a function that returns pseudorandom # values between 0 <= X < ``num``. It doesn't need to be random in the # sense that the output value is cryptographically secure or difficult # to calculate, it only needs to be random in the sense that all output # values should be equally likely to appear. # For this reason, Pollard suggested using ``f(x) = (x**2 - 1) % num`` # However, the success of Pollard's algorithm isn't guaranteed and is # determined in part by the initial seed and the chosen random function. # To make retries easier, we will instead use ``f(x) = (x**2 + C) % num`` # where ``C`` is a value that we can modify between each attempt. def rand_fn(SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int ) -> int: return (pow(_snake_case , 2 ) + step) % modulus for _ in range(_snake_case ): # These track the position within the cycle detection logic. __lowerCAmelCase: str = seed __lowerCAmelCase: int = seed while True: # At each iteration, the tortoise moves one step and the hare moves two. __lowerCAmelCase: List[str] = rand_fn(_snake_case , _snake_case , _snake_case ) __lowerCAmelCase: Dict = rand_fn(_snake_case , _snake_case , _snake_case ) __lowerCAmelCase: int = rand_fn(_snake_case , _snake_case , _snake_case ) # At some point both the tortoise and the hare will enter a cycle whose # length ``p`` is a divisor of ``num``. Once in that cycle, at some point # the tortoise and hare will end up on the same value modulo ``p``. # We can detect when this happens because the position difference between # the tortoise and the hare will share a common divisor with ``num``. __lowerCAmelCase: List[str] = gcd(hare - tortoise , _snake_case ) if divisor == 1: # No common divisor yet, just keep searching. continue else: # We found a common divisor! if divisor == num: # Unfortunately, the divisor is ``num`` itself and is useless. break else: # The divisor is a nontrivial factor of ``num``! return divisor # If we made it here, then this attempt failed. # We need to pick a new starting seed for the tortoise and hare # in addition to a new step value for the random function. # To keep this example implementation deterministic, the # new values will be generated based on currently available # values instead of using something like ``random.randint``. # We can use the hare's position as the new seed. # This is actually what Richard Brent's the "optimized" variant does. __lowerCAmelCase: int = hare # The new step value for the random function can just be incremented. # At first the results will be similar to what the old function would # have produced, but the value will quickly diverge after a bit. step += 1 # We haven't found a divisor within the requested number of attempts. # We were unlucky or ``num`` itself is actually prime. return None if __name__ == "__main__": import argparse _a = argparse.ArgumentParser() parser.add_argument( '''num''', type=int, help='''The value to find a divisor of''', ) parser.add_argument( '''--attempts''', type=int, default=3, help='''The number of attempts before giving up''', ) _a = parser.parse_args() _a = pollard_rho(args.num, attempts=args.attempts) if divisor is None: print(f"{args.num} is probably prime") else: _a = args.num // divisor print(f"{args.num} = {divisor} * {quotient}")
322
import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipaConfig, BlipaForConditionalGeneration, BlipaProcessor, BlipaVisionConfig, BlipImageProcessor, OPTConfig, TaConfig, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def lowerCAmelCase_ ( ) -> str: '''simple docstring''' __magic_name__ : int = "https://storage.googleapis.com/sfr-vision-language-research/LAVIS/assets/merlion.png" __magic_name__ : Union[str, Any] = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ).convert("RGB" ) return image def lowerCAmelCase_ ( _snake_case : str ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : List[str] = [] # fmt: off # vision encoder rename_keys.append(("visual_encoder.cls_token", "vision_model.embeddings.class_embedding") ) rename_keys.append(("visual_encoder.pos_embed", "vision_model.embeddings.position_embedding") ) rename_keys.append(("visual_encoder.patch_embed.proj.weight", "vision_model.embeddings.patch_embedding.weight") ) rename_keys.append(("visual_encoder.patch_embed.proj.bias", "vision_model.embeddings.patch_embedding.bias") ) rename_keys.append(("ln_vision.weight", "vision_model.post_layernorm.weight") ) rename_keys.append(("ln_vision.bias", "vision_model.post_layernorm.bias") ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((F'''visual_encoder.blocks.{i}.norm1.weight''', F'''vision_model.encoder.layers.{i}.layer_norm1.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm1.bias''', F'''vision_model.encoder.layers.{i}.layer_norm1.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm2.weight''', F'''vision_model.encoder.layers.{i}.layer_norm2.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm2.bias''', F'''vision_model.encoder.layers.{i}.layer_norm2.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.qkv.weight''', F'''vision_model.encoder.layers.{i}.self_attn.qkv.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.proj.weight''', F'''vision_model.encoder.layers.{i}.self_attn.projection.weight''',) ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.proj.bias''', F'''vision_model.encoder.layers.{i}.self_attn.projection.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc1.weight''', F'''vision_model.encoder.layers.{i}.mlp.fc1.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc1.bias''', F'''vision_model.encoder.layers.{i}.mlp.fc1.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc2.weight''', F'''vision_model.encoder.layers.{i}.mlp.fc2.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc2.bias''', F'''vision_model.encoder.layers.{i}.mlp.fc2.bias''') ) # QFormer rename_keys.append(("Qformer.bert.embeddings.LayerNorm.weight", "qformer.layernorm.weight") ) rename_keys.append(("Qformer.bert.embeddings.LayerNorm.bias", "qformer.layernorm.bias") ) # fmt: on return rename_keys def lowerCAmelCase_ ( _snake_case : Any , _snake_case : Union[str, Any] , _snake_case : Optional[Any] ) -> int: '''simple docstring''' __magic_name__ : Tuple = dct.pop(_snake_case ) __magic_name__ : int = val def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : Optional[Any] ) -> Dict: '''simple docstring''' for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases __magic_name__ : List[Any] = state_dict.pop(F'''visual_encoder.blocks.{i}.attn.q_bias''' ) __magic_name__ : Optional[Any] = state_dict.pop(F'''visual_encoder.blocks.{i}.attn.v_bias''' ) # next, set bias in the state dict __magic_name__ : Optional[int] = torch.cat((q_bias, torch.zeros_like(_snake_case , requires_grad=_snake_case ), v_bias) ) __magic_name__ : Union[str, Any] = qkv_bias def lowerCAmelCase_ ( _snake_case : Dict , _snake_case : str ) -> int: '''simple docstring''' __magic_name__ : List[Any] = 364 if "coco" in model_name else 224 __magic_name__ : Union[str, Any] = BlipaVisionConfig(image_size=_snake_case ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "opt-2.7b" in model_name: __magic_name__ : List[str] = OPTConfig.from_pretrained("facebook/opt-2.7b" , eos_token_id=_snake_case ).to_dict() elif "opt-6.7b" in model_name: __magic_name__ : Any = OPTConfig.from_pretrained("facebook/opt-6.7b" , eos_token_id=_snake_case ).to_dict() elif "t5-xl" in model_name: __magic_name__ : Dict = TaConfig.from_pretrained("google/flan-t5-xl" , dense_act_fn="gelu" , bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: __magic_name__ : int = TaConfig.from_pretrained("google/flan-t5-xxl" , dense_act_fn="gelu" , bos_token_id=1 ).to_dict() __magic_name__ : List[Any] = BlipaConfig(vision_config=_snake_case , text_config=_snake_case ) return config, image_size @torch.no_grad() def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : str=None , _snake_case : Dict=False ) -> List[Any]: '''simple docstring''' __magic_name__ : Optional[int] = ( AutoTokenizer.from_pretrained("facebook/opt-2.7b" ) if "opt" in model_name else AutoTokenizer.from_pretrained("google/flan-t5-xl" ) ) __magic_name__ : List[Any] = tokenizer("\n" , add_special_tokens=_snake_case ).input_ids[0] __magic_name__ , __magic_name__ : Tuple = get_blipa_config(_snake_case , eos_token_id=_snake_case ) __magic_name__ : Union[str, Any] = BlipaForConditionalGeneration(_snake_case ).eval() __magic_name__ : Any = { "blip2-opt-2.7b": ("blip2_opt", "pretrain_opt2.7b"), "blip2-opt-6.7b": ("blip2_opt", "pretrain_opt6.7b"), "blip2-opt-2.7b-coco": ("blip2_opt", "caption_coco_opt2.7b"), "blip2-opt-6.7b-coco": ("blip2_opt", "caption_coco_opt6.7b"), "blip2-flan-t5-xl": ("blip2_t5", "pretrain_flant5xl"), "blip2-flan-t5-xl-coco": ("blip2_t5", "caption_coco_flant5xl"), "blip2-flan-t5-xxl": ("blip2_t5", "pretrain_flant5xxl"), } __magic_name__ , __magic_name__ : Union[str, Any] = model_name_to_original[model_name] # load original model print("Loading original model..." ) __magic_name__ : Union[str, Any] = "cuda" if torch.cuda.is_available() else "cpu" __magic_name__ , __magic_name__ , __magic_name__ : Optional[Any] = load_model_and_preprocess( name=_snake_case , model_type=_snake_case , is_eval=_snake_case , device=_snake_case ) original_model.eval() print("Done!" ) # update state dict keys __magic_name__ : Dict = original_model.state_dict() __magic_name__ : str = create_rename_keys(_snake_case ) for src, dest in rename_keys: rename_key(_snake_case , _snake_case , _snake_case ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): __magic_name__ : Any = state_dict.pop(_snake_case ) if key.startswith("Qformer.bert" ): __magic_name__ : Optional[int] = key.replace("Qformer.bert" , "qformer" ) if "attention.self" in key: __magic_name__ : Any = key.replace("self" , "attention" ) if "opt_proj" in key: __magic_name__ : Union[str, Any] = key.replace("opt_proj" , "language_projection" ) if "t5_proj" in key: __magic_name__ : Optional[int] = key.replace("t5_proj" , "language_projection" ) if key.startswith("opt" ): __magic_name__ : List[str] = key.replace("opt" , "language" ) if key.startswith("t5" ): __magic_name__ : Tuple = key.replace("t5" , "language" ) __magic_name__ : Dict = val # read in qv biases read_in_q_v_bias(_snake_case , _snake_case ) __magic_name__ , __magic_name__ : Tuple = hf_model.load_state_dict(_snake_case , strict=_snake_case ) assert len(_snake_case ) == 0 assert unexpected_keys == ["qformer.embeddings.position_ids"] __magic_name__ : List[Any] = load_demo_image() __magic_name__ : Tuple = vis_processors["eval"](_snake_case ).unsqueeze(0 ).to(_snake_case ) __magic_name__ : Dict = tokenizer(["\n"] , return_tensors="pt" ).input_ids.to(_snake_case ) # create processor __magic_name__ : Optional[Any] = BlipImageProcessor( size={"height": image_size, "width": image_size} , image_mean=_snake_case , image_std=_snake_case ) __magic_name__ : Dict = BlipaProcessor(image_processor=_snake_case , tokenizer=_snake_case ) __magic_name__ : Union[str, Any] = processor(images=_snake_case , return_tensors="pt" ).pixel_values.to(_snake_case ) # make sure processor creates exact same pixel values assert torch.allclose(_snake_case , _snake_case ) original_model.to(_snake_case ) hf_model.to(_snake_case ) with torch.no_grad(): if "opt" in model_name: __magic_name__ : List[Any] = original_model({"image": original_pixel_values, "text_input": [""]} ).logits __magic_name__ : Optional[int] = hf_model(_snake_case , _snake_case ).logits else: __magic_name__ : int = original_model( {"image": original_pixel_values, "text_input": ["\n"], "text_output": ["\n"]} ).logits __magic_name__ : Tuple = input_ids.masked_fill(input_ids == tokenizer.pad_token_id , -100 ) __magic_name__ : List[str] = hf_model(_snake_case , _snake_case , labels=_snake_case ).logits assert original_logits.shape == logits.shape print("First values of original logits:" , original_logits[0, :3, :3] ) print("First values of HF logits:" , logits[0, :3, :3] ) # assert values if model_name == "blip2-flan-t5-xl": __magic_name__ : List[str] = torch.tensor( [[-41.5_850, -4.4_440, -8.9_922], [-47.4_322, -5.9_143, -1.7_340]] , device=_snake_case ) assert torch.allclose(logits[0, :3, :3] , _snake_case , atol=1E-4 ) elif model_name == "blip2-flan-t5-xl-coco": __magic_name__ : Tuple = torch.tensor( [[-57.0_109, -9.8_967, -12.6_280], [-68.6_578, -12.7_191, -10.5_065]] , device=_snake_case ) else: # cast to same type __magic_name__ : str = logits.dtype assert torch.allclose(original_logits.to(_snake_case ) , _snake_case , atol=1E-2 ) print("Looks ok!" ) print("Generating a caption..." ) __magic_name__ : Optional[int] = "" __magic_name__ : Dict = tokenizer(_snake_case , return_tensors="pt" ).input_ids.to(_snake_case ) __magic_name__ : int = original_model.generate({"image": original_pixel_values} ) __magic_name__ : Optional[Any] = hf_model.generate( _snake_case , _snake_case , do_sample=_snake_case , num_beams=5 , max_length=30 , min_length=1 , top_p=0.9 , repetition_penalty=1.0 , length_penalty=1.0 , temperature=1 , ) print("Original generation:" , _snake_case ) __magic_name__ : Tuple = input_ids.shape[1] __magic_name__ : int = processor.batch_decode(outputs[:, prompt_length:] , skip_special_tokens=_snake_case ) __magic_name__ : Union[str, Any] = [text.strip() for text in output_text] print("HF generation:" , _snake_case ) if pytorch_dump_folder_path is not None: processor.save_pretrained(_snake_case ) hf_model.save_pretrained(_snake_case ) if push_to_hub: processor.push_to_hub(F'''nielsr/{model_name}''' ) hf_model.push_to_hub(F'''nielsr/{model_name}''' ) if __name__ == "__main__": snake_case : Any = argparse.ArgumentParser() snake_case : Union[str, Any] = [ "blip2-opt-2.7b", "blip2-opt-6.7b", "blip2-opt-2.7b-coco", "blip2-opt-6.7b-coco", "blip2-flan-t5-xl", "blip2-flan-t5-xl-coco", "blip2-flan-t5-xxl", ] parser.add_argument( "--model_name", default="blip2-opt-2.7b", choices=choices, type=str, help="Path to hf config.json of model to convert", ) parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model and processor to the hub after converting", ) snake_case : int = parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
281
0
import argparse from copy import deepcopy import numpy as np from datasets import ClassLabel, DatasetDict, load_dataset from evaluate import load from transformers import ( AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, Trainer, TrainerCallback, TrainingArguments, set_seed, ) def _UpperCamelCase ( ) -> str: __UpperCAmelCase : int = argparse.ArgumentParser() parser.add_argument("--model_ckpt", type=_snake_case, default="microsoft/unixcoder-base-nine" ) parser.add_argument("--num_epochs", type=_snake_case, default=5 ) parser.add_argument("--batch_size", type=_snake_case, default=6 ) parser.add_argument("--gradient_accumulation_steps", type=_snake_case, default=1 ) parser.add_argument("--freeze", type=_snake_case, default=_snake_case ) parser.add_argument("--learning_rate", type=_snake_case, default=5e-4 ) parser.add_argument("--seed", type=_snake_case, default=0 ) parser.add_argument("--lr_scheduler_type", type=_snake_case, default="cosine" ) parser.add_argument("--num_warmup_steps", type=_snake_case, default=10 ) parser.add_argument("--weight_decay", type=_snake_case, default=0.01 ) parser.add_argument("--output_dir", type=_snake_case, default="./results" ) return parser.parse_args() _snake_case = load('''accuracy''') def _UpperCamelCase ( snake_case__ ) -> int: __UpperCAmelCase : Any = eval_pred __UpperCAmelCase : List[str] = np.argmax(_snake_case, axis=1 ) return metric.compute(predictions=_snake_case, references=_snake_case ) class _snake_case ( _lowercase ): def __init__( self: List[Any] , __lowerCamelCase: Any ) -> Tuple: super().__init__() __UpperCAmelCase : Tuple = trainer def _lowerCamelCase ( self: Tuple , __lowerCamelCase: Dict , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: List[Any] , **__lowerCamelCase: str ) -> Any: if control.should_evaluate: __UpperCAmelCase : Any = deepcopy(_a ) self._trainer.evaluate(eval_dataset=self._trainer.train_dataset , metric_key_prefix="train" ) return control_copy def _UpperCamelCase ( ) -> Any: __UpperCAmelCase : Optional[Any] = get_args() set_seed(args.seed ) __UpperCAmelCase : Union[str, Any] = load_dataset("codeparrot/codecomplex", split="train" ) __UpperCAmelCase : Union[str, Any] = dataset.train_test_split(test_size=0.2 ) __UpperCAmelCase : Tuple = train_test["test"].train_test_split(test_size=0.5 ) __UpperCAmelCase : Optional[Any] = DatasetDict( { "train": train_test["train"], "test": test_validation["train"], "valid": test_validation["test"], } ) print("Loading tokenizer and model" ) __UpperCAmelCase : Dict = AutoTokenizer.from_pretrained(args.model_ckpt ) __UpperCAmelCase : Union[str, Any] = tokenizer.eos_token __UpperCAmelCase : str = AutoModelForSequenceClassification.from_pretrained(args.model_ckpt, num_labels=7 ) __UpperCAmelCase : Tuple = model.config.eos_token_id if args.freeze: for param in model.roberta.parameters(): __UpperCAmelCase : Dict = False __UpperCAmelCase : str = ClassLabel(num_classes=7, names=list(set(train_test_validation["train"]["complexity"] ) ) ) def tokenize(snake_case__ ): __UpperCAmelCase : Dict = tokenizer(example["src"], truncation=_snake_case, max_length=1024 ) __UpperCAmelCase : Union[str, Any] = labels.straint(example["complexity"] ) return { "input_ids": inputs["input_ids"], "attention_mask": inputs["attention_mask"], "label": label, } __UpperCAmelCase : Optional[int] = train_test_validation.map( _snake_case, batched=_snake_case, remove_columns=train_test_validation["train"].column_names, ) __UpperCAmelCase : str = DataCollatorWithPadding(tokenizer=_snake_case ) __UpperCAmelCase : Any = TrainingArguments( output_dir=args.output_dir, learning_rate=args.learning_rate, lr_scheduler_type=args.lr_scheduler_type, evaluation_strategy="epoch", save_strategy="epoch", logging_strategy="epoch", per_device_train_batch_size=args.batch_size, per_device_eval_batch_size=args.batch_size, num_train_epochs=args.num_epochs, gradient_accumulation_steps=args.gradient_accumulation_steps, weight_decay=0.01, metric_for_best_model="accuracy", run_name="complexity-java", report_to="wandb", ) __UpperCAmelCase : Dict = Trainer( model=_snake_case, args=_snake_case, train_dataset=tokenized_datasets["train"], eval_dataset=tokenized_datasets["valid"], tokenizer=_snake_case, data_collator=_snake_case, compute_metrics=_snake_case, ) print("Training..." ) trainer.add_callback(CustomCallback(_snake_case ) ) trainer.train() if __name__ == "__main__": main()
157
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging snake_case : Dict = logging.get_logger(__name__) snake_case : Union[str, Any] = { "vocab_file": "vocab.txt", "merges_file": "bpe.codes", } snake_case : Dict = { "vocab_file": { "vinai/phobert-base": "https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt", "vinai/phobert-large": "https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt", }, "merges_file": { "vinai/phobert-base": "https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes", "vinai/phobert-large": "https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes", }, } snake_case : Union[str, Any] = { "vinai/phobert-base": 256, "vinai/phobert-large": 256, } def lowerCAmelCase_ ( _snake_case : str ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : List[str] = set() __magic_name__ : Any = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __magic_name__ : int = char __magic_name__ : List[str] = set(_snake_case ) return pairs class _snake_case ( snake_case ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , _a , _a , _a="<s>" , _a="</s>" , _a="</s>" , _a="<s>" , _a="<unk>" , _a="<pad>" , _a="<mask>" , **_a , ): super().__init__( bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , cls_token=_a , pad_token=_a , mask_token=_a , **_a , ) __magic_name__ : Dict = vocab_file __magic_name__ : Tuple = merges_file __magic_name__ : List[Any] = {} __magic_name__ : List[Any] = 0 __magic_name__ : Tuple = 1 __magic_name__ : int = 2 __magic_name__ : Union[str, Any] = 3 self.add_from_file(_a ) __magic_name__ : Optional[int] = {v: k for k, v in self.encoder.items()} with open(_a , encoding="utf-8" ) as merges_handle: __magic_name__ : List[str] = merges_handle.read().split("\n" )[:-1] __magic_name__ : Union[str, Any] = [tuple(merge.split()[:-1] ) for merge in merges] __magic_name__ : Union[str, Any] = dict(zip(_a , range(len(_a ) ) ) ) __magic_name__ : Optional[int] = {} def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __magic_name__ : Optional[Any] = [self.cls_token_id] __magic_name__ : Union[str, Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self , _a , _a = None , _a = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a , token_ids_a=_a , already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): __magic_name__ : Optional[Any] = [self.sep_token_id] __magic_name__ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def SCREAMING_SNAKE_CASE ( self ): return len(self.encoder ) def SCREAMING_SNAKE_CASE ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def SCREAMING_SNAKE_CASE ( self , _a ): if token in self.cache: return self.cache[token] __magic_name__ : List[Any] = tuple(_a ) __magic_name__ : List[Any] = tuple(list(word[:-1] ) + [word[-1] + "</w>"] ) __magic_name__ : Any = get_pairs(_a ) if not pairs: return token while True: __magic_name__ : str = min(_a , key=lambda _a : self.bpe_ranks.get(_a , float("inf" ) ) ) if bigram not in self.bpe_ranks: break __magic_name__ , __magic_name__ : List[str] = bigram __magic_name__ : List[str] = [] __magic_name__ : List[str] = 0 while i < len(_a ): try: __magic_name__ : Any = word.index(_a , _a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) __magic_name__ : Tuple = j if word[i] == first and i < len(_a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __magic_name__ : Union[str, Any] = tuple(_a ) __magic_name__ : Optional[int] = new_word if len(_a ) == 1: break else: __magic_name__ : List[Any] = get_pairs(_a ) __magic_name__ : Optional[int] = "@@ ".join(_a ) __magic_name__ : Tuple = word[:-4] __magic_name__ : str = word return word def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Optional[Any] = [] __magic_name__ : Dict = re.findall(r"\S+\n?" , _a ) for token in words: split_tokens.extend(list(self.bpe(_a ).split(" " ) ) ) return split_tokens def SCREAMING_SNAKE_CASE ( self , _a ): return self.encoder.get(_a , self.encoder.get(self.unk_token ) ) def SCREAMING_SNAKE_CASE ( self , _a ): return self.decoder.get(_a , self.unk_token ) def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Tuple = " ".join(_a ).replace("@@ " , "" ).strip() return out_string def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): if not os.path.isdir(_a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __magic_name__ : Optional[int] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) __magic_name__ : Union[str, Any] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ): copyfile(self.vocab_file , _a ) if os.path.abspath(self.merges_file ) != os.path.abspath(_a ): copyfile(self.merges_file , _a ) return out_vocab_file, out_merge_file def SCREAMING_SNAKE_CASE ( self , _a ): if isinstance(_a , _a ): try: with open(_a , "r" , encoding="utf-8" ) as fd: self.add_from_file(_a ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(f'''Incorrect encoding detected in {f}, please rebuild the dataset''' ) return __magic_name__ : List[Any] = f.readlines() for lineTmp in lines: __magic_name__ : Optional[Any] = lineTmp.strip() __magic_name__ : Union[str, Any] = line.rfind(" " ) if idx == -1: raise ValueError("Incorrect dictionary format, expected '<token> <cnt>'" ) __magic_name__ : Optional[int] = line[:idx] __magic_name__ : Dict = len(self.encoder )
281
0
"""simple docstring""" import json import os import sys import tempfile import unittest from pathlib import Path from shutil import copyfile from huggingface_hub import HfFolder, Repository, create_repo, delete_repo from requests.exceptions import HTTPError import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, PROCESSOR_MAPPING, TOKENIZER_MAPPING, AutoConfig, AutoFeatureExtractor, AutoProcessor, AutoTokenizer, BertTokenizer, ProcessorMixin, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaProcessor, ) from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test from transformers.tokenization_utils import TOKENIZER_CONFIG_FILE from transformers.utils import FEATURE_EXTRACTOR_NAME, is_tokenizers_available sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 from test_module.custom_processing import CustomProcessor # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 lowerCamelCase__ = get_tests_dir("""fixtures/dummy_feature_extractor_config.json""") lowerCamelCase__ = get_tests_dir("""fixtures/vocab.json""") lowerCamelCase__ = get_tests_dir("""fixtures""") class A__ ( unittest.TestCase): A_ : Any = ['[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'bla', 'blou'] def __lowerCamelCase ( self ): __lowerCAmelCase : List[str] = 0 def __lowerCamelCase ( self ): __lowerCAmelCase : Any = AutoProcessor.from_pretrained('facebook/wav2vec2-base-960h' ) self.assertIsInstance(_a , _a ) def __lowerCamelCase ( self ): with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : Tuple = WavaVecaConfig() __lowerCAmelCase : int = AutoProcessor.from_pretrained('facebook/wav2vec2-base-960h' ) # save in new folder model_config.save_pretrained(_a ) processor.save_pretrained(_a ) __lowerCAmelCase : int = AutoProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def __lowerCamelCase ( self ): with tempfile.TemporaryDirectory() as tmpdirname: # copy relevant files copyfile(_a , os.path.join(_a , _a ) ) copyfile(_a , os.path.join(_a , 'vocab.json' ) ) __lowerCAmelCase : List[Any] = AutoProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def __lowerCamelCase ( self ): with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : Optional[int] = WavaVecaFeatureExtractor() __lowerCAmelCase : Dict = AutoTokenizer.from_pretrained('facebook/wav2vec2-base-960h' ) __lowerCAmelCase : Dict = WavaVecaProcessor(_a , _a ) # save in new folder processor.save_pretrained(_a ) # drop `processor_class` in tokenizer with open(os.path.join(_a , _a ) , 'r' ) as f: __lowerCAmelCase : Optional[Any] = json.load(_a ) config_dict.pop('processor_class' ) with open(os.path.join(_a , _a ) , 'w' ) as f: f.write(json.dumps(_a ) ) __lowerCAmelCase : Any = AutoProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def __lowerCamelCase ( self ): with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : Dict = WavaVecaFeatureExtractor() __lowerCAmelCase : Dict = AutoTokenizer.from_pretrained('facebook/wav2vec2-base-960h' ) __lowerCAmelCase : List[str] = WavaVecaProcessor(_a , _a ) # save in new folder processor.save_pretrained(_a ) # drop `processor_class` in feature extractor with open(os.path.join(_a , _a ) , 'r' ) as f: __lowerCAmelCase : List[str] = json.load(_a ) config_dict.pop('processor_class' ) with open(os.path.join(_a , _a ) , 'w' ) as f: f.write(json.dumps(_a ) ) __lowerCAmelCase : int = AutoProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def __lowerCamelCase ( self ): with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : int = WavaVecaConfig(processor_class='Wav2Vec2Processor' ) model_config.save_pretrained(_a ) # copy relevant files copyfile(_a , os.path.join(_a , 'vocab.json' ) ) # create emtpy sample processor with open(os.path.join(_a , _a ) , 'w' ) as f: f.write('{}' ) __lowerCAmelCase : str = AutoProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def __lowerCamelCase ( self ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(_a ): __lowerCAmelCase : Tuple = AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(_a ): __lowerCAmelCase : Any = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=_a ) __lowerCAmelCase : Any = AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' , trust_remote_code=_a ) self.assertTrue(processor.special_attribute_present ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) __lowerCAmelCase : str = processor.feature_extractor self.assertTrue(feature_extractor.special_attribute_present ) self.assertEqual(feature_extractor.__class__.__name__ , 'NewFeatureExtractor' ) __lowerCAmelCase : Dict = processor.tokenizer self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) # Test we can also load the slow version __lowerCAmelCase : str = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=_a , use_fast=_a ) __lowerCAmelCase : int = new_processor.tokenizer self.assertTrue(new_tokenizer.special_attribute_present ) self.assertEqual(new_tokenizer.__class__.__name__ , 'NewTokenizer' ) else: self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) def __lowerCamelCase ( self ): try: AutoConfig.register('custom' , _a ) AutoFeatureExtractor.register(_a , _a ) AutoTokenizer.register(_a , slow_tokenizer_class=_a ) AutoProcessor.register(_a , _a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_a ): AutoProcessor.register(_a , _a ) # Now that the config is registered, it can be used as any other config with the auto-API __lowerCAmelCase : str = CustomFeatureExtractor.from_pretrained(_a ) with tempfile.TemporaryDirectory() as tmp_dir: __lowerCAmelCase : Dict = os.path.join(_a , 'vocab.txt' ) with open(_a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) __lowerCAmelCase : int = CustomTokenizer(_a ) __lowerCAmelCase : Optional[Any] = CustomProcessor(_a , _a ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained(_a ) __lowerCAmelCase : Optional[Any] = AutoProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def __lowerCamelCase ( self ): class A__ ( _lowerCamelCase): A_ : List[Any] = False class A__ ( _lowerCamelCase): A_ : List[str] = False class A__ ( _lowerCamelCase): A_ : str = 'AutoFeatureExtractor' A_ : int = 'AutoTokenizer' A_ : int = False try: AutoConfig.register('custom' , _a ) AutoFeatureExtractor.register(_a , _a ) AutoTokenizer.register(_a , slow_tokenizer_class=_a ) AutoProcessor.register(_a , _a ) # If remote code is not set, the default is to use local classes. __lowerCAmelCase : Optional[Any] = AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote code is disabled, we load the local ones. __lowerCAmelCase : int = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=_a ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub. __lowerCAmelCase : Optional[int] = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=_a ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertTrue(processor.special_attribute_present ) self.assertTrue(processor.feature_extractor.special_attribute_present ) self.assertTrue(processor.tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def __lowerCamelCase ( self ): __lowerCAmelCase : Dict = AutoProcessor.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(processor.__class__.__name__ , 'BertTokenizerFast' ) def __lowerCamelCase ( self ): __lowerCAmelCase : Tuple = AutoProcessor.from_pretrained('hf-internal-testing/tiny-random-convnext' ) self.assertEqual(processor.__class__.__name__ , 'ConvNextImageProcessor' ) @is_staging_test class A__ ( unittest.TestCase): A_ : Dict = ['[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'bla', 'blou'] @classmethod def __lowerCamelCase ( cls ): __lowerCAmelCase : List[str] = TOKEN HfFolder.save_token(_a ) @classmethod def __lowerCamelCase ( cls ): try: delete_repo(token=cls._token , repo_id='test-processor' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-processor-org' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-processor' ) except HTTPError: pass def __lowerCamelCase ( self ): __lowerCAmelCase : Optional[Any] = WavaVecaProcessor.from_pretrained(_a ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(_a , 'test-processor' ) , push_to_hub=_a , use_auth_token=self._token ) __lowerCAmelCase : int = WavaVecaProcessor.from_pretrained(f"{USER}/test-processor" ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(_a , getattr(new_processor.feature_extractor , _a ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def __lowerCamelCase ( self ): __lowerCAmelCase : Optional[Any] = WavaVecaProcessor.from_pretrained(_a ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(_a , 'test-processor-org' ) , push_to_hub=_a , use_auth_token=self._token , organization='valid_org' , ) __lowerCAmelCase : Dict = WavaVecaProcessor.from_pretrained('valid_org/test-processor-org' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(_a , getattr(new_processor.feature_extractor , _a ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def __lowerCamelCase ( self ): CustomFeatureExtractor.register_for_auto_class() CustomTokenizer.register_for_auto_class() CustomProcessor.register_for_auto_class() __lowerCAmelCase : str = CustomFeatureExtractor.from_pretrained(_a ) with tempfile.TemporaryDirectory() as tmp_dir: __lowerCAmelCase : List[str] = os.path.join(_a , 'vocab.txt' ) with open(_a , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) __lowerCAmelCase : List[Any] = CustomTokenizer(_a ) __lowerCAmelCase : Union[str, Any] = CustomProcessor(_a , _a ) with tempfile.TemporaryDirectory() as tmp_dir: create_repo(f"{USER}/test-dynamic-processor" , token=self._token ) __lowerCAmelCase : Optional[int] = Repository(_a , clone_from=f"{USER}/test-dynamic-processor" , token=self._token ) processor.save_pretrained(_a ) # This has added the proper auto_map field to the feature extractor config self.assertDictEqual( processor.feature_extractor.auto_map , { 'AutoFeatureExtractor': 'custom_feature_extraction.CustomFeatureExtractor', 'AutoProcessor': 'custom_processing.CustomProcessor', } , ) # This has added the proper auto_map field to the tokenizer config with open(os.path.join(_a , 'tokenizer_config.json' ) ) as f: __lowerCAmelCase : Optional[int] = json.load(_a ) self.assertDictEqual( tokenizer_config['auto_map'] , { 'AutoTokenizer': ['custom_tokenization.CustomTokenizer', None], 'AutoProcessor': 'custom_processing.CustomProcessor', } , ) # The code has been copied from fixtures self.assertTrue(os.path.isfile(os.path.join(_a , 'custom_feature_extraction.py' ) ) ) self.assertTrue(os.path.isfile(os.path.join(_a , 'custom_tokenization.py' ) ) ) self.assertTrue(os.path.isfile(os.path.join(_a , 'custom_processing.py' ) ) ) repo.push_to_hub() __lowerCAmelCase : List[str] = AutoProcessor.from_pretrained(f"{USER}/test-dynamic-processor" , trust_remote_code=_a ) # Can't make an isinstance check because the new_processor is from the CustomProcessor class of a dynamic module self.assertEqual(new_processor.__class__.__name__ , 'CustomProcessor' )
86
from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_ ( _snake_case : str = "laptop" ) -> DataFrame: '''simple docstring''' __magic_name__ : Tuple = F'''https://www.amazon.in/laptop/s?k={product}''' __magic_name__ : Dict = { "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36", "Accept-Language": "en-US, en;q=0.5", } __magic_name__ : Tuple = BeautifulSoup(requests.get(_snake_case , headers=_snake_case ).text ) # Initialize a Pandas dataframe with the column titles __magic_name__ : int = DataFrame( columns=[ "Product Title", "Product Link", "Current Price of the product", "Product Rating", "MRP of the product", "Discount", ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( "div" , attrs={"class": "s-result-item", "data-component-type": "s-search-result"} , ) , soup.find_all("div" , attrs={"class": "a-row a-size-base a-color-base"} ) , ): try: __magic_name__ : Dict = item.ha.text __magic_name__ : Optional[int] = "https://www.amazon.in/" + item.ha.a["href"] __magic_name__ : Optional[Any] = item.find("span" , attrs={"class": "a-offscreen"} ).text try: __magic_name__ : Union[str, Any] = item.find("span" , attrs={"class": "a-icon-alt"} ).text except AttributeError: __magic_name__ : Dict = "Not available" try: __magic_name__ : Optional[int] = ( "₹" + item.find( "span" , attrs={"class": "a-price a-text-price"} ).text.split("₹" )[1] ) except AttributeError: __magic_name__ : List[str] = "" try: __magic_name__ : int = float( ( ( float(product_mrp.strip("₹" ).replace("," , "" ) ) - float(product_price.strip("₹" ).replace("," , "" ) ) ) / float(product_mrp.strip("₹" ).replace("," , "" ) ) ) * 100 ) except ValueError: __magic_name__ : str = float("nan" ) except AttributeError: pass __magic_name__ : Optional[int] = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] __magic_name__ : Optional[Any] = " " __magic_name__ : str = " " data_frame.index += 1 return data_frame if __name__ == "__main__": snake_case : Any = "headphones" get_amazon_product_data(product).to_csv(F"Amazon Product Data for {product}.csv")
281
0
import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class __snake_case ( lowerCamelCase_ , unittest.TestCase ): lowerCAmelCase_ = CTRLTokenizer lowerCAmelCase_ = False lowerCAmelCase_ = False def __a ( self : Any ): """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt SCREAMING_SNAKE_CASE__ = ["adapt", "re@@", "a@@", "apt", "c@@", "t", "<unk>"] SCREAMING_SNAKE_CASE__ = dict(zip(_a , range(len(_a ) ) ) ) SCREAMING_SNAKE_CASE__ = ["#version: 0.2", "a p", "ap t</w>", "r e", "a d", "ad apt</w>", ""] SCREAMING_SNAKE_CASE__ = {"unk_token": "<unk>"} SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(_a ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(_a ) ) def __a ( self : Tuple , **_lowercase : Optional[Any] ): """simple docstring""" kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname , **_a ) def __a ( self : Union[str, Any] , _lowercase : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = "adapt react readapt apt" SCREAMING_SNAKE_CASE__ = "adapt react readapt apt" return input_text, output_text def __a ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE__ = CTRLTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) SCREAMING_SNAKE_CASE__ = "adapt react readapt apt" SCREAMING_SNAKE_CASE__ = "adapt re@@ a@@ c@@ t re@@ adapt apt".split() SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(_a ) self.assertListEqual(_a , _a ) SCREAMING_SNAKE_CASE__ = tokens + [tokenizer.unk_token] SCREAMING_SNAKE_CASE__ = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , _a )
219
from __future__ import annotations class _snake_case : def __init__( self , _a ): __magic_name__ : Optional[Any] = data __magic_name__ : Node | None = None __magic_name__ : Node | None = None def lowerCAmelCase_ ( _snake_case : Node | None ) -> None: # In Order traversal of the tree '''simple docstring''' if tree: display(tree.left ) print(tree.data ) display(tree.right ) def lowerCAmelCase_ ( _snake_case : Node | None ) -> int: '''simple docstring''' return 1 + max(depth_of_tree(tree.left ) , depth_of_tree(tree.right ) ) if tree else 0 def lowerCAmelCase_ ( _snake_case : Node ) -> bool: '''simple docstring''' if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def lowerCAmelCase_ ( ) -> None: # Main function for testing. '''simple docstring''' __magic_name__ : int = Node(1 ) __magic_name__ : Union[str, Any] = Node(2 ) __magic_name__ : Tuple = Node(3 ) __magic_name__ : Optional[Any] = Node(4 ) __magic_name__ : Union[str, Any] = Node(5 ) __magic_name__ : Any = Node(6 ) __magic_name__ : int = Node(7 ) __magic_name__ : List[str] = Node(8 ) __magic_name__ : Union[str, Any] = Node(9 ) print(is_full_binary_tree(_snake_case ) ) print(depth_of_tree(_snake_case ) ) print("Tree is: " ) display(_snake_case ) if __name__ == "__main__": main()
281
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig class lowercase ( A__ ): """simple docstring""" _a = 'bert-generation' def __init__( self , UpperCamelCase_=50358 , UpperCamelCase_=1024 , UpperCamelCase_=24 , UpperCamelCase_=16 , UpperCamelCase_=4096 , UpperCamelCase_="gelu" , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_=512 , UpperCamelCase_=0.02 , UpperCamelCase_=1e-12 , UpperCamelCase_=0 , UpperCamelCase_=2 , UpperCamelCase_=1 , UpperCamelCase_="absolute" , UpperCamelCase_=True , **UpperCamelCase_ , ): '''simple docstring''' super().__init__(pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , **_a ) UpperCamelCase__ :str = vocab_size UpperCamelCase__ :Dict = hidden_size UpperCamelCase__ :str = num_hidden_layers UpperCamelCase__ :List[str] = num_attention_heads UpperCamelCase__ :Optional[int] = hidden_act UpperCamelCase__ :Tuple = intermediate_size UpperCamelCase__ :Dict = hidden_dropout_prob UpperCamelCase__ :List[Any] = attention_probs_dropout_prob UpperCamelCase__ :List[str] = max_position_embeddings UpperCamelCase__ :Optional[int] = initializer_range UpperCamelCase__ :List[Any] = layer_norm_eps UpperCamelCase__ :Optional[Any] = position_embedding_type UpperCamelCase__ :List[Any] = use_cache
97
def lowerCAmelCase_ ( _snake_case : str , _snake_case : str ) -> bool: '''simple docstring''' __magic_name__ : Union[str, Any] = len(_snake_case ) + 1 __magic_name__ : List[str] = len(_snake_case ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length i of input_string matches with prefix string of length j of # given pattern. # "dp" stands for dynamic programming. __magic_name__ : str = [[0 for i in range(_snake_case )] for j in range(_snake_case )] # since string of zero length match pattern of zero length __magic_name__ : Optional[int] = 1 # since pattern of zero length will never match with string of non-zero length for i in range(1 , _snake_case ): __magic_name__ : Optional[int] = 0 # since string of zero length will match with pattern where there # is at least one * alternatively for j in range(1 , _snake_case ): __magic_name__ : Union[str, Any] = dp[0][j - 2] if pattern[j - 1] == "*" else 0 # now using bottom-up approach to find for all remaining lengths for i in range(1 , _snake_case ): for j in range(1 , _snake_case ): if input_string[i - 1] == pattern[j - 1] or pattern[j - 1] == ".": __magic_name__ : Optional[int] = dp[i - 1][j - 1] elif pattern[j - 1] == "*": if dp[i][j - 2] == 1: __magic_name__ : Optional[Any] = 1 elif pattern[j - 2] in (input_string[i - 1], "."): __magic_name__ : List[Any] = dp[i - 1][j] else: __magic_name__ : Union[str, Any] = 0 else: __magic_name__ : Dict = 0 return bool(dp[-1][-1] ) if __name__ == "__main__": import doctest doctest.testmod() # inputing the strings # input_string = input("input a string :") # pattern = input("input a pattern :") snake_case : Optional[Any] = "aab" snake_case : List[str] = "c*a*b" # using function to check whether given string matches the given pattern if match_pattern(input_string, pattern): print(F"{input_string} matches the given pattern {pattern}") else: print(F"{input_string} does not match with the given pattern {pattern}")
281
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() A: Optional[Any] = logging.get_logger(__name__) def _snake_case ( UpperCamelCase : Union[str, Any] , UpperCamelCase : Union[str, Any]=False ): UpperCAmelCase : Union[str, Any] = [] # fmt: off # stem: rename_keys.append(("""cls_token""", """vit.embeddings.cls_token""") ) rename_keys.append(("""pos_embed""", """vit.embeddings.position_embeddings""") ) rename_keys.append(("""patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight""") ) rename_keys.append(("""patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias""") ) # backbone rename_keys.append(("""patch_embed.backbone.stem.conv.weight""", """vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight""") ) rename_keys.append(("""patch_embed.backbone.stem.norm.weight""", """vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight""") ) rename_keys.append(("""patch_embed.backbone.stem.norm.bias""", """vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias""") ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias") ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"blocks.{i}.norm1.weight", F"vit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((F"blocks.{i}.norm1.bias", F"vit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append((F"blocks.{i}.attn.proj.weight", F"vit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((F"blocks.{i}.attn.proj.bias", F"vit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((F"blocks.{i}.norm2.weight", F"vit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((F"blocks.{i}.norm2.bias", F"vit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((F"blocks.{i}.mlp.fc1.weight", F"vit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((F"blocks.{i}.mlp.fc1.bias", F"vit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((F"blocks.{i}.mlp.fc2.weight", F"vit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((F"blocks.{i}.mlp.fc2.bias", F"vit.encoder.layer.{i}.output.dense.bias") ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" UpperCAmelCase : int = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) # fmt: on return rename_keys def _snake_case ( UpperCamelCase : Any , UpperCamelCase : Any , UpperCamelCase : Dict=False ): for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase : int = "" else: UpperCAmelCase : Union[str, Any] = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase : Optional[Any] = state_dict.pop(F"blocks.{i}.attn.qkv.weight" ) UpperCAmelCase : int = state_dict.pop(F"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase : Dict = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase : List[str] = in_proj_bias[: config.hidden_size] UpperCAmelCase : List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase : Optional[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase : int = in_proj_bias[-config.hidden_size :] def _snake_case ( UpperCamelCase : List[str] ): UpperCAmelCase : List[str] = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(_snake_case , _snake_case ) def _snake_case ( UpperCamelCase : Optional[int] , UpperCamelCase : int , UpperCamelCase : Union[str, Any] ): UpperCAmelCase : int = dct.pop(_snake_case ) UpperCAmelCase : List[Any] = val def _snake_case ( ): UpperCAmelCase : List[str] = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase : List[str] = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return im @torch.no_grad() def _snake_case ( UpperCamelCase : Optional[int] , UpperCamelCase : Any , UpperCamelCase : int=False ): UpperCAmelCase : List[str] = BitConfig( global_padding="""same""" , layer_type="""bottleneck""" , depths=(3, 4, 9) , out_features=["""stage3"""] , embedding_dynamic_padding=_snake_case , ) UpperCAmelCase : List[str] = ViTHybridConfig(backbone_config=_snake_case , image_size=384 , num_labels=1000 ) UpperCAmelCase : str = False # load original model from timm UpperCAmelCase : Union[str, Any] = timm.create_model(_snake_case , pretrained=_snake_case ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase : List[Any] = timm_model.state_dict() if base_model: remove_classification_head_(_snake_case ) UpperCAmelCase : Tuple = create_rename_keys(_snake_case , _snake_case ) for src, dest in rename_keys: rename_key(_snake_case , _snake_case , _snake_case ) read_in_q_k_v(_snake_case , _snake_case , _snake_case ) UpperCAmelCase : List[str] = "huggingface/label-files" UpperCAmelCase : int = "imagenet-1k-id2label.json" UpperCAmelCase : Optional[int] = json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase : int = {int(_snake_case ): v for k, v in idalabel.items()} UpperCAmelCase : List[str] = idalabel UpperCAmelCase : List[str] = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": UpperCAmelCase : List[str] = ViTHybridModel(_snake_case ).eval() else: UpperCAmelCase : str = ViTHybridForImageClassification(_snake_case ).eval() model.load_state_dict(_snake_case ) # create image processor UpperCAmelCase : List[Any] = create_transform(**resolve_data_config({} , model=_snake_case ) ) UpperCAmelCase : int = transform.transforms UpperCAmelCase : List[str] = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } UpperCAmelCase : int = ViTHybridImageProcessor( do_resize=_snake_case , size={"""shortest_edge""": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_snake_case , crop_size={"""height""": timm_transforms[1].size[0], """width""": timm_transforms[1].size[1]} , do_normalize=_snake_case , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) UpperCAmelCase : List[Any] = prepare_img() UpperCAmelCase : Any = transform(_snake_case ).unsqueeze(0 ) UpperCAmelCase : Tuple = processor(_snake_case , return_tensors="""pt""" ).pixel_values # verify pixel values assert torch.allclose(_snake_case , _snake_case ) # verify logits with torch.no_grad(): UpperCAmelCase : Optional[int] = model(_snake_case ) UpperCAmelCase : List[str] = outputs.logits print("""Predicted class:""" , logits.argmax(-1 ).item() ) if base_model: UpperCAmelCase : List[str] = timm_model.forward_features(_snake_case ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_snake_case , outputs.pooler_output , atol=1e-3 ) else: UpperCAmelCase : Any = timm_model(_snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_snake_case , outputs.logits , atol=1e-3 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: Path(_snake_case ).mkdir(exist_ok=_snake_case ) print(F"Saving model {vit_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(_snake_case ) print(F"Saving processor to {pytorch_dump_folder_path}" ) processor.save_pretrained(_snake_case ) if push_to_hub: print(F"Pushing model and processor to the hub {vit_name}" ) model.push_to_hub(F"ybelkada/{vit_name}" ) processor.push_to_hub(F"ybelkada/{vit_name}" ) if __name__ == "__main__": A: Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--vit_name", default="vit_base_r50_s16_384", type=str, help="Name of the hybrid ViT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to upload the model to the HuggingFace hub." ) A: List[Any] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
109
import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class _snake_case : @staticmethod def SCREAMING_SNAKE_CASE ( *_a , **_a ): pass def lowerCAmelCase_ ( _snake_case : Image ) -> str: '''simple docstring''' __magic_name__ : Optional[int] = hashlib.mda(image.tobytes() ) return m.hexdigest()[:10] def lowerCAmelCase_ ( _snake_case : Image ) -> Dict: '''simple docstring''' __magic_name__ : List[Any] = np.array(_snake_case ) __magic_name__ : Optional[int] = npimg.shape return {"hash": hashimage(_snake_case ), "shape": shape} @is_pipeline_test @require_vision @require_torch class _snake_case ( unittest.TestCase ): UpperCamelCase__ = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) UpperCamelCase__ = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def SCREAMING_SNAKE_CASE ( self , _a , _a , _a ): __magic_name__ : Dict = MaskGenerationPipeline(model=_a , image_processor=_a ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self , _a , _a ): pass @require_tf @unittest.skip("Image segmentation not implemented in TF" ) def SCREAMING_SNAKE_CASE ( self ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = pipeline("mask-generation" , model="facebook/sam-vit-huge" ) __magic_name__ : str = image_segmenter("http://images.cocodataset.org/val2017/000000039769.jpg" , points_per_batch=256 ) # Shortening by hashing __magic_name__ : Dict = [] for i, o in enumerate(outputs["masks"] ): new_outupt += [{"mask": mask_to_test_readable(_a ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(_a , decimals=4 ) , [ {"mask": {"hash": "115ad19f5f", "shape": (480, 640)}, "scores": 1.04_44}, {"mask": {"hash": "6affa964c6", "shape": (480, 640)}, "scores": 1.0_21}, {"mask": {"hash": "dfe28a0388", "shape": (480, 640)}, "scores": 1.01_67}, {"mask": {"hash": "c0a5f4a318", "shape": (480, 640)}, "scores": 1.01_32}, {"mask": {"hash": "fe8065c197", "shape": (480, 640)}, "scores": 1.00_53}, {"mask": {"hash": "e2d0b7a0b7", "shape": (480, 640)}, "scores": 0.99_67}, {"mask": {"hash": "453c7844bd", "shape": (480, 640)}, "scores": 0.9_93}, {"mask": {"hash": "3d44f2926d", "shape": (480, 640)}, "scores": 0.99_09}, {"mask": {"hash": "64033ddc3f", "shape": (480, 640)}, "scores": 0.98_79}, {"mask": {"hash": "801064ff79", "shape": (480, 640)}, "scores": 0.98_34}, {"mask": {"hash": "6172f276ef", "shape": (480, 640)}, "scores": 0.97_16}, {"mask": {"hash": "b49e60e084", "shape": (480, 640)}, "scores": 0.96_12}, {"mask": {"hash": "a811e775fd", "shape": (480, 640)}, "scores": 0.95_99}, {"mask": {"hash": "a6a8ebcf4b", "shape": (480, 640)}, "scores": 0.95_52}, {"mask": {"hash": "9d8257e080", "shape": (480, 640)}, "scores": 0.95_32}, {"mask": {"hash": "32de6454a8", "shape": (480, 640)}, "scores": 0.95_16}, {"mask": {"hash": "af3d4af2c8", "shape": (480, 640)}, "scores": 0.94_99}, {"mask": {"hash": "3c6db475fb", "shape": (480, 640)}, "scores": 0.94_83}, {"mask": {"hash": "c290813fb9", "shape": (480, 640)}, "scores": 0.94_64}, {"mask": {"hash": "b6f0b8f606", "shape": (480, 640)}, "scores": 0.9_43}, {"mask": {"hash": "92ce16bfdf", "shape": (480, 640)}, "scores": 0.9_43}, {"mask": {"hash": "c749b25868", "shape": (480, 640)}, "scores": 0.94_08}, {"mask": {"hash": "efb6cab859", "shape": (480, 640)}, "scores": 0.93_35}, {"mask": {"hash": "1ff2eafb30", "shape": (480, 640)}, "scores": 0.93_26}, {"mask": {"hash": "788b798e24", "shape": (480, 640)}, "scores": 0.92_62}, {"mask": {"hash": "abea804f0e", "shape": (480, 640)}, "scores": 0.89_99}, {"mask": {"hash": "7b9e8ddb73", "shape": (480, 640)}, "scores": 0.89_86}, {"mask": {"hash": "cd24047c8a", "shape": (480, 640)}, "scores": 0.89_84}, {"mask": {"hash": "6943e6bcbd", "shape": (480, 640)}, "scores": 0.88_73}, {"mask": {"hash": "b5f47c9191", "shape": (480, 640)}, "scores": 0.88_71} ] , ) # fmt: on @require_torch @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : str = "facebook/sam-vit-huge" __magic_name__ : str = pipeline("mask-generation" , model=_a ) __magic_name__ : Tuple = image_segmenter( "http://images.cocodataset.org/val2017/000000039769.jpg" , pred_iou_thresh=1 , points_per_batch=256 ) # Shortening by hashing __magic_name__ : Any = [] for i, o in enumerate(outputs["masks"] ): new_outupt += [{"mask": mask_to_test_readable(_a ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(_a , decimals=4 ) , [ {"mask": {"hash": "115ad19f5f", "shape": (480, 640)}, "scores": 1.04_44}, {"mask": {"hash": "6affa964c6", "shape": (480, 640)}, "scores": 1.02_10}, {"mask": {"hash": "dfe28a0388", "shape": (480, 640)}, "scores": 1.01_67}, {"mask": {"hash": "c0a5f4a318", "shape": (480, 640)}, "scores": 1.01_32}, {"mask": {"hash": "fe8065c197", "shape": (480, 640)}, "scores": 1.00_53}, ] , )
281
0
import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipaConfig, BlipaForConditionalGeneration, BlipaProcessor, BlipaVisionConfig, BlipImageProcessor, OPTConfig, TaConfig, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def __UpperCamelCase ( ): lowerCAmelCase_ = "https://storage.googleapis.com/sfr-vision-language-research/LAVIS/assets/merlion.png" lowerCAmelCase_ = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ).convert('''RGB''' ) return image def __UpperCamelCase ( _A ): lowerCAmelCase_ = [] # fmt: off # vision encoder rename_keys.append(('''visual_encoder.cls_token''', '''vision_model.embeddings.class_embedding''') ) rename_keys.append(('''visual_encoder.pos_embed''', '''vision_model.embeddings.position_embedding''') ) rename_keys.append(('''visual_encoder.patch_embed.proj.weight''', '''vision_model.embeddings.patch_embedding.weight''') ) rename_keys.append(('''visual_encoder.patch_embed.proj.bias''', '''vision_model.embeddings.patch_embedding.bias''') ) rename_keys.append(('''ln_vision.weight''', '''vision_model.post_layernorm.weight''') ) rename_keys.append(('''ln_vision.bias''', '''vision_model.post_layernorm.bias''') ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((f"visual_encoder.blocks.{i}.norm1.weight", f"vision_model.encoder.layers.{i}.layer_norm1.weight") ) rename_keys.append((f"visual_encoder.blocks.{i}.norm1.bias", f"vision_model.encoder.layers.{i}.layer_norm1.bias") ) rename_keys.append((f"visual_encoder.blocks.{i}.norm2.weight", f"vision_model.encoder.layers.{i}.layer_norm2.weight") ) rename_keys.append((f"visual_encoder.blocks.{i}.norm2.bias", f"vision_model.encoder.layers.{i}.layer_norm2.bias") ) rename_keys.append((f"visual_encoder.blocks.{i}.attn.qkv.weight", f"vision_model.encoder.layers.{i}.self_attn.qkv.weight") ) rename_keys.append((f"visual_encoder.blocks.{i}.attn.proj.weight", f"vision_model.encoder.layers.{i}.self_attn.projection.weight",) ) rename_keys.append((f"visual_encoder.blocks.{i}.attn.proj.bias", f"vision_model.encoder.layers.{i}.self_attn.projection.bias") ) rename_keys.append((f"visual_encoder.blocks.{i}.mlp.fc1.weight", f"vision_model.encoder.layers.{i}.mlp.fc1.weight") ) rename_keys.append((f"visual_encoder.blocks.{i}.mlp.fc1.bias", f"vision_model.encoder.layers.{i}.mlp.fc1.bias") ) rename_keys.append((f"visual_encoder.blocks.{i}.mlp.fc2.weight", f"vision_model.encoder.layers.{i}.mlp.fc2.weight") ) rename_keys.append((f"visual_encoder.blocks.{i}.mlp.fc2.bias", f"vision_model.encoder.layers.{i}.mlp.fc2.bias") ) # QFormer rename_keys.append(('''Qformer.bert.embeddings.LayerNorm.weight''', '''qformer.layernorm.weight''') ) rename_keys.append(('''Qformer.bert.embeddings.LayerNorm.bias''', '''qformer.layernorm.bias''') ) # fmt: on return rename_keys def __UpperCamelCase ( _A , _A , _A ): lowerCAmelCase_ = dct.pop(_snake_case ) lowerCAmelCase_ = val def __UpperCamelCase ( _A , _A ): for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases lowerCAmelCase_ = state_dict.pop(f"visual_encoder.blocks.{i}.attn.q_bias" ) lowerCAmelCase_ = state_dict.pop(f"visual_encoder.blocks.{i}.attn.v_bias" ) # next, set bias in the state dict lowerCAmelCase_ = torch.cat((q_bias, torch.zeros_like(_snake_case , requires_grad=_snake_case ), v_bias) ) lowerCAmelCase_ = qkv_bias def __UpperCamelCase ( _A , _A ): lowerCAmelCase_ = 364 if "coco" in model_name else 224 lowerCAmelCase_ = BlipaVisionConfig(image_size=_snake_case ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "opt-2.7b" in model_name: lowerCAmelCase_ = OPTConfig.from_pretrained('''facebook/opt-2.7b''' , eos_token_id=_snake_case ).to_dict() elif "opt-6.7b" in model_name: lowerCAmelCase_ = OPTConfig.from_pretrained('''facebook/opt-6.7b''' , eos_token_id=_snake_case ).to_dict() elif "t5-xl" in model_name: lowerCAmelCase_ = TaConfig.from_pretrained('''google/flan-t5-xl''' , dense_act_fn='''gelu''' , bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: lowerCAmelCase_ = TaConfig.from_pretrained('''google/flan-t5-xxl''' , dense_act_fn='''gelu''' , bos_token_id=1 ).to_dict() lowerCAmelCase_ = BlipaConfig(vision_config=_snake_case , text_config=_snake_case ) return config, image_size @torch.no_grad() def __UpperCamelCase ( _A , _A=None , _A=False ): lowerCAmelCase_ = ( AutoTokenizer.from_pretrained('''facebook/opt-2.7b''' ) if "opt" in model_name else AutoTokenizer.from_pretrained('''google/flan-t5-xl''' ) ) lowerCAmelCase_ = tokenizer('''\n''' , add_special_tokens=_snake_case ).input_ids[0] lowerCAmelCase_ = get_blipa_config(_snake_case , eos_token_id=_snake_case ) lowerCAmelCase_ = BlipaForConditionalGeneration(_snake_case ).eval() lowerCAmelCase_ = { "blip2-opt-2.7b": ("blip2_opt", "pretrain_opt2.7b"), "blip2-opt-6.7b": ("blip2_opt", "pretrain_opt6.7b"), "blip2-opt-2.7b-coco": ("blip2_opt", "caption_coco_opt2.7b"), "blip2-opt-6.7b-coco": ("blip2_opt", "caption_coco_opt6.7b"), "blip2-flan-t5-xl": ("blip2_t5", "pretrain_flant5xl"), "blip2-flan-t5-xl-coco": ("blip2_t5", "caption_coco_flant5xl"), "blip2-flan-t5-xxl": ("blip2_t5", "pretrain_flant5xxl"), } lowerCAmelCase_ = model_name_to_original[model_name] # load original model print('''Loading original model...''' ) lowerCAmelCase_ = "cuda" if torch.cuda.is_available() else "cpu" lowerCAmelCase_ = load_model_and_preprocess( name=_snake_case , model_type=_snake_case , is_eval=_snake_case , device=_snake_case ) original_model.eval() print('''Done!''' ) # update state dict keys lowerCAmelCase_ = original_model.state_dict() lowerCAmelCase_ = create_rename_keys(_snake_case ) for src, dest in rename_keys: rename_key(_snake_case , _snake_case , _snake_case ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): lowerCAmelCase_ = state_dict.pop(_snake_case ) if key.startswith('''Qformer.bert''' ): lowerCAmelCase_ = key.replace('''Qformer.bert''' , '''qformer''' ) if "attention.self" in key: lowerCAmelCase_ = key.replace('''self''' , '''attention''' ) if "opt_proj" in key: lowerCAmelCase_ = key.replace('''opt_proj''' , '''language_projection''' ) if "t5_proj" in key: lowerCAmelCase_ = key.replace('''t5_proj''' , '''language_projection''' ) if key.startswith('''opt''' ): lowerCAmelCase_ = key.replace('''opt''' , '''language''' ) if key.startswith('''t5''' ): lowerCAmelCase_ = key.replace('''t5''' , '''language''' ) lowerCAmelCase_ = val # read in qv biases read_in_q_v_bias(_snake_case , _snake_case ) lowerCAmelCase_ = hf_model.load_state_dict(_snake_case , strict=_snake_case ) assert len(_snake_case ) == 0 assert unexpected_keys == ["qformer.embeddings.position_ids"] lowerCAmelCase_ = load_demo_image() lowerCAmelCase_ = vis_processors["eval"](_snake_case ).unsqueeze(0 ).to(_snake_case ) lowerCAmelCase_ = tokenizer(['''\n'''] , return_tensors='''pt''' ).input_ids.to(_snake_case ) # create processor lowerCAmelCase_ = BlipImageProcessor( size={'''height''': image_size, '''width''': image_size} , image_mean=_snake_case , image_std=_snake_case ) lowerCAmelCase_ = BlipaProcessor(image_processor=_snake_case , tokenizer=_snake_case ) lowerCAmelCase_ = processor(images=_snake_case , return_tensors='''pt''' ).pixel_values.to(_snake_case ) # make sure processor creates exact same pixel values assert torch.allclose(_snake_case , _snake_case ) original_model.to(_snake_case ) hf_model.to(_snake_case ) with torch.no_grad(): if "opt" in model_name: lowerCAmelCase_ = original_model({'''image''': original_pixel_values, '''text_input''': ['''''']} ).logits lowerCAmelCase_ = hf_model(_snake_case , _snake_case ).logits else: lowerCAmelCase_ = original_model( {'''image''': original_pixel_values, '''text_input''': ['''\n'''], '''text_output''': ['''\n''']} ).logits lowerCAmelCase_ = input_ids.masked_fill(input_ids == tokenizer.pad_token_id , -100 ) lowerCAmelCase_ = hf_model(_snake_case , _snake_case , labels=_snake_case ).logits assert original_logits.shape == logits.shape print('''First values of original logits:''' , original_logits[0, :3, :3] ) print('''First values of HF logits:''' , logits[0, :3, :3] ) # assert values if model_name == "blip2-flan-t5-xl": lowerCAmelCase_ = torch.tensor( [[-4_1.5_8_5_0, -4.4_4_4_0, -8.9_9_2_2], [-4_7.4_3_2_2, -5.9_1_4_3, -1.7_3_4_0]] , device=_snake_case ) assert torch.allclose(logits[0, :3, :3] , _snake_case , atol=1E-4 ) elif model_name == "blip2-flan-t5-xl-coco": lowerCAmelCase_ = torch.tensor( [[-5_7.0_1_0_9, -9.8_9_6_7, -1_2.6_2_8_0], [-6_8.6_5_7_8, -1_2.7_1_9_1, -1_0.5_0_6_5]] , device=_snake_case ) else: # cast to same type lowerCAmelCase_ = logits.dtype assert torch.allclose(original_logits.to(_snake_case ) , _snake_case , atol=1E-2 ) print('''Looks ok!''' ) print('''Generating a caption...''' ) lowerCAmelCase_ = "" lowerCAmelCase_ = tokenizer(_snake_case , return_tensors='''pt''' ).input_ids.to(_snake_case ) lowerCAmelCase_ = original_model.generate({'''image''': original_pixel_values} ) lowerCAmelCase_ = hf_model.generate( _snake_case , _snake_case , do_sample=_snake_case , num_beams=5 , max_length=30 , min_length=1 , top_p=0.9 , repetition_penalty=1.0 , length_penalty=1.0 , temperature=1 , ) print('''Original generation:''' , _snake_case ) lowerCAmelCase_ = input_ids.shape[1] lowerCAmelCase_ = processor.batch_decode(outputs[:, prompt_length:] , skip_special_tokens=_snake_case ) lowerCAmelCase_ = [text.strip() for text in output_text] print('''HF generation:''' , _snake_case ) if pytorch_dump_folder_path is not None: processor.save_pretrained(_snake_case ) hf_model.save_pretrained(_snake_case ) if push_to_hub: processor.push_to_hub(f"nielsr/{model_name}" ) hf_model.push_to_hub(f"nielsr/{model_name}" ) if __name__ == "__main__": _A = argparse.ArgumentParser() _A = [ "blip2-opt-2.7b", "blip2-opt-6.7b", "blip2-opt-2.7b-coco", "blip2-opt-6.7b-coco", "blip2-flan-t5-xl", "blip2-flan-t5-xl-coco", "blip2-flan-t5-xxl", ] parser.add_argument( '''--model_name''', default='''blip2-opt-2.7b''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model and processor to the hub after converting''', ) _A = parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
278
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets snake_case : List[Any] = "\\n@inproceedings{lin-2004-rouge,\n title = \"{ROUGE}: A Package for Automatic Evaluation of Summaries\",\n author = \"Lin, Chin-Yew\",\n booktitle = \"Text Summarization Branches Out\",\n month = jul,\n year = \"2004\",\n address = \"Barcelona, Spain\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W04-1013\",\n pages = \"74--81\",\n}\n" snake_case : Any = "\\nROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for\nevaluating automatic summarization and machine translation software in natural language processing.\nThe metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation.\n\nNote that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters.\n\nThis metrics is a wrapper around Google Research reimplementation of ROUGE:\nhttps://github.com/google-research/google-research/tree/master/rouge\n" snake_case : str = "\nCalculates average rouge scores for a list of hypotheses and references\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n rouge_types: A list of rouge types to calculate.\n Valid names:\n `\"rouge{n}\"` (e.g. `\"rouge1\"`, `\"rouge2\"`) where: {n} is the n-gram based scoring,\n `\"rougeL\"`: Longest common subsequence based scoring.\n `\"rougeLSum\"`: rougeLsum splits text using `\"\n\"`.\n See details in https://github.com/huggingface/datasets/issues/617\n use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes.\n use_aggregator: Return aggregates if this is set to True\nReturns:\n rouge1: rouge_1 (precision, recall, f1),\n rouge2: rouge_2 (precision, recall, f1),\n rougeL: rouge_l (precision, recall, f1),\n rougeLsum: rouge_lsum (precision, recall, f1)\nExamples:\n\n >>> rouge = datasets.load_metric('rouge')\n >>> predictions = [\"hello there\", \"general kenobi\"]\n >>> references = [\"hello there\", \"general kenobi\"]\n >>> results = rouge.compute(predictions=predictions, references=references)\n >>> print(list(results.keys()))\n ['rouge1', 'rouge2', 'rougeL', 'rougeLsum']\n >>> print(results[\"rouge1\"])\n AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0))\n >>> print(results[\"rouge1\"].mid.fmeasure)\n 1.0\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _snake_case ( datasets.Metric ): def SCREAMING_SNAKE_CASE ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/google-research/google-research/tree/master/rouge"] , reference_urls=[ "https://en.wikipedia.org/wiki/ROUGE_(metric)", "https://github.com/google-research/google-research/tree/master/rouge", ] , ) def SCREAMING_SNAKE_CASE ( self , _a , _a , _a=None , _a=True , _a=False ): if rouge_types is None: __magic_name__ : str = ["rouge1", "rouge2", "rougeL", "rougeLsum"] __magic_name__ : List[str] = rouge_scorer.RougeScorer(rouge_types=_a , use_stemmer=_a ) if use_aggregator: __magic_name__ : Dict = scoring.BootstrapAggregator() else: __magic_name__ : str = [] for ref, pred in zip(_a , _a ): __magic_name__ : Union[str, Any] = scorer.score(_a , _a ) if use_aggregator: aggregator.add_scores(_a ) else: scores.append(_a ) if use_aggregator: __magic_name__ : Any = aggregator.aggregate() else: __magic_name__ : List[Any] = {} for key in scores[0]: __magic_name__ : str = [score[key] for score in scores] return result
281
0
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator class lowerCAmelCase : '''simple docstring''' def __init__( self , lowerCAmelCase__ ) -> int: SCREAMING_SNAKE_CASE = value SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None class lowerCAmelCase : '''simple docstring''' def __init__( self , lowerCAmelCase__ ) -> List[str]: SCREAMING_SNAKE_CASE = tree def __A ( self , lowerCAmelCase__ ) -> List[Any]: if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self ) -> Tuple: yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
113
snake_case : Optional[int] = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/" def lowerCAmelCase_ ( _snake_case : bytes ) -> bytes: '''simple docstring''' if not isinstance(_snake_case , _snake_case ): __magic_name__ : Tuple = F'''a bytes-like object is required, not \'{data.__class__.__name__}\'''' raise TypeError(_snake_case ) __magic_name__ : Optional[int] = "".join(bin(_snake_case )[2:].zfill(8 ) for byte in data ) __magic_name__ : List[Any] = len(_snake_case ) % 6 != 0 if padding_needed: # The padding that will be added later __magic_name__ : List[str] = B"=" * ((6 - len(_snake_case ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(_snake_case ) % 6) else: __magic_name__ : List[str] = B"" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(_snake_case ) , 6 ) ).encode() + padding ) def lowerCAmelCase_ ( _snake_case : str ) -> bytes: '''simple docstring''' if not isinstance(_snake_case , _snake_case ) and not isinstance(_snake_case , _snake_case ): __magic_name__ : List[str] = ( "argument should be a bytes-like object or ASCII string, " F'''not \'{encoded_data.__class__.__name__}\'''' ) raise TypeError(_snake_case ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(_snake_case , _snake_case ): try: __magic_name__ : List[Any] = encoded_data.decode("utf-8" ) except UnicodeDecodeError: raise ValueError("base64 encoded data should only contain ASCII characters" ) __magic_name__ : List[str] = encoded_data.count("=" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(_snake_case ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one __magic_name__ : Optional[int] = encoded_data[:-padding] __magic_name__ : Dict = "".join( bin(B64_CHARSET.index(_snake_case ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: __magic_name__ : Union[str, Any] = "".join( bin(B64_CHARSET.index(_snake_case ) )[2:].zfill(6 ) for char in encoded_data ) __magic_name__ : List[Any] = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(_snake_case ) , 8 ) ] return bytes(_snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
281
0
"""simple docstring""" from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class lowerCAmelCase__ ( lowercase ): '''simple docstring''' @slow @require_torch def A_ ( self ): _lowerCamelCase : List[str] = EncoderDecoderModel.from_encoder_decoder_pretrained('prajjwal1/bert-tiny' , 'prajjwal1/bert-tiny' ) _lowerCamelCase : Dict = BertTokenizer.from_pretrained('bert-base-uncased' ) _lowerCamelCase : Optional[Any] = bertabert.config.encoder.vocab_size _lowerCamelCase : int = tokenizer.sep_token_id _lowerCamelCase : Any = tokenizer.cls_token_id _lowerCamelCase : Tuple = 128 _lowerCamelCase : Dict = datasets.load_dataset('cnn_dailymail' , '3.0.0' , split='train[:1%]' ) _lowerCamelCase : List[str] = datasets.load_dataset('cnn_dailymail' , '3.0.0' , split='validation[:1%]' ) _lowerCamelCase : int = train_dataset.select(range(32 ) ) _lowerCamelCase : int = val_dataset.select(range(16 ) ) _lowerCamelCase : Dict = 4 def _map_to_encoder_decoder_inputs(lowercase ): # Tokenizer will automatically set [BOS] <text> [EOS] _lowerCamelCase : Optional[int] = tokenizer(batch['article'] , padding='max_length' , truncation=_a , max_length=512 ) _lowerCamelCase : Union[str, Any] = tokenizer(batch['highlights'] , padding='max_length' , truncation=_a , max_length=128 ) _lowerCamelCase : Union[str, Any] = inputs.input_ids _lowerCamelCase : List[Any] = inputs.attention_mask _lowerCamelCase : List[Any] = outputs.input_ids _lowerCamelCase : List[Any] = outputs.input_ids.copy() _lowerCamelCase : List[Any] = [ [-100 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch["labels"] ] _lowerCamelCase : int = outputs.attention_mask assert all(len(_a ) == 512 for x in inputs.input_ids ) assert all(len(_a ) == 128 for x in outputs.input_ids ) return batch def _compute_metrics(lowercase ): _lowerCamelCase : List[Any] = pred.label_ids _lowerCamelCase : List[Any] = pred.predictions # all unnecessary tokens are removed _lowerCamelCase : Any = tokenizer.batch_decode(_a , skip_special_tokens=_a ) _lowerCamelCase : Any = tokenizer.batch_decode(_a , skip_special_tokens=_a ) _lowerCamelCase : str = sum([int(pred_str[i] == label_str[i] ) for i in range(len(_a ) )] ) / len(_a ) return {"accuracy": accuracy} # map train dataset _lowerCamelCase : Tuple = train_dataset.map( _map_to_encoder_decoder_inputs , batched=_a , batch_size=_a , remove_columns=['article', 'highlights'] , ) train_dataset.set_format( type='torch' , columns=['input_ids', 'attention_mask', 'decoder_input_ids', 'decoder_attention_mask', 'labels'] , ) # same for validation dataset _lowerCamelCase : Union[str, Any] = val_dataset.map( _map_to_encoder_decoder_inputs , batched=_a , batch_size=_a , remove_columns=['article', 'highlights'] , ) val_dataset.set_format( type='torch' , columns=['input_ids', 'attention_mask', 'decoder_input_ids', 'decoder_attention_mask', 'labels'] , ) _lowerCamelCase : List[Any] = self.get_auto_remove_tmp_dir() _lowerCamelCase : str = SeqaSeqTrainingArguments( output_dir=_a , per_device_train_batch_size=_a , per_device_eval_batch_size=_a , predict_with_generate=_a , evaluation_strategy='steps' , do_train=_a , do_eval=_a , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer _lowerCamelCase : List[Any] = SeqaSeqTrainer( model=_a , args=_a , compute_metrics=_compute_metrics , train_dataset=_a , eval_dataset=_a , tokenizer=_a , ) # start training trainer.train()
96
import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class _snake_case ( unittest.TestCase ): def __init__( self , _a , _a=13 , _a=7 , _a=True , _a=True , _a=True , _a=True , _a=99 , _a=32 , _a=5 , _a=4 , _a=37 , _a="gelu" , _a=0.1 , _a=0.1 , _a=512 , _a=16 , _a=2 , _a=0.02 , _a=4 , ): __magic_name__ : List[Any] = parent __magic_name__ : Optional[Any] = batch_size __magic_name__ : Dict = seq_length __magic_name__ : Union[str, Any] = is_training __magic_name__ : Optional[Any] = use_attention_mask __magic_name__ : Optional[Any] = use_token_type_ids __magic_name__ : int = use_labels __magic_name__ : List[Any] = vocab_size __magic_name__ : Union[str, Any] = hidden_size __magic_name__ : Optional[Any] = num_hidden_layers __magic_name__ : int = num_attention_heads __magic_name__ : Any = intermediate_size __magic_name__ : List[Any] = hidden_act __magic_name__ : List[Any] = hidden_dropout_prob __magic_name__ : Optional[int] = attention_probs_dropout_prob __magic_name__ : List[Any] = max_position_embeddings __magic_name__ : Tuple = type_vocab_size __magic_name__ : List[str] = type_sequence_label_size __magic_name__ : Dict = initializer_range __magic_name__ : List[Any] = num_choices def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __magic_name__ : List[Any] = None if self.use_attention_mask: __magic_name__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) __magic_name__ : str = None if self.use_token_type_ids: __magic_name__ : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __magic_name__ : List[str] = RobertaPreLayerNormConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_a , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : int = self.prepare_config_and_inputs() __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : List[Any] = config_and_inputs __magic_name__ : List[str] = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = self.prepare_config_and_inputs() __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Union[str, Any] = config_and_inputs __magic_name__ : Tuple = True __magic_name__ : int = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __magic_name__ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class _snake_case ( snake_case , unittest.TestCase ): UpperCamelCase__ = True UpperCamelCase__ = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[Any] = FlaxRobertaPreLayerNormModelTester(self ) @slow def SCREAMING_SNAKE_CASE ( self ): for model_class_name in self.all_model_classes: __magic_name__ : Optional[Any] = model_class_name.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=_a ) __magic_name__ : Dict = model(np.ones((1, 1) ) ) self.assertIsNotNone(_a ) @require_flax class _snake_case ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = FlaxRobertaPreLayerNormForMaskedLM.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=_a ) __magic_name__ : Union[str, Any] = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) __magic_name__ : List[str] = model(_a )[0] __magic_name__ : str = [1, 11, 50_265] self.assertEqual(list(output.shape ) , _a ) # compare the actual values for a slice. __magic_name__ : List[str] = np.array( [[[40.48_80, 18.01_99, -5.23_67], [-1.88_77, -4.08_85, 10.70_85], [-2.26_13, -5.61_10, 7.26_65]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , _a , atol=1e-4 ) ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[str] = FlaxRobertaPreLayerNormModel.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=_a ) __magic_name__ : Tuple = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) __magic_name__ : Tuple = model(_a )[0] # compare the actual values for a slice. __magic_name__ : Dict = np.array( [[[0.02_08, -0.03_56, 0.02_37], [-0.15_69, -0.04_11, -0.26_26], [0.18_79, 0.01_25, -0.00_89]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , _a , atol=1e-4 ) )
281
0
"""simple docstring""" import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## SCREAMING_SNAKE_CASE__ = 16 SCREAMING_SNAKE_CASE__ = 32 def lowerCAmelCase__ ( _UpperCamelCase : Accelerator , _UpperCamelCase : int = 1_6 ) -> Tuple: """simple docstring""" snake_case = AutoTokenizer.from_pretrained('bert-base-cased' ) snake_case = load_dataset('glue' , 'mrpc' ) def tokenize_function(_UpperCamelCase : Any ): # max_length=None => use the model max length (it's actually the default) snake_case = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=_snake_case , max_length=_snake_case ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): snake_case = datasets.map( _snake_case , batched=_snake_case , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library snake_case = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(_UpperCamelCase : Tuple ): # On TPU it's best to pad everything to the same length or training will be very slow. snake_case = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": snake_case = 1_6 elif accelerator.mixed_precision != "no": snake_case = 8 else: snake_case = None return tokenizer.pad( _snake_case , padding='longest' , max_length=_snake_case , pad_to_multiple_of=_snake_case , return_tensors='pt' , ) # Instantiate dataloaders. snake_case = DataLoader( tokenized_datasets['train'] , shuffle=_snake_case , collate_fn=_snake_case , batch_size=_snake_case ) snake_case = DataLoader( tokenized_datasets['validation'] , shuffle=_snake_case , collate_fn=_snake_case , batch_size=_snake_case ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("TESTING_MOCKED_DATALOADERS", None) == "1": from accelerate.test_utils.training import mocked_dataloaders SCREAMING_SNAKE_CASE__ = mocked_dataloaders # noqa: F811 def lowerCAmelCase__ ( _UpperCamelCase : List[str] , _UpperCamelCase : List[str] ) -> Union[str, Any]: """simple docstring""" if os.environ.get('TESTING_MOCKED_DATALOADERS' , _snake_case ) == "1": snake_case = 2 # Initialize accelerator snake_case = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs snake_case = config["lr"] snake_case = int(config['num_epochs'] ) snake_case = int(config['seed'] ) snake_case = int(config['batch_size'] ) snake_case = evaluate.load('glue' , 'mrpc' ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=_snake_case ) def inner_training_loop(_UpperCamelCase : int ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(_snake_case ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) snake_case = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=_snake_case ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). snake_case = model.to(accelerator.device ) # Instantiate optimizer snake_case = AdamW(params=model.parameters() , lr=_snake_case ) snake_case = get_dataloaders(_snake_case , _snake_case ) # Instantiate scheduler snake_case = get_linear_schedule_with_warmup( optimizer=_snake_case , num_warmup_steps=1_0_0 , num_training_steps=(len(_snake_case ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. snake_case = accelerator.prepare( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) # Now we train the model for epoch in range(_snake_case ): model.train() for step, batch in enumerate(_snake_case ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) snake_case = model(**_snake_case ) snake_case = outputs.loss accelerator.backward(_snake_case ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(_snake_case ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): snake_case = model(**_snake_case ) snake_case = outputs.logits.argmax(dim=-1 ) snake_case = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=_snake_case , references=_snake_case , ) snake_case = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , _snake_case ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def lowerCAmelCase__ ( ) -> List[Any]: """simple docstring""" snake_case = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=_snake_case , default=_snake_case , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) snake_case = parser.parse_args() snake_case = {"lr": 2e-5, "num_epochs": 3, "seed": 4_2, "batch_size": 1_6} training_function(_snake_case , _snake_case ) if __name__ == "__main__": main()
150
def lowerCAmelCase_ ( _snake_case : list[list[int | float]] ) -> int: '''simple docstring''' __magic_name__ : Any = len(_snake_case ) __magic_name__ : Optional[Any] = len(matrix[0] ) __magic_name__ : Union[str, Any] = min(_snake_case , _snake_case ) for row in range(_snake_case ): # Check if diagonal element is not zero if matrix[row][row] != 0: # Eliminate all the elements below the diagonal for col in range(row + 1 , _snake_case ): __magic_name__ : Optional[Any] = matrix[col][row] / matrix[row][row] for i in range(_snake_case , _snake_case ): matrix[col][i] -= multiplier * matrix[row][i] else: # Find a non-zero diagonal element to swap rows __magic_name__ : str = True for i in range(row + 1 , _snake_case ): if matrix[i][row] != 0: __magic_name__ , __magic_name__ : List[str] = matrix[i], matrix[row] __magic_name__ : Union[str, Any] = False break if reduce: rank -= 1 for i in range(_snake_case ): __magic_name__ : Any = matrix[i][rank] # Reduce the row pointer by one to stay on the same row row -= 1 return rank if __name__ == "__main__": import doctest doctest.testmod()
281
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase : Optional[Any] = logging.get_logger(__name__) UpperCAmelCase : List[Any] = { "facebook/data2vec-vision-base-ft": ( "https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json" ), } class lowerCamelCase__ ( A ): """simple docstring""" __a = """data2vec-vision""" def __init__( self : Optional[int] , UpperCamelCase : List[str]=768 , UpperCamelCase : List[Any]=12 , UpperCamelCase : int=12 , UpperCamelCase : str=3_072 , UpperCamelCase : Tuple="gelu" , UpperCamelCase : str=0.0 , UpperCamelCase : List[Any]=0.0 , UpperCamelCase : Union[str, Any]=0.02 , UpperCamelCase : Optional[int]=1e-1_2 , UpperCamelCase : List[Any]=224 , UpperCamelCase : int=16 , UpperCamelCase : List[Any]=3 , UpperCamelCase : Dict=False , UpperCamelCase : Any=False , UpperCamelCase : Optional[Any]=False , UpperCamelCase : str=False , UpperCamelCase : List[str]=0.1 , UpperCamelCase : List[Any]=0.1 , UpperCamelCase : int=True , UpperCamelCase : Optional[int]=[3, 5, 7, 11] , UpperCamelCase : List[Any]=[1, 2, 3, 6] , UpperCamelCase : Union[str, Any]=True , UpperCamelCase : Dict=0.4 , UpperCamelCase : Optional[int]=256 , UpperCamelCase : str=1 , UpperCamelCase : str=False , UpperCamelCase : str=255 , **UpperCamelCase : Any , ): '''simple docstring''' super().__init__(**_a ) __UpperCAmelCase : Optional[int] = hidden_size __UpperCAmelCase : Union[str, Any] = num_hidden_layers __UpperCAmelCase : List[str] = num_attention_heads __UpperCAmelCase : int = intermediate_size __UpperCAmelCase : Tuple = hidden_act __UpperCAmelCase : str = hidden_dropout_prob __UpperCAmelCase : Union[str, Any] = attention_probs_dropout_prob __UpperCAmelCase : str = initializer_range __UpperCAmelCase : Optional[Any] = layer_norm_eps __UpperCAmelCase : Tuple = image_size __UpperCAmelCase : str = patch_size __UpperCAmelCase : Tuple = num_channels __UpperCAmelCase : Optional[Any] = use_mask_token __UpperCAmelCase : Any = use_absolute_position_embeddings __UpperCAmelCase : Union[str, Any] = use_relative_position_bias __UpperCAmelCase : str = use_shared_relative_position_bias __UpperCAmelCase : Tuple = layer_scale_init_value __UpperCAmelCase : List[str] = drop_path_rate __UpperCAmelCase : Any = use_mean_pooling # decode head attributes (semantic segmentation) __UpperCAmelCase : Tuple = out_indices __UpperCAmelCase : int = pool_scales # auxiliary head attributes (semantic segmentation) __UpperCAmelCase : str = use_auxiliary_head __UpperCAmelCase : Union[str, Any] = auxiliary_loss_weight __UpperCAmelCase : Optional[Any] = auxiliary_channels __UpperCAmelCase : Dict = auxiliary_num_convs __UpperCAmelCase : Optional[int] = auxiliary_concat_input __UpperCAmelCase : Dict = semantic_loss_ignore_index class lowerCamelCase__ ( A ): """simple docstring""" __a = version.parse("""1.11""" ) @property def lowerCamelCase__ ( self : int ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def lowerCamelCase__ ( self : List[str] ): '''simple docstring''' return 1e-4
115
import argparse import collections import json import os import re import string import sys import numpy as np snake_case : Dict = re.compile(R"\b(a|an|the)\b", re.UNICODE) snake_case : Optional[int] = None def lowerCAmelCase_ ( ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : Any = argparse.ArgumentParser("Official evaluation script for SQuAD version 2.0." ) parser.add_argument("data_file" , metavar="data.json" , help="Input data JSON file." ) parser.add_argument("pred_file" , metavar="pred.json" , help="Model predictions." ) parser.add_argument( "--out-file" , "-o" , metavar="eval.json" , help="Write accuracy metrics to file (default is stdout)." ) parser.add_argument( "--na-prob-file" , "-n" , metavar="na_prob.json" , help="Model estimates of probability of no answer." ) parser.add_argument( "--na-prob-thresh" , "-t" , type=_snake_case , default=1.0 , help="Predict \"\" if no-answer probability exceeds this (default = 1.0)." , ) parser.add_argument( "--out-image-dir" , "-p" , metavar="out_images" , default=_snake_case , help="Save precision-recall curves to directory." ) parser.add_argument("--verbose" , "-v" , action="store_true" ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def lowerCAmelCase_ ( _snake_case : Optional[Any] ) -> Tuple: '''simple docstring''' __magic_name__ : Optional[int] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: __magic_name__ : str = bool(qa["answers"]["text"] ) return qid_to_has_ans def lowerCAmelCase_ ( _snake_case : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' def remove_articles(_snake_case : List[str] ): return ARTICLES_REGEX.sub(" " , _snake_case ) def white_space_fix(_snake_case : Optional[int] ): return " ".join(text.split() ) def remove_punc(_snake_case : Optional[int] ): __magic_name__ : Dict = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_snake_case : str ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_snake_case ) ) ) ) def lowerCAmelCase_ ( _snake_case : Any ) -> Optional[Any]: '''simple docstring''' if not s: return [] return normalize_answer(_snake_case ).split() def lowerCAmelCase_ ( _snake_case : str , _snake_case : Dict ) -> Tuple: '''simple docstring''' return int(normalize_answer(_snake_case ) == normalize_answer(_snake_case ) ) def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : int ) -> str: '''simple docstring''' __magic_name__ : Any = get_tokens(_snake_case ) __magic_name__ : Optional[int] = get_tokens(_snake_case ) __magic_name__ : Tuple = collections.Counter(_snake_case ) & collections.Counter(_snake_case ) __magic_name__ : Tuple = sum(common.values() ) if len(_snake_case ) == 0 or len(_snake_case ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 __magic_name__ : Dict = 1.0 * num_same / len(_snake_case ) __magic_name__ : Optional[Any] = 1.0 * num_same / len(_snake_case ) __magic_name__ : List[Any] = (2 * precision * recall) / (precision + recall) return fa def lowerCAmelCase_ ( _snake_case : Optional[Any] , _snake_case : List[Any] ) -> List[Any]: '''simple docstring''' __magic_name__ : Union[str, Any] = {} __magic_name__ : int = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: __magic_name__ : Union[str, Any] = qa["id"] __magic_name__ : Any = [t for t in qa["answers"]["text"] if normalize_answer(_snake_case )] if not gold_answers: # For unanswerable questions, only correct answer is empty string __magic_name__ : Tuple = [""] if qid not in preds: print(F'''Missing prediction for {qid}''' ) continue __magic_name__ : Any = preds[qid] # Take max over all gold answers __magic_name__ : List[Any] = max(compute_exact(_snake_case , _snake_case ) for a in gold_answers ) __magic_name__ : int = max(compute_fa(_snake_case , _snake_case ) for a in gold_answers ) return exact_scores, fa_scores def lowerCAmelCase_ ( _snake_case : Optional[Any] , _snake_case : List[Any] , _snake_case : Optional[int] , _snake_case : Dict ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : str = {} for qid, s in scores.items(): __magic_name__ : Dict = na_probs[qid] > na_prob_thresh if pred_na: __magic_name__ : str = float(not qid_to_has_ans[qid] ) else: __magic_name__ : Optional[int] = s return new_scores def lowerCAmelCase_ ( _snake_case : List[Any] , _snake_case : List[str] , _snake_case : Tuple=None ) -> Tuple: '''simple docstring''' if not qid_list: __magic_name__ : Any = len(_snake_case ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores.values() ) / total), ("f1", 100.0 * sum(fa_scores.values() ) / total), ("total", total), ] ) else: __magic_name__ : Tuple = len(_snake_case ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores[k] for k in qid_list ) / total), ("f1", 100.0 * sum(fa_scores[k] for k in qid_list ) / total), ("total", total), ] ) def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : str , _snake_case : str ) -> Dict: '''simple docstring''' for k in new_eval: __magic_name__ : int = new_eval[k] def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Dict , _snake_case : Optional[Any] , _snake_case : Union[str, Any] ) -> str: '''simple docstring''' plt.step(_snake_case , _snake_case , color="b" , alpha=0.2 , where="post" ) plt.fill_between(_snake_case , _snake_case , step="post" , alpha=0.2 , color="b" ) plt.xlabel("Recall" ) plt.ylabel("Precision" ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(_snake_case ) plt.savefig(_snake_case ) plt.clf() def lowerCAmelCase_ ( _snake_case : Dict , _snake_case : Any , _snake_case : Optional[int] , _snake_case : List[Any] , _snake_case : Optional[int]=None , _snake_case : int=None ) -> str: '''simple docstring''' __magic_name__ : Union[str, Any] = sorted(_snake_case , key=lambda _snake_case : na_probs[k] ) __magic_name__ : Optional[int] = 0.0 __magic_name__ : str = 1.0 __magic_name__ : str = 0.0 __magic_name__ : List[str] = [1.0] __magic_name__ : str = [0.0] __magic_name__ : Optional[Any] = 0.0 for i, qid in enumerate(_snake_case ): if qid_to_has_ans[qid]: true_pos += scores[qid] __magic_name__ : List[str] = true_pos / float(i + 1 ) __magic_name__ : Any = true_pos / float(_snake_case ) if i == len(_snake_case ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(_snake_case ) recalls.append(_snake_case ) if out_image: plot_pr_curve(_snake_case , _snake_case , _snake_case , _snake_case ) return {"ap": 100.0 * avg_prec} def lowerCAmelCase_ ( _snake_case : Tuple , _snake_case : Optional[Any] , _snake_case : Optional[int] , _snake_case : Optional[Any] , _snake_case : Any , _snake_case : List[Any] ) -> Union[str, Any]: '''simple docstring''' if out_image_dir and not os.path.exists(_snake_case ): os.makedirs(_snake_case ) __magic_name__ : Any = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return __magic_name__ : str = make_precision_recall_eval( _snake_case , _snake_case , _snake_case , _snake_case , out_image=os.path.join(_snake_case , "pr_exact.png" ) , title="Precision-Recall curve for Exact Match score" , ) __magic_name__ : Union[str, Any] = make_precision_recall_eval( _snake_case , _snake_case , _snake_case , _snake_case , out_image=os.path.join(_snake_case , "pr_f1.png" ) , title="Precision-Recall curve for F1 score" , ) __magic_name__ : str = {k: float(_snake_case ) for k, v in qid_to_has_ans.items()} __magic_name__ : str = make_precision_recall_eval( _snake_case , _snake_case , _snake_case , _snake_case , out_image=os.path.join(_snake_case , "pr_oracle.png" ) , title="Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)" , ) merge_eval(_snake_case , _snake_case , "pr_exact" ) merge_eval(_snake_case , _snake_case , "pr_f1" ) merge_eval(_snake_case , _snake_case , "pr_oracle" ) def lowerCAmelCase_ ( _snake_case : int , _snake_case : Optional[Any] , _snake_case : List[str] , _snake_case : Optional[Any] ) -> Dict: '''simple docstring''' if not qid_list: return __magic_name__ : Dict = [na_probs[k] for k in qid_list] __magic_name__ : str = np.ones_like(_snake_case ) / float(len(_snake_case ) ) plt.hist(_snake_case , weights=_snake_case , bins=20 , range=(0.0, 1.0) ) plt.xlabel("Model probability of no-answer" ) plt.ylabel("Proportion of dataset" ) plt.title(F'''Histogram of no-answer probability: {name}''' ) plt.savefig(os.path.join(_snake_case , F'''na_prob_hist_{name}.png''' ) ) plt.clf() def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Tuple , _snake_case : List[str] , _snake_case : Dict ) -> List[Any]: '''simple docstring''' __magic_name__ : Union[str, Any] = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) __magic_name__ : List[str] = num_no_ans __magic_name__ : Dict = cur_score __magic_name__ : Dict = 0.0 __magic_name__ : Any = sorted(_snake_case , key=lambda _snake_case : na_probs[k] ) for i, qid in enumerate(_snake_case ): if qid not in scores: continue if qid_to_has_ans[qid]: __magic_name__ : Union[str, Any] = scores[qid] else: if preds[qid]: __magic_name__ : List[Any] = -1 else: __magic_name__ : Optional[int] = 0 cur_score += diff if cur_score > best_score: __magic_name__ : Optional[int] = cur_score __magic_name__ : List[Any] = na_probs[qid] return 100.0 * best_score / len(_snake_case ), best_thresh def lowerCAmelCase_ ( _snake_case : int , _snake_case : str , _snake_case : List[str] , _snake_case : Tuple , _snake_case : List[Any] , _snake_case : Dict ) -> Optional[Any]: '''simple docstring''' __magic_name__ , __magic_name__ : List[str] = find_best_thresh(_snake_case , _snake_case , _snake_case , _snake_case ) __magic_name__ , __magic_name__ : int = find_best_thresh(_snake_case , _snake_case , _snake_case , _snake_case ) __magic_name__ : Optional[int] = best_exact __magic_name__ : List[Any] = exact_thresh __magic_name__ : Dict = best_fa __magic_name__ : Any = fa_thresh def lowerCAmelCase_ ( ) -> int: '''simple docstring''' with open(OPTS.data_file ) as f: __magic_name__ : Optional[Any] = json.load(_snake_case ) __magic_name__ : List[Any] = dataset_json["data"] with open(OPTS.pred_file ) as f: __magic_name__ : Optional[Any] = json.load(_snake_case ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: __magic_name__ : Any = json.load(_snake_case ) else: __magic_name__ : Any = {k: 0.0 for k in preds} __magic_name__ : str = make_qid_to_has_ans(_snake_case ) # maps qid to True/False __magic_name__ : Tuple = [k for k, v in qid_to_has_ans.items() if v] __magic_name__ : Optional[Any] = [k for k, v in qid_to_has_ans.items() if not v] __magic_name__ , __magic_name__ : Union[str, Any] = get_raw_scores(_snake_case , _snake_case ) __magic_name__ : Optional[Any] = apply_no_ans_threshold(_snake_case , _snake_case , _snake_case , OPTS.na_prob_thresh ) __magic_name__ : Optional[Any] = apply_no_ans_threshold(_snake_case , _snake_case , _snake_case , OPTS.na_prob_thresh ) __magic_name__ : List[Any] = make_eval_dict(_snake_case , _snake_case ) if has_ans_qids: __magic_name__ : int = make_eval_dict(_snake_case , _snake_case , qid_list=_snake_case ) merge_eval(_snake_case , _snake_case , "HasAns" ) if no_ans_qids: __magic_name__ : List[Any] = make_eval_dict(_snake_case , _snake_case , qid_list=_snake_case ) merge_eval(_snake_case , _snake_case , "NoAns" ) if OPTS.na_prob_file: find_all_best_thresh(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case , OPTS.out_image_dir ) histogram_na_prob(_snake_case , _snake_case , OPTS.out_image_dir , "hasAns" ) histogram_na_prob(_snake_case , _snake_case , OPTS.out_image_dir , "noAns" ) if OPTS.out_file: with open(OPTS.out_file , "w" ) as f: json.dump(_snake_case , _snake_case ) else: print(json.dumps(_snake_case , indent=2 ) ) if __name__ == "__main__": snake_case : int = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("Agg") import matplotlib.pyplot as plt main()
281
0
import os import pytest from transformers.dynamic_module_utils import get_imports _a = "\nimport os\n" _a = "\ndef foo():\n import os\n return False\n" _a = "\ndef foo():\n def bar():\n if True:\n import os\n return False\n return bar()\n" _a = "\nimport os\n\ntry:\n import bar\nexcept ImportError:\n raise ValueError()\n" _a = "\nimport os\n\ndef foo():\n try:\n import bar\n except ImportError:\n raise ValueError()\n" _a = "\nimport os\n\ntry:\n import bar\nexcept (ImportError, AttributeError):\n raise ValueError()\n" _a = "\nimport os\n\ntry:\n import bar\nexcept ImportError as e:\n raise ValueError()\n" _a = "\nimport os\n\ntry:\n import bar\nexcept:\n raise ValueError()\n" _a = "\nimport os\n\ntry:\n import bar\n import baz\nexcept ImportError:\n raise ValueError()\n" _a = "\nimport os\n\ntry:\n import bar\n import baz\nexcept ImportError:\n x = 1\n raise ValueError()\n" _a = [ TOP_LEVEL_IMPORT, IMPORT_IN_FUNCTION, DEEPLY_NESTED_IMPORT, TOP_LEVEL_TRY_IMPORT, GENERIC_EXCEPT_IMPORT, MULTILINE_TRY_IMPORT, MULTILINE_BOTH_IMPORT, MULTIPLE_EXCEPTS_IMPORT, EXCEPT_AS_IMPORT, TRY_IMPORT_IN_FUNCTION, ] @pytest.mark.parametrize('case' , _snake_case ) def _a ( SCREAMING_SNAKE_CASE : Dict , SCREAMING_SNAKE_CASE : List[str] ) -> Any: """simple docstring""" __lowerCAmelCase: Optional[Any] = os.path.join(_snake_case , 'test_file.py' ) with open(_snake_case , 'w' ) as _tmp_file: _tmp_file.write(_snake_case ) __lowerCAmelCase: Tuple = get_imports(_snake_case ) assert parsed_imports == ["os"]
322
import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin snake_case : str = "▁" snake_case : List[Any] = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class _snake_case ( snake_case , unittest.TestCase ): UpperCamelCase__ = BigBirdTokenizer UpperCamelCase__ = BigBirdTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True def SCREAMING_SNAKE_CASE ( self ): super().setUp() __magic_name__ : Optional[Any] = self.tokenizer_class(_a , keep_accents=_a ) tokenizer.save_pretrained(self.tmpdirname ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Union[str, Any] = "<s>" __magic_name__ : Dict = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_a ) , _a ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_a ) , _a ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "[MASK]" ) self.assertEqual(len(_a ) , 1_004 ) def SCREAMING_SNAKE_CASE ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def SCREAMING_SNAKE_CASE ( self ): if not self.test_rust_tokenizer: return __magic_name__ : Dict = self.get_tokenizer() __magic_name__ : str = self.get_rust_tokenizer() __magic_name__ : Any = "I was born in 92000, and this is falsé." __magic_name__ : Dict = tokenizer.tokenize(_a ) __magic_name__ : Any = rust_tokenizer.tokenize(_a ) self.assertListEqual(_a , _a ) __magic_name__ : List[Any] = tokenizer.encode(_a , add_special_tokens=_a ) __magic_name__ : List[str] = rust_tokenizer.encode(_a , add_special_tokens=_a ) self.assertListEqual(_a , _a ) __magic_name__ : str = self.get_rust_tokenizer() __magic_name__ : Dict = tokenizer.encode(_a ) __magic_name__ : Optional[int] = rust_tokenizer.encode(_a ) self.assertListEqual(_a , _a ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = BigBirdTokenizer(_a , keep_accents=_a ) __magic_name__ : str = tokenizer.tokenize("This is a test" ) self.assertListEqual(_a , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_a ) , [285, 46, 10, 170, 382] , ) __magic_name__ : Dict = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( _a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) __magic_name__ : Union[str, Any] = tokenizer.convert_tokens_to_ids(_a ) self.assertListEqual( _a , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) __magic_name__ : int = tokenizer.convert_ids_to_tokens(_a ) self.assertListEqual( _a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def SCREAMING_SNAKE_CASE ( self ): return BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Any = "Hello World!" __magic_name__ : Dict = [65, 18_536, 2_260, 101, 66] self.assertListEqual(_a , self.big_tokenizer.encode(_a ) ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) # fmt: off __magic_name__ : List[str] = [65, 871, 419, 358, 946, 991, 2_521, 452, 358, 1_357, 387, 7_751, 3_536, 112, 985, 456, 126, 865, 938, 5_400, 5_734, 458, 1_368, 467, 786, 2_462, 5_246, 1_159, 633, 865, 4_519, 457, 582, 852, 2_557, 427, 916, 508, 405, 34_324, 497, 391, 408, 11_342, 1_244, 385, 100, 938, 985, 456, 574, 362, 12_597, 3_200, 3_129, 1_172, 66] # noqa: E231 # fmt: on self.assertListEqual(_a , self.big_tokenizer.encode(_a ) ) @require_torch @slow def SCREAMING_SNAKE_CASE ( self ): import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence __magic_name__ : Optional[Any] = list(self.big_tokenizer.get_vocab().keys() )[:10] __magic_name__ : List[Any] = " ".join(_a ) __magic_name__ : Any = self.big_tokenizer.encode_plus(_a , return_tensors="pt" , return_token_type_ids=_a ) __magic_name__ : Union[str, Any] = self.big_tokenizer.batch_encode_plus( [sequence + " " + sequence] , return_tensors="pt" , return_token_type_ids=_a ) __magic_name__ : List[str] = BigBirdConfig(attention_type="original_full" ) __magic_name__ : Optional[int] = BigBirdModel(_a ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_a ) model(**_a ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : int = BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) __magic_name__ : int = tokenizer.decode(tokenizer("Paris is the [MASK]." ).input_ids ) self.assertTrue(decoded_text == "[CLS] Paris is the[MASK].[SEP]" ) @slow def SCREAMING_SNAKE_CASE ( self ): # fmt: off __magic_name__ : Optional[Any] = {"input_ids": [[65, 39_286, 458, 36_335, 2_001, 456, 13_073, 13_266, 455, 113, 7_746, 1_741, 11_157, 391, 13_073, 13_266, 455, 113, 3_967, 35_412, 113, 4_936, 109, 3_870, 2_377, 113, 30_084, 45_720, 458, 134, 17_496, 112, 503, 11_672, 113, 118, 112, 5_665, 13_347, 38_687, 112, 1_496, 31_389, 112, 3_268, 47_264, 134, 962, 112, 16_377, 8_035, 23_130, 430, 12_169, 15_518, 28_592, 458, 146, 41_697, 109, 391, 12_169, 15_518, 16_689, 458, 146, 41_358, 109, 452, 726, 4_034, 111, 763, 35_412, 5_082, 388, 1_903, 111, 9_051, 391, 2_870, 48_918, 1_900, 1_123, 550, 998, 112, 9_586, 15_985, 455, 391, 410, 22_955, 37_636, 114, 66], [65, 448, 17_496, 419, 3_663, 385, 763, 113, 27_533, 2_870, 3_283, 13_043, 1_639, 24_713, 523, 656, 24_013, 18_550, 2_521, 517, 27_014, 21_244, 420, 1_212, 1_465, 391, 927, 4_833, 388, 578, 11_786, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 484, 2_169, 7_687, 21_932, 18_146, 726, 363, 17_032, 3_391, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_a , model_name="google/bigbird-roberta-base" , revision="215c99f1600e06f83acce68422f2035b2b5c3510" , )
281
0
import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class _snake_case ( tf.keras.layers.Layer ): def __init__( self: str , __lowerCamelCase: Optional[int] , __lowerCamelCase: int , __lowerCamelCase: Any = None , __lowerCamelCase: Dict = None ) -> List[Any]: super().__init__() __UpperCAmelCase : Tuple = pad_token_id __UpperCAmelCase : Union[str, Any] = max_length __UpperCAmelCase : Dict = vocab __UpperCAmelCase : List[Any] = merges __UpperCAmelCase : Any = BytePairTokenizer(_a , _a , sequence_length=_a ) @classmethod def _lowerCamelCase ( cls: List[Any] , __lowerCamelCase: int , *__lowerCamelCase: Tuple , **__lowerCamelCase: int ) -> List[str]: __UpperCAmelCase : Optional[int] = [" ".join(_a ) for m in tokenizer.bpe_ranks.keys()] __UpperCAmelCase : List[str] = tokenizer.get_vocab() return cls(_a , _a , *_a , **_a ) @classmethod def _lowerCamelCase ( cls: Union[str, Any] , __lowerCamelCase: Dict , *__lowerCamelCase: Optional[Any] , **__lowerCamelCase: Any ) -> Tuple: __UpperCAmelCase : Any = GPTaTokenizer.from_pretrained(_a , *_a , **_a ) return cls.from_tokenizer(_a , *_a , **_a ) @classmethod def _lowerCamelCase ( cls: Dict , __lowerCamelCase: Any ) -> List[str]: return cls(**_a ) def _lowerCamelCase ( self: Tuple ) -> Optional[Any]: return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def _lowerCamelCase ( self: Optional[int] , __lowerCamelCase: Optional[int] , __lowerCamelCase: Optional[int] = None ) -> str: __UpperCAmelCase : int = self.tf_tokenizer(_a ) __UpperCAmelCase : List[str] = tf.ones_like(_a ) if self.pad_token_id is not None: # pad the tokens up to max length __UpperCAmelCase : Any = max_length if max_length is not None else self.max_length if max_length is not None: __UpperCAmelCase : List[str] = pad_model_inputs( _a , max_seq_length=_a , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
157
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging snake_case : int = logging.get_logger(__name__) snake_case : List[str] = {"vocab_file": "spiece.model"} snake_case : List[str] = { "vocab_file": { "albert-base-v1": "https://huggingface.co/albert-base-v1/resolve/main/spiece.model", "albert-large-v1": "https://huggingface.co/albert-large-v1/resolve/main/spiece.model", "albert-xlarge-v1": "https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model", "albert-xxlarge-v1": "https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model", "albert-base-v2": "https://huggingface.co/albert-base-v2/resolve/main/spiece.model", "albert-large-v2": "https://huggingface.co/albert-large-v2/resolve/main/spiece.model", "albert-xlarge-v2": "https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model", "albert-xxlarge-v2": "https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model", } } snake_case : Tuple = { "albert-base-v1": 512, "albert-large-v1": 512, "albert-xlarge-v1": 512, "albert-xxlarge-v1": 512, "albert-base-v2": 512, "albert-large-v2": 512, "albert-xlarge-v2": 512, "albert-xxlarge-v2": 512, } snake_case : List[str] = "▁" class _snake_case ( snake_case ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , _a , _a=True , _a=True , _a=False , _a="[CLS]" , _a="[SEP]" , _a="<unk>" , _a="[SEP]" , _a="<pad>" , _a="[CLS]" , _a="[MASK]" , _a = None , **_a , ): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. __magic_name__ : str = ( AddedToken(_a , lstrip=_a , rstrip=_a , normalized=_a ) if isinstance(_a , _a ) else mask_token ) __magic_name__ : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=_a , remove_space=_a , keep_accents=_a , bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , pad_token=_a , cls_token=_a , mask_token=_a , sp_model_kwargs=self.sp_model_kwargs , **_a , ) __magic_name__ : Dict = do_lower_case __magic_name__ : Tuple = remove_space __magic_name__ : Union[str, Any] = keep_accents __magic_name__ : Tuple = vocab_file __magic_name__ : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_a ) @property def SCREAMING_SNAKE_CASE ( self ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[str] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): __magic_name__ : List[str] = self.__dict__.copy() __magic_name__ : Any = None return state def __setstate__( self , _a ): __magic_name__ : Union[str, Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): __magic_name__ : str = {} __magic_name__ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self , _a ): if self.remove_space: __magic_name__ : List[Any] = " ".join(inputs.strip().split() ) else: __magic_name__ : str = inputs __magic_name__ : int = outputs.replace("``" , "\"" ).replace("''" , "\"" ) if not self.keep_accents: __magic_name__ : str = unicodedata.normalize("NFKD" , _a ) __magic_name__ : Tuple = "".join([c for c in outputs if not unicodedata.combining(_a )] ) if self.do_lower_case: __magic_name__ : int = outputs.lower() return outputs def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Optional[Any] = self.preprocess_text(_a ) __magic_name__ : Dict = self.sp_model.encode(_a , out_type=_a ) __magic_name__ : Any = [] for piece in pieces: if len(_a ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): __magic_name__ : Optional[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(_a , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __magic_name__ : List[str] = cur_pieces[1:] else: __magic_name__ : Optional[int] = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(_a ) else: new_pieces.append(_a ) return new_pieces def SCREAMING_SNAKE_CASE ( self , _a ): return self.sp_model.PieceToId(_a ) def SCREAMING_SNAKE_CASE ( self , _a ): return self.sp_model.IdToPiece(_a ) def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Any = [] __magic_name__ : Union[str, Any] = "" __magic_name__ : int = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(_a ) + token __magic_name__ : List[Any] = True __magic_name__ : Optional[int] = [] else: current_sub_tokens.append(_a ) __magic_name__ : Optional[Any] = False out_string += self.sp_model.decode(_a ) return out_string.strip() def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): __magic_name__ : List[str] = [self.sep_token_id] __magic_name__ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self , _a , _a = None , _a = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a , token_ids_a=_a , already_has_special_tokens=_a ) if token_ids_a is not None: return [1] + ([0] * len(_a )) + [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1] def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): __magic_name__ : Optional[int] = [self.sep_token_id] __magic_name__ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): if not os.path.isdir(_a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __magic_name__ : List[str] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _a ) elif not os.path.isfile(self.vocab_file ): with open(_a , "wb" ) as fi: __magic_name__ : Union[str, Any] = self.sp_model.serialized_model_proto() fi.write(_a ) return (out_vocab_file,)
281
0
"""simple docstring""" lowerCamelCase__ = "Input must be a string of 8 numbers plus letter" lowerCamelCase__ = "TRWAGMYFPDXBNJZSQVHLCKE" def __lowerCAmelCase (_UpperCamelCase ): if not isinstance(_snake_case , _snake_case ): __lowerCAmelCase : List[str] = F"Expected string as input, found {type(_snake_case ).__name__}" raise TypeError(_snake_case ) __lowerCAmelCase : int = spanish_id.replace('-' , '' ).upper() if len(_snake_case ) != 9: raise ValueError(_snake_case ) try: __lowerCAmelCase : Optional[int] = int(spanish_id_clean[0:8] ) __lowerCAmelCase : Optional[int] = spanish_id_clean[8] except ValueError as ex: raise ValueError(_snake_case ) from ex if letter.isdigit(): raise ValueError(_snake_case ) return letter == LOOKUP_LETTERS[number % 23] if __name__ == "__main__": import doctest doctest.testmod()
86
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : Tuple , _snake_case : List[Any] , _snake_case : Optional[Any] ) -> Optional[Any]: '''simple docstring''' if isinstance(_snake_case , _snake_case ): __magic_name__ : Union[str, Any] = np.full((len(_snake_case ), sequence_length, 2) , _snake_case ) else: __magic_name__ : List[Any] = np.full((len(_snake_case ), sequence_length) , _snake_case ) for i, tensor in enumerate(_snake_case ): if padding_side == "right": if isinstance(_snake_case , _snake_case ): __magic_name__ : Optional[Any] = tensor[:sequence_length] else: __magic_name__ : Union[str, Any] = tensor[:sequence_length] else: if isinstance(_snake_case , _snake_case ): __magic_name__ : List[Any] = tensor[:sequence_length] else: __magic_name__ : Optional[Any] = tensor[:sequence_length] return out_tensor.tolist() def lowerCAmelCase_ ( _snake_case : Optional[int] ) -> Tuple: '''simple docstring''' __magic_name__ : Union[str, Any] = ord(_snake_case ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 123 and cp <= 126): return True __magic_name__ : Any = unicodedata.category(_snake_case ) if cat.startswith("P" ): return True return False @dataclass class _snake_case ( snake_case ): UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = -100 UpperCamelCase__ = "pt" def SCREAMING_SNAKE_CASE ( self , _a ): import torch __magic_name__ : List[str] = "label" if "label" in features[0].keys() else "labels" __magic_name__ : Union[str, Any] = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __magic_name__ : Optional[int] = self.tokenizer.pad( _a , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="pt" if labels is None else None , ) if labels is None: return batch __magic_name__ : Dict = torch.tensor(batch["entity_ids"] ).shape[1] __magic_name__ : List[Any] = self.tokenizer.padding_side if padding_side == "right": __magic_name__ : str = [ list(_a ) + [self.label_pad_token_id] * (sequence_length - len(_a )) for label in labels ] else: __magic_name__ : int = [ [self.label_pad_token_id] * (sequence_length - len(_a )) + list(_a ) for label in labels ] __magic_name__ : Dict = [feature["ner_tags"] for feature in features] __magic_name__ : List[Any] = padding_tensor(_a , -1 , _a , _a ) __magic_name__ : Any = [feature["original_entity_spans"] for feature in features] __magic_name__ : Any = padding_tensor(_a , (-1, -1) , _a , _a ) __magic_name__ : List[Any] = {k: torch.tensor(_a , dtype=torch.intaa ) for k, v in batch.items()} return batch
281
0
import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home __lowerCamelCase : str = HUGGINGFACE_HUB_CACHE __lowerCamelCase : Tuple = "config.json" __lowerCamelCase : int = "diffusion_pytorch_model.bin" __lowerCamelCase : Dict = "diffusion_flax_model.msgpack" __lowerCamelCase : List[str] = "model.onnx" __lowerCamelCase : str = "diffusion_pytorch_model.safetensors" __lowerCamelCase : Tuple = "weights.pb" __lowerCamelCase : Optional[Any] = "https://huggingface.co" __lowerCamelCase : Dict = default_cache_path __lowerCamelCase : Any = "diffusers_modules" __lowerCamelCase : Optional[Any] = os.getenv('''HF_MODULES_CACHE''', os.path.join(hf_cache_home, '''modules''')) __lowerCamelCase : Any = ["fp16", "non-ema"] __lowerCamelCase : Any = ".self_attn"
219
import math def lowerCAmelCase_ ( _snake_case : float , _snake_case : float ) -> float: '''simple docstring''' return math.pow(_snake_case , 2 ) - a def lowerCAmelCase_ ( _snake_case : float ) -> float: '''simple docstring''' return 2 * x def lowerCAmelCase_ ( _snake_case : float ) -> float: '''simple docstring''' __magic_name__ : Optional[int] = 2.0 while start <= a: __magic_name__ : str = math.pow(_snake_case , 2 ) return start def lowerCAmelCase_ ( _snake_case : float , _snake_case : int = 9999 , _snake_case : float = 0.00_000_000_000_001 ) -> float: '''simple docstring''' if a < 0: raise ValueError("math domain error" ) __magic_name__ : Optional[int] = get_initial_point(_snake_case ) for _ in range(_snake_case ): __magic_name__ : int = value __magic_name__ : str = value - fx(_snake_case , _snake_case ) / fx_derivative(_snake_case ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
281
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import DistilBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.distilbert.modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertModel, ) class lowercase : """simple docstring""" def __init__( self , UpperCamelCase_ , ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = parent UpperCamelCase__ :Optional[Any] = 13 UpperCamelCase__ :List[Any] = 7 UpperCamelCase__ :int = True UpperCamelCase__ :int = True UpperCamelCase__ :Dict = False UpperCamelCase__ :int = True UpperCamelCase__ :List[str] = 99 UpperCamelCase__ :Tuple = 32 UpperCamelCase__ :Tuple = 2 UpperCamelCase__ :str = 4 UpperCamelCase__ :Union[str, Any] = 37 UpperCamelCase__ :Union[str, Any] = "gelu" UpperCamelCase__ :int = 0.1 UpperCamelCase__ :Optional[Any] = 0.1 UpperCamelCase__ :List[Any] = 512 UpperCamelCase__ :int = 16 UpperCamelCase__ :Optional[Any] = 2 UpperCamelCase__ :Any = 0.02 UpperCamelCase__ :Union[str, Any] = 3 UpperCamelCase__ :Tuple = 4 UpperCamelCase__ :Tuple = None def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase__ :str = None if self.use_input_mask: UpperCamelCase__ :Optional[int] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase__ :Tuple = None UpperCamelCase__ :int = None UpperCamelCase__ :Dict = None if self.use_labels: UpperCamelCase__ :Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase__ :Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase__ :int = ids_tensor([self.batch_size] , self.num_choices ) UpperCamelCase__ :Optional[Any] = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = TFDistilBertModel(config=_a ) UpperCamelCase__ :Dict = {"input_ids": input_ids, "attention_mask": input_mask} UpperCamelCase__ :str = model(_a ) UpperCamelCase__ :List[str] = [input_ids, input_mask] UpperCamelCase__ :Dict = model(_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :List[str] = TFDistilBertForMaskedLM(config=_a ) UpperCamelCase__ :Any = {"input_ids": input_ids, "attention_mask": input_mask} UpperCamelCase__ :List[str] = model(_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :str = TFDistilBertForQuestionAnswering(config=_a ) UpperCamelCase__ :Optional[Any] = { "input_ids": input_ids, "attention_mask": input_mask, } UpperCamelCase__ :str = model(_a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :List[str] = self.num_labels UpperCamelCase__ :List[Any] = TFDistilBertForSequenceClassification(_a ) UpperCamelCase__ :Optional[Any] = {"input_ids": input_ids, "attention_mask": input_mask} UpperCamelCase__ :int = model(_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :List[Any] = self.num_choices UpperCamelCase__ :int = TFDistilBertForMultipleChoice(_a ) UpperCamelCase__ :int = tf.tile(tf.expand_dims(_a , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase__ :Optional[int] = tf.tile(tf.expand_dims(_a , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase__ :Optional[int] = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, } UpperCamelCase__ :Optional[Any] = model(_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :Dict = self.num_labels UpperCamelCase__ :List[str] = TFDistilBertForTokenClassification(_a ) UpperCamelCase__ :int = {"input_ids": input_ids, "attention_mask": input_mask} UpperCamelCase__ :List[str] = model(_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :List[str] = self.prepare_config_and_inputs() (UpperCamelCase__) :int = config_and_inputs UpperCamelCase__ :str = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class lowercase ( A__ , A__ , unittest.TestCase ): """simple docstring""" _a = ( ( TFDistilBertModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertForMultipleChoice, ) if is_tf_available() else None ) _a = ( { 'feature-extraction': TFDistilBertModel, 'fill-mask': TFDistilBertForMaskedLM, 'question-answering': TFDistilBertForQuestionAnswering, 'text-classification': TFDistilBertForSequenceClassification, 'token-classification': TFDistilBertForTokenClassification, 'zero-shot': TFDistilBertForSequenceClassification, } if is_tf_available() else {} ) _a = False _a = False def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :List[str] = TFDistilBertModelTester(self ) UpperCamelCase__ :List[str] = ConfigTester(self , config_class=_a , dim=37 ) def lowerCAmelCase__ ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_a ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_a ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_a ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_a ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_a ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_a ) @slow def lowerCAmelCase__ ( self ): '''simple docstring''' for model_name in list(TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1] ): UpperCamelCase__ :int = TFDistilBertModel.from_pretrained(_a ) self.assertIsNotNone(_a ) @require_tf class lowercase ( unittest.TestCase ): """simple docstring""" @slow def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = TFDistilBertModel.from_pretrained('''distilbert-base-uncased''' ) UpperCamelCase__ :Union[str, Any] = tf.constant([[0, 1, 2, 3, 4, 5]] ) UpperCamelCase__ :List[Any] = model(_a )[0] UpperCamelCase__ :Optional[Any] = [1, 6, 768] self.assertEqual(output.shape , _a ) UpperCamelCase__ :List[str] = tf.constant( [ [ [0.19261885, -0.13732955, 0.4119799], [0.22150156, -0.07422661, 0.39037204], [0.22756018, -0.0896414, 0.3701467], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , _a , atol=1e-4 )
97
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class _snake_case : UpperCamelCase__ = LEDConfig UpperCamelCase__ = {} UpperCamelCase__ = 'gelu' def __init__( self , _a , _a=13 , _a=7 , _a=True , _a=False , _a=99 , _a=32 , _a=2 , _a=4 , _a=37 , _a=0.1 , _a=0.1 , _a=20 , _a=2 , _a=1 , _a=0 , _a=4 , ): __magic_name__ : int = parent __magic_name__ : Optional[int] = batch_size __magic_name__ : Tuple = seq_length __magic_name__ : List[Any] = is_training __magic_name__ : Dict = use_labels __magic_name__ : Optional[Any] = vocab_size __magic_name__ : int = hidden_size __magic_name__ : Optional[int] = num_hidden_layers __magic_name__ : Optional[int] = num_attention_heads __magic_name__ : Tuple = intermediate_size __magic_name__ : Any = hidden_dropout_prob __magic_name__ : Optional[int] = attention_probs_dropout_prob __magic_name__ : List[str] = max_position_embeddings __magic_name__ : Any = eos_token_id __magic_name__ : str = pad_token_id __magic_name__ : int = bos_token_id __magic_name__ : Optional[int] = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after __magic_name__ : Tuple = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests __magic_name__ : Tuple = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __magic_name__ : int = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __magic_name__ : int = tf.concat([input_ids, eos_tensor] , axis=1 ) __magic_name__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __magic_name__ : Dict = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) __magic_name__ : List[str] = prepare_led_inputs_dict(_a , _a , _a ) __magic_name__ : Union[str, Any] = tf.concat( [tf.zeros_like(_a )[:, :-1], tf.ones_like(_a )[:, -1:]] , axis=-1 , ) __magic_name__ : List[Any] = global_attention_mask return config, inputs_dict def SCREAMING_SNAKE_CASE ( self , _a , _a ): __magic_name__ : Dict = TFLEDModel(config=_a ).get_decoder() __magic_name__ : Optional[int] = inputs_dict["input_ids"] __magic_name__ : Union[str, Any] = input_ids[:1, :] __magic_name__ : str = inputs_dict["attention_mask"][:1, :] __magic_name__ : int = 1 # first forward pass __magic_name__ : Tuple = model(_a , attention_mask=_a , use_cache=_a ) __magic_name__ , __magic_name__ : str = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __magic_name__ : Optional[int] = ids_tensor((self.batch_size, 3) , config.vocab_size ) __magic_name__ : Any = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __magic_name__ : Optional[Any] = tf.concat([input_ids, next_tokens] , axis=-1 ) __magic_name__ : List[Any] = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __magic_name__ : List[str] = model(_a , attention_mask=_a )[0] __magic_name__ : Dict = model(_a , attention_mask=_a , past_key_values=_a )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __magic_name__ : List[Any] = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __magic_name__ : Union[str, Any] = output_from_no_past[:, -3:, random_slice_idx] __magic_name__ : List[str] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_a , _a , rtol=1e-3 ) def lowerCAmelCase_ ( _snake_case : Any , _snake_case : List[Any] , _snake_case : Any , _snake_case : str=None , _snake_case : List[str]=None , _snake_case : int=None , _snake_case : Any=None , ) -> int: '''simple docstring''' if attention_mask is None: __magic_name__ : str = tf.cast(tf.math.not_equal(_snake_case , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: __magic_name__ : List[Any] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: __magic_name__ : Dict = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: __magic_name__ : Any = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class _snake_case ( snake_case , snake_case , unittest.TestCase ): UpperCamelCase__ = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () UpperCamelCase__ = (TFLEDForConditionalGeneration,) if is_tf_available() else () UpperCamelCase__ = ( { 'conversational': TFLEDForConditionalGeneration, 'feature-extraction': TFLEDModel, 'summarization': TFLEDForConditionalGeneration, 'text2text-generation': TFLEDForConditionalGeneration, 'translation': TFLEDForConditionalGeneration, } if is_tf_available() else {} ) UpperCamelCase__ = True UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = TFLEDModelTester(self ) __magic_name__ : List[Any] = ConfigTester(self , config_class=_a ) def SCREAMING_SNAKE_CASE ( self ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_a ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ , __magic_name__ : int = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ : List[str] = tf.zeros_like(inputs_dict["attention_mask"] ) __magic_name__ : Optional[Any] = 2 __magic_name__ : Tuple = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict["global_attention_mask"] , ) __magic_name__ : Any = True __magic_name__ : str = self.model_tester.seq_length __magic_name__ : Dict = self.model_tester.encoder_seq_length def check_decoder_attentions_output(_a ): __magic_name__ : str = outputs.decoder_attentions self.assertEqual(len(_a ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(_a ): __magic_name__ : Any = [t.numpy() for t in outputs.encoder_attentions] __magic_name__ : Tuple = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(_a ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(_a ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: __magic_name__ : Union[str, Any] = True __magic_name__ : List[str] = False __magic_name__ : Tuple = False __magic_name__ : Optional[int] = model_class(_a ) __magic_name__ : str = model(self._prepare_for_class(_a , _a ) ) __magic_name__ : Any = len(_a ) self.assertEqual(config.output_hidden_states , _a ) check_encoder_attentions_output(_a ) if self.is_encoder_decoder: __magic_name__ : Tuple = model_class(_a ) __magic_name__ : Optional[Any] = model(self._prepare_for_class(_a , _a ) ) self.assertEqual(config.output_hidden_states , _a ) check_decoder_attentions_output(_a ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] __magic_name__ : Dict = True __magic_name__ : str = model_class(_a ) __magic_name__ : Any = model(self._prepare_for_class(_a , _a ) ) self.assertEqual(config.output_hidden_states , _a ) check_encoder_attentions_output(_a ) # Check attention is always last and order is fine __magic_name__ : Union[str, Any] = True __magic_name__ : Union[str, Any] = True __magic_name__ : List[str] = model_class(_a ) __magic_name__ : Any = model(self._prepare_for_class(_a , _a ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(_a ) ) self.assertEqual(model.config.output_hidden_states , _a ) check_encoder_attentions_output(_a ) @unittest.skip("LED keeps using potentially symbolic tensors in conditionals and breaks tracing." ) def SCREAMING_SNAKE_CASE ( self ): pass def SCREAMING_SNAKE_CASE ( self ): # TODO: Head-masking not yet implement pass def lowerCAmelCase_ ( _snake_case : int ) -> Optional[int]: '''simple docstring''' return tf.constant(_snake_case , dtype=tf.intaa ) snake_case : Optional[int] = 1E-4 @slow @require_tf class _snake_case ( unittest.TestCase ): def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[Any] = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ).led # change to intended input here __magic_name__ : Optional[int] = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) __magic_name__ : str = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) __magic_name__ : Any = prepare_led_inputs_dict(model.config , _a , _a ) __magic_name__ : List[Any] = model(**_a )[0] __magic_name__ : List[str] = (1, 1_024, 768) self.assertEqual(output.shape , _a ) # change to expected output here __magic_name__ : int = tf.convert_to_tensor( [[2.30_50, 2.82_79, 0.65_31], [-1.84_57, -0.14_55, -3.56_61], [-1.01_86, 0.45_86, -2.20_43]] , ) tf.debugging.assert_near(output[:, :3, :3] , _a , atol=1e-3 ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Tuple = TFLEDForConditionalGeneration.from_pretrained("allenai/led-base-16384" ) # change to intended input here __magic_name__ : int = _long_tensor([512 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) __magic_name__ : Tuple = _long_tensor([128 * [0, 31_414, 232, 328, 740, 1_140, 12_695, 69]] ) __magic_name__ : Optional[Any] = prepare_led_inputs_dict(model.config , _a , _a ) __magic_name__ : Union[str, Any] = model(**_a )[0] __magic_name__ : Optional[int] = (1, 1_024, model.config.vocab_size) self.assertEqual(output.shape , _a ) # change to expected output here __magic_name__ : str = tf.convert_to_tensor( [[33.65_07, 6.45_72, 16.80_89], [5.87_39, -2.42_38, 11.29_02], [-3.21_39, -4.31_49, 4.27_83]] , ) tf.debugging.assert_near(output[:, :3, :3] , _a , atol=1e-3 , rtol=1e-3 )
281
0
"""simple docstring""" # flake8: noqa # Lint as: python3 A: Optional[Any] = [ "VerificationMode", "Version", "disable_progress_bar", "enable_progress_bar", "is_progress_bar_enabled", "experimental", ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
109
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() snake_case : Optional[Any] = logging.get_logger(__name__) def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Union[str, Any]=False ) -> List[str]: '''simple docstring''' __magic_name__ : Union[str, Any] = [] # fmt: off # stem: rename_keys.append(("cls_token", "vit.embeddings.cls_token") ) rename_keys.append(("pos_embed", "vit.embeddings.position_embeddings") ) rename_keys.append(("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight") ) rename_keys.append(("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias") ) # backbone rename_keys.append(("patch_embed.backbone.stem.conv.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.bias", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias") ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight''') ) rename_keys.append((F'''patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias''', F'''vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias''') ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''vit.encoder.layer.{i}.output.dense.bias''') ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" __magic_name__ : int = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) # fmt: on return rename_keys def lowerCAmelCase_ ( _snake_case : Any , _snake_case : Any , _snake_case : Dict=False ) -> int: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: __magic_name__ : int = "" else: __magic_name__ : Union[str, Any] = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __magic_name__ : Optional[Any] = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) __magic_name__ : int = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict __magic_name__ : Dict = in_proj_weight[ : config.hidden_size, : ] __magic_name__ : List[str] = in_proj_bias[: config.hidden_size] __magic_name__ : List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __magic_name__ : Optional[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __magic_name__ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] __magic_name__ : int = in_proj_bias[-config.hidden_size :] def lowerCAmelCase_ ( _snake_case : List[str] ) -> List[str]: '''simple docstring''' __magic_name__ : List[str] = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(_snake_case , _snake_case ) def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : int , _snake_case : Union[str, Any] ) -> Optional[int]: '''simple docstring''' __magic_name__ : int = dct.pop(_snake_case ) __magic_name__ : List[Any] = val def lowerCAmelCase_ ( ) -> Dict: '''simple docstring''' __magic_name__ : List[str] = "http://images.cocodataset.org/val2017/000000039769.jpg" __magic_name__ : List[str] = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return im @torch.no_grad() def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : Any , _snake_case : int=False ) -> Dict: '''simple docstring''' __magic_name__ : List[str] = BitConfig( global_padding="same" , layer_type="bottleneck" , depths=(3, 4, 9) , out_features=["stage3"] , embedding_dynamic_padding=_snake_case , ) __magic_name__ : List[str] = ViTHybridConfig(backbone_config=_snake_case , image_size=384 , num_labels=1000 ) __magic_name__ : str = False # load original model from timm __magic_name__ : Union[str, Any] = timm.create_model(_snake_case , pretrained=_snake_case ) timm_model.eval() # load state_dict of original model, remove and rename some keys __magic_name__ : List[Any] = timm_model.state_dict() if base_model: remove_classification_head_(_snake_case ) __magic_name__ : Tuple = create_rename_keys(_snake_case , _snake_case ) for src, dest in rename_keys: rename_key(_snake_case , _snake_case , _snake_case ) read_in_q_k_v(_snake_case , _snake_case , _snake_case ) __magic_name__ : List[str] = "huggingface/label-files" __magic_name__ : int = "imagenet-1k-id2label.json" __magic_name__ : Optional[int] = json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type="dataset" ) , "r" ) ) __magic_name__ : int = {int(_snake_case ): v for k, v in idalabel.items()} __magic_name__ : List[str] = idalabel __magic_name__ : List[str] = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": __magic_name__ : List[str] = ViTHybridModel(_snake_case ).eval() else: __magic_name__ : str = ViTHybridForImageClassification(_snake_case ).eval() model.load_state_dict(_snake_case ) # create image processor __magic_name__ : List[Any] = create_transform(**resolve_data_config({} , model=_snake_case ) ) __magic_name__ : int = transform.transforms __magic_name__ : List[str] = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } __magic_name__ : int = ViTHybridImageProcessor( do_resize=_snake_case , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_snake_case , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=_snake_case , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) __magic_name__ : List[Any] = prepare_img() __magic_name__ : Any = transform(_snake_case ).unsqueeze(0 ) __magic_name__ : Tuple = processor(_snake_case , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(_snake_case , _snake_case ) # verify logits with torch.no_grad(): __magic_name__ : Optional[int] = model(_snake_case ) __magic_name__ : List[str] = outputs.logits print("Predicted class:" , logits.argmax(-1 ).item() ) if base_model: __magic_name__ : List[str] = timm_model.forward_features(_snake_case ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_snake_case , outputs.pooler_output , atol=1E-3 ) else: __magic_name__ : Any = timm_model(_snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_snake_case , outputs.logits , atol=1E-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(_snake_case ).mkdir(exist_ok=_snake_case ) print(F'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(_snake_case ) print(F'''Saving processor to {pytorch_dump_folder_path}''' ) processor.save_pretrained(_snake_case ) if push_to_hub: print(F'''Pushing model and processor to the hub {vit_name}''' ) model.push_to_hub(F'''ybelkada/{vit_name}''' ) processor.push_to_hub(F'''ybelkada/{vit_name}''' ) if __name__ == "__main__": snake_case : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--vit_name", default="vit_base_r50_s16_384", type=str, help="Name of the hybrid ViT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to upload the model to the HuggingFace hub." ) snake_case : List[Any] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
281
0
import argparse import math import traceback import dateutil.parser as date_parser import requests def __UpperCamelCase ( _A ): lowerCAmelCase_ = {} lowerCAmelCase_ = job["started_at"] lowerCAmelCase_ = job["completed_at"] lowerCAmelCase_ = date_parser.parse(_snake_case ) lowerCAmelCase_ = date_parser.parse(_snake_case ) lowerCAmelCase_ = round((end_datetime - start_datetime).total_seconds() / 6_0.0 ) lowerCAmelCase_ = start lowerCAmelCase_ = end lowerCAmelCase_ = duration_in_min return job_info def __UpperCamelCase ( _A , _A=None ): lowerCAmelCase_ = None if token is not None: lowerCAmelCase_ = {"Accept": "application/vnd.github+json", "Authorization": f"Bearer {token}"} lowerCAmelCase_ = f"https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100" lowerCAmelCase_ = requests.get(_snake_case , headers=_snake_case ).json() lowerCAmelCase_ = {} try: job_time.update({job['''name''']: extract_time_from_single_job(_snake_case ) for job in result['''jobs''']} ) lowerCAmelCase_ = math.ceil((result['''total_count'''] - 100) / 100 ) for i in range(_snake_case ): lowerCAmelCase_ = requests.get(url + f"&page={i + 2}" , headers=_snake_case ).json() job_time.update({job['''name''']: extract_time_from_single_job(_snake_case ) for job in result['''jobs''']} ) return job_time except Exception: print(f"Unknown error, could not fetch links:\n{traceback.format_exc()}" ) return {} if __name__ == "__main__": _A = argparse.ArgumentParser() # Required parameters parser.add_argument('''--workflow_run_id''', type=str, required=True, help='''A GitHub Actions workflow run id.''') _A = parser.parse_args() _A = get_job_time(args.workflow_run_id) _A = dict(sorted(job_time.items(), key=lambda item: item[1]["duration"], reverse=True)) for k, v in job_time.items(): print(f"{k}: {v['duration']}")
278
# This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny model through reduction of a normal pre-trained model, but keeping the # full vocab, merges file, and thus also resulting in a larger model due to a large vocab size. # This gives ~3MB in total for all files. # # If you want a 50 times smaller than this see `fsmt-make-super-tiny-model.py`, which is slightly more complicated # # # It will be used then as "stas/tiny-wmt19-en-de" # Build from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration snake_case : List[str] = "facebook/wmt19-en-de" snake_case : Dict = FSMTTokenizer.from_pretrained(mname) # get the correct vocab sizes, etc. from the master model snake_case : List[str] = FSMTConfig.from_pretrained(mname) config.update( dict( d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) ) snake_case : int = FSMTForConditionalGeneration(config) print(F"num of params {tiny_model.num_parameters()}") # Test snake_case : Optional[Any] = tokenizer(["Making tiny model"], return_tensors="pt") snake_case : List[str] = tiny_model(**batch) print("test output:", len(outputs.logits[0])) # Save snake_case : Dict = "tiny-wmt19-en-de" tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F"Generated {mname_tiny}") # Upload # transformers-cli upload tiny-wmt19-en-de
281
0
"""simple docstring""" from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class lowerCAmelCase : '''simple docstring''' def __init__( self , lowerCAmelCase__ , ) -> Union[str, Any]: SCREAMING_SNAKE_CASE = parent SCREAMING_SNAKE_CASE = 13 SCREAMING_SNAKE_CASE = 7 SCREAMING_SNAKE_CASE = 30 SCREAMING_SNAKE_CASE = self.seq_length + self.mem_len SCREAMING_SNAKE_CASE = 15 SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = 99 SCREAMING_SNAKE_CASE = [10, 50, 80] SCREAMING_SNAKE_CASE = 32 SCREAMING_SNAKE_CASE = 32 SCREAMING_SNAKE_CASE = 4 SCREAMING_SNAKE_CASE = 8 SCREAMING_SNAKE_CASE = 128 SCREAMING_SNAKE_CASE = 2 SCREAMING_SNAKE_CASE = 2 SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = 1 SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = 3 SCREAMING_SNAKE_CASE = self.vocab_size - 1 SCREAMING_SNAKE_CASE = 0.01 def __A ( self ) -> List[Any]: SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE = None if self.use_labels: SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE = TransfoXLConfig( vocab_size=self.vocab_size , mem_len=self.mem_len , clamp_len=self.clamp_len , cutoffs=self.cutoffs , d_model=self.hidden_size , d_embed=self.d_embed , n_head=self.num_attention_heads , d_head=self.d_head , d_inner=self.d_inner , div_val=self.div_val , n_layer=self.num_hidden_layers , eos_token_id=self.eos_token_id , pad_token_id=self.vocab_size - 1 , init_range=self.init_range , num_labels=self.num_labels , ) return (config, input_ids_a, input_ids_a, lm_labels) def __A ( self ) -> int: random.seed(self.seed ) tf.random.set_seed(self.seed ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: SCREAMING_SNAKE_CASE = TFTransfoXLModel(_a ) SCREAMING_SNAKE_CASE = model(_a ).to_tuple() SCREAMING_SNAKE_CASE = {"input_ids": input_ids_a, "mems": mems_a} SCREAMING_SNAKE_CASE = model(_a ).to_tuple() self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: SCREAMING_SNAKE_CASE = TFTransfoXLLMHeadModel(_a ) SCREAMING_SNAKE_CASE = model(_a ).to_tuple() SCREAMING_SNAKE_CASE = {"input_ids": input_ids_a, "labels": lm_labels} SCREAMING_SNAKE_CASE = model(_a ).to_tuple() SCREAMING_SNAKE_CASE = model([input_ids_a, mems_a] ).to_tuple() SCREAMING_SNAKE_CASE = {"input_ids": input_ids_a, "mems": mems_a, "labels": lm_labels} SCREAMING_SNAKE_CASE = model(_a ).to_tuple() self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: SCREAMING_SNAKE_CASE = TFTransfoXLForSequenceClassification(_a ) SCREAMING_SNAKE_CASE = model(_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __A ( self ) -> Tuple: SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() (SCREAMING_SNAKE_CASE) = config_and_inputs SCREAMING_SNAKE_CASE = {"input_ids": input_ids_a} return config, inputs_dict @require_tf class lowerCAmelCase ( lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Dict = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE_ : Tuple = () if is_tf_available() else () SCREAMING_SNAKE_CASE_ : Union[str, Any] = ( { """feature-extraction""": TFTransfoXLModel, """text-classification""": TFTransfoXLForSequenceClassification, """text-generation""": TFTransfoXLLMHeadModel, """zero-shot""": TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented SCREAMING_SNAKE_CASE_ : str = False SCREAMING_SNAKE_CASE_ : Union[str, Any] = False SCREAMING_SNAKE_CASE_ : int = False SCREAMING_SNAKE_CASE_ : int = False def __A ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Union[str, Any]: if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def __A ( self ) -> Optional[int]: SCREAMING_SNAKE_CASE = TFTransfoXLModelTester(self ) SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=_a , d_embed=37 ) def __A ( self ) -> Optional[Any]: self.config_tester.run_common_tests() def __A ( self ) -> Optional[Any]: self.model_tester.set_seed() SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*_a ) def __A ( self ) -> Union[str, Any]: self.model_tester.set_seed() SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*_a ) def __A ( self ) -> str: SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*_a ) def __A ( self ) -> str: SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE = model_class(_a ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: SCREAMING_SNAKE_CASE = model.get_output_embeddings() assert isinstance(_a , tf.keras.layers.Layer ) SCREAMING_SNAKE_CASE = model.get_bias() assert name is None else: SCREAMING_SNAKE_CASE = model.get_output_embeddings() assert x is None SCREAMING_SNAKE_CASE = model.get_bias() assert name is None def __A ( self ) -> Tuple: # TODO JP: Make TransfoXL XLA compliant pass @slow def __A ( self ) -> Optional[Any]: for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE = TFTransfoXLModel.from_pretrained(_a ) self.assertIsNotNone(_a ) @unittest.skip(reason='This model doesn\'t play well with fit() due to not returning a single loss.' ) def __A ( self ) -> Dict: pass @require_tf class lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @unittest.skip('Skip test until #12651 is resolved.' ) @slow def __A ( self ) -> Optional[int]: SCREAMING_SNAKE_CASE = TFTransfoXLLMHeadModel.from_pretrained('transfo-xl-wt103' ) # fmt: off SCREAMING_SNAKE_CASE = tf.convert_to_tensor([[33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0]] , dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off SCREAMING_SNAKE_CASE = [33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0,33,1,1_857,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,28,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> SCREAMING_SNAKE_CASE = model.generate(_a , max_length=200 , do_sample=_a ) self.assertListEqual(output_ids[0].numpy().tolist() , _a )
113
import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s", datefmt="%m/%d/%Y %H:%M:%S", level=logging.INFO ) snake_case : Optional[int] = logging.getLogger(__name__) def lowerCAmelCase_ ( _snake_case : Dict , _snake_case : Union[str, Any] ) -> Tuple: '''simple docstring''' __magic_name__ : List[str] = np.argmax(_snake_case , axis=1 ) return np.sum(outputs == labels ) def lowerCAmelCase_ ( _snake_case : Optional[Any] ) -> Dict: '''simple docstring''' with open(_snake_case , encoding="utf_8" ) as f: __magic_name__ : List[str] = csv.reader(_snake_case ) __magic_name__ : List[Any] = [] next(_snake_case ) # skip the first line for line in tqdm(_snake_case ): output.append((" ".join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def lowerCAmelCase_ ( _snake_case : str , _snake_case : Tuple , _snake_case : Union[str, Any] , _snake_case : List[Any] , _snake_case : Tuple , _snake_case : Optional[int] ) -> int: '''simple docstring''' __magic_name__ : Optional[int] = [] for dataset in encoded_datasets: __magic_name__ : Union[str, Any] = len(_snake_case ) __magic_name__ : Dict = np.zeros((n_batch, 2, input_len) , dtype=np.intaa ) __magic_name__ : List[str] = np.zeros((n_batch, 2) , dtype=np.intaa ) __magic_name__ : Optional[int] = np.full((n_batch, 2, input_len) , fill_value=-100 , dtype=np.intaa ) __magic_name__ : int = np.zeros((n_batch,) , dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(_snake_case ): __magic_name__ : Dict = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __magic_name__ : Dict = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __magic_name__ : str = with_conta __magic_name__ : Tuple = with_conta __magic_name__ : Union[str, Any] = len(_snake_case ) - 1 __magic_name__ : int = len(_snake_case ) - 1 __magic_name__ : Optional[Any] = with_conta __magic_name__ : Optional[Any] = with_conta __magic_name__ : Optional[int] = mc_label __magic_name__ : str = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(_snake_case ) for t in all_inputs ) ) return tensor_datasets def lowerCAmelCase_ ( ) -> List[Any]: '''simple docstring''' __magic_name__ : Any = argparse.ArgumentParser() parser.add_argument("--model_name" , type=_snake_case , default="openai-gpt" , help="pretrained model name" ) parser.add_argument("--do_train" , action="store_true" , help="Whether to run training." ) parser.add_argument("--do_eval" , action="store_true" , help="Whether to run eval on the dev set." ) parser.add_argument( "--output_dir" , default=_snake_case , type=_snake_case , required=_snake_case , help="The output directory where the model predictions and checkpoints will be written." , ) parser.add_argument("--train_dataset" , type=_snake_case , default="" ) parser.add_argument("--eval_dataset" , type=_snake_case , default="" ) parser.add_argument("--seed" , type=_snake_case , default=42 ) parser.add_argument("--num_train_epochs" , type=_snake_case , default=3 ) parser.add_argument("--train_batch_size" , type=_snake_case , default=8 ) parser.add_argument("--eval_batch_size" , type=_snake_case , default=16 ) parser.add_argument("--adam_epsilon" , default=1E-8 , type=_snake_case , help="Epsilon for Adam optimizer." ) parser.add_argument("--max_grad_norm" , type=_snake_case , default=1 ) parser.add_argument( "--max_steps" , default=-1 , type=_snake_case , help=( "If > 0: set total number of training steps to perform. Override num_train_epochs." ) , ) parser.add_argument( "--gradient_accumulation_steps" , type=_snake_case , default=1 , help="Number of updates steps to accumulate before performing a backward/update pass." , ) parser.add_argument("--learning_rate" , type=_snake_case , default=6.25E-5 ) parser.add_argument("--warmup_steps" , default=0 , type=_snake_case , help="Linear warmup over warmup_steps." ) parser.add_argument("--lr_schedule" , type=_snake_case , default="warmup_linear" ) parser.add_argument("--weight_decay" , type=_snake_case , default=0.01 ) parser.add_argument("--lm_coef" , type=_snake_case , default=0.9 ) parser.add_argument("--n_valid" , type=_snake_case , default=374 ) parser.add_argument("--server_ip" , type=_snake_case , default="" , help="Can be used for distant debugging." ) parser.add_argument("--server_port" , type=_snake_case , default="" , help="Can be used for distant debugging." ) __magic_name__ : List[Any] = parser.parse_args() print(_snake_case ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print("Waiting for debugger attach" ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=_snake_case ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) __magic_name__ : Dict = torch.device("cuda" if torch.cuda.is_available() else "cpu" ) __magic_name__ : Optional[int] = torch.cuda.device_count() logger.info("device: {}, n_gpu {}".format(_snake_case , _snake_case ) ) if not args.do_train and not args.do_eval: raise ValueError("At least one of `do_train` or `do_eval` must be True." ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset __magic_name__ : List[Any] = ["_start_", "_delimiter_", "_classify_"] __magic_name__ : Optional[int] = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(_snake_case ) __magic_name__ : Optional[Any] = tokenizer.convert_tokens_to_ids(_snake_case ) __magic_name__ : List[str] = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(_snake_case ) ) model.to(_snake_case ) # Load and encode the datasets def tokenize_and_encode(_snake_case : str ): if isinstance(_snake_case , _snake_case ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(_snake_case ) ) elif isinstance(_snake_case , _snake_case ): return obj return [tokenize_and_encode(_snake_case ) for o in obj] logger.info("Encoding dataset..." ) __magic_name__ : Optional[int] = load_rocstories_dataset(args.train_dataset ) __magic_name__ : str = load_rocstories_dataset(args.eval_dataset ) __magic_name__ : int = (train_dataset, eval_dataset) __magic_name__ : List[str] = tokenize_and_encode(_snake_case ) # Compute the max input length for the Transformer __magic_name__ : Optional[Any] = model.config.n_positions // 2 - 2 __magic_name__ : Optional[int] = max( len(story[:max_length] ) + max(len(conta[:max_length] ) , len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) __magic_name__ : List[str] = min(_snake_case , model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders __magic_name__ : List[Any] = pre_process_datasets(_snake_case , _snake_case , _snake_case , *_snake_case ) __magic_name__ , __magic_name__ : Optional[int] = tensor_datasets[0], tensor_datasets[1] __magic_name__ : Tuple = TensorDataset(*_snake_case ) __magic_name__ : Union[str, Any] = RandomSampler(_snake_case ) __magic_name__ : Dict = DataLoader(_snake_case , sampler=_snake_case , batch_size=args.train_batch_size ) __magic_name__ : Any = TensorDataset(*_snake_case ) __magic_name__ : Optional[Any] = SequentialSampler(_snake_case ) __magic_name__ : int = DataLoader(_snake_case , sampler=_snake_case , batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: __magic_name__ : Tuple = args.max_steps __magic_name__ : List[str] = args.max_steps // (len(_snake_case ) // args.gradient_accumulation_steps) + 1 else: __magic_name__ : List[str] = len(_snake_case ) // args.gradient_accumulation_steps * args.num_train_epochs __magic_name__ : str = list(model.named_parameters() ) __magic_name__ : Dict = ["bias", "LayerNorm.bias", "LayerNorm.weight"] __magic_name__ : str = [ { "params": [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], "weight_decay": args.weight_decay, }, {"params": [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], "weight_decay": 0.0}, ] __magic_name__ : str = AdamW(_snake_case , lr=args.learning_rate , eps=args.adam_epsilon ) __magic_name__ : List[str] = get_linear_schedule_with_warmup( _snake_case , num_warmup_steps=args.warmup_steps , num_training_steps=_snake_case ) if args.do_train: __magic_name__ , __magic_name__ , __magic_name__ : Union[str, Any] = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) , desc="Epoch" ): __magic_name__ : List[str] = 0 __magic_name__ : Tuple = 0 __magic_name__ : Dict = tqdm(_snake_case , desc="Training" ) for step, batch in enumerate(_snake_case ): __magic_name__ : Optional[Any] = tuple(t.to(_snake_case ) for t in batch ) __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Dict = batch __magic_name__ : Optional[Any] = model(_snake_case , mc_token_ids=_snake_case , lm_labels=_snake_case , mc_labels=_snake_case ) __magic_name__ : Optional[Any] = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() __magic_name__ : List[str] = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 __magic_name__ : int = "Training loss: {:.2e} lr: {:.2e}".format(_snake_case , scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer __magic_name__ : Dict = model.module if hasattr(_snake_case , "module" ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` __magic_name__ : List[Any] = os.path.join(args.output_dir , _snake_case ) __magic_name__ : Dict = os.path.join(args.output_dir , _snake_case ) torch.save(model_to_save.state_dict() , _snake_case ) model_to_save.config.to_json_file(_snake_case ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned __magic_name__ : Dict = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) __magic_name__ : Optional[int] = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(_snake_case ) if args.do_eval: model.eval() __magic_name__ , __magic_name__ : Any = 0, 0 __magic_name__ , __magic_name__ : Union[str, Any] = 0, 0 for batch in tqdm(_snake_case , desc="Evaluating" ): __magic_name__ : int = tuple(t.to(_snake_case ) for t in batch ) __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Union[str, Any] = batch with torch.no_grad(): __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Dict = model( _snake_case , mc_token_ids=_snake_case , lm_labels=_snake_case , mc_labels=_snake_case ) __magic_name__ : Tuple = mc_logits.detach().cpu().numpy() __magic_name__ : Any = mc_labels.to("cpu" ).numpy() __magic_name__ : str = accuracy(_snake_case , _snake_case ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 __magic_name__ : Tuple = eval_loss / nb_eval_steps __magic_name__ : List[Any] = eval_accuracy / nb_eval_examples __magic_name__ : int = tr_loss / nb_tr_steps if args.do_train else None __magic_name__ : Any = {"eval_loss": eval_loss, "eval_accuracy": eval_accuracy, "train_loss": train_loss} __magic_name__ : int = os.path.join(args.output_dir , "eval_results.txt" ) with open(_snake_case , "w" ) as writer: logger.info("***** Eval results *****" ) for key in sorted(result.keys() ): logger.info(" %s = %s" , _snake_case , str(result[key] ) ) writer.write("%s = %s\n" % (key, str(result[key] )) ) if __name__ == "__main__": main()
281
0
"""simple docstring""" from __future__ import annotations from typing import Any def _snake_case ( lowercase__ ): create_state_space_tree(_snake_case , [] , 0 ) def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): if index == len(_snake_case ): print(_snake_case ) return create_state_space_tree(_snake_case , _snake_case , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(_snake_case , _snake_case , index + 1 ) current_subsequence.pop() if __name__ == "__main__": lowercase__ = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["""A""", """B""", """C"""]) generate_all_subsequences(seq)
96
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
281
0
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import warnings from typing import List from unittest.mock import Mock import torch from torch.utils.data import DataLoader, IterableDataset, TensorDataset from accelerate.accelerator import Accelerator from accelerate.utils.dataclasses import DistributedType class lowerCAmelCase_ ( lowerCAmelCase ): """simple docstring""" def __init__( self , lowerCAmelCase ): """simple docstring""" snake_case = data def __iter__( self ): """simple docstring""" for element in self.data: yield element def lowerCAmelCase__ ( _UpperCamelCase : List[str]=True ) -> Union[str, Any]: """simple docstring""" snake_case = Accelerator(even_batches=_snake_case ) assert accelerator.num_processes == 2, "this script expects that two GPUs are available" return accelerator def lowerCAmelCase__ ( _UpperCamelCase : Accelerator , _UpperCamelCase : int , _UpperCamelCase : int , _UpperCamelCase : bool = False ) -> str: """simple docstring""" if iterable: snake_case = DummyIterableDataset(torch.as_tensor(range(_snake_case ) ) ) else: snake_case = TensorDataset(torch.as_tensor(range(_snake_case ) ) ) snake_case = DataLoader(_snake_case , batch_size=_snake_case ) snake_case = accelerator.prepare(_snake_case ) return dl def lowerCAmelCase__ ( _UpperCamelCase : Accelerator , _UpperCamelCase : int , _UpperCamelCase : int , _UpperCamelCase : List[int] , _UpperCamelCase : List[int] , ) -> List[Any]: """simple docstring""" snake_case = create_dataloader(accelerator=_snake_case , dataset_size=_snake_case , batch_size=_snake_case ) snake_case = [len(batch[0] ) for batch in dl] if accelerator.process_index == 0: assert batch_sizes == process_0_expected_batch_sizes elif accelerator.process_index == 1: assert batch_sizes == process_1_expected_batch_sizes def lowerCAmelCase__ ( ) -> List[str]: """simple docstring""" snake_case = create_accelerator() # without padding, we would expect a different number of batches verify_dataloader_batch_sizes( _snake_case , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1, 1] , ) # without padding, we would expect the same number of batches, but different sizes verify_dataloader_batch_sizes( _snake_case , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 2] , ) def lowerCAmelCase__ ( ) -> Union[str, Any]: """simple docstring""" snake_case = create_accelerator(even_batches=_snake_case ) verify_dataloader_batch_sizes( _snake_case , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1] , ) verify_dataloader_batch_sizes( _snake_case , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 1] , ) def lowerCAmelCase__ ( ) -> Tuple: """simple docstring""" snake_case = create_accelerator(even_batches=_snake_case ) snake_case = torch.nn.Linear(1 , 1 ) snake_case = accelerator.prepare(_snake_case ) snake_case = create_dataloader(_snake_case , dataset_size=3 , batch_size=1 ) snake_case = [] with accelerator.join_uneven_inputs([ddp_model] ): for batch_idx, batch in enumerate(_snake_case ): snake_case = ddp_model(batch[0].float() ) snake_case = output.sum() loss.backward() batch_idxs.append(_snake_case ) accelerator.wait_for_everyone() if accelerator.process_index == 0: assert batch_idxs == [0, 1] elif accelerator.process_index == 1: assert batch_idxs == [0] def lowerCAmelCase__ ( _UpperCamelCase : str ) -> str: """simple docstring""" with warnings.catch_warnings(record=_snake_case ) as w: with accelerator.join_uneven_inputs([Mock()] ): pass assert issubclass(w[-1].category , _snake_case ) assert "only supported for multi-GPU" in str(w[-1].message ) def lowerCAmelCase__ ( ) -> Optional[Any]: """simple docstring""" snake_case = True snake_case = False snake_case = create_accelerator(even_batches=_snake_case ) snake_case = torch.nn.Linear(1 , 1 ) snake_case = accelerator.prepare(_snake_case ) snake_case = create_dataloader(_snake_case , dataset_size=3 , batch_size=1 ) snake_case = create_dataloader(_snake_case , dataset_size=3 , batch_size=1 ) with accelerator.join_uneven_inputs([ddp_model] , even_batches=_snake_case ): snake_case = train_dl.batch_sampler.even_batches snake_case = valid_dl.batch_sampler.even_batches assert train_dl_overridden_value == overridden_even_batches assert valid_dl_overridden_value == overridden_even_batches assert train_dl.batch_sampler.even_batches == default_even_batches assert valid_dl.batch_sampler.even_batches == default_even_batches def lowerCAmelCase__ ( ) -> List[str]: """simple docstring""" snake_case = True snake_case = False snake_case = create_accelerator(even_batches=_snake_case ) snake_case = torch.nn.Linear(1 , 1 ) snake_case = accelerator.prepare(_snake_case ) create_dataloader(_snake_case , dataset_size=3 , batch_size=1 , iterable=_snake_case ) snake_case = create_dataloader(_snake_case , dataset_size=3 , batch_size=1 ) with warnings.catch_warnings(): warnings.filterwarnings('ignore' ) try: with accelerator.join_uneven_inputs([ddp_model] , even_batches=_snake_case ): snake_case = batch_dl.batch_sampler.even_batches except AttributeError: # ensure attribute error is not raised when processing iterable dl raise AssertionError assert batch_dl_overridden_value == overridden_even_batches assert batch_dl.batch_sampler.even_batches == default_even_batches def lowerCAmelCase__ ( ) -> str: """simple docstring""" snake_case = create_accelerator() snake_case = torch.nn.Linear(1 , 1 ) snake_case = accelerator.prepare(_snake_case ) create_dataloader(_snake_case , dataset_size=3 , batch_size=1 , iterable=_snake_case ) with warnings.catch_warnings(record=_snake_case ) as w: with accelerator.join_uneven_inputs([ddp_model] , even_batches=_snake_case ): pass assert issubclass(w[-1].category , _snake_case ) assert "only supported for map-style datasets" in str(w[-1].message ) def lowerCAmelCase__ ( ) -> List[str]: """simple docstring""" snake_case = create_accelerator() accelerator.print('Test that even_batches variable ensures uniform batches across processes' ) test_default_ensures_even_batch_sizes() accelerator.print('Run tests with even_batches disabled' ) test_can_disable_even_batches() accelerator.print('Test joining uneven inputs' ) test_can_join_uneven_inputs() accelerator.print('Test overriding even_batches when joining uneven inputs' ) test_join_can_override_even_batches() accelerator.print('Test overriding even_batches for mixed dataloader types' ) test_join_can_override_for_mixed_type_dataloaders() accelerator.print('Test overriding even_batches raises a warning for iterable dataloaders' ) test_join_raises_warning_for_iterable_when_overriding_even_batches() accelerator.print('Test join with non DDP distributed raises warning' ) snake_case = accelerator.state.distributed_type snake_case = DistributedType.FSDP test_join_raises_warning_for_non_ddp_distributed(_snake_case ) snake_case = original_state if __name__ == "__main__": main()
150
import importlib import os import fsspec import pytest from fsspec import register_implementation from fsspec.registry import _registry as _fsspec_registry from datasets.filesystems import COMPRESSION_FILESYSTEMS, HfFileSystem, extract_path_from_uri, is_remote_filesystem from .utils import require_lza, require_zstandard def lowerCAmelCase_ ( _snake_case : List[Any] ) -> List[Any]: '''simple docstring''' assert "mock" in _fsspec_registry assert "bz2" in _fsspec_registry def lowerCAmelCase_ ( ) -> Tuple: '''simple docstring''' assert "mock" not in _fsspec_registry assert "bz2" in _fsspec_registry def lowerCAmelCase_ ( ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : Dict = "mock-s3-bucket" __magic_name__ : Any = F'''s3://{mock_bucket}''' __magic_name__ : str = extract_path_from_uri(_snake_case ) assert dataset_path.startswith("s3://" ) is False __magic_name__ : Tuple = "./local/path" __magic_name__ : Optional[Any] = extract_path_from_uri(_snake_case ) assert dataset_path == new_dataset_path def lowerCAmelCase_ ( _snake_case : List[str] ) -> Optional[Any]: '''simple docstring''' __magic_name__ : str = is_remote_filesystem(_snake_case ) assert is_remote is True __magic_name__ : Optional[int] = fsspec.filesystem("file" ) __magic_name__ : int = is_remote_filesystem(_snake_case ) assert is_remote is False @pytest.mark.parametrize("compression_fs_class" , _snake_case ) def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : Optional[Any] , _snake_case : int , _snake_case : Tuple , _snake_case : Any , _snake_case : Union[str, Any] , _snake_case : Any ) -> int: '''simple docstring''' __magic_name__ : Any = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_file, "bz2": bza_file, "lz4": lza_file} __magic_name__ : str = input_paths[compression_fs_class.protocol] if input_path is None: __magic_name__ : Dict = F'''for \'{compression_fs_class.protocol}\' compression protocol, ''' if compression_fs_class.protocol == "lz4": reason += require_lza.kwargs["reason"] elif compression_fs_class.protocol == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(_snake_case ) __magic_name__ : str = fsspec.filesystem(compression_fs_class.protocol , fo=_snake_case ) assert isinstance(_snake_case , _snake_case ) __magic_name__ : int = os.path.basename(_snake_case ) __magic_name__ : Optional[int] = expected_filename[: expected_filename.rindex("." )] assert fs.glob("*" ) == [expected_filename] with fs.open(_snake_case , "r" , encoding="utf-8" ) as f, open(_snake_case , encoding="utf-8" ) as expected_file: assert f.read() == expected_file.read() @pytest.mark.parametrize("protocol" , ["zip", "gzip"] ) def lowerCAmelCase_ ( _snake_case : List[Any] , _snake_case : Optional[Any] , _snake_case : Optional[Any] ) -> str: '''simple docstring''' __magic_name__ : int = {"zip": zip_jsonl_path, "gzip": jsonl_gz_path} __magic_name__ : int = compressed_file_paths[protocol] __magic_name__ : Tuple = "dataset.jsonl" __magic_name__ : List[str] = F'''{protocol}://{member_file_path}::{compressed_file_path}''' __magic_name__ , *__magic_name__ : Optional[Any] = fsspec.get_fs_token_paths(_snake_case ) assert fs.isfile(_snake_case ) assert not fs.isfile("non_existing_" + member_file_path ) @pytest.mark.integration def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Dict , _snake_case : List[str] , _snake_case : Tuple ) -> str: '''simple docstring''' __magic_name__ : int = hf_api.dataset_info(_snake_case , token=_snake_case ) __magic_name__ : Optional[Any] = HfFileSystem(repo_info=_snake_case , token=_snake_case ) assert sorted(hffs.glob("*" ) ) == [".gitattributes", "data"] assert hffs.isdir("data" ) assert hffs.isfile(".gitattributes" ) and hffs.isfile("data/text_data.txt" ) with open(_snake_case ) as f: assert hffs.open("data/text_data.txt" , "r" ).read() == f.read() def lowerCAmelCase_ ( ) -> Optional[int]: '''simple docstring''' __magic_name__ : Optional[Any] = "bz2" # Import module import datasets.filesystems # Overwrite protocol and reload register_implementation(_snake_case , _snake_case , clobber=_snake_case ) with pytest.warns(_snake_case ) as warning_info: importlib.reload(datasets.filesystems ) assert len(_snake_case ) == 1 assert ( str(warning_info[0].message ) == F'''A filesystem protocol was already set for {protocol} and will be overwritten.''' )
281
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging UpperCAmelCase : Optional[int] = logging.get_logger(__name__) if is_vision_available(): import PIL class lowerCamelCase__ ( A ): """simple docstring""" __a = ["""pixel_values"""] def __init__( self : Any , UpperCamelCase : Dict = True , UpperCamelCase : Dict = None , UpperCamelCase : int = PILImageResampling.BICUBIC , UpperCamelCase : Optional[Any] = True , UpperCamelCase : int = None , UpperCamelCase : Optional[int] = True , UpperCamelCase : List[Any] = 1 / 255 , UpperCamelCase : Optional[int] = True , UpperCamelCase : Optional[int] = None , UpperCamelCase : int = None , UpperCamelCase : Tuple = True , **UpperCamelCase : Optional[int] , ): '''simple docstring''' super().__init__(**_a ) __UpperCAmelCase : Tuple = size if size is not None else {"shortest_edge": 224} __UpperCAmelCase : int = get_size_dict(_a , default_to_square=_a ) __UpperCAmelCase : Dict = crop_size if crop_size is not None else {"height": 224, "width": 224} __UpperCAmelCase : Tuple = get_size_dict(_a , default_to_square=_a , param_name="""crop_size""" ) __UpperCAmelCase : Optional[Any] = do_resize __UpperCAmelCase : str = size __UpperCAmelCase : str = resample __UpperCAmelCase : Tuple = do_center_crop __UpperCAmelCase : Tuple = crop_size __UpperCAmelCase : List[str] = do_rescale __UpperCAmelCase : Any = rescale_factor __UpperCAmelCase : Union[str, Any] = do_normalize __UpperCAmelCase : Union[str, Any] = image_mean if image_mean is not None else OPENAI_CLIP_MEAN __UpperCAmelCase : Optional[int] = image_std if image_std is not None else OPENAI_CLIP_STD __UpperCAmelCase : Optional[Any] = do_convert_rgb def lowerCamelCase__ ( self : Optional[int] , UpperCamelCase : List[Any] , UpperCamelCase : Optional[Any] , UpperCamelCase : Union[str, Any] = PILImageResampling.BICUBIC , UpperCamelCase : Tuple = None , **UpperCamelCase : Dict , ): '''simple docstring''' __UpperCAmelCase : List[Any] = get_size_dict(_a , default_to_square=_a ) if "shortest_edge" not in size: raise ValueError(f'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) __UpperCAmelCase : Dict = get_resize_output_image_size(_a , size=size["""shortest_edge"""] , default_to_square=_a ) return resize(_a , size=_a , resample=_a , data_format=_a , **_a ) def lowerCamelCase__ ( self : Dict , UpperCamelCase : Tuple , UpperCamelCase : str , UpperCamelCase : Any = None , **UpperCamelCase : Tuple , ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] = get_size_dict(_a ) if "height" not in size or "width" not in size: raise ValueError(f'''The `size` parameter must contain the keys (height, width). Got {size.keys()}''' ) return center_crop(_a , size=(size["""height"""], size["""width"""]) , data_format=_a , **_a ) def lowerCamelCase__ ( self : int , UpperCamelCase : List[str] , UpperCamelCase : str , UpperCamelCase : Any = None , **UpperCamelCase : Tuple , ): '''simple docstring''' return rescale(_a , scale=_a , data_format=_a , **_a ) def lowerCamelCase__ ( self : Dict , UpperCamelCase : Any , UpperCamelCase : Any , UpperCamelCase : Optional[int] , UpperCamelCase : Union[str, Any] = None , **UpperCamelCase : Dict , ): '''simple docstring''' return normalize(_a , mean=_a , std=_a , data_format=_a , **_a ) def lowerCamelCase__ ( self : Union[str, Any] , UpperCamelCase : Union[str, Any] , UpperCamelCase : Optional[Any] = None , UpperCamelCase : Optional[Any] = None , UpperCamelCase : Dict = None , UpperCamelCase : Optional[int] = None , UpperCamelCase : Optional[Any] = None , UpperCamelCase : int = None , UpperCamelCase : Any = None , UpperCamelCase : Optional[int] = None , UpperCamelCase : int = None , UpperCamelCase : List[Any] = None , UpperCamelCase : List[str] = None , UpperCamelCase : int = None , UpperCamelCase : List[str] = ChannelDimension.FIRST , **UpperCamelCase : str , ): '''simple docstring''' __UpperCAmelCase : Dict = do_resize if do_resize is not None else self.do_resize __UpperCAmelCase : str = size if size is not None else self.size __UpperCAmelCase : Tuple = get_size_dict(_a , param_name="""size""" , default_to_square=_a ) __UpperCAmelCase : Optional[Any] = resample if resample is not None else self.resample __UpperCAmelCase : Union[str, Any] = do_center_crop if do_center_crop is not None else self.do_center_crop __UpperCAmelCase : Optional[int] = crop_size if crop_size is not None else self.crop_size __UpperCAmelCase : Dict = get_size_dict(_a , param_name="""crop_size""" , default_to_square=_a ) __UpperCAmelCase : Optional[Any] = do_rescale if do_rescale is not None else self.do_rescale __UpperCAmelCase : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __UpperCAmelCase : Optional[int] = do_normalize if do_normalize is not None else self.do_normalize __UpperCAmelCase : List[Any] = image_mean if image_mean is not None else self.image_mean __UpperCAmelCase : List[Any] = image_std if image_std is not None else self.image_std __UpperCAmelCase : List[str] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb __UpperCAmelCase : List[Any] = make_list_of_images(_a ) if not valid_images(_a ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # PIL RGBA images are converted to RGB if do_convert_rgb: __UpperCAmelCase : Optional[Any] = [convert_to_rgb(_a ) for image in images] # All transformations expect numpy arrays. __UpperCAmelCase : Union[str, Any] = [to_numpy_array(_a ) for image in images] if do_resize: __UpperCAmelCase : List[Any] = [self.resize(image=_a , size=_a , resample=_a ) for image in images] if do_center_crop: __UpperCAmelCase : Tuple = [self.center_crop(image=_a , size=_a ) for image in images] if do_rescale: __UpperCAmelCase : Union[str, Any] = [self.rescale(image=_a , scale=_a ) for image in images] if do_normalize: __UpperCAmelCase : List[str] = [self.normalize(image=_a , mean=_a , std=_a ) for image in images] __UpperCAmelCase : Dict = [to_channel_dimension_format(_a , _a ) for image in images] __UpperCAmelCase : Any = {"pixel_values": images} return BatchFeature(data=_a , tensor_type=_a )
115
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case : Dict = logging.get_logger(__name__) snake_case : List[Any] = { "YituTech/conv-bert-base": "https://huggingface.co/YituTech/conv-bert-base/resolve/main/config.json", "YituTech/conv-bert-medium-small": ( "https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/config.json" ), "YituTech/conv-bert-small": "https://huggingface.co/YituTech/conv-bert-small/resolve/main/config.json", # See all ConvBERT models at https://huggingface.co/models?filter=convbert } class _snake_case ( snake_case ): UpperCamelCase__ = 'convbert' def __init__( self , _a=30_522 , _a=768 , _a=12 , _a=12 , _a=3_072 , _a="gelu" , _a=0.1 , _a=0.1 , _a=512 , _a=2 , _a=0.02 , _a=1e-12 , _a=1 , _a=0 , _a=2 , _a=768 , _a=2 , _a=9 , _a=1 , _a=None , **_a , ): super().__init__( pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , **_a , ) __magic_name__ : Tuple = vocab_size __magic_name__ : List[Any] = hidden_size __magic_name__ : Union[str, Any] = num_hidden_layers __magic_name__ : List[Any] = num_attention_heads __magic_name__ : str = intermediate_size __magic_name__ : Any = hidden_act __magic_name__ : List[Any] = hidden_dropout_prob __magic_name__ : Optional[int] = attention_probs_dropout_prob __magic_name__ : Tuple = max_position_embeddings __magic_name__ : str = type_vocab_size __magic_name__ : List[str] = initializer_range __magic_name__ : Tuple = layer_norm_eps __magic_name__ : List[Any] = embedding_size __magic_name__ : List[Any] = head_ratio __magic_name__ : str = conv_kernel_size __magic_name__ : Dict = num_groups __magic_name__ : str = classifier_dropout class _snake_case ( snake_case ): @property def SCREAMING_SNAKE_CASE ( self ): if self.task == "multiple-choice": __magic_name__ : Dict = {0: "batch", 1: "choice", 2: "sequence"} else: __magic_name__ : Dict = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis), ] )
281
0
import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class A_ ( snake_case__ ): _lowercase : int = ['image_processor', 'tokenizer'] _lowercase : Optional[int] = 'AutoImageProcessor' _lowercase : List[Any] = 'AutoTokenizer' def __init__( self : List[str] , UpperCAmelCase : Union[str, Any]=None , UpperCAmelCase : Optional[Any]=None , **UpperCAmelCase : int ) -> str: __lowerCAmelCase: Optional[int] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , _a , ) __lowerCAmelCase: Union[str, Any] = kwargs.pop('feature_extractor' ) __lowerCAmelCase: List[str] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(_a , _a ) __lowerCAmelCase: Tuple = self.image_processor __lowerCAmelCase: Union[str, Any] = False def __call__( self : str , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : List[Any] ) -> int: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*_a , **_a ) __lowerCAmelCase: str = kwargs.pop('images' , _a ) __lowerCAmelCase: Dict = kwargs.pop('text' , _a ) if len(_a ) > 0: __lowerCAmelCase: List[Any] = args[0] __lowerCAmelCase: Any = args[1:] if images is None and text is None: raise ValueError('You need to specify either an `images` or `text` input to process.' ) if images is not None: __lowerCAmelCase: Tuple = self.image_processor(_a , *_a , **_a ) if text is not None: __lowerCAmelCase: str = self.tokenizer(_a , **_a ) if text is None: return inputs elif images is None: return encodings else: __lowerCAmelCase: List[str] = encodings["input_ids"] return inputs def UpperCAmelCase ( self : int , *UpperCAmelCase : Dict , **UpperCAmelCase : Union[str, Any] ) -> Tuple: return self.tokenizer.batch_decode(*_a , **_a ) def UpperCAmelCase ( self : List[str] , *UpperCAmelCase : Tuple , **UpperCAmelCase : Optional[Any] ) -> Union[str, Any]: return self.tokenizer.decode(*_a , **_a ) @contextmanager def UpperCAmelCase ( self : Any ) -> Tuple: warnings.warn( '`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ' 'labels by using the argument `text` of the regular `__call__` method (either in the same call as ' 'your images inputs, or in a separate call.' ) __lowerCAmelCase: Any = True __lowerCAmelCase: Dict = self.tokenizer yield __lowerCAmelCase: Any = self.image_processor __lowerCAmelCase: Optional[int] = False def UpperCAmelCase ( self : int , UpperCAmelCase : int , UpperCAmelCase : Union[str, Any]=False , UpperCAmelCase : int=None ) -> List[str]: if added_vocab is None: __lowerCAmelCase: Optional[int] = self.tokenizer.get_added_vocab() __lowerCAmelCase: int = {} while tokens: __lowerCAmelCase: Optional[Any] = re.search(R'<s_(.*?)>' , _a , re.IGNORECASE ) if start_token is None: break __lowerCAmelCase: List[str] = start_token.group(1 ) __lowerCAmelCase: Optional[Any] = re.search(RF'''</s_{key}>''' , _a , re.IGNORECASE ) __lowerCAmelCase: Union[str, Any] = start_token.group() if end_token is None: __lowerCAmelCase: str = tokens.replace(_a , '' ) else: __lowerCAmelCase: Optional[int] = end_token.group() __lowerCAmelCase: Dict = re.escape(_a ) __lowerCAmelCase: List[str] = re.escape(_a ) __lowerCAmelCase: str = re.search(F'''{start_token_escaped}(.*?){end_token_escaped}''' , _a , re.IGNORECASE ) if content is not None: __lowerCAmelCase: int = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node __lowerCAmelCase: List[str] = self.tokenajson(_a , is_inner_value=_a , added_vocab=_a ) if value: if len(_a ) == 1: __lowerCAmelCase: Any = value[0] __lowerCAmelCase: int = value else: # leaf nodes __lowerCAmelCase: str = [] for leaf in content.split(R'<sep/>' ): __lowerCAmelCase: Optional[Any] = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": __lowerCAmelCase: Optional[Any] = leaf[1:-2] # for categorical special tokens output[key].append(_a ) if len(output[key] ) == 1: __lowerCAmelCase: Any = output[key][0] __lowerCAmelCase: Optional[Any] = tokens[tokens.find(_a ) + len(_a ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:] , is_inner_value=_a , added_vocab=_a ) if len(_a ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def UpperCAmelCase ( self : List[str] ) -> List[Any]: warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , _a , ) return self.image_processor_class @property def UpperCAmelCase ( self : List[Any] ) -> Optional[Any]: warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , _a , ) return self.image_processor
322
import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipaConfig, BlipaForConditionalGeneration, BlipaProcessor, BlipaVisionConfig, BlipImageProcessor, OPTConfig, TaConfig, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def lowerCAmelCase_ ( ) -> str: '''simple docstring''' __magic_name__ : int = "https://storage.googleapis.com/sfr-vision-language-research/LAVIS/assets/merlion.png" __magic_name__ : Union[str, Any] = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ).convert("RGB" ) return image def lowerCAmelCase_ ( _snake_case : str ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : List[str] = [] # fmt: off # vision encoder rename_keys.append(("visual_encoder.cls_token", "vision_model.embeddings.class_embedding") ) rename_keys.append(("visual_encoder.pos_embed", "vision_model.embeddings.position_embedding") ) rename_keys.append(("visual_encoder.patch_embed.proj.weight", "vision_model.embeddings.patch_embedding.weight") ) rename_keys.append(("visual_encoder.patch_embed.proj.bias", "vision_model.embeddings.patch_embedding.bias") ) rename_keys.append(("ln_vision.weight", "vision_model.post_layernorm.weight") ) rename_keys.append(("ln_vision.bias", "vision_model.post_layernorm.bias") ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((F'''visual_encoder.blocks.{i}.norm1.weight''', F'''vision_model.encoder.layers.{i}.layer_norm1.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm1.bias''', F'''vision_model.encoder.layers.{i}.layer_norm1.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm2.weight''', F'''vision_model.encoder.layers.{i}.layer_norm2.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.norm2.bias''', F'''vision_model.encoder.layers.{i}.layer_norm2.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.qkv.weight''', F'''vision_model.encoder.layers.{i}.self_attn.qkv.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.proj.weight''', F'''vision_model.encoder.layers.{i}.self_attn.projection.weight''',) ) rename_keys.append((F'''visual_encoder.blocks.{i}.attn.proj.bias''', F'''vision_model.encoder.layers.{i}.self_attn.projection.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc1.weight''', F'''vision_model.encoder.layers.{i}.mlp.fc1.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc1.bias''', F'''vision_model.encoder.layers.{i}.mlp.fc1.bias''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc2.weight''', F'''vision_model.encoder.layers.{i}.mlp.fc2.weight''') ) rename_keys.append((F'''visual_encoder.blocks.{i}.mlp.fc2.bias''', F'''vision_model.encoder.layers.{i}.mlp.fc2.bias''') ) # QFormer rename_keys.append(("Qformer.bert.embeddings.LayerNorm.weight", "qformer.layernorm.weight") ) rename_keys.append(("Qformer.bert.embeddings.LayerNorm.bias", "qformer.layernorm.bias") ) # fmt: on return rename_keys def lowerCAmelCase_ ( _snake_case : Any , _snake_case : Union[str, Any] , _snake_case : Optional[Any] ) -> int: '''simple docstring''' __magic_name__ : Tuple = dct.pop(_snake_case ) __magic_name__ : int = val def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : Optional[Any] ) -> Dict: '''simple docstring''' for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases __magic_name__ : List[Any] = state_dict.pop(F'''visual_encoder.blocks.{i}.attn.q_bias''' ) __magic_name__ : Optional[Any] = state_dict.pop(F'''visual_encoder.blocks.{i}.attn.v_bias''' ) # next, set bias in the state dict __magic_name__ : Optional[int] = torch.cat((q_bias, torch.zeros_like(_snake_case , requires_grad=_snake_case ), v_bias) ) __magic_name__ : Union[str, Any] = qkv_bias def lowerCAmelCase_ ( _snake_case : Dict , _snake_case : str ) -> int: '''simple docstring''' __magic_name__ : List[Any] = 364 if "coco" in model_name else 224 __magic_name__ : Union[str, Any] = BlipaVisionConfig(image_size=_snake_case ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "opt-2.7b" in model_name: __magic_name__ : List[str] = OPTConfig.from_pretrained("facebook/opt-2.7b" , eos_token_id=_snake_case ).to_dict() elif "opt-6.7b" in model_name: __magic_name__ : Any = OPTConfig.from_pretrained("facebook/opt-6.7b" , eos_token_id=_snake_case ).to_dict() elif "t5-xl" in model_name: __magic_name__ : Dict = TaConfig.from_pretrained("google/flan-t5-xl" , dense_act_fn="gelu" , bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: __magic_name__ : int = TaConfig.from_pretrained("google/flan-t5-xxl" , dense_act_fn="gelu" , bos_token_id=1 ).to_dict() __magic_name__ : List[Any] = BlipaConfig(vision_config=_snake_case , text_config=_snake_case ) return config, image_size @torch.no_grad() def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : str=None , _snake_case : Dict=False ) -> List[Any]: '''simple docstring''' __magic_name__ : Optional[int] = ( AutoTokenizer.from_pretrained("facebook/opt-2.7b" ) if "opt" in model_name else AutoTokenizer.from_pretrained("google/flan-t5-xl" ) ) __magic_name__ : List[Any] = tokenizer("\n" , add_special_tokens=_snake_case ).input_ids[0] __magic_name__ , __magic_name__ : Tuple = get_blipa_config(_snake_case , eos_token_id=_snake_case ) __magic_name__ : Union[str, Any] = BlipaForConditionalGeneration(_snake_case ).eval() __magic_name__ : Any = { "blip2-opt-2.7b": ("blip2_opt", "pretrain_opt2.7b"), "blip2-opt-6.7b": ("blip2_opt", "pretrain_opt6.7b"), "blip2-opt-2.7b-coco": ("blip2_opt", "caption_coco_opt2.7b"), "blip2-opt-6.7b-coco": ("blip2_opt", "caption_coco_opt6.7b"), "blip2-flan-t5-xl": ("blip2_t5", "pretrain_flant5xl"), "blip2-flan-t5-xl-coco": ("blip2_t5", "caption_coco_flant5xl"), "blip2-flan-t5-xxl": ("blip2_t5", "pretrain_flant5xxl"), } __magic_name__ , __magic_name__ : Union[str, Any] = model_name_to_original[model_name] # load original model print("Loading original model..." ) __magic_name__ : Union[str, Any] = "cuda" if torch.cuda.is_available() else "cpu" __magic_name__ , __magic_name__ , __magic_name__ : Optional[Any] = load_model_and_preprocess( name=_snake_case , model_type=_snake_case , is_eval=_snake_case , device=_snake_case ) original_model.eval() print("Done!" ) # update state dict keys __magic_name__ : Dict = original_model.state_dict() __magic_name__ : str = create_rename_keys(_snake_case ) for src, dest in rename_keys: rename_key(_snake_case , _snake_case , _snake_case ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): __magic_name__ : Any = state_dict.pop(_snake_case ) if key.startswith("Qformer.bert" ): __magic_name__ : Optional[int] = key.replace("Qformer.bert" , "qformer" ) if "attention.self" in key: __magic_name__ : Any = key.replace("self" , "attention" ) if "opt_proj" in key: __magic_name__ : Union[str, Any] = key.replace("opt_proj" , "language_projection" ) if "t5_proj" in key: __magic_name__ : Optional[int] = key.replace("t5_proj" , "language_projection" ) if key.startswith("opt" ): __magic_name__ : List[str] = key.replace("opt" , "language" ) if key.startswith("t5" ): __magic_name__ : Tuple = key.replace("t5" , "language" ) __magic_name__ : Dict = val # read in qv biases read_in_q_v_bias(_snake_case , _snake_case ) __magic_name__ , __magic_name__ : Tuple = hf_model.load_state_dict(_snake_case , strict=_snake_case ) assert len(_snake_case ) == 0 assert unexpected_keys == ["qformer.embeddings.position_ids"] __magic_name__ : List[Any] = load_demo_image() __magic_name__ : Tuple = vis_processors["eval"](_snake_case ).unsqueeze(0 ).to(_snake_case ) __magic_name__ : Dict = tokenizer(["\n"] , return_tensors="pt" ).input_ids.to(_snake_case ) # create processor __magic_name__ : Optional[Any] = BlipImageProcessor( size={"height": image_size, "width": image_size} , image_mean=_snake_case , image_std=_snake_case ) __magic_name__ : Dict = BlipaProcessor(image_processor=_snake_case , tokenizer=_snake_case ) __magic_name__ : Union[str, Any] = processor(images=_snake_case , return_tensors="pt" ).pixel_values.to(_snake_case ) # make sure processor creates exact same pixel values assert torch.allclose(_snake_case , _snake_case ) original_model.to(_snake_case ) hf_model.to(_snake_case ) with torch.no_grad(): if "opt" in model_name: __magic_name__ : List[Any] = original_model({"image": original_pixel_values, "text_input": [""]} ).logits __magic_name__ : Optional[int] = hf_model(_snake_case , _snake_case ).logits else: __magic_name__ : int = original_model( {"image": original_pixel_values, "text_input": ["\n"], "text_output": ["\n"]} ).logits __magic_name__ : Tuple = input_ids.masked_fill(input_ids == tokenizer.pad_token_id , -100 ) __magic_name__ : List[str] = hf_model(_snake_case , _snake_case , labels=_snake_case ).logits assert original_logits.shape == logits.shape print("First values of original logits:" , original_logits[0, :3, :3] ) print("First values of HF logits:" , logits[0, :3, :3] ) # assert values if model_name == "blip2-flan-t5-xl": __magic_name__ : List[str] = torch.tensor( [[-41.5_850, -4.4_440, -8.9_922], [-47.4_322, -5.9_143, -1.7_340]] , device=_snake_case ) assert torch.allclose(logits[0, :3, :3] , _snake_case , atol=1E-4 ) elif model_name == "blip2-flan-t5-xl-coco": __magic_name__ : Tuple = torch.tensor( [[-57.0_109, -9.8_967, -12.6_280], [-68.6_578, -12.7_191, -10.5_065]] , device=_snake_case ) else: # cast to same type __magic_name__ : str = logits.dtype assert torch.allclose(original_logits.to(_snake_case ) , _snake_case , atol=1E-2 ) print("Looks ok!" ) print("Generating a caption..." ) __magic_name__ : Optional[int] = "" __magic_name__ : Dict = tokenizer(_snake_case , return_tensors="pt" ).input_ids.to(_snake_case ) __magic_name__ : int = original_model.generate({"image": original_pixel_values} ) __magic_name__ : Optional[Any] = hf_model.generate( _snake_case , _snake_case , do_sample=_snake_case , num_beams=5 , max_length=30 , min_length=1 , top_p=0.9 , repetition_penalty=1.0 , length_penalty=1.0 , temperature=1 , ) print("Original generation:" , _snake_case ) __magic_name__ : Tuple = input_ids.shape[1] __magic_name__ : int = processor.batch_decode(outputs[:, prompt_length:] , skip_special_tokens=_snake_case ) __magic_name__ : Union[str, Any] = [text.strip() for text in output_text] print("HF generation:" , _snake_case ) if pytorch_dump_folder_path is not None: processor.save_pretrained(_snake_case ) hf_model.save_pretrained(_snake_case ) if push_to_hub: processor.push_to_hub(F'''nielsr/{model_name}''' ) hf_model.push_to_hub(F'''nielsr/{model_name}''' ) if __name__ == "__main__": snake_case : Any = argparse.ArgumentParser() snake_case : Union[str, Any] = [ "blip2-opt-2.7b", "blip2-opt-6.7b", "blip2-opt-2.7b-coco", "blip2-opt-6.7b-coco", "blip2-flan-t5-xl", "blip2-flan-t5-xl-coco", "blip2-flan-t5-xxl", ] parser.add_argument( "--model_name", default="blip2-opt-2.7b", choices=choices, type=str, help="Path to hf config.json of model to convert", ) parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model and processor to the hub after converting", ) snake_case : int = parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
281
0
def _UpperCamelCase ( snake_case__ ) -> str: __UpperCAmelCase : Tuple = len(_snake_case ) for i in range(length - 1 ): __UpperCAmelCase : Dict = i for k in range(i + 1, _snake_case ): if collection[k] < collection[least]: __UpperCAmelCase : Tuple = k if least != i: __UpperCAmelCase : List[str] = (collection[i], collection[least]) return collection if __name__ == "__main__": _snake_case = input('''Enter numbers separated by a comma:\n''').strip() _snake_case = [int(item) for item in user_input.split(''',''')] print(selection_sort(unsorted))
157
import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging snake_case : Dict = logging.get_logger(__name__) snake_case : Union[str, Any] = { "vocab_file": "vocab.txt", "merges_file": "bpe.codes", } snake_case : Dict = { "vocab_file": { "vinai/phobert-base": "https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt", "vinai/phobert-large": "https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt", }, "merges_file": { "vinai/phobert-base": "https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes", "vinai/phobert-large": "https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes", }, } snake_case : Union[str, Any] = { "vinai/phobert-base": 256, "vinai/phobert-large": 256, } def lowerCAmelCase_ ( _snake_case : str ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : List[str] = set() __magic_name__ : Any = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __magic_name__ : int = char __magic_name__ : List[str] = set(_snake_case ) return pairs class _snake_case ( snake_case ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , _a , _a , _a="<s>" , _a="</s>" , _a="</s>" , _a="<s>" , _a="<unk>" , _a="<pad>" , _a="<mask>" , **_a , ): super().__init__( bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , cls_token=_a , pad_token=_a , mask_token=_a , **_a , ) __magic_name__ : Dict = vocab_file __magic_name__ : Tuple = merges_file __magic_name__ : List[Any] = {} __magic_name__ : List[Any] = 0 __magic_name__ : Tuple = 1 __magic_name__ : int = 2 __magic_name__ : Union[str, Any] = 3 self.add_from_file(_a ) __magic_name__ : Optional[int] = {v: k for k, v in self.encoder.items()} with open(_a , encoding="utf-8" ) as merges_handle: __magic_name__ : List[str] = merges_handle.read().split("\n" )[:-1] __magic_name__ : Union[str, Any] = [tuple(merge.split()[:-1] ) for merge in merges] __magic_name__ : Union[str, Any] = dict(zip(_a , range(len(_a ) ) ) ) __magic_name__ : Optional[int] = {} def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __magic_name__ : Optional[Any] = [self.cls_token_id] __magic_name__ : Union[str, Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self , _a , _a = None , _a = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a , token_ids_a=_a , already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): __magic_name__ : Optional[Any] = [self.sep_token_id] __magic_name__ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def SCREAMING_SNAKE_CASE ( self ): return len(self.encoder ) def SCREAMING_SNAKE_CASE ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def SCREAMING_SNAKE_CASE ( self , _a ): if token in self.cache: return self.cache[token] __magic_name__ : List[Any] = tuple(_a ) __magic_name__ : List[Any] = tuple(list(word[:-1] ) + [word[-1] + "</w>"] ) __magic_name__ : Any = get_pairs(_a ) if not pairs: return token while True: __magic_name__ : str = min(_a , key=lambda _a : self.bpe_ranks.get(_a , float("inf" ) ) ) if bigram not in self.bpe_ranks: break __magic_name__ , __magic_name__ : List[str] = bigram __magic_name__ : List[str] = [] __magic_name__ : List[str] = 0 while i < len(_a ): try: __magic_name__ : Any = word.index(_a , _a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) __magic_name__ : Tuple = j if word[i] == first and i < len(_a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __magic_name__ : Union[str, Any] = tuple(_a ) __magic_name__ : Optional[int] = new_word if len(_a ) == 1: break else: __magic_name__ : List[Any] = get_pairs(_a ) __magic_name__ : Optional[int] = "@@ ".join(_a ) __magic_name__ : Tuple = word[:-4] __magic_name__ : str = word return word def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Optional[Any] = [] __magic_name__ : Dict = re.findall(r"\S+\n?" , _a ) for token in words: split_tokens.extend(list(self.bpe(_a ).split(" " ) ) ) return split_tokens def SCREAMING_SNAKE_CASE ( self , _a ): return self.encoder.get(_a , self.encoder.get(self.unk_token ) ) def SCREAMING_SNAKE_CASE ( self , _a ): return self.decoder.get(_a , self.unk_token ) def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Tuple = " ".join(_a ).replace("@@ " , "" ).strip() return out_string def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): if not os.path.isdir(_a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __magic_name__ : Optional[int] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) __magic_name__ : Union[str, Any] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ): copyfile(self.vocab_file , _a ) if os.path.abspath(self.merges_file ) != os.path.abspath(_a ): copyfile(self.merges_file , _a ) return out_vocab_file, out_merge_file def SCREAMING_SNAKE_CASE ( self , _a ): if isinstance(_a , _a ): try: with open(_a , "r" , encoding="utf-8" ) as fd: self.add_from_file(_a ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(f'''Incorrect encoding detected in {f}, please rebuild the dataset''' ) return __magic_name__ : List[Any] = f.readlines() for lineTmp in lines: __magic_name__ : Optional[Any] = lineTmp.strip() __magic_name__ : Union[str, Any] = line.rfind(" " ) if idx == -1: raise ValueError("Incorrect dictionary format, expected '<token> <cnt>'" ) __magic_name__ : Optional[int] = line[:idx] __magic_name__ : Dict = len(self.encoder )
281
0
"""simple docstring""" import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument( """--txt2img_unclip""", default="""kakaobrain/karlo-v1-alpha""", type=str, required=False, help="""The pretrained txt2img unclip.""", ) lowerCamelCase__ = parser.parse_args() lowerCamelCase__ = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) lowerCamelCase__ = CLIPImageProcessor() lowerCamelCase__ = CLIPVisionModelWithProjection.from_pretrained("""openai/clip-vit-large-patch14""") lowerCamelCase__ = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
86
from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_ ( _snake_case : str = "laptop" ) -> DataFrame: '''simple docstring''' __magic_name__ : Tuple = F'''https://www.amazon.in/laptop/s?k={product}''' __magic_name__ : Dict = { "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36", "Accept-Language": "en-US, en;q=0.5", } __magic_name__ : Tuple = BeautifulSoup(requests.get(_snake_case , headers=_snake_case ).text ) # Initialize a Pandas dataframe with the column titles __magic_name__ : int = DataFrame( columns=[ "Product Title", "Product Link", "Current Price of the product", "Product Rating", "MRP of the product", "Discount", ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( "div" , attrs={"class": "s-result-item", "data-component-type": "s-search-result"} , ) , soup.find_all("div" , attrs={"class": "a-row a-size-base a-color-base"} ) , ): try: __magic_name__ : Dict = item.ha.text __magic_name__ : Optional[int] = "https://www.amazon.in/" + item.ha.a["href"] __magic_name__ : Optional[Any] = item.find("span" , attrs={"class": "a-offscreen"} ).text try: __magic_name__ : Union[str, Any] = item.find("span" , attrs={"class": "a-icon-alt"} ).text except AttributeError: __magic_name__ : Dict = "Not available" try: __magic_name__ : Optional[int] = ( "₹" + item.find( "span" , attrs={"class": "a-price a-text-price"} ).text.split("₹" )[1] ) except AttributeError: __magic_name__ : List[str] = "" try: __magic_name__ : int = float( ( ( float(product_mrp.strip("₹" ).replace("," , "" ) ) - float(product_price.strip("₹" ).replace("," , "" ) ) ) / float(product_mrp.strip("₹" ).replace("," , "" ) ) ) * 100 ) except ValueError: __magic_name__ : str = float("nan" ) except AttributeError: pass __magic_name__ : Optional[int] = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] __magic_name__ : Optional[Any] = " " __magic_name__ : str = " " data_frame.index += 1 return data_frame if __name__ == "__main__": snake_case : Any = "headphones" get_amazon_product_data(product).to_csv(F"Amazon Product Data for {product}.csv")
281
0
from abc import ABC, abstractmethod from typing import List, Optional class __snake_case ( lowerCamelCase_ ): def __init__( self : Tuple ): """simple docstring""" self.test() def __a ( self : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = False while not completed: if counter == 1: self.reset() SCREAMING_SNAKE_CASE__ = self.advance() if not self.does_advance(_a ): raise Exception( """Custom Constraint is not defined correctly. self.does_advance(self.advance()) must be true.""" ) SCREAMING_SNAKE_CASE__ = self.update(_a ) counter += 1 if counter > 1_00_00: raise Exception("""update() does not fulfill the constraint.""" ) if self.remaining() != 0: raise Exception("""Custom Constraint is not defined correctly.""" ) @abstractmethod def __a ( self : Optional[int] ): """simple docstring""" raise NotImplementedError( f"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def __a ( self : str , _lowercase : str ): """simple docstring""" raise NotImplementedError( f"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def __a ( self : str , _lowercase : str ): """simple docstring""" raise NotImplementedError( f"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def __a ( self : Dict ): """simple docstring""" raise NotImplementedError( f"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def __a ( self : int ): """simple docstring""" raise NotImplementedError( f"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def __a ( self : Optional[int] , _lowercase : Dict=False ): """simple docstring""" raise NotImplementedError( f"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) class __snake_case ( lowerCamelCase_ ): def __init__( self : Dict , _lowercase : Optional[Any] ): """simple docstring""" super(_a , self ).__init__() if not isinstance(_a , _a ) or len(_a ) == 0: raise ValueError(f"""`token_ids` has to be a non-empty list, but is {token_ids}.""" ) if any((not isinstance(_a , _a ) or token_id < 0) for token_id in token_ids ): raise ValueError(f"""Each list in `token_ids` has to be a list of positive integers, but is {token_ids}.""" ) SCREAMING_SNAKE_CASE__ = token_ids SCREAMING_SNAKE_CASE__ = len(self.token_ids ) SCREAMING_SNAKE_CASE__ = -1 # the index of the currently fulfilled step SCREAMING_SNAKE_CASE__ = False def __a ( self : List[str] ): """simple docstring""" if self.completed: return None return self.token_ids[self.fulfilled_idx + 1] def __a ( self : List[Any] , _lowercase : Union[str, Any] ): """simple docstring""" if not isinstance(_a , _a ): raise ValueError(f"""`token_id` has to be an `int`, but is {token_id} of type {type(_a )}""" ) if self.completed: return False return token_id == self.token_ids[self.fulfilled_idx + 1] def __a ( self : Tuple , _lowercase : Optional[int] ): """simple docstring""" if not isinstance(_a , _a ): raise ValueError(f"""`token_id` has to be an `int`, but is {token_id} of type {type(_a )}""" ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False if self.does_advance(_a ): self.fulfilled_idx += 1 SCREAMING_SNAKE_CASE__ = True if self.fulfilled_idx == (self.seqlen - 1): SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = completed else: # failed to make progress. SCREAMING_SNAKE_CASE__ = True self.reset() return stepped, completed, reset def __a ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = 0 def __a ( self : Dict ): """simple docstring""" return self.seqlen - (self.fulfilled_idx + 1) def __a ( self : Dict , _lowercase : int=False ): """simple docstring""" SCREAMING_SNAKE_CASE__ = PhrasalConstraint(self.token_ids ) if stateful: SCREAMING_SNAKE_CASE__ = self.seqlen SCREAMING_SNAKE_CASE__ = self.fulfilled_idx SCREAMING_SNAKE_CASE__ = self.completed return new_constraint class __snake_case : def __init__( self : int , _lowercase : Any , _lowercase : Optional[Any]=True ): """simple docstring""" SCREAMING_SNAKE_CASE__ = max([len(_a ) for one in nested_token_ids] ) SCREAMING_SNAKE_CASE__ = {} for token_ids in nested_token_ids: SCREAMING_SNAKE_CASE__ = root for tidx, token_id in enumerate(_a ): if token_id not in level: SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = level[token_id] if no_subsets and self.has_subsets(_a , _a ): raise ValueError( """Each list in `nested_token_ids` can't be a complete subset of another list, but is""" f""" {nested_token_ids}.""" ) SCREAMING_SNAKE_CASE__ = root def __a ( self : Dict , _lowercase : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.trie for current_token in current_seq: SCREAMING_SNAKE_CASE__ = start[current_token] SCREAMING_SNAKE_CASE__ = list(start.keys() ) return next_tokens def __a ( self : Tuple , _lowercase : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.next_tokens(_a ) return len(_a ) == 0 def __a ( self : List[Any] , _lowercase : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE__ = list(root.values() ) if len(_a ) == 0: return 1 else: return sum([self.count_leaves(_a ) for nn in next_nodes] ) def __a ( self : Tuple , _lowercase : List[Any] , _lowercase : List[str] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.count_leaves(_a ) return len(_a ) != leaf_count class __snake_case ( lowerCamelCase_ ): def __init__( self : int , _lowercase : str ): """simple docstring""" super(_a , self ).__init__() if not isinstance(_a , _a ) or len(_a ) == 0: raise ValueError(f"""`nested_token_ids` has to be a non-empty list, but is {nested_token_ids}.""" ) if any(not isinstance(_a , _a ) for token_ids in nested_token_ids ): raise ValueError(f"""`nested_token_ids` has to be a list of lists, but is {nested_token_ids}.""" ) if any( any((not isinstance(_a , _a ) or token_id < 0) for token_id in token_ids ) for token_ids in nested_token_ids ): raise ValueError( f"""Each list in `nested_token_ids` has to be a list of positive integers, but is {nested_token_ids}.""" ) SCREAMING_SNAKE_CASE__ = DisjunctiveTrie(_a ) SCREAMING_SNAKE_CASE__ = nested_token_ids SCREAMING_SNAKE_CASE__ = self.trie.max_height SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = False def __a ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.trie.next_tokens(self.current_seq ) if len(_a ) == 0: return None else: return token_list def __a ( self : Dict , _lowercase : List[Any] ): """simple docstring""" if not isinstance(_a , _a ): raise ValueError(f"""`token_id` is supposed to be type `int`, but is {token_id} of type {type(_a )}""" ) SCREAMING_SNAKE_CASE__ = self.trie.next_tokens(self.current_seq ) return token_id in next_tokens def __a ( self : List[str] , _lowercase : Union[str, Any] ): """simple docstring""" if not isinstance(_a , _a ): raise ValueError(f"""`token_id` is supposed to be type `int`, but is {token_id} of type {type(_a )}""" ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False if self.does_advance(_a ): self.current_seq.append(_a ) SCREAMING_SNAKE_CASE__ = True else: SCREAMING_SNAKE_CASE__ = True self.reset() SCREAMING_SNAKE_CASE__ = self.trie.reached_leaf(self.current_seq ) SCREAMING_SNAKE_CASE__ = completed return stepped, completed, reset def __a ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = [] def __a ( self : Dict ): """simple docstring""" if self.completed: # since this can be completed without reaching max height return 0 else: return self.seqlen - len(self.current_seq ) def __a ( self : Any , _lowercase : List[str]=False ): """simple docstring""" SCREAMING_SNAKE_CASE__ = DisjunctiveConstraint(self.token_ids ) if stateful: SCREAMING_SNAKE_CASE__ = self.seqlen SCREAMING_SNAKE_CASE__ = self.current_seq SCREAMING_SNAKE_CASE__ = self.completed return new_constraint class __snake_case : def __init__( self : List[str] , _lowercase : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = constraints # max # of steps required to fulfill a given constraint SCREAMING_SNAKE_CASE__ = max([c.seqlen for c in constraints] ) SCREAMING_SNAKE_CASE__ = len(_a ) SCREAMING_SNAKE_CASE__ = False self.init_state() def __a ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = [constraint.copy(stateful=_a ) for constraint in self.constraints] def __a ( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE__ = 0 if self.inprogress_constraint: # extra points for having a constraint mid-fulfilled add += self.max_seqlen - self.inprogress_constraint.remaining() return (len(self.complete_constraints ) * self.max_seqlen) + add def __a ( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE__ = [] if self.inprogress_constraint is None: for constraint in self.pending_constraints: # "pending" == "unfulfilled yet" SCREAMING_SNAKE_CASE__ = constraint.advance() if isinstance(_a , _a ): token_list.append(_a ) elif isinstance(_a , _a ): token_list.extend(_a ) else: SCREAMING_SNAKE_CASE__ = self.inprogress_constraint.advance() if isinstance(_a , _a ): token_list.append(_a ) elif isinstance(_a , _a ): token_list.extend(_a ) if len(_a ) == 0: return None else: return token_list def __a ( self : Union[str, Any] , _lowercase : str ): """simple docstring""" self.init_state() if token_ids is not None: for token in token_ids: # completes or steps **one** constraint SCREAMING_SNAKE_CASE__ = self.add(_a ) # the entire list of constraints are fulfilled if self.completed: break def __a ( self : Any , _lowercase : int ): """simple docstring""" if not isinstance(_a , _a ): raise ValueError(f"""`token_id` should be an `int`, but is `{token_id}`.""" ) SCREAMING_SNAKE_CASE__ = False, False if self.completed: SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = False return complete, stepped if self.inprogress_constraint is not None: # In the middle of fulfilling a constraint. If the `token_id` *does* makes an incremental progress to current # job, simply update the state SCREAMING_SNAKE_CASE__ = self.inprogress_constraint.update(_a ) if reset: # 1. If the next token breaks the progress, then we must restart. # e.g. constraint = "I love pies" and sequence so far is "I love" but `token_id` == "books". # But that doesn't mean we self.init_state(), since we only reset the state for this particular # constraint, not the full list of constraints. self.pending_constraints.append(self.inprogress_constraint.copy(stateful=_a ) ) SCREAMING_SNAKE_CASE__ = None if complete: # 2. If the next token completes the constraint, move it to completed list, set # inprogress to None. If there are no pending constraints either, then this full list of constraints # is complete. self.complete_constraints.append(self.inprogress_constraint ) SCREAMING_SNAKE_CASE__ = None if len(self.pending_constraints ) == 0: # we're done! SCREAMING_SNAKE_CASE__ = True else: # Not in the middle of fulfilling a constraint. So does this `token_id` helps us step towards any of our list # of constraints? for cidx, pending_constraint in enumerate(self.pending_constraints ): if pending_constraint.does_advance(_a ): SCREAMING_SNAKE_CASE__ = pending_constraint.update(_a ) if not stepped: raise Exception( """`constraint.update(token_id)` is not yielding incremental progress, """ """even though `constraint.does_advance(token_id)` is true.""" ) if complete: self.complete_constraints.append(_a ) SCREAMING_SNAKE_CASE__ = None if not complete and stepped: SCREAMING_SNAKE_CASE__ = pending_constraint if complete or stepped: # If we made any progress at all, then it's at least not a "pending constraint". SCREAMING_SNAKE_CASE__ = ( self.pending_constraints[:cidx] + self.pending_constraints[cidx + 1 :] ) if len(self.pending_constraints ) == 0 and self.inprogress_constraint is None: # If there's no longer any pending after this and no inprogress either, then we must be # complete. SCREAMING_SNAKE_CASE__ = True break # prevent accidentally stepping through multiple constraints with just one token. return complete, stepped def __a ( self : Tuple , _lowercase : List[Any]=True ): """simple docstring""" SCREAMING_SNAKE_CASE__ = ConstraintListState(self.constraints ) # we actually never though self.constraints objects # throughout this process. So it's at initialization state. if stateful: SCREAMING_SNAKE_CASE__ = [ constraint.copy(stateful=_a ) for constraint in self.complete_constraints ] if self.inprogress_constraint is not None: SCREAMING_SNAKE_CASE__ = self.inprogress_constraint.copy(stateful=_a ) SCREAMING_SNAKE_CASE__ = [constraint.copy() for constraint in self.pending_constraints] return new_state
219
from __future__ import annotations class _snake_case : def __init__( self , _a ): __magic_name__ : Optional[Any] = data __magic_name__ : Node | None = None __magic_name__ : Node | None = None def lowerCAmelCase_ ( _snake_case : Node | None ) -> None: # In Order traversal of the tree '''simple docstring''' if tree: display(tree.left ) print(tree.data ) display(tree.right ) def lowerCAmelCase_ ( _snake_case : Node | None ) -> int: '''simple docstring''' return 1 + max(depth_of_tree(tree.left ) , depth_of_tree(tree.right ) ) if tree else 0 def lowerCAmelCase_ ( _snake_case : Node ) -> bool: '''simple docstring''' if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def lowerCAmelCase_ ( ) -> None: # Main function for testing. '''simple docstring''' __magic_name__ : int = Node(1 ) __magic_name__ : Union[str, Any] = Node(2 ) __magic_name__ : Tuple = Node(3 ) __magic_name__ : Optional[Any] = Node(4 ) __magic_name__ : Union[str, Any] = Node(5 ) __magic_name__ : Any = Node(6 ) __magic_name__ : int = Node(7 ) __magic_name__ : List[str] = Node(8 ) __magic_name__ : Union[str, Any] = Node(9 ) print(is_full_binary_tree(_snake_case ) ) print(depth_of_tree(_snake_case ) ) print("Tree is: " ) display(_snake_case ) if __name__ == "__main__": main()
281
0
'''simple docstring''' __snake_case = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} __snake_case = ["a", "b", "c", "d", "e"] def a ( __a , __a , __a ) -> int: '''simple docstring''' UpperCamelCase__ :Any = start # add current to visited visited.append(_snake_case ) UpperCamelCase__ :int = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: UpperCamelCase__ :str = topological_sort(_snake_case , _snake_case , _snake_case ) # if all neighbors visited add current to sort sort.append(_snake_case ) # if all vertices haven't been visited select a new one to visit if len(_snake_case ) != len(_snake_case ): for vertice in vertices: if vertice not in visited: UpperCamelCase__ :Optional[int] = topological_sort(_snake_case , _snake_case , _snake_case ) # return sort return sort if __name__ == "__main__": __snake_case = topological_sort('''a''', [], []) print(sort)
97
def lowerCAmelCase_ ( _snake_case : str , _snake_case : str ) -> bool: '''simple docstring''' __magic_name__ : Union[str, Any] = len(_snake_case ) + 1 __magic_name__ : List[str] = len(_snake_case ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length i of input_string matches with prefix string of length j of # given pattern. # "dp" stands for dynamic programming. __magic_name__ : str = [[0 for i in range(_snake_case )] for j in range(_snake_case )] # since string of zero length match pattern of zero length __magic_name__ : Optional[int] = 1 # since pattern of zero length will never match with string of non-zero length for i in range(1 , _snake_case ): __magic_name__ : Optional[int] = 0 # since string of zero length will match with pattern where there # is at least one * alternatively for j in range(1 , _snake_case ): __magic_name__ : Union[str, Any] = dp[0][j - 2] if pattern[j - 1] == "*" else 0 # now using bottom-up approach to find for all remaining lengths for i in range(1 , _snake_case ): for j in range(1 , _snake_case ): if input_string[i - 1] == pattern[j - 1] or pattern[j - 1] == ".": __magic_name__ : Optional[int] = dp[i - 1][j - 1] elif pattern[j - 1] == "*": if dp[i][j - 2] == 1: __magic_name__ : Optional[Any] = 1 elif pattern[j - 2] in (input_string[i - 1], "."): __magic_name__ : List[Any] = dp[i - 1][j] else: __magic_name__ : Union[str, Any] = 0 else: __magic_name__ : Dict = 0 return bool(dp[-1][-1] ) if __name__ == "__main__": import doctest doctest.testmod() # inputing the strings # input_string = input("input a string :") # pattern = input("input a pattern :") snake_case : Optional[Any] = "aab" snake_case : List[str] = "c*a*b" # using function to check whether given string matches the given pattern if match_pattern(input_string, pattern): print(F"{input_string} matches the given pattern {pattern}") else: print(F"{input_string} does not match with the given pattern {pattern}")
281
0
"""simple docstring""" from collections import OrderedDict from typing import List, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A: List[Any] = logging.get_logger(__name__) A: Optional[int] = { "google/efficientnet-b7": "https://huggingface.co/google/efficientnet-b7/resolve/main/config.json", } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ ): __lowerCAmelCase : Any = 'efficientnet' def __init__( self , _SCREAMING_SNAKE_CASE = 3 , _SCREAMING_SNAKE_CASE = 600 , _SCREAMING_SNAKE_CASE = 2.0 , _SCREAMING_SNAKE_CASE = 3.1 , _SCREAMING_SNAKE_CASE = 8 , _SCREAMING_SNAKE_CASE = [3, 3, 5, 3, 5, 5, 3] , _SCREAMING_SNAKE_CASE = [32, 16, 24, 40, 80, 112, 192] , _SCREAMING_SNAKE_CASE = [16, 24, 40, 80, 112, 192, 320] , _SCREAMING_SNAKE_CASE = [] , _SCREAMING_SNAKE_CASE = [1, 2, 2, 2, 1, 2, 1] , _SCREAMING_SNAKE_CASE = [1, 2, 2, 3, 3, 4, 1] , _SCREAMING_SNAKE_CASE = [1, 6, 6, 6, 6, 6, 6] , _SCREAMING_SNAKE_CASE = 0.25 , _SCREAMING_SNAKE_CASE = "swish" , _SCREAMING_SNAKE_CASE = 2560 , _SCREAMING_SNAKE_CASE = "mean" , _SCREAMING_SNAKE_CASE = 0.02 , _SCREAMING_SNAKE_CASE = 0.001 , _SCREAMING_SNAKE_CASE = 0.99 , _SCREAMING_SNAKE_CASE = 0.5 , _SCREAMING_SNAKE_CASE = 0.2 , **_SCREAMING_SNAKE_CASE , ) -> str: '''simple docstring''' super().__init__(**_a ) UpperCAmelCase : Optional[int] = num_channels UpperCAmelCase : int = image_size UpperCAmelCase : List[Any] = width_coefficient UpperCAmelCase : Dict = depth_coefficient UpperCAmelCase : Union[str, Any] = depth_divisor UpperCAmelCase : str = kernel_sizes UpperCAmelCase : List[Any] = in_channels UpperCAmelCase : Optional[Any] = out_channels UpperCAmelCase : Any = depthwise_padding UpperCAmelCase : Dict = strides UpperCAmelCase : Dict = num_block_repeats UpperCAmelCase : Optional[int] = expand_ratios UpperCAmelCase : int = squeeze_expansion_ratio UpperCAmelCase : Tuple = hidden_act UpperCAmelCase : str = hidden_dim UpperCAmelCase : Dict = pooling_type UpperCAmelCase : Any = initializer_range UpperCAmelCase : Tuple = batch_norm_eps UpperCAmelCase : List[str] = batch_norm_momentum UpperCAmelCase : Any = dropout_rate UpperCAmelCase : List[Any] = drop_connect_rate UpperCAmelCase : Any = sum(_a ) * 4 class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ ): __lowerCAmelCase : Optional[int] = version.parse('1.11' ) @property def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: '''simple docstring''' return 1E-5
109
import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class _snake_case : @staticmethod def SCREAMING_SNAKE_CASE ( *_a , **_a ): pass def lowerCAmelCase_ ( _snake_case : Image ) -> str: '''simple docstring''' __magic_name__ : Optional[int] = hashlib.mda(image.tobytes() ) return m.hexdigest()[:10] def lowerCAmelCase_ ( _snake_case : Image ) -> Dict: '''simple docstring''' __magic_name__ : List[Any] = np.array(_snake_case ) __magic_name__ : Optional[int] = npimg.shape return {"hash": hashimage(_snake_case ), "shape": shape} @is_pipeline_test @require_vision @require_torch class _snake_case ( unittest.TestCase ): UpperCamelCase__ = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) UpperCamelCase__ = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def SCREAMING_SNAKE_CASE ( self , _a , _a , _a ): __magic_name__ : Dict = MaskGenerationPipeline(model=_a , image_processor=_a ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self , _a , _a ): pass @require_tf @unittest.skip("Image segmentation not implemented in TF" ) def SCREAMING_SNAKE_CASE ( self ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = pipeline("mask-generation" , model="facebook/sam-vit-huge" ) __magic_name__ : str = image_segmenter("http://images.cocodataset.org/val2017/000000039769.jpg" , points_per_batch=256 ) # Shortening by hashing __magic_name__ : Dict = [] for i, o in enumerate(outputs["masks"] ): new_outupt += [{"mask": mask_to_test_readable(_a ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(_a , decimals=4 ) , [ {"mask": {"hash": "115ad19f5f", "shape": (480, 640)}, "scores": 1.04_44}, {"mask": {"hash": "6affa964c6", "shape": (480, 640)}, "scores": 1.0_21}, {"mask": {"hash": "dfe28a0388", "shape": (480, 640)}, "scores": 1.01_67}, {"mask": {"hash": "c0a5f4a318", "shape": (480, 640)}, "scores": 1.01_32}, {"mask": {"hash": "fe8065c197", "shape": (480, 640)}, "scores": 1.00_53}, {"mask": {"hash": "e2d0b7a0b7", "shape": (480, 640)}, "scores": 0.99_67}, {"mask": {"hash": "453c7844bd", "shape": (480, 640)}, "scores": 0.9_93}, {"mask": {"hash": "3d44f2926d", "shape": (480, 640)}, "scores": 0.99_09}, {"mask": {"hash": "64033ddc3f", "shape": (480, 640)}, "scores": 0.98_79}, {"mask": {"hash": "801064ff79", "shape": (480, 640)}, "scores": 0.98_34}, {"mask": {"hash": "6172f276ef", "shape": (480, 640)}, "scores": 0.97_16}, {"mask": {"hash": "b49e60e084", "shape": (480, 640)}, "scores": 0.96_12}, {"mask": {"hash": "a811e775fd", "shape": (480, 640)}, "scores": 0.95_99}, {"mask": {"hash": "a6a8ebcf4b", "shape": (480, 640)}, "scores": 0.95_52}, {"mask": {"hash": "9d8257e080", "shape": (480, 640)}, "scores": 0.95_32}, {"mask": {"hash": "32de6454a8", "shape": (480, 640)}, "scores": 0.95_16}, {"mask": {"hash": "af3d4af2c8", "shape": (480, 640)}, "scores": 0.94_99}, {"mask": {"hash": "3c6db475fb", "shape": (480, 640)}, "scores": 0.94_83}, {"mask": {"hash": "c290813fb9", "shape": (480, 640)}, "scores": 0.94_64}, {"mask": {"hash": "b6f0b8f606", "shape": (480, 640)}, "scores": 0.9_43}, {"mask": {"hash": "92ce16bfdf", "shape": (480, 640)}, "scores": 0.9_43}, {"mask": {"hash": "c749b25868", "shape": (480, 640)}, "scores": 0.94_08}, {"mask": {"hash": "efb6cab859", "shape": (480, 640)}, "scores": 0.93_35}, {"mask": {"hash": "1ff2eafb30", "shape": (480, 640)}, "scores": 0.93_26}, {"mask": {"hash": "788b798e24", "shape": (480, 640)}, "scores": 0.92_62}, {"mask": {"hash": "abea804f0e", "shape": (480, 640)}, "scores": 0.89_99}, {"mask": {"hash": "7b9e8ddb73", "shape": (480, 640)}, "scores": 0.89_86}, {"mask": {"hash": "cd24047c8a", "shape": (480, 640)}, "scores": 0.89_84}, {"mask": {"hash": "6943e6bcbd", "shape": (480, 640)}, "scores": 0.88_73}, {"mask": {"hash": "b5f47c9191", "shape": (480, 640)}, "scores": 0.88_71} ] , ) # fmt: on @require_torch @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : str = "facebook/sam-vit-huge" __magic_name__ : str = pipeline("mask-generation" , model=_a ) __magic_name__ : Tuple = image_segmenter( "http://images.cocodataset.org/val2017/000000039769.jpg" , pred_iou_thresh=1 , points_per_batch=256 ) # Shortening by hashing __magic_name__ : Any = [] for i, o in enumerate(outputs["masks"] ): new_outupt += [{"mask": mask_to_test_readable(_a ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(_a , decimals=4 ) , [ {"mask": {"hash": "115ad19f5f", "shape": (480, 640)}, "scores": 1.04_44}, {"mask": {"hash": "6affa964c6", "shape": (480, 640)}, "scores": 1.02_10}, {"mask": {"hash": "dfe28a0388", "shape": (480, 640)}, "scores": 1.01_67}, {"mask": {"hash": "c0a5f4a318", "shape": (480, 640)}, "scores": 1.01_32}, {"mask": {"hash": "fe8065c197", "shape": (480, 640)}, "scores": 1.00_53}, ] , )
281
0
import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput _A = logging.get_logger(__name__) # pylint: disable=invalid-name def __UpperCamelCase ( _A ): warnings.warn( '''The preprocess method is deprecated and will be removed in a future version. Please''' ''' use VaeImageProcessor.preprocess instead''' , _snake_case , ) if isinstance(_snake_case , torch.Tensor ): return image elif isinstance(_snake_case , PIL.Image.Image ): lowerCAmelCase_ = [image] if isinstance(image[0] , PIL.Image.Image ): lowerCAmelCase_ = image[0].size lowerCAmelCase_ = (x - x % 8 for x in (w, h)) # resize to integer multiple of 8 lowerCAmelCase_ = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) )[None, :] for i in image] lowerCAmelCase_ = np.concatenate(_snake_case , axis=0 ) lowerCAmelCase_ = np.array(_snake_case ).astype(np.floataa ) / 2_5_5.0 lowerCAmelCase_ = image.transpose(0 , 3 , 1 , 2 ) lowerCAmelCase_ = 2.0 * image - 1.0 lowerCAmelCase_ = torch.from_numpy(_snake_case ) elif isinstance(image[0] , torch.Tensor ): lowerCAmelCase_ = torch.cat(_snake_case , dim=0 ) return image def __UpperCamelCase ( _A ): if isinstance(_snake_case , torch.Tensor ): return mask elif isinstance(_snake_case , PIL.Image.Image ): lowerCAmelCase_ = [mask] if isinstance(mask[0] , PIL.Image.Image ): lowerCAmelCase_ = mask[0].size lowerCAmelCase_ = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 lowerCAmelCase_ = [np.array(m.convert('''L''' ).resize((w, h) , resample=PIL_INTERPOLATION['''nearest'''] ) )[None, :] for m in mask] lowerCAmelCase_ = np.concatenate(_snake_case , axis=0 ) lowerCAmelCase_ = mask.astype(np.floataa ) / 2_5_5.0 lowerCAmelCase_ = 0 lowerCAmelCase_ = 1 lowerCAmelCase_ = torch.from_numpy(_snake_case ) elif isinstance(mask[0] , torch.Tensor ): lowerCAmelCase_ = torch.cat(_snake_case , dim=0 ) return mask class A ( __UpperCAmelCase ): __snake_case = 42 __snake_case = 42 def __init__( self, UpperCamelCase__, UpperCamelCase__ ): """simple docstring""" super().__init__() self.register_modules(unet=_a, scheduler=_a ) @torch.no_grad() def __call__( self, UpperCamelCase__, UpperCamelCase__, UpperCamelCase__ = 250, UpperCamelCase__ = 0.0, UpperCamelCase__ = 10, UpperCamelCase__ = 10, UpperCamelCase__ = None, UpperCamelCase__ = "pil", UpperCamelCase__ = True, ): """simple docstring""" lowerCAmelCase_ = image lowerCAmelCase_ = _preprocess_image(_a ) lowerCAmelCase_ = original_image.to(device=self.device, dtype=self.unet.dtype ) lowerCAmelCase_ = _preprocess_mask(_a ) lowerCAmelCase_ = mask_image.to(device=self.device, dtype=self.unet.dtype ) lowerCAmelCase_ = original_image.shape[0] # sample gaussian noise to begin the loop if isinstance(_a, _a ) and len(_a ) != batch_size: raise ValueError( f"You have passed a list of generators of length {len(_a )}, but requested an effective batch" f" size of {batch_size}. Make sure the batch size matches the length of the generators." ) lowerCAmelCase_ = original_image.shape lowerCAmelCase_ = randn_tensor(_a, generator=_a, device=self.device, dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(_a, _a, _a, self.device ) lowerCAmelCase_ = eta lowerCAmelCase_ = self.scheduler.timesteps[0] + 1 lowerCAmelCase_ = generator[0] if isinstance(_a, _a ) else generator for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): if t < t_last: # predict the noise residual lowerCAmelCase_ = self.unet(_a, _a ).sample # compute previous image: x_t -> x_t-1 lowerCAmelCase_ = self.scheduler.step(_a, _a, _a, _a, _a, _a ).prev_sample else: # compute the reverse: x_t-1 -> x_t lowerCAmelCase_ = self.scheduler.undo_step(_a, _a, _a ) lowerCAmelCase_ = t lowerCAmelCase_ = (image / 2 + 0.5).clamp(0, 1 ) lowerCAmelCase_ = image.cpu().permute(0, 2, 3, 1 ).numpy() if output_type == "pil": lowerCAmelCase_ = self.numpy_to_pil(_a ) if not return_dict: return (image,) return ImagePipelineOutput(images=_a )
278
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets snake_case : List[Any] = "\\n@inproceedings{lin-2004-rouge,\n title = \"{ROUGE}: A Package for Automatic Evaluation of Summaries\",\n author = \"Lin, Chin-Yew\",\n booktitle = \"Text Summarization Branches Out\",\n month = jul,\n year = \"2004\",\n address = \"Barcelona, Spain\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W04-1013\",\n pages = \"74--81\",\n}\n" snake_case : Any = "\\nROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for\nevaluating automatic summarization and machine translation software in natural language processing.\nThe metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation.\n\nNote that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters.\n\nThis metrics is a wrapper around Google Research reimplementation of ROUGE:\nhttps://github.com/google-research/google-research/tree/master/rouge\n" snake_case : str = "\nCalculates average rouge scores for a list of hypotheses and references\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n rouge_types: A list of rouge types to calculate.\n Valid names:\n `\"rouge{n}\"` (e.g. `\"rouge1\"`, `\"rouge2\"`) where: {n} is the n-gram based scoring,\n `\"rougeL\"`: Longest common subsequence based scoring.\n `\"rougeLSum\"`: rougeLsum splits text using `\"\n\"`.\n See details in https://github.com/huggingface/datasets/issues/617\n use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes.\n use_aggregator: Return aggregates if this is set to True\nReturns:\n rouge1: rouge_1 (precision, recall, f1),\n rouge2: rouge_2 (precision, recall, f1),\n rougeL: rouge_l (precision, recall, f1),\n rougeLsum: rouge_lsum (precision, recall, f1)\nExamples:\n\n >>> rouge = datasets.load_metric('rouge')\n >>> predictions = [\"hello there\", \"general kenobi\"]\n >>> references = [\"hello there\", \"general kenobi\"]\n >>> results = rouge.compute(predictions=predictions, references=references)\n >>> print(list(results.keys()))\n ['rouge1', 'rouge2', 'rougeL', 'rougeLsum']\n >>> print(results[\"rouge1\"])\n AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0))\n >>> print(results[\"rouge1\"].mid.fmeasure)\n 1.0\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _snake_case ( datasets.Metric ): def SCREAMING_SNAKE_CASE ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/google-research/google-research/tree/master/rouge"] , reference_urls=[ "https://en.wikipedia.org/wiki/ROUGE_(metric)", "https://github.com/google-research/google-research/tree/master/rouge", ] , ) def SCREAMING_SNAKE_CASE ( self , _a , _a , _a=None , _a=True , _a=False ): if rouge_types is None: __magic_name__ : str = ["rouge1", "rouge2", "rougeL", "rougeLsum"] __magic_name__ : List[str] = rouge_scorer.RougeScorer(rouge_types=_a , use_stemmer=_a ) if use_aggregator: __magic_name__ : Dict = scoring.BootstrapAggregator() else: __magic_name__ : str = [] for ref, pred in zip(_a , _a ): __magic_name__ : Union[str, Any] = scorer.score(_a , _a ) if use_aggregator: aggregator.add_scores(_a ) else: scores.append(_a ) if use_aggregator: __magic_name__ : Any = aggregator.aggregate() else: __magic_name__ : List[Any] = {} for key in scores[0]: __magic_name__ : str = [score[key] for score in scores] return result
281
0
"""simple docstring""" import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = { "RUCAIBox/mvp": "https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json", } class lowerCAmelCase ( lowerCamelCase_ ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Union[str, Any] = """mvp""" SCREAMING_SNAKE_CASE_ : Tuple = ["""past_key_values"""] SCREAMING_SNAKE_CASE_ : Any = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , lowerCAmelCase__=50_267 , lowerCAmelCase__=1_024 , lowerCAmelCase__=12 , lowerCAmelCase__=4_096 , lowerCAmelCase__=16 , lowerCAmelCase__=12 , lowerCAmelCase__=4_096 , lowerCAmelCase__=16 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__="gelu" , lowerCAmelCase__=1_024 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.02 , lowerCAmelCase__=0.0 , lowerCAmelCase__=False , lowerCAmelCase__=True , lowerCAmelCase__=1 , lowerCAmelCase__=0 , lowerCAmelCase__=2 , lowerCAmelCase__=True , lowerCAmelCase__=2 , lowerCAmelCase__=2 , lowerCAmelCase__=False , lowerCAmelCase__=100 , lowerCAmelCase__=800 , **lowerCAmelCase__ , ) -> Union[str, Any]: SCREAMING_SNAKE_CASE = vocab_size SCREAMING_SNAKE_CASE = max_position_embeddings SCREAMING_SNAKE_CASE = d_model SCREAMING_SNAKE_CASE = encoder_ffn_dim SCREAMING_SNAKE_CASE = encoder_layers SCREAMING_SNAKE_CASE = encoder_attention_heads SCREAMING_SNAKE_CASE = decoder_ffn_dim SCREAMING_SNAKE_CASE = decoder_layers SCREAMING_SNAKE_CASE = decoder_attention_heads SCREAMING_SNAKE_CASE = dropout SCREAMING_SNAKE_CASE = attention_dropout SCREAMING_SNAKE_CASE = activation_dropout SCREAMING_SNAKE_CASE = activation_function SCREAMING_SNAKE_CASE = init_std SCREAMING_SNAKE_CASE = encoder_layerdrop SCREAMING_SNAKE_CASE = decoder_layerdrop SCREAMING_SNAKE_CASE = classifier_dropout SCREAMING_SNAKE_CASE = use_cache SCREAMING_SNAKE_CASE = encoder_layers SCREAMING_SNAKE_CASE = scale_embedding # scale factor will be sqrt(d_model) if True SCREAMING_SNAKE_CASE = use_prompt SCREAMING_SNAKE_CASE = prompt_length SCREAMING_SNAKE_CASE = prompt_mid_dim super().__init__( pad_token_id=_a , bos_token_id=_a , eos_token_id=_a , is_encoder_decoder=_a , decoder_start_token_id=_a , forced_eos_token_id=_a , **_a , ) if self.forced_bos_token_id is None and kwargs.get('force_bos_token_to_be_generated' , _a ): SCREAMING_SNAKE_CASE = self.bos_token_id warnings.warn( F'Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ' 'The config can simply be saved and uploaded again to be fixed.' )
113
snake_case : Optional[int] = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/" def lowerCAmelCase_ ( _snake_case : bytes ) -> bytes: '''simple docstring''' if not isinstance(_snake_case , _snake_case ): __magic_name__ : Tuple = F'''a bytes-like object is required, not \'{data.__class__.__name__}\'''' raise TypeError(_snake_case ) __magic_name__ : Optional[int] = "".join(bin(_snake_case )[2:].zfill(8 ) for byte in data ) __magic_name__ : List[Any] = len(_snake_case ) % 6 != 0 if padding_needed: # The padding that will be added later __magic_name__ : List[str] = B"=" * ((6 - len(_snake_case ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(_snake_case ) % 6) else: __magic_name__ : List[str] = B"" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(_snake_case ) , 6 ) ).encode() + padding ) def lowerCAmelCase_ ( _snake_case : str ) -> bytes: '''simple docstring''' if not isinstance(_snake_case , _snake_case ) and not isinstance(_snake_case , _snake_case ): __magic_name__ : List[str] = ( "argument should be a bytes-like object or ASCII string, " F'''not \'{encoded_data.__class__.__name__}\'''' ) raise TypeError(_snake_case ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(_snake_case , _snake_case ): try: __magic_name__ : List[Any] = encoded_data.decode("utf-8" ) except UnicodeDecodeError: raise ValueError("base64 encoded data should only contain ASCII characters" ) __magic_name__ : List[str] = encoded_data.count("=" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(_snake_case ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one __magic_name__ : Optional[int] = encoded_data[:-padding] __magic_name__ : Dict = "".join( bin(B64_CHARSET.index(_snake_case ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: __magic_name__ : Union[str, Any] = "".join( bin(B64_CHARSET.index(_snake_case ) )[2:].zfill(6 ) for char in encoded_data ) __magic_name__ : List[Any] = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(_snake_case ) , 8 ) ] return bytes(_snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
281
0
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): _lowerCamelCase : Tuple = 0 def A_ ( self ): _lowerCamelCase : Dict = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32' ) self.assertIsInstance(_a , _a ) def A_ ( self ): with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : Tuple = Path(_a ) / "preprocessor_config.json" _lowerCamelCase : str = Path(_a ) / "config.json" json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(_a , 'w' ) ) _lowerCamelCase : List[str] = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def A_ ( self ): # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : Dict = Path(_a ) / "preprocessor_config.json" _lowerCamelCase : List[str] = Path(_a ) / "config.json" json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(_a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(_a , 'w' ) ) _lowerCamelCase : List[Any] = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def A_ ( self ): with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : List[str] = CLIPConfig() # Create a dummy config file with image_proceesor_type _lowerCamelCase : Tuple = Path(_a ) / "preprocessor_config.json" _lowerCamelCase : int = Path(_a ) / "config.json" json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(_a , 'w' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally _lowerCamelCase : Tuple = AutoImageProcessor.from_pretrained(_a ).to_dict() config_dict.pop('image_processor_type' ) _lowerCamelCase : List[Any] = CLIPImageProcessor(**_a ) # save in new folder model_config.save_pretrained(_a ) config.save_pretrained(_a ) _lowerCamelCase : Tuple = AutoImageProcessor.from_pretrained(_a ) # make sure private variable is not incorrectly saved _lowerCamelCase : int = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(_a , _a ) def A_ ( self ): with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : Optional[int] = Path(_a ) / "preprocessor_config.json" json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(_a , 'w' ) , ) _lowerCamelCase : Optional[int] = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def A_ ( self ): with self.assertRaisesRegex( _a , 'clip-base is not a local folder and is not a valid model identifier' ): _lowerCamelCase : Union[str, Any] = AutoImageProcessor.from_pretrained('clip-base' ) def A_ ( self ): with self.assertRaisesRegex( _a , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): _lowerCamelCase : List[Any] = AutoImageProcessor.from_pretrained(_a , revision='aaaaaa' ) def A_ ( self ): with self.assertRaisesRegex( _a , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): _lowerCamelCase : str = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model' ) def A_ ( self ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(_a ): _lowerCamelCase : List[str] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(_a ): _lowerCamelCase : Union[str, Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_a ) _lowerCamelCase : Any = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_a ) _lowerCamelCase : List[str] = AutoImageProcessor.from_pretrained(_a , trust_remote_code=_a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor' ) def A_ ( self ): try: AutoConfig.register('custom' , _a ) AutoImageProcessor.register(_a , _a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_a ): AutoImageProcessor.register(_a , _a ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : Dict = Path(_a ) / "preprocessor_config.json" _lowerCamelCase : Union[str, Any] = Path(_a ) / "config.json" json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(_a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(_a , 'w' ) ) _lowerCamelCase : Dict = CustomImageProcessor.from_pretrained(_a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_a ) _lowerCamelCase : Any = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def A_ ( self ): class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = True try: AutoConfig.register('custom' , _a ) AutoImageProcessor.register(_a , _a ) # If remote code is not set, the default is to use local _lowerCamelCase : Optional[int] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. _lowerCamelCase : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub _lowerCamelCase : List[Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=_a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(not hasattr(_a , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
96
import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class _snake_case ( unittest.TestCase ): def __init__( self , _a , _a=13 , _a=7 , _a=True , _a=True , _a=True , _a=True , _a=99 , _a=32 , _a=5 , _a=4 , _a=37 , _a="gelu" , _a=0.1 , _a=0.1 , _a=512 , _a=16 , _a=2 , _a=0.02 , _a=4 , ): __magic_name__ : List[Any] = parent __magic_name__ : Optional[Any] = batch_size __magic_name__ : Dict = seq_length __magic_name__ : Union[str, Any] = is_training __magic_name__ : Optional[Any] = use_attention_mask __magic_name__ : Optional[Any] = use_token_type_ids __magic_name__ : int = use_labels __magic_name__ : List[Any] = vocab_size __magic_name__ : Union[str, Any] = hidden_size __magic_name__ : Optional[Any] = num_hidden_layers __magic_name__ : int = num_attention_heads __magic_name__ : Any = intermediate_size __magic_name__ : List[Any] = hidden_act __magic_name__ : List[Any] = hidden_dropout_prob __magic_name__ : Optional[int] = attention_probs_dropout_prob __magic_name__ : List[Any] = max_position_embeddings __magic_name__ : Tuple = type_vocab_size __magic_name__ : List[str] = type_sequence_label_size __magic_name__ : Dict = initializer_range __magic_name__ : List[Any] = num_choices def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __magic_name__ : List[Any] = None if self.use_attention_mask: __magic_name__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) __magic_name__ : str = None if self.use_token_type_ids: __magic_name__ : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __magic_name__ : List[str] = RobertaPreLayerNormConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_a , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : int = self.prepare_config_and_inputs() __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : List[Any] = config_and_inputs __magic_name__ : List[str] = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = self.prepare_config_and_inputs() __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ : Union[str, Any] = config_and_inputs __magic_name__ : Tuple = True __magic_name__ : int = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __magic_name__ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class _snake_case ( snake_case , unittest.TestCase ): UpperCamelCase__ = True UpperCamelCase__ = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[Any] = FlaxRobertaPreLayerNormModelTester(self ) @slow def SCREAMING_SNAKE_CASE ( self ): for model_class_name in self.all_model_classes: __magic_name__ : Optional[Any] = model_class_name.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=_a ) __magic_name__ : Dict = model(np.ones((1, 1) ) ) self.assertIsNotNone(_a ) @require_flax class _snake_case ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = FlaxRobertaPreLayerNormForMaskedLM.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=_a ) __magic_name__ : Union[str, Any] = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) __magic_name__ : List[str] = model(_a )[0] __magic_name__ : str = [1, 11, 50_265] self.assertEqual(list(output.shape ) , _a ) # compare the actual values for a slice. __magic_name__ : List[str] = np.array( [[[40.48_80, 18.01_99, -5.23_67], [-1.88_77, -4.08_85, 10.70_85], [-2.26_13, -5.61_10, 7.26_65]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , _a , atol=1e-4 ) ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[str] = FlaxRobertaPreLayerNormModel.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=_a ) __magic_name__ : Tuple = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) __magic_name__ : Tuple = model(_a )[0] # compare the actual values for a slice. __magic_name__ : Dict = np.array( [[[0.02_08, -0.03_56, 0.02_37], [-0.15_69, -0.04_11, -0.26_26], [0.18_79, 0.01_25, -0.00_89]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , _a , atol=1e-4 ) )
281
0
"""simple docstring""" import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase_ ( lowerCAmelCase , unittest.TestCase ): """simple docstring""" _lowerCAmelCase : Optional[Any] = LxmertTokenizer _lowerCAmelCase : Any = LxmertTokenizerFast _lowerCAmelCase : Any = True _lowerCAmelCase : List[Any] = True def snake_case ( self ): """simple docstring""" super().setUp() snake_case = [ "[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def snake_case ( self , lowerCAmelCase ): """simple docstring""" snake_case = "UNwant\u00E9d,running" snake_case = "unwanted, running" return input_text, output_text def snake_case ( self ): """simple docstring""" snake_case = self.tokenizer_class(self.vocab_file ) snake_case = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(_a , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , [7, 4, 5, 10, 8, 9] ) def snake_case ( self ): """simple docstring""" if not self.test_rust_tokenizer: return snake_case = self.get_tokenizer() snake_case = self.get_rust_tokenizer() snake_case = "I was born in 92000, and this is falsé." snake_case = tokenizer.tokenize(_a ) snake_case = rust_tokenizer.tokenize(_a ) self.assertListEqual(_a , _a ) snake_case = tokenizer.encode(_a , add_special_tokens=_a ) snake_case = rust_tokenizer.encode(_a , add_special_tokens=_a ) self.assertListEqual(_a , _a ) snake_case = self.get_rust_tokenizer() snake_case = tokenizer.encode(_a ) snake_case = rust_tokenizer.encode(_a ) self.assertListEqual(_a , _a )
150
def lowerCAmelCase_ ( _snake_case : list[list[int | float]] ) -> int: '''simple docstring''' __magic_name__ : Any = len(_snake_case ) __magic_name__ : Optional[Any] = len(matrix[0] ) __magic_name__ : Union[str, Any] = min(_snake_case , _snake_case ) for row in range(_snake_case ): # Check if diagonal element is not zero if matrix[row][row] != 0: # Eliminate all the elements below the diagonal for col in range(row + 1 , _snake_case ): __magic_name__ : Optional[Any] = matrix[col][row] / matrix[row][row] for i in range(_snake_case , _snake_case ): matrix[col][i] -= multiplier * matrix[row][i] else: # Find a non-zero diagonal element to swap rows __magic_name__ : str = True for i in range(row + 1 , _snake_case ): if matrix[i][row] != 0: __magic_name__ , __magic_name__ : List[str] = matrix[i], matrix[row] __magic_name__ : Union[str, Any] = False break if reduce: rank -= 1 for i in range(_snake_case ): __magic_name__ : Any = matrix[i][rank] # Reduce the row pointer by one to stay on the same row row -= 1 return rank if __name__ == "__main__": import doctest doctest.testmod()
281
0
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING UpperCAmelCase : int = logging.get_logger(__name__) @add_end_docstrings(A ) class lowerCamelCase__ ( A ): """simple docstring""" def __init__( self : int , *UpperCamelCase : Dict , **UpperCamelCase : Tuple ): '''simple docstring''' super().__init__(*_a , **_a ) requires_backends(self , """vision""" ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == """tf""" else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def lowerCamelCase__ ( self : List[Any] , UpperCamelCase : Dict=None ): '''simple docstring''' __UpperCAmelCase : Dict = {} if top_k is not None: __UpperCAmelCase : Optional[Any] = top_k return {}, {}, postprocess_params def __call__( self : Union[str, Any] , UpperCamelCase : int , **UpperCamelCase : Union[str, Any] ): '''simple docstring''' return super().__call__(_a , **_a ) def lowerCamelCase__ ( self : List[str] , UpperCamelCase : Tuple ): '''simple docstring''' __UpperCAmelCase : Dict = load_image(_a ) __UpperCAmelCase : List[Any] = self.image_processor(images=_a , return_tensors=self.framework ) return model_inputs def lowerCamelCase__ ( self : int , UpperCamelCase : Tuple ): '''simple docstring''' __UpperCAmelCase : Any = self.model(**_a ) return model_outputs def lowerCamelCase__ ( self : Dict , UpperCamelCase : Optional[int] , UpperCamelCase : List[Any]=5 ): '''simple docstring''' if top_k > self.model.config.num_labels: __UpperCAmelCase : Optional[int] = self.model.config.num_labels if self.framework == "pt": __UpperCAmelCase : Any = model_outputs.logits.softmax(-1 )[0] __UpperCAmelCase : List[Any] = probs.topk(_a ) elif self.framework == "tf": __UpperCAmelCase : int = stable_softmax(model_outputs.logits , axis=-1 )[0] __UpperCAmelCase : List[str] = tf.math.top_k(_a , k=_a ) __UpperCAmelCase : Optional[int] = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f'''Unsupported framework: {self.framework}''' ) __UpperCAmelCase : Dict = scores.tolist() __UpperCAmelCase : str = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(_a , _a )]
115
import argparse import collections import json import os import re import string import sys import numpy as np snake_case : Dict = re.compile(R"\b(a|an|the)\b", re.UNICODE) snake_case : Optional[int] = None def lowerCAmelCase_ ( ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : Any = argparse.ArgumentParser("Official evaluation script for SQuAD version 2.0." ) parser.add_argument("data_file" , metavar="data.json" , help="Input data JSON file." ) parser.add_argument("pred_file" , metavar="pred.json" , help="Model predictions." ) parser.add_argument( "--out-file" , "-o" , metavar="eval.json" , help="Write accuracy metrics to file (default is stdout)." ) parser.add_argument( "--na-prob-file" , "-n" , metavar="na_prob.json" , help="Model estimates of probability of no answer." ) parser.add_argument( "--na-prob-thresh" , "-t" , type=_snake_case , default=1.0 , help="Predict \"\" if no-answer probability exceeds this (default = 1.0)." , ) parser.add_argument( "--out-image-dir" , "-p" , metavar="out_images" , default=_snake_case , help="Save precision-recall curves to directory." ) parser.add_argument("--verbose" , "-v" , action="store_true" ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def lowerCAmelCase_ ( _snake_case : Optional[Any] ) -> Tuple: '''simple docstring''' __magic_name__ : Optional[int] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: __magic_name__ : str = bool(qa["answers"]["text"] ) return qid_to_has_ans def lowerCAmelCase_ ( _snake_case : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' def remove_articles(_snake_case : List[str] ): return ARTICLES_REGEX.sub(" " , _snake_case ) def white_space_fix(_snake_case : Optional[int] ): return " ".join(text.split() ) def remove_punc(_snake_case : Optional[int] ): __magic_name__ : Dict = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_snake_case : str ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_snake_case ) ) ) ) def lowerCAmelCase_ ( _snake_case : Any ) -> Optional[Any]: '''simple docstring''' if not s: return [] return normalize_answer(_snake_case ).split() def lowerCAmelCase_ ( _snake_case : str , _snake_case : Dict ) -> Tuple: '''simple docstring''' return int(normalize_answer(_snake_case ) == normalize_answer(_snake_case ) ) def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : int ) -> str: '''simple docstring''' __magic_name__ : Any = get_tokens(_snake_case ) __magic_name__ : Optional[int] = get_tokens(_snake_case ) __magic_name__ : Tuple = collections.Counter(_snake_case ) & collections.Counter(_snake_case ) __magic_name__ : Tuple = sum(common.values() ) if len(_snake_case ) == 0 or len(_snake_case ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 __magic_name__ : Dict = 1.0 * num_same / len(_snake_case ) __magic_name__ : Optional[Any] = 1.0 * num_same / len(_snake_case ) __magic_name__ : List[Any] = (2 * precision * recall) / (precision + recall) return fa def lowerCAmelCase_ ( _snake_case : Optional[Any] , _snake_case : List[Any] ) -> List[Any]: '''simple docstring''' __magic_name__ : Union[str, Any] = {} __magic_name__ : int = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: __magic_name__ : Union[str, Any] = qa["id"] __magic_name__ : Any = [t for t in qa["answers"]["text"] if normalize_answer(_snake_case )] if not gold_answers: # For unanswerable questions, only correct answer is empty string __magic_name__ : Tuple = [""] if qid not in preds: print(F'''Missing prediction for {qid}''' ) continue __magic_name__ : Any = preds[qid] # Take max over all gold answers __magic_name__ : List[Any] = max(compute_exact(_snake_case , _snake_case ) for a in gold_answers ) __magic_name__ : int = max(compute_fa(_snake_case , _snake_case ) for a in gold_answers ) return exact_scores, fa_scores def lowerCAmelCase_ ( _snake_case : Optional[Any] , _snake_case : List[Any] , _snake_case : Optional[int] , _snake_case : Dict ) -> Union[str, Any]: '''simple docstring''' __magic_name__ : str = {} for qid, s in scores.items(): __magic_name__ : Dict = na_probs[qid] > na_prob_thresh if pred_na: __magic_name__ : str = float(not qid_to_has_ans[qid] ) else: __magic_name__ : Optional[int] = s return new_scores def lowerCAmelCase_ ( _snake_case : List[Any] , _snake_case : List[str] , _snake_case : Tuple=None ) -> Tuple: '''simple docstring''' if not qid_list: __magic_name__ : Any = len(_snake_case ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores.values() ) / total), ("f1", 100.0 * sum(fa_scores.values() ) / total), ("total", total), ] ) else: __magic_name__ : Tuple = len(_snake_case ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores[k] for k in qid_list ) / total), ("f1", 100.0 * sum(fa_scores[k] for k in qid_list ) / total), ("total", total), ] ) def lowerCAmelCase_ ( _snake_case : Optional[int] , _snake_case : str , _snake_case : str ) -> Dict: '''simple docstring''' for k in new_eval: __magic_name__ : int = new_eval[k] def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Dict , _snake_case : Optional[Any] , _snake_case : Union[str, Any] ) -> str: '''simple docstring''' plt.step(_snake_case , _snake_case , color="b" , alpha=0.2 , where="post" ) plt.fill_between(_snake_case , _snake_case , step="post" , alpha=0.2 , color="b" ) plt.xlabel("Recall" ) plt.ylabel("Precision" ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(_snake_case ) plt.savefig(_snake_case ) plt.clf() def lowerCAmelCase_ ( _snake_case : Dict , _snake_case : Any , _snake_case : Optional[int] , _snake_case : List[Any] , _snake_case : Optional[int]=None , _snake_case : int=None ) -> str: '''simple docstring''' __magic_name__ : Union[str, Any] = sorted(_snake_case , key=lambda _snake_case : na_probs[k] ) __magic_name__ : Optional[int] = 0.0 __magic_name__ : str = 1.0 __magic_name__ : str = 0.0 __magic_name__ : List[str] = [1.0] __magic_name__ : str = [0.0] __magic_name__ : Optional[Any] = 0.0 for i, qid in enumerate(_snake_case ): if qid_to_has_ans[qid]: true_pos += scores[qid] __magic_name__ : List[str] = true_pos / float(i + 1 ) __magic_name__ : Any = true_pos / float(_snake_case ) if i == len(_snake_case ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(_snake_case ) recalls.append(_snake_case ) if out_image: plot_pr_curve(_snake_case , _snake_case , _snake_case , _snake_case ) return {"ap": 100.0 * avg_prec} def lowerCAmelCase_ ( _snake_case : Tuple , _snake_case : Optional[Any] , _snake_case : Optional[int] , _snake_case : Optional[Any] , _snake_case : Any , _snake_case : List[Any] ) -> Union[str, Any]: '''simple docstring''' if out_image_dir and not os.path.exists(_snake_case ): os.makedirs(_snake_case ) __magic_name__ : Any = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return __magic_name__ : str = make_precision_recall_eval( _snake_case , _snake_case , _snake_case , _snake_case , out_image=os.path.join(_snake_case , "pr_exact.png" ) , title="Precision-Recall curve for Exact Match score" , ) __magic_name__ : Union[str, Any] = make_precision_recall_eval( _snake_case , _snake_case , _snake_case , _snake_case , out_image=os.path.join(_snake_case , "pr_f1.png" ) , title="Precision-Recall curve for F1 score" , ) __magic_name__ : str = {k: float(_snake_case ) for k, v in qid_to_has_ans.items()} __magic_name__ : str = make_precision_recall_eval( _snake_case , _snake_case , _snake_case , _snake_case , out_image=os.path.join(_snake_case , "pr_oracle.png" ) , title="Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)" , ) merge_eval(_snake_case , _snake_case , "pr_exact" ) merge_eval(_snake_case , _snake_case , "pr_f1" ) merge_eval(_snake_case , _snake_case , "pr_oracle" ) def lowerCAmelCase_ ( _snake_case : int , _snake_case : Optional[Any] , _snake_case : List[str] , _snake_case : Optional[Any] ) -> Dict: '''simple docstring''' if not qid_list: return __magic_name__ : Dict = [na_probs[k] for k in qid_list] __magic_name__ : str = np.ones_like(_snake_case ) / float(len(_snake_case ) ) plt.hist(_snake_case , weights=_snake_case , bins=20 , range=(0.0, 1.0) ) plt.xlabel("Model probability of no-answer" ) plt.ylabel("Proportion of dataset" ) plt.title(F'''Histogram of no-answer probability: {name}''' ) plt.savefig(os.path.join(_snake_case , F'''na_prob_hist_{name}.png''' ) ) plt.clf() def lowerCAmelCase_ ( _snake_case : Union[str, Any] , _snake_case : Tuple , _snake_case : List[str] , _snake_case : Dict ) -> List[Any]: '''simple docstring''' __magic_name__ : Union[str, Any] = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) __magic_name__ : List[str] = num_no_ans __magic_name__ : Dict = cur_score __magic_name__ : Dict = 0.0 __magic_name__ : Any = sorted(_snake_case , key=lambda _snake_case : na_probs[k] ) for i, qid in enumerate(_snake_case ): if qid not in scores: continue if qid_to_has_ans[qid]: __magic_name__ : Union[str, Any] = scores[qid] else: if preds[qid]: __magic_name__ : List[Any] = -1 else: __magic_name__ : Optional[int] = 0 cur_score += diff if cur_score > best_score: __magic_name__ : Optional[int] = cur_score __magic_name__ : List[Any] = na_probs[qid] return 100.0 * best_score / len(_snake_case ), best_thresh def lowerCAmelCase_ ( _snake_case : int , _snake_case : str , _snake_case : List[str] , _snake_case : Tuple , _snake_case : List[Any] , _snake_case : Dict ) -> Optional[Any]: '''simple docstring''' __magic_name__ , __magic_name__ : List[str] = find_best_thresh(_snake_case , _snake_case , _snake_case , _snake_case ) __magic_name__ , __magic_name__ : int = find_best_thresh(_snake_case , _snake_case , _snake_case , _snake_case ) __magic_name__ : Optional[int] = best_exact __magic_name__ : List[Any] = exact_thresh __magic_name__ : Dict = best_fa __magic_name__ : Any = fa_thresh def lowerCAmelCase_ ( ) -> int: '''simple docstring''' with open(OPTS.data_file ) as f: __magic_name__ : Optional[Any] = json.load(_snake_case ) __magic_name__ : List[Any] = dataset_json["data"] with open(OPTS.pred_file ) as f: __magic_name__ : Optional[Any] = json.load(_snake_case ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: __magic_name__ : Any = json.load(_snake_case ) else: __magic_name__ : Any = {k: 0.0 for k in preds} __magic_name__ : str = make_qid_to_has_ans(_snake_case ) # maps qid to True/False __magic_name__ : Tuple = [k for k, v in qid_to_has_ans.items() if v] __magic_name__ : Optional[Any] = [k for k, v in qid_to_has_ans.items() if not v] __magic_name__ , __magic_name__ : Union[str, Any] = get_raw_scores(_snake_case , _snake_case ) __magic_name__ : Optional[Any] = apply_no_ans_threshold(_snake_case , _snake_case , _snake_case , OPTS.na_prob_thresh ) __magic_name__ : Optional[Any] = apply_no_ans_threshold(_snake_case , _snake_case , _snake_case , OPTS.na_prob_thresh ) __magic_name__ : List[Any] = make_eval_dict(_snake_case , _snake_case ) if has_ans_qids: __magic_name__ : int = make_eval_dict(_snake_case , _snake_case , qid_list=_snake_case ) merge_eval(_snake_case , _snake_case , "HasAns" ) if no_ans_qids: __magic_name__ : List[Any] = make_eval_dict(_snake_case , _snake_case , qid_list=_snake_case ) merge_eval(_snake_case , _snake_case , "NoAns" ) if OPTS.na_prob_file: find_all_best_thresh(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case , OPTS.out_image_dir ) histogram_na_prob(_snake_case , _snake_case , OPTS.out_image_dir , "hasAns" ) histogram_na_prob(_snake_case , _snake_case , OPTS.out_image_dir , "noAns" ) if OPTS.out_file: with open(OPTS.out_file , "w" ) as f: json.dump(_snake_case , _snake_case ) else: print(json.dumps(_snake_case , indent=2 ) ) if __name__ == "__main__": snake_case : int = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("Agg") import matplotlib.pyplot as plt main()
281
0
import unittest from queue import Empty from threading import Thread from transformers import AutoTokenizer, TextIteratorStreamer, TextStreamer, is_torch_available from transformers.testing_utils import CaptureStdout, require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers import AutoModelForCausalLM @require_torch class A_ ( unittest.TestCase ): def UpperCAmelCase ( self : List[Any] ) -> Dict: __lowerCAmelCase: Dict = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) __lowerCAmelCase: Optional[Any] = AutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ).to(_a ) __lowerCAmelCase: str = -1 __lowerCAmelCase: Any = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(_a ) __lowerCAmelCase: Any = model.generate(_a , max_new_tokens=1_0 , do_sample=_a ) __lowerCAmelCase: str = tokenizer.decode(greedy_ids[0] ) with CaptureStdout() as cs: __lowerCAmelCase: List[str] = TextStreamer(_a ) model.generate(_a , max_new_tokens=1_0 , do_sample=_a , streamer=_a ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer __lowerCAmelCase: Any = cs.out[:-1] self.assertEqual(_a , _a ) def UpperCAmelCase ( self : List[str] ) -> Any: __lowerCAmelCase: List[str] = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) __lowerCAmelCase: List[str] = AutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ).to(_a ) __lowerCAmelCase: List[str] = -1 __lowerCAmelCase: str = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(_a ) __lowerCAmelCase: Optional[Any] = model.generate(_a , max_new_tokens=1_0 , do_sample=_a ) __lowerCAmelCase: str = tokenizer.decode(greedy_ids[0] ) __lowerCAmelCase: Dict = TextIteratorStreamer(_a ) __lowerCAmelCase: List[Any] = {"input_ids": input_ids, "max_new_tokens": 1_0, "do_sample": False, "streamer": streamer} __lowerCAmelCase: Any = Thread(target=model.generate , kwargs=_a ) thread.start() __lowerCAmelCase: Dict = "" for new_text in streamer: streamer_text += new_text self.assertEqual(_a , _a ) def UpperCAmelCase ( self : List[str] ) -> Optional[Any]: __lowerCAmelCase: Union[str, Any] = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) __lowerCAmelCase: List[Any] = AutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ).to(_a ) __lowerCAmelCase: str = -1 __lowerCAmelCase: List[str] = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(_a ) __lowerCAmelCase: int = model.generate(_a , max_new_tokens=1_0 , do_sample=_a ) __lowerCAmelCase: List[Any] = greedy_ids[:, input_ids.shape[1] :] __lowerCAmelCase: int = tokenizer.decode(new_greedy_ids[0] ) with CaptureStdout() as cs: __lowerCAmelCase: str = TextStreamer(_a , skip_prompt=_a ) model.generate(_a , max_new_tokens=1_0 , do_sample=_a , streamer=_a ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer __lowerCAmelCase: Dict = cs.out[:-1] self.assertEqual(_a , _a ) def UpperCAmelCase ( self : Optional[Any] ) -> str: # Tests that we can pass `decode_kwargs` to the streamer to control how the tokens are decoded. Must be tested # with actual models -- the dummy models' tokenizers are not aligned with their models, and # `skip_special_tokens=True` has no effect on them __lowerCAmelCase: Tuple = AutoTokenizer.from_pretrained('distilgpt2' ) __lowerCAmelCase: List[Any] = AutoModelForCausalLM.from_pretrained('distilgpt2' ).to(_a ) __lowerCAmelCase: List[str] = -1 __lowerCAmelCase: Optional[Any] = torch.ones((1, 5) , device=_a ).long() * model.config.bos_token_id with CaptureStdout() as cs: __lowerCAmelCase: Union[str, Any] = TextStreamer(_a , skip_special_tokens=_a ) model.generate(_a , max_new_tokens=1 , do_sample=_a , streamer=_a ) # The prompt contains a special token, so the streamer should not print it. As such, the output text, when # re-tokenized, must only contain one token __lowerCAmelCase: Union[str, Any] = cs.out[:-1] # Remove the final "\n" __lowerCAmelCase: List[str] = tokenizer(_a , return_tensors='pt' ) self.assertEqual(streamer_text_tokenized.input_ids.shape , (1, 1) ) def UpperCAmelCase ( self : Dict ) -> Dict: __lowerCAmelCase: Optional[Any] = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) __lowerCAmelCase: Optional[int] = AutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ).to(_a ) __lowerCAmelCase: Optional[Any] = -1 __lowerCAmelCase: int = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(_a ) __lowerCAmelCase: List[str] = TextIteratorStreamer(_a , timeout=0.001 ) __lowerCAmelCase: Optional[int] = {"input_ids": input_ids, "max_new_tokens": 1_0, "do_sample": False, "streamer": streamer} __lowerCAmelCase: str = Thread(target=model.generate , kwargs=_a ) thread.start() # The streamer will timeout after 0.001 seconds, so an exception will be raised with self.assertRaises(_a ): __lowerCAmelCase: str = "" for new_text in streamer: streamer_text += new_text
322
import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin snake_case : str = "▁" snake_case : List[Any] = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class _snake_case ( snake_case , unittest.TestCase ): UpperCamelCase__ = BigBirdTokenizer UpperCamelCase__ = BigBirdTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True def SCREAMING_SNAKE_CASE ( self ): super().setUp() __magic_name__ : Optional[Any] = self.tokenizer_class(_a , keep_accents=_a ) tokenizer.save_pretrained(self.tmpdirname ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Union[str, Any] = "<s>" __magic_name__ : Dict = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_a ) , _a ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_a ) , _a ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "[MASK]" ) self.assertEqual(len(_a ) , 1_004 ) def SCREAMING_SNAKE_CASE ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_000 ) def SCREAMING_SNAKE_CASE ( self ): if not self.test_rust_tokenizer: return __magic_name__ : Dict = self.get_tokenizer() __magic_name__ : str = self.get_rust_tokenizer() __magic_name__ : Any = "I was born in 92000, and this is falsé." __magic_name__ : Dict = tokenizer.tokenize(_a ) __magic_name__ : Any = rust_tokenizer.tokenize(_a ) self.assertListEqual(_a , _a ) __magic_name__ : List[Any] = tokenizer.encode(_a , add_special_tokens=_a ) __magic_name__ : List[str] = rust_tokenizer.encode(_a , add_special_tokens=_a ) self.assertListEqual(_a , _a ) __magic_name__ : str = self.get_rust_tokenizer() __magic_name__ : Dict = tokenizer.encode(_a ) __magic_name__ : Optional[int] = rust_tokenizer.encode(_a ) self.assertListEqual(_a , _a ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Optional[int] = BigBirdTokenizer(_a , keep_accents=_a ) __magic_name__ : str = tokenizer.tokenize("This is a test" ) self.assertListEqual(_a , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_a ) , [285, 46, 10, 170, 382] , ) __magic_name__ : Dict = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( _a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) __magic_name__ : Union[str, Any] = tokenizer.convert_tokens_to_ids(_a ) self.assertListEqual( _a , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) __magic_name__ : int = tokenizer.convert_ids_to_tokens(_a ) self.assertListEqual( _a , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def SCREAMING_SNAKE_CASE ( self ): return BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Any = "Hello World!" __magic_name__ : Dict = [65, 18_536, 2_260, 101, 66] self.assertListEqual(_a , self.big_tokenizer.encode(_a ) ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : Dict = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) # fmt: off __magic_name__ : List[str] = [65, 871, 419, 358, 946, 991, 2_521, 452, 358, 1_357, 387, 7_751, 3_536, 112, 985, 456, 126, 865, 938, 5_400, 5_734, 458, 1_368, 467, 786, 2_462, 5_246, 1_159, 633, 865, 4_519, 457, 582, 852, 2_557, 427, 916, 508, 405, 34_324, 497, 391, 408, 11_342, 1_244, 385, 100, 938, 985, 456, 574, 362, 12_597, 3_200, 3_129, 1_172, 66] # noqa: E231 # fmt: on self.assertListEqual(_a , self.big_tokenizer.encode(_a ) ) @require_torch @slow def SCREAMING_SNAKE_CASE ( self ): import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence __magic_name__ : Optional[Any] = list(self.big_tokenizer.get_vocab().keys() )[:10] __magic_name__ : List[Any] = " ".join(_a ) __magic_name__ : Any = self.big_tokenizer.encode_plus(_a , return_tensors="pt" , return_token_type_ids=_a ) __magic_name__ : Union[str, Any] = self.big_tokenizer.batch_encode_plus( [sequence + " " + sequence] , return_tensors="pt" , return_token_type_ids=_a ) __magic_name__ : List[str] = BigBirdConfig(attention_type="original_full" ) __magic_name__ : Optional[int] = BigBirdModel(_a ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_a ) model(**_a ) @slow def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : int = BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) __magic_name__ : int = tokenizer.decode(tokenizer("Paris is the [MASK]." ).input_ids ) self.assertTrue(decoded_text == "[CLS] Paris is the[MASK].[SEP]" ) @slow def SCREAMING_SNAKE_CASE ( self ): # fmt: off __magic_name__ : Optional[Any] = {"input_ids": [[65, 39_286, 458, 36_335, 2_001, 456, 13_073, 13_266, 455, 113, 7_746, 1_741, 11_157, 391, 13_073, 13_266, 455, 113, 3_967, 35_412, 113, 4_936, 109, 3_870, 2_377, 113, 30_084, 45_720, 458, 134, 17_496, 112, 503, 11_672, 113, 118, 112, 5_665, 13_347, 38_687, 112, 1_496, 31_389, 112, 3_268, 47_264, 134, 962, 112, 16_377, 8_035, 23_130, 430, 12_169, 15_518, 28_592, 458, 146, 41_697, 109, 391, 12_169, 15_518, 16_689, 458, 146, 41_358, 109, 452, 726, 4_034, 111, 763, 35_412, 5_082, 388, 1_903, 111, 9_051, 391, 2_870, 48_918, 1_900, 1_123, 550, 998, 112, 9_586, 15_985, 455, 391, 410, 22_955, 37_636, 114, 66], [65, 448, 17_496, 419, 3_663, 385, 763, 113, 27_533, 2_870, 3_283, 13_043, 1_639, 24_713, 523, 656, 24_013, 18_550, 2_521, 517, 27_014, 21_244, 420, 1_212, 1_465, 391, 927, 4_833, 388, 578, 11_786, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 484, 2_169, 7_687, 21_932, 18_146, 726, 363, 17_032, 3_391, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_a , model_name="google/bigbird-roberta-base" , revision="215c99f1600e06f83acce68422f2035b2b5c3510" , )
281
0
import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser _snake_case = re.compile(r'''\s+''') def _UpperCamelCase ( snake_case__ ) -> Dict: return {"hash": hashlib.mda(re.sub(_snake_case, "", example["content"] ).encode("utf-8" ) ).hexdigest()} def _UpperCamelCase ( snake_case__ ) -> List[str]: __UpperCAmelCase : str = [len(_snake_case ) for line in example["content"].splitlines()] return {"line_mean": np.mean(_snake_case ), "line_max": max(_snake_case )} def _UpperCamelCase ( snake_case__ ) -> Tuple: __UpperCAmelCase : List[Any] = np.mean([c.isalnum() for c in example["content"]] ) return {"alpha_frac": alpha_frac} def _UpperCamelCase ( snake_case__, snake_case__ ) -> Optional[Any]: if example["hash"] in uniques: uniques.remove(example["hash"] ) return True else: return False def _UpperCamelCase ( snake_case__, snake_case__=5 ) -> Any: __UpperCAmelCase : List[Any] = ["auto-generated", "autogenerated", "automatically generated"] __UpperCAmelCase : Optional[int] = example["content"].splitlines() for _, line in zip(range(_snake_case ), _snake_case ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def _UpperCamelCase ( snake_case__, snake_case__=5, snake_case__=0.05 ) -> Tuple: __UpperCAmelCase : List[Any] = ["unit tests", "test file", "configuration file"] __UpperCAmelCase : Any = example["content"].splitlines() __UpperCAmelCase : Optional[Any] = 0 __UpperCAmelCase : Tuple = 0 # first test for _, line in zip(range(_snake_case ), _snake_case ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test __UpperCAmelCase : Union[str, Any] = example["content"].count("\n" ) __UpperCAmelCase : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count("config" ) count_test += line.lower().count("test" ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def _UpperCamelCase ( snake_case__ ) -> Optional[Any]: __UpperCAmelCase : Optional[Any] = ["def ", "class ", "for ", "while "] __UpperCAmelCase : Optional[Any] = example["content"].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def _UpperCamelCase ( snake_case__, snake_case__=4 ) -> Union[str, Any]: __UpperCAmelCase : Any = example["content"].splitlines() __UpperCAmelCase : Dict = 0 for line in lines: counter += line.lower().count("=" ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def _UpperCamelCase ( snake_case__ ) -> List[Any]: __UpperCAmelCase : Any = tokenizer(example["content"], truncation=_snake_case )["input_ids"] __UpperCAmelCase : List[Any] = len(example["content"] ) / len(_snake_case ) return {"ratio": ratio} def _UpperCamelCase ( snake_case__ ) -> str: __UpperCAmelCase : str = {} results.update(get_hash(_snake_case ) ) results.update(line_stats(_snake_case ) ) results.update(alpha_stats(_snake_case ) ) results.update(char_token_ratio(_snake_case ) ) results.update(is_autogenerated(_snake_case ) ) results.update(is_config_or_test(_snake_case ) ) results.update(has_no_keywords(_snake_case ) ) results.update(has_few_assignments(_snake_case ) ) return results def _UpperCamelCase ( snake_case__, snake_case__, snake_case__ ) -> Union[str, Any]: if not check_uniques(_snake_case, _snake_case ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def _UpperCamelCase ( snake_case__ ) -> List[Any]: with open(_snake_case, "rb" ) as f_in: with gzip.open(str(_snake_case ) + ".gz", "wb", compresslevel=6 ) as f_out: shutil.copyfileobj(_snake_case, _snake_case ) os.unlink(_snake_case ) # Settings _snake_case = HfArgumentParser(PreprocessingArguments) _snake_case = parser.parse_args() if args.num_workers is None: _snake_case = multiprocessing.cpu_count() _snake_case = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset _snake_case = time.time() _snake_case = load_dataset(args.dataset_name, split='''train''') print(F'Time to load dataset: {time.time()-t_start:.2f}') # Run preprocessing _snake_case = time.time() _snake_case = ds.map(preprocess, num_proc=args.num_workers) print(F'Time to preprocess dataset: {time.time()-t_start:.2f}') # Deduplicate hashes _snake_case = set(ds.unique('''hash''')) _snake_case = len(uniques) / len(ds) print(F'Fraction of duplicates: {1-frac:.2%}') # Deduplicate data and apply heuristics _snake_case = time.time() _snake_case = ds.filter(filter, fn_kwargs={'''uniques''': uniques, '''args''': args}) print(F'Time to filter dataset: {time.time()-t_start:.2f}') print(F'Size of filtered dataset: {len(ds_filter)}') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: _snake_case = time.time() _snake_case = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(F'Time to deduplicate dataset: {time.time()-t_start:.2f}') print(F'Size of deduplicate dataset: {len(ds_filter)}') # Save data in batches of samples_per_file _snake_case = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / '''duplicate_clusters.json''', '''w''') as f: json.dump(duplicate_clusters, f) _snake_case = output_dir / "data" data_dir.mkdir(exist_ok=True) _snake_case = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): _snake_case = str(data_dir / F'file-{file_number+1:012}.json') _snake_case = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(F'Time to save dataset: {time.time()-t_start:.2f}')
157
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging snake_case : int = logging.get_logger(__name__) snake_case : List[str] = {"vocab_file": "spiece.model"} snake_case : List[str] = { "vocab_file": { "albert-base-v1": "https://huggingface.co/albert-base-v1/resolve/main/spiece.model", "albert-large-v1": "https://huggingface.co/albert-large-v1/resolve/main/spiece.model", "albert-xlarge-v1": "https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model", "albert-xxlarge-v1": "https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model", "albert-base-v2": "https://huggingface.co/albert-base-v2/resolve/main/spiece.model", "albert-large-v2": "https://huggingface.co/albert-large-v2/resolve/main/spiece.model", "albert-xlarge-v2": "https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model", "albert-xxlarge-v2": "https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model", } } snake_case : Tuple = { "albert-base-v1": 512, "albert-large-v1": 512, "albert-xlarge-v1": 512, "albert-xxlarge-v1": 512, "albert-base-v2": 512, "albert-large-v2": 512, "albert-xlarge-v2": 512, "albert-xxlarge-v2": 512, } snake_case : List[str] = "▁" class _snake_case ( snake_case ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , _a , _a=True , _a=True , _a=False , _a="[CLS]" , _a="[SEP]" , _a="<unk>" , _a="[SEP]" , _a="<pad>" , _a="[CLS]" , _a="[MASK]" , _a = None , **_a , ): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. __magic_name__ : str = ( AddedToken(_a , lstrip=_a , rstrip=_a , normalized=_a ) if isinstance(_a , _a ) else mask_token ) __magic_name__ : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=_a , remove_space=_a , keep_accents=_a , bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , pad_token=_a , cls_token=_a , mask_token=_a , sp_model_kwargs=self.sp_model_kwargs , **_a , ) __magic_name__ : Dict = do_lower_case __magic_name__ : Tuple = remove_space __magic_name__ : Union[str, Any] = keep_accents __magic_name__ : Tuple = vocab_file __magic_name__ : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_a ) @property def SCREAMING_SNAKE_CASE ( self ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self ): __magic_name__ : List[str] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): __magic_name__ : List[str] = self.__dict__.copy() __magic_name__ : Any = None return state def __setstate__( self , _a ): __magic_name__ : Union[str, Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): __magic_name__ : str = {} __magic_name__ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self , _a ): if self.remove_space: __magic_name__ : List[Any] = " ".join(inputs.strip().split() ) else: __magic_name__ : str = inputs __magic_name__ : int = outputs.replace("``" , "\"" ).replace("''" , "\"" ) if not self.keep_accents: __magic_name__ : str = unicodedata.normalize("NFKD" , _a ) __magic_name__ : Tuple = "".join([c for c in outputs if not unicodedata.combining(_a )] ) if self.do_lower_case: __magic_name__ : int = outputs.lower() return outputs def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Optional[Any] = self.preprocess_text(_a ) __magic_name__ : Dict = self.sp_model.encode(_a , out_type=_a ) __magic_name__ : Any = [] for piece in pieces: if len(_a ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): __magic_name__ : Optional[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(_a , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __magic_name__ : List[str] = cur_pieces[1:] else: __magic_name__ : Optional[int] = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(_a ) else: new_pieces.append(_a ) return new_pieces def SCREAMING_SNAKE_CASE ( self , _a ): return self.sp_model.PieceToId(_a ) def SCREAMING_SNAKE_CASE ( self , _a ): return self.sp_model.IdToPiece(_a ) def SCREAMING_SNAKE_CASE ( self , _a ): __magic_name__ : Any = [] __magic_name__ : Union[str, Any] = "" __magic_name__ : int = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(_a ) + token __magic_name__ : List[Any] = True __magic_name__ : Optional[int] = [] else: current_sub_tokens.append(_a ) __magic_name__ : Optional[Any] = False out_string += self.sp_model.decode(_a ) return out_string.strip() def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): __magic_name__ : List[str] = [self.sep_token_id] __magic_name__ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self , _a , _a = None , _a = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a , token_ids_a=_a , already_has_special_tokens=_a ) if token_ids_a is not None: return [1] + ([0] * len(_a )) + [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1] def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): __magic_name__ : Optional[int] = [self.sep_token_id] __magic_name__ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self , _a , _a = None ): if not os.path.isdir(_a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __magic_name__ : List[str] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _a ) elif not os.path.isfile(self.vocab_file ): with open(_a , "wb" ) as fi: __magic_name__ : Union[str, Any] = self.sp_model.serialized_model_proto() fi.write(_a ) return (out_vocab_file,)
281
0
import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor _lowerCamelCase : Dict = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def __init__( self : Tuple , *lowercase : List[Any] , **lowercase : Optional[int] ): '''simple docstring''' warnings.warn( 'The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use BeitImageProcessor instead.' , lowercase , ) super().__init__(*lowercase , **lowercase )
282
from collections.abc import Sequence def a_ ( __lowercase : Sequence[float] , __lowercase : float ) -> float: return sum(c * (x**i) for i, c in enumerate(__lowercase ) ) def a_ ( __lowercase : Sequence[float] , __lowercase : float ) -> float: _snake_case = 0.0 for coeff in reversed(__lowercase ): _snake_case = result * x + coeff return result if __name__ == "__main__": _lowerCamelCase : Optional[Any] = (0.0, 0.0, 5.0, 9.3, 7.0) _lowerCamelCase : Optional[int] = 1_0.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
282
1
import torch from diffusers import StableDiffusionPipeline _lowerCamelCase : int = '''path-to-your-trained-model''' _lowerCamelCase : str = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to('''cuda''') _lowerCamelCase : List[Any] = '''A photo of sks dog in a bucket''' _lowerCamelCase : Optional[int] = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save('''dog-bucket.png''')
282
import unittest from transformers import LiltConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, ) from transformers.models.lilt.modeling_lilt import LILT_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : Dict , lowercase : str , lowercase : List[str]=13 , lowercase : Any=7 , lowercase : Dict=True , lowercase : str=True , lowercase : List[Any]=True , lowercase : Any=True , lowercase : Tuple=99 , lowercase : str=24 , lowercase : str=2 , lowercase : Any=6 , lowercase : Dict=37 , lowercase : List[str]="gelu" , lowercase : Dict=0.1 , lowercase : Tuple=0.1 , lowercase : Optional[Any]=512 , lowercase : List[Any]=16 , lowercase : str=2 , lowercase : int=0.02 , lowercase : List[Any]=3 , lowercase : List[Any]=None , lowercase : int=1_000 , ): '''simple docstring''' _snake_case = parent _snake_case = batch_size _snake_case = seq_length _snake_case = is_training _snake_case = use_input_mask _snake_case = use_token_type_ids _snake_case = use_labels _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = type_sequence_label_size _snake_case = initializer_range _snake_case = num_labels _snake_case = scope _snake_case = range_bbox def A ( self : List[Any] ): '''simple docstring''' _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: _snake_case = bbox[i, j, 3] _snake_case = bbox[i, j, 1] _snake_case = t if bbox[i, j, 2] < bbox[i, j, 0]: _snake_case = bbox[i, j, 2] _snake_case = bbox[i, j, 0] _snake_case = t _snake_case = None if self.use_input_mask: _snake_case = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _snake_case = None if self.use_token_type_ids: _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _snake_case = None _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _snake_case = self.get_config() return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels def A ( self : List[str] ): '''simple docstring''' return LiltConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def A ( self : str , lowercase : Tuple , lowercase : Tuple , lowercase : str , lowercase : Any , lowercase : Union[str, Any] , lowercase : List[str] , lowercase : str , ): '''simple docstring''' _snake_case = LiltModel(config=lowercase ) model.to(lowercase ) model.eval() _snake_case = model(lowercase , bbox=lowercase , attention_mask=lowercase , token_type_ids=lowercase ) _snake_case = model(lowercase , bbox=lowercase , token_type_ids=lowercase ) _snake_case = model(lowercase , bbox=lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A ( self : List[Any] , lowercase : int , lowercase : int , lowercase : Any , lowercase : Optional[int] , lowercase : Union[str, Any] , lowercase : Optional[Any] , lowercase : Optional[int] , ): '''simple docstring''' _snake_case = self.num_labels _snake_case = LiltForTokenClassification(config=lowercase ) model.to(lowercase ) model.eval() _snake_case = model( lowercase , bbox=lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : List[str] , lowercase : Union[str, Any] , lowercase : str , lowercase : Dict , lowercase : Optional[int] , lowercase : List[str] , lowercase : int , lowercase : int , ): '''simple docstring''' _snake_case = LiltForQuestionAnswering(config=lowercase ) model.to(lowercase ) model.eval() _snake_case = model( lowercase , bbox=lowercase , attention_mask=lowercase , token_type_ids=lowercase , start_positions=lowercase , end_positions=lowercase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = self.prepare_config_and_inputs() ( ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ) = config_and_inputs _snake_case = { 'input_ids': input_ids, 'bbox': bbox, 'token_type_ids': token_type_ids, 'attention_mask': input_mask, } return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ,UpperCAmelCase ,UpperCAmelCase ,unittest.TestCase ): '''simple docstring''' _UpperCAmelCase : List[Any] = ( ( LiltModel, LiltForSequenceClassification, LiltForTokenClassification, LiltForQuestionAnswering, ) if is_torch_available() else () ) _UpperCAmelCase : List[str] = ( { "feature-extraction": LiltModel, "question-answering": LiltForQuestionAnswering, "text-classification": LiltForSequenceClassification, "token-classification": LiltForTokenClassification, "zero-shot": LiltForSequenceClassification, } if is_torch_available() else {} ) _UpperCAmelCase : Optional[Any] = False _UpperCAmelCase : Union[str, Any] = False def A ( self : Dict , lowercase : Dict , lowercase : Optional[int] , lowercase : Optional[int] , lowercase : List[str] , lowercase : Tuple ): '''simple docstring''' return True def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = LiltModelTester(self ) _snake_case = ConfigTester(self , config_class=lowercase , hidden_size=37 ) def A ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def A ( self : Dict ): '''simple docstring''' _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def A ( self : List[Any] ): '''simple docstring''' _snake_case = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _snake_case = type self.model_tester.create_and_check_model(*lowercase ) def A ( self : Any ): '''simple docstring''' _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase ) def A ( self : Any ): '''simple docstring''' _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase ) @slow def A ( self : Union[str, Any] ): '''simple docstring''' for model_name in LILT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = LiltModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) @require_torch @slow class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' def A ( self : Tuple ): '''simple docstring''' _snake_case = LiltModel.from_pretrained('SCUT-DLVCLab/lilt-roberta-en-base' ).to(lowercase ) _snake_case = torch.tensor([[1, 2]] , device=lowercase ) _snake_case = torch.tensor([[[1, 2, 3, 4], [5, 6, 7, 8]]] , device=lowercase ) # forward pass with torch.no_grad(): _snake_case = model(input_ids=lowercase , bbox=lowercase ) _snake_case = torch.Size([1, 2, 768] ) _snake_case = torch.tensor( [[-0.0653, 0.0950, -0.0061], [-0.0545, 0.0926, -0.0324]] , device=lowercase , ) self.assertTrue(outputs.last_hidden_state.shape , lowercase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :, :3] , lowercase , atol=1E-3 ) )
282
1
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore _lowerCamelCase : Dict = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" _lowerCamelCase : str = [file for file in filepaths if file != file.lower()] if upper_files: print(F'{len(upper_files)} files contain uppercase characters:') print('''\n'''.join(upper_files) + '''\n''') _lowerCamelCase : List[str] = [file for file in filepaths if ''' ''' in file] if space_files: print(F'{len(space_files)} files contain space characters:') print('''\n'''.join(space_files) + '''\n''') _lowerCamelCase : List[Any] = [file for file in filepaths if '''-''' in file] if hyphen_files: print(F'{len(hyphen_files)} files contain hyphen characters:') print('''\n'''.join(hyphen_files) + '''\n''') _lowerCamelCase : Any = [file for file in filepaths if os.sep not in file] if nodir_files: print(F'{len(nodir_files)} files are not in a directory:') print('''\n'''.join(nodir_files) + '''\n''') _lowerCamelCase : List[Any] = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
282
from __future__ import annotations import time from collections.abc import Sequence from random import randint from matplotlib import pyplot as plt def a_ ( __lowercase : Sequence[float] , __lowercase : int , __lowercase : int ) -> tuple[int | None, int | None, float]: if not arr: return None, None, 0 if low == high: return low, high, arr[low] _snake_case = (low + high) // 2 _snake_case , _snake_case , _snake_case = max_subarray(__lowercase , __lowercase , __lowercase ) _snake_case , _snake_case , _snake_case = max_subarray(__lowercase , mid + 1 , __lowercase ) _snake_case , _snake_case , _snake_case = max_cross_sum(__lowercase , __lowercase , __lowercase , __lowercase ) if left_sum >= right_sum and left_sum >= cross_sum: return left_low, left_high, left_sum elif right_sum >= left_sum and right_sum >= cross_sum: return right_low, right_high, right_sum return cross_left, cross_right, cross_sum def a_ ( __lowercase : Sequence[float] , __lowercase : int , __lowercase : int , __lowercase : int ) -> tuple[int, int, float]: _snake_case , _snake_case = float('-inf' ), -1 _snake_case , _snake_case = float('-inf' ), -1 _snake_case = 0 for i in range(__lowercase , low - 1 , -1 ): summ += arr[i] if summ > left_sum: _snake_case = summ _snake_case = i _snake_case = 0 for i in range(mid + 1 , high + 1 ): summ += arr[i] if summ > right_sum: _snake_case = summ _snake_case = i return max_left, max_right, (left_sum + right_sum) def a_ ( __lowercase : int ) -> float: _snake_case = [randint(1 , __lowercase ) for _ in range(__lowercase )] _snake_case = time.time() max_subarray(__lowercase , 0 , input_size - 1 ) _snake_case = time.time() return end - start def a_ ( ) -> None: _snake_case = [10, 100, 1_000, 10_000, 50_000, 100_000, 200_000, 300_000, 400_000, 500_000] _snake_case = [time_max_subarray(__lowercase ) for input_size in input_sizes] print('No of Inputs\t\tTime Taken' ) for input_size, runtime in zip(__lowercase , __lowercase ): print(__lowercase , '\t\t' , __lowercase ) plt.plot(__lowercase , __lowercase ) plt.xlabel('Number of Inputs' ) plt.ylabel('Time taken in seconds' ) plt.show() if __name__ == "__main__": from doctest import testmod testmod()
282
1
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _lowerCamelCase : int = logging.getLogger(__name__) def a_ ( ) -> Union[str, Any]: _snake_case = argparse.ArgumentParser( description='Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset.' ) parser.add_argument( '--dataset_name' , type=__lowercase , default='wikitext' , help='Name of the training. Explore datasets at: hf.co/datasets.' , ) parser.add_argument( '--dataset_config' , type=__lowercase , default='wikitext-103-raw-v1' , help='Configuration name of the dataset.' ) parser.add_argument( '--tokenizer_name_or_path' , type=__lowercase , default='sayakpaul/unigram-tokenizer-wikitext' , help='Tokenizer identifier. Can be a local filepath or a Hub identifier.' , ) parser.add_argument( '--shard_size' , type=__lowercase , default=1_000 , help='Number of entries to go in a single shard.' , ) parser.add_argument('--split' , type=__lowercase , default='train' , choices=['train', 'test', 'validation'] ) parser.add_argument( '--limit' , default=__lowercase , type=__lowercase , help='Limit the number of shards (used for debugging).' , ) parser.add_argument( '--max_length' , type=__lowercase , default=512 , help='Maximum sequence length. For training on TPUs, it helps to have a maximum' ' sequence length that is a multiple of 8.' , ) parser.add_argument( '--output_dir' , default='tf-tpu' , type=__lowercase , help='Output directory where the TFRecord shards will be saved. If the' ' path is appended with `gs://` (\'gs://tf-tpu\', for example) then the TFRecord' ' shards will be directly saved to a Google Cloud Storage bucket.' , ) _snake_case = parser.parse_args() return args def a_ ( __lowercase : Any ) -> str: def fn(__lowercase : Dict ): return tokenizer(examples['text'] ) return fn def a_ ( __lowercase : Tuple ) -> int: _snake_case = [] for i in range(len(tokenized_data['input_ids'] ) ): _snake_case = { 'input_ids': tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data['input_ids'][i] ) ), 'attention_mask': tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data['attention_mask'][i] ) ), } _snake_case = tf.train.Features(feature=__lowercase ) _snake_case = tf.train.Example(features=__lowercase ) _snake_case = example.SerializeToString() records.append(__lowercase ) return records def a_ ( __lowercase : Dict ) -> Any: _snake_case = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: _snake_case = min(len(__lowercase ) , args.limit ) _snake_case = dataset.select(range(__lowercase ) ) print(f'''Limiting the dataset to {args.limit} entries.''' ) _snake_case = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) _snake_case = os.path.join(args.output_dir , args.split ) if not os.path.exists(__lowercase ): os.makedirs(__lowercase ) else: _snake_case = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. _snake_case = tokenize_function(__lowercase ) _snake_case = dataset.map(__lowercase , batched=__lowercase , num_proc=4 , remove_columns=['text'] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(__lowercase : Optional[int] ): # Concatenate all texts. _snake_case = {k: sum(examples[k] , [] ) for k in examples.keys()} _snake_case = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 _snake_case = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. _snake_case = { k: [t[i : i + args.max_length] for i in range(0 , __lowercase , args.max_length )] for k, t in concatenated_examples.items() } return result _snake_case = dataset_tokenized.map(__lowercase , batched=__lowercase , batch_size=1_000 , num_proc=4 ) _snake_case = 0 _snake_case = 0 for shard in range(0 , len(__lowercase ) , args.shard_size ): _snake_case = grouped_dataset[shard : shard + args.shard_size] _snake_case = len(dataset_snapshot['input_ids'] ) _snake_case = os.path.join(__lowercase , f'''dataset-{shard_count}-{records_containing}.tfrecord''' ) _snake_case = get_serialized_examples(__lowercase ) with tf.io.TFRecordWriter(__lowercase ) as out_file: for i in range(len(__lowercase ) ): _snake_case = serialized_examples[i] out_file.write(__lowercase ) print('Wrote file {} containing {} records'.format(__lowercase , __lowercase ) ) shard_count += 1 total_records += records_containing with open(f'''split-{args.split}-records-count.txt''' , 'w' ) as f: print(f'''Total {args.split} records: {total_records}''' , file=__lowercase ) if __name__ == "__main__": _lowerCamelCase : Union[str, Any] = parse_args() main(args)
282
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' def A ( self : List[Any] , lowercase : Dict ): '''simple docstring''' for model_result in results.values(): for batch_size, sequence_length in zip(model_result['bs'] , model_result['ss'] ): _snake_case = model_result['result'][batch_size][sequence_length] self.assertIsNotNone(lowercase ) def A ( self : str ): '''simple docstring''' _snake_case = 'sshleifer/tiny-gpt2' _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) _snake_case = PyTorchBenchmark(lowercase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def A ( self : Any ): '''simple docstring''' _snake_case = 'sgugger/tiny-distilbert-classification' _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , only_pretrain_model=lowercase , ) _snake_case = PyTorchBenchmark(lowercase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def A ( self : Optional[int] ): '''simple docstring''' _snake_case = 'sshleifer/tiny-gpt2' _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , torchscript=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) _snake_case = PyTorchBenchmark(lowercase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == 'cpu' , 'Cant do half precision' ) def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = 'sshleifer/tiny-gpt2' _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , fpaa=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) _snake_case = PyTorchBenchmark(lowercase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def A ( self : str ): '''simple docstring''' _snake_case = 'sshleifer/tiny-gpt2' _snake_case = AutoConfig.from_pretrained(lowercase ) # set architectures equal to `None` _snake_case = None _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) _snake_case = PyTorchBenchmark(lowercase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = 'sshleifer/tiny-gpt2' _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) _snake_case = PyTorchBenchmark(lowercase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == 'cpu' , 'Can\'t do half precision' ) def A ( self : str ): '''simple docstring''' _snake_case = 'sshleifer/tiny-gpt2' _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , fpaa=lowercase , multi_process=lowercase , ) _snake_case = PyTorchBenchmark(lowercase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def A ( self : Tuple ): '''simple docstring''' _snake_case = 'sshleifer/tiny-gpt2' _snake_case = AutoConfig.from_pretrained(lowercase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) _snake_case = PyTorchBenchmark(lowercase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def A ( self : Union[str, Any] ): '''simple docstring''' _snake_case = 'sshleifer/tinier_bart' _snake_case = AutoConfig.from_pretrained(lowercase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) _snake_case = PyTorchBenchmark(lowercase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def A ( self : Dict ): '''simple docstring''' _snake_case = 'sshleifer/tiny-gpt2' _snake_case = AutoConfig.from_pretrained(lowercase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) _snake_case = PyTorchBenchmark(lowercase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def A ( self : Dict ): '''simple docstring''' _snake_case = 'sshleifer/tinier_bart' _snake_case = AutoConfig.from_pretrained(lowercase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowercase , ) _snake_case = PyTorchBenchmark(lowercase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = 'sshleifer/tiny-gpt2' with tempfile.TemporaryDirectory() as tmp_dir: _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , save_to_csv=lowercase , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(lowercase , 'inf_time.csv' ) , train_memory_csv_file=os.path.join(lowercase , 'train_mem.csv' ) , inference_memory_csv_file=os.path.join(lowercase , 'inf_mem.csv' ) , train_time_csv_file=os.path.join(lowercase , 'train_time.csv' ) , env_info_csv_file=os.path.join(lowercase , 'env.csv' ) , multi_process=lowercase , ) _snake_case = PyTorchBenchmark(lowercase ) benchmark.run() self.assertTrue(Path(os.path.join(lowercase , 'inf_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(lowercase , 'train_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(lowercase , 'inf_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(lowercase , 'train_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(lowercase , 'env.csv' ) ).exists() ) def A ( self : Union[str, Any] ): '''simple docstring''' _snake_case = 'sshleifer/tiny-gpt2' def _check_summary_is_not_empty(lowercase : Optional[Any] ): self.assertTrue(hasattr(lowercase , 'sequential' ) ) self.assertTrue(hasattr(lowercase , 'cumulative' ) ) self.assertTrue(hasattr(lowercase , 'current' ) ) self.assertTrue(hasattr(lowercase , 'total' ) ) with tempfile.TemporaryDirectory() as tmp_dir: _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowercase , inference=lowercase , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(lowercase , 'log.txt' ) , log_print=lowercase , trace_memory_line_by_line=lowercase , multi_process=lowercase , ) _snake_case = PyTorchBenchmark(lowercase ) _snake_case = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(lowercase , 'log.txt' ) ).exists() )
282
1
import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaProcessor, is_apex_available, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): _lowerCamelCase : Any = True from torch.cuda.amp import autocast _lowerCamelCase : Any = logging.getLogger(__name__) def a_ ( __lowercase : int=None , __lowercase : Dict=None ) -> Any: return field(default_factory=lambda: default , metadata=__lowercase ) @dataclass class SCREAMING_SNAKE_CASE__ : '''simple docstring''' _UpperCAmelCase : str = field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) _UpperCAmelCase : Optional[str] = field( default=UpperCAmelCase ,metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} ,) _UpperCAmelCase : Optional[bool] = field( default=UpperCAmelCase ,metadata={"help": "Whether to freeze the feature extractor layers of the model."} ) _UpperCAmelCase : Optional[float] = field( default=0.1 ,metadata={"help": "The dropout ratio for the attention probabilities."} ) _UpperCAmelCase : Optional[float] = field( default=0.1 ,metadata={"help": "The dropout ratio for activations inside the fully connected layer."} ) _UpperCAmelCase : Optional[float] = field( default=0.1 ,metadata={ "help": "The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler." } ,) _UpperCAmelCase : Optional[float] = field( default=0.1 ,metadata={"help": "The dropout probabilitiy for all 1D convolutional layers in feature extractor."} ,) _UpperCAmelCase : Optional[float] = field( default=0.05 ,metadata={ "help": ( "Propability of each feature vector along the time axis to be chosen as the start of the vector" "span to be masked. Approximately ``mask_time_prob * sequence_length // mask_time_length`` feature" "vectors will be masked along the time axis. This is only relevant if ``apply_spec_augment is True``." ) } ,) _UpperCAmelCase : Optional[float] = field(default=0.0 ,metadata={"help": "The LayerDrop probability."} ) @dataclass class SCREAMING_SNAKE_CASE__ : '''simple docstring''' _UpperCAmelCase : Optional[str] = field( default=UpperCAmelCase ,metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} ) _UpperCAmelCase : Optional[str] = field( default="train+validation" ,metadata={ "help": "The name of the training data set split to use (via the datasets library). Defaults to 'train'" } ,) _UpperCAmelCase : bool = field( default=UpperCAmelCase ,metadata={"help": "Overwrite the cached preprocessed datasets or not."} ) _UpperCAmelCase : Optional[int] = field( default=UpperCAmelCase ,metadata={"help": "The number of processes to use for the preprocessing."} ,) _UpperCAmelCase : Optional[int] = field( default=UpperCAmelCase ,metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } ,) _UpperCAmelCase : Optional[int] = field( default=UpperCAmelCase ,metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of validation examples to this " "value if set." ) } ,) _UpperCAmelCase : List[str] = list_field( default=[",", "?", ".", "!", "-", ";", ":", "\"\"", "%", "'", "\"", "�"] ,metadata={"help": "A list of characters to remove from the transcripts."} ,) @dataclass class SCREAMING_SNAKE_CASE__ : '''simple docstring''' _UpperCAmelCase : WavaVecaProcessor _UpperCAmelCase : Union[bool, str] = True _UpperCAmelCase : Optional[int] = None _UpperCAmelCase : Optional[int] = None _UpperCAmelCase : Optional[int] = None _UpperCAmelCase : Optional[int] = None def __call__( self : Tuple , lowercase : List[Dict[str, Union[List[int], torch.Tensor]]] ): '''simple docstring''' _snake_case = [{'input_values': feature['input_values']} for feature in features] _snake_case = [{'input_ids': feature['labels']} for feature in features] _snake_case = self.processor.pad( lowercase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='pt' , ) _snake_case = self.processor.pad( labels=lowercase , padding=self.padding , max_length=self.max_length_labels , pad_to_multiple_of=self.pad_to_multiple_of_labels , return_tensors='pt' , ) # replace padding with -100 to ignore loss correctly _snake_case = labels_batch['input_ids'].masked_fill(labels_batch.attention_mask.ne(1 ) , -100 ) _snake_case = labels return batch class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def A ( self : List[Any] , lowercase : nn.Module , lowercase : Dict[str, Union[torch.Tensor, Any]] ): '''simple docstring''' model.train() _snake_case = self._prepare_inputs(lowercase ) if self.use_amp: with autocast(): _snake_case = self.compute_loss(lowercase , lowercase ) else: _snake_case = self.compute_loss(lowercase , lowercase ) if self.args.n_gpu > 1: if model.module.config.ctc_loss_reduction == "mean": _snake_case = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": _snake_case = loss.sum() / (inputs['labels'] >= 0).sum() else: raise ValueError(f'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: _snake_case = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(lowercase ).backward() elif self.use_apex: with amp.scale_loss(lowercase , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(lowercase ) else: loss.backward() return loss.detach() def a_ ( ) -> Dict: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _snake_case = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _snake_case , _snake_case , _snake_case = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _snake_case , _snake_case , _snake_case = parser.parse_args_into_dataclasses() # Detecting last checkpoint. _snake_case = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _snake_case = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None: logger.info( f'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( f'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + f'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info('Training/evaluation parameters %s' , __lowercase ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: _snake_case = datasets.load_dataset( 'common_voice' , data_args.dataset_config_name , split=data_args.train_split_name ) _snake_case = datasets.load_dataset('common_voice' , data_args.dataset_config_name , split='test' ) # Create and save tokenizer _snake_case = f'''[{''.join(data_args.chars_to_ignore )}]''' def remove_special_characters(__lowercase : Dict ): _snake_case = re.sub(__lowercase , '' , batch['sentence'] ).lower() + ' ' return batch _snake_case = train_dataset.map(__lowercase , remove_columns=['sentence'] ) _snake_case = eval_dataset.map(__lowercase , remove_columns=['sentence'] ) def extract_all_chars(__lowercase : int ): _snake_case = ' '.join(batch['text'] ) _snake_case = list(set(__lowercase ) ) return {"vocab": [vocab], "all_text": [all_text]} _snake_case = train_dataset.map( __lowercase , batched=__lowercase , batch_size=-1 , keep_in_memory=__lowercase , remove_columns=train_dataset.column_names , ) _snake_case = train_dataset.map( __lowercase , batched=__lowercase , batch_size=-1 , keep_in_memory=__lowercase , remove_columns=eval_dataset.column_names , ) _snake_case = list(set(vocab_train['vocab'][0] ) | set(vocab_test['vocab'][0] ) ) _snake_case = {v: k for k, v in enumerate(__lowercase )} _snake_case = vocab_dict[' '] del vocab_dict[" "] _snake_case = len(__lowercase ) _snake_case = len(__lowercase ) with open('vocab.json' , 'w' ) as vocab_file: json.dump(__lowercase , __lowercase ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _snake_case = WavaVecaCTCTokenizer( 'vocab.json' , unk_token='[UNK]' , pad_token='[PAD]' , word_delimiter_token='|' , ) _snake_case = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0.0 , do_normalize=__lowercase , return_attention_mask=__lowercase ) _snake_case = WavaVecaProcessor(feature_extractor=__lowercase , tokenizer=__lowercase ) _snake_case = WavaVecaForCTC.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , activation_dropout=model_args.activation_dropout , attention_dropout=model_args.attention_dropout , hidden_dropout=model_args.hidden_dropout , feat_proj_dropout=model_args.feat_proj_dropout , mask_time_prob=model_args.mask_time_prob , gradient_checkpointing=training_args.gradient_checkpointing , layerdrop=model_args.layerdrop , ctc_loss_reduction='mean' , pad_token_id=processor.tokenizer.pad_token_id , vocab_size=len(processor.tokenizer ) , ) if data_args.max_train_samples is not None: _snake_case = min(len(__lowercase ) , data_args.max_train_samples ) _snake_case = train_dataset.select(range(__lowercase ) ) if data_args.max_val_samples is not None: _snake_case = eval_dataset.select(range(data_args.max_val_samples ) ) _snake_case = torchaudio.transforms.Resample(48_000 , 16_000 ) # Preprocessing the datasets. # We need to read the aduio files as arrays and tokenize the targets. def speech_file_to_array_fn(__lowercase : Union[str, Any] ): _snake_case , _snake_case = torchaudio.load(batch['path'] ) _snake_case = resampler(__lowercase ).squeeze().numpy() _snake_case = 16_000 _snake_case = batch['text'] return batch _snake_case = train_dataset.map( __lowercase , remove_columns=train_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) _snake_case = eval_dataset.map( __lowercase , remove_columns=eval_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) def prepare_dataset(__lowercase : Optional[int] ): # check that all files have the correct sampling rate assert ( len(set(batch['sampling_rate'] ) ) == 1 ), f'''Make sure all inputs have the same sampling rate of {processor.feature_extractor.sampling_rate}.''' _snake_case = processor( audio=batch['speech'] , text=batch['target_text'] , sampling_rate=batch['sampling_rate'][0] ) batch.update(__lowercase ) return batch _snake_case = train_dataset.map( __lowercase , remove_columns=train_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=__lowercase , num_proc=data_args.preprocessing_num_workers , ) _snake_case = eval_dataset.map( __lowercase , remove_columns=eval_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=__lowercase , num_proc=data_args.preprocessing_num_workers , ) # Metric _snake_case = datasets.load_metric('wer' ) def compute_metrics(__lowercase : Any ): _snake_case = pred.predictions _snake_case = np.argmax(__lowercase , axis=-1 ) _snake_case = processor.tokenizer.pad_token_id _snake_case = processor.batch_decode(__lowercase ) # we do not want to group tokens when computing the metrics _snake_case = processor.batch_decode(pred.label_ids , group_tokens=__lowercase ) _snake_case = wer_metric.compute(predictions=__lowercase , references=__lowercase ) return {"wer": wer} if model_args.freeze_feature_extractor: model.freeze_feature_extractor() # Data collator _snake_case = DataCollatorCTCWithPadding(processor=__lowercase , padding=__lowercase ) # Initialize our Trainer _snake_case = CTCTrainer( model=__lowercase , data_collator=__lowercase , args=__lowercase , compute_metrics=__lowercase , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=processor.feature_extractor , ) # Training if training_args.do_train: if last_checkpoint is not None: _snake_case = last_checkpoint elif os.path.isdir(model_args.model_name_or_path ): _snake_case = model_args.model_name_or_path else: _snake_case = None # Save the feature_extractor and the tokenizer if is_main_process(training_args.local_rank ): processor.save_pretrained(training_args.output_dir ) _snake_case = trainer.train(resume_from_checkpoint=__lowercase ) trainer.save_model() _snake_case = train_result.metrics _snake_case = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(__lowercase ) ) _snake_case = min(__lowercase , len(__lowercase ) ) trainer.log_metrics('train' , __lowercase ) trainer.save_metrics('train' , __lowercase ) trainer.save_state() # Evaluation _snake_case = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) _snake_case = trainer.evaluate() _snake_case = data_args.max_val_samples if data_args.max_val_samples is not None else len(__lowercase ) _snake_case = min(__lowercase , len(__lowercase ) ) trainer.log_metrics('eval' , __lowercase ) trainer.save_metrics('eval' , __lowercase ) return results if __name__ == "__main__": main()
282
from __future__ import annotations from typing import Any class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : Tuple , lowercase : int , lowercase : int , lowercase : float = 0 ): '''simple docstring''' _snake_case , _snake_case = row, column _snake_case = [[default_value for c in range(lowercase )] for r in range(lowercase )] def __str__( self : int ): '''simple docstring''' _snake_case = f'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier _snake_case = 0 for row_vector in self.array: for obj in row_vector: _snake_case = max(lowercase , len(str(lowercase ) ) ) _snake_case = f'''%{max_element_length}s''' # Make string and return def single_line(lowercase : list[float] ) -> str: nonlocal string_format_identifier _snake_case = '[' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase ) for row_vector in self.array ) return s def __repr__( self : Dict ): '''simple docstring''' return str(self ) def A ( self : str , lowercase : tuple[int, int] ): '''simple docstring''' if not (isinstance(lowercase , (list, tuple) ) and len(lowercase ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Dict , lowercase : tuple[int, int] ): '''simple docstring''' assert self.validate_indicies(lowercase ) return self.array[loc[0]][loc[1]] def __setitem__( self : str , lowercase : tuple[int, int] , lowercase : float ): '''simple docstring''' assert self.validate_indicies(lowercase ) _snake_case = value def __add__( self : str , lowercase : Matrix ): '''simple docstring''' assert isinstance(lowercase , lowercase ) assert self.row == another.row and self.column == another.column # Add _snake_case = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): _snake_case = self[r, c] + another[r, c] return result def __neg__( self : Tuple ): '''simple docstring''' _snake_case = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): _snake_case = -self[r, c] return result def __sub__( self : List[str] , lowercase : Matrix ): '''simple docstring''' return self + (-another) def __mul__( self : Dict , lowercase : int | float | Matrix ): '''simple docstring''' if isinstance(lowercase , (int, float) ): # Scalar multiplication _snake_case = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): _snake_case = self[r, c] * another return result elif isinstance(lowercase , lowercase ): # Matrix multiplication assert self.column == another.row _snake_case = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: _snake_case = f'''Unsupported type given for another ({type(lowercase )})''' raise TypeError(lowercase ) def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): _snake_case = self[r, c] return result def A ( self : List[Any] , lowercase : Matrix , lowercase : Matrix ): '''simple docstring''' assert isinstance(lowercase , lowercase ) and isinstance(lowercase , lowercase ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate _snake_case = v.transpose() _snake_case = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def a_ ( ) -> None: # a^(-1) _snake_case = Matrix(3 , 3 , 0 ) for i in range(3 ): _snake_case = 1 print(f'''a^(-1) is {ainv}''' ) # u, v _snake_case = Matrix(3 , 1 , 0 ) _snake_case , _snake_case , _snake_case = 1, 2, -3 _snake_case = Matrix(3 , 1 , 0 ) _snake_case , _snake_case , _snake_case = 4, -2, 5 print(f'''u is {u}''' ) print(f'''v is {v}''' ) print(f'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(f'''(a + uv^T)^(-1) is {ainv.sherman_morrison(__lowercase , __lowercase )}''' ) def a_ ( ) -> None: import doctest doctest.testmod() testa()
282
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowerCamelCase : List[str] = { '''configuration_vivit''': ['''VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''VivitConfig'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Dict = ['''VivitImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Union[str, Any] = [ '''VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''VivitModel''', '''VivitPreTrainedModel''', '''VivitForVideoClassification''', ] if TYPE_CHECKING: from .configuration_vivit import VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, VivitConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_vivit import VivitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vivit import ( VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST, VivitForVideoClassification, VivitModel, VivitPreTrainedModel, ) else: import sys _lowerCamelCase : Dict = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
282
import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor _lowerCamelCase : Dict = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def __init__( self : Tuple , *lowercase : Optional[int] , **lowercase : Any ): '''simple docstring''' warnings.warn( 'The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use ChineseCLIPImageProcessor instead.' , lowercase , ) super().__init__(*lowercase , **lowercase )
282
1
import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def a_ ( __lowercase : Tuple ) -> Tuple: _snake_case = [] embed.append( ( f'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight''', f'''stage{idx}.patch_embed.proj.weight''', ) ) embed.append( ( f'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias''', f'''stage{idx}.patch_embed.proj.bias''', ) ) embed.append( ( f'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight''', f'''stage{idx}.patch_embed.norm.weight''', ) ) embed.append( ( f'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias''', f'''stage{idx}.patch_embed.norm.bias''', ) ) return embed def a_ ( __lowercase : Tuple , __lowercase : List[str] ) -> Optional[int]: _snake_case = [] attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight''', f'''stage{idx}.blocks.{cnt}.attn.proj_q.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias''', f'''stage{idx}.blocks.{cnt}.attn.proj_q.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight''', f'''stage{idx}.blocks.{cnt}.attn.proj_k.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias''', f'''stage{idx}.blocks.{cnt}.attn.proj_k.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight''', f'''stage{idx}.blocks.{cnt}.attn.proj_v.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias''', f'''stage{idx}.blocks.{cnt}.attn.proj_v.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight''', f'''stage{idx}.blocks.{cnt}.attn.proj.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias''', f'''stage{idx}.blocks.{cnt}.attn.proj.bias''', ) ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight''', f'''stage{idx}.blocks.{cnt}.mlp.fc1.weight''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias''', f'''stage{idx}.blocks.{cnt}.mlp.fc1.bias''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight''', f'''stage{idx}.blocks.{cnt}.mlp.fc2.weight''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias''', f'''stage{idx}.blocks.{cnt}.mlp.fc2.bias''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight''', f'''stage{idx}.blocks.{cnt}.norm1.weight''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias''', f'''stage{idx}.blocks.{cnt}.norm1.bias''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight''', f'''stage{idx}.blocks.{cnt}.norm2.weight''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias''', f'''stage{idx}.blocks.{cnt}.norm2.bias''') ) return attention_weights def a_ ( __lowercase : Dict ) -> Optional[int]: _snake_case = [] token.append((f'''cvt.encoder.stages.{idx}.cls_token''', 'stage2.cls_token') ) return token def a_ ( ) -> List[str]: _snake_case = [] head.append(('layernorm.weight', 'norm.weight') ) head.append(('layernorm.bias', 'norm.bias') ) head.append(('classifier.weight', 'head.weight') ) head.append(('classifier.bias', 'head.bias') ) return head def a_ ( __lowercase : Tuple , __lowercase : Tuple , __lowercase : List[Any] , __lowercase : Dict ) -> Optional[Any]: _snake_case = 'imagenet-1k-id2label.json' _snake_case = 1_000 _snake_case = 'huggingface/label-files' _snake_case = num_labels _snake_case = json.load(open(cached_download(hf_hub_url(__lowercase , __lowercase , repo_type='dataset' ) ) , 'r' ) ) _snake_case = {int(__lowercase ): v for k, v in idalabel.items()} _snake_case = idalabel _snake_case = {v: k for k, v in idalabel.items()} _snake_case = _snake_case = CvtConfig(num_labels=__lowercase , idalabel=__lowercase , labelaid=__lowercase ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit('/' , 1 )[-1][4:6] == "13": _snake_case = [1, 2, 10] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit('/' , 1 )[-1][4:6] == "21": _snake_case = [1, 4, 16] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: _snake_case = [2, 2, 20] _snake_case = [3, 12, 16] _snake_case = [192, 768, 1_024] _snake_case = CvtForImageClassification(__lowercase ) _snake_case = AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' ) _snake_case = image_size _snake_case = torch.load(__lowercase , map_location=torch.device('cpu' ) ) _snake_case = OrderedDict() _snake_case = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: _snake_case = list_of_state_dict + cls_token(__lowercase ) _snake_case = list_of_state_dict + embeddings(__lowercase ) for cnt in range(config.depth[idx] ): _snake_case = list_of_state_dict + attention(__lowercase , __lowercase ) _snake_case = list_of_state_dict + final() for gg in list_of_state_dict: print(__lowercase ) for i in range(len(__lowercase ) ): _snake_case = original_weights[list_of_state_dict[i][1]] model.load_state_dict(__lowercase ) model.save_pretrained(__lowercase ) image_processor.save_pretrained(__lowercase ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": _lowerCamelCase : int = argparse.ArgumentParser() parser.add_argument( '''--cvt_model''', default='''cvt-w24''', type=str, help='''Name of the cvt model you\'d like to convert.''', ) parser.add_argument( '''--image_size''', default=384, type=int, help='''Input Image Size''', ) parser.add_argument( '''--cvt_file_name''', default=r'''cvtmodels\CvT-w24-384x384-IN-22k.pth''', type=str, help='''Input Image Size''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) _lowerCamelCase : int = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
282
def a_ ( __lowercase : str ) -> int: _snake_case = hex_num.strip() if not hex_num: raise ValueError('No value was passed to the function' ) _snake_case = hex_num[0] == '-' if is_negative: _snake_case = hex_num[1:] try: _snake_case = int(__lowercase , 16 ) except ValueError: raise ValueError('Invalid value was passed to the function' ) _snake_case = '' while int_num > 0: _snake_case = str(int_num % 2 ) + bin_str int_num >>= 1 return int(('-' + bin_str) if is_negative else bin_str ) if __name__ == "__main__": import doctest doctest.testmod()
282
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase : List[Any] = logging.get_logger(__name__) _lowerCamelCase : int = { '''xlm-mlm-en-2048''': '''https://huggingface.co/xlm-mlm-en-2048/resolve/main/config.json''', '''xlm-mlm-ende-1024''': '''https://huggingface.co/xlm-mlm-ende-1024/resolve/main/config.json''', '''xlm-mlm-enfr-1024''': '''https://huggingface.co/xlm-mlm-enfr-1024/resolve/main/config.json''', '''xlm-mlm-enro-1024''': '''https://huggingface.co/xlm-mlm-enro-1024/resolve/main/config.json''', '''xlm-mlm-tlm-xnli15-1024''': '''https://huggingface.co/xlm-mlm-tlm-xnli15-1024/resolve/main/config.json''', '''xlm-mlm-xnli15-1024''': '''https://huggingface.co/xlm-mlm-xnli15-1024/resolve/main/config.json''', '''xlm-clm-enfr-1024''': '''https://huggingface.co/xlm-clm-enfr-1024/resolve/main/config.json''', '''xlm-clm-ende-1024''': '''https://huggingface.co/xlm-clm-ende-1024/resolve/main/config.json''', '''xlm-mlm-17-1280''': '''https://huggingface.co/xlm-mlm-17-1280/resolve/main/config.json''', '''xlm-mlm-100-1280''': '''https://huggingface.co/xlm-mlm-100-1280/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _UpperCAmelCase : Optional[int] = "xlm" _UpperCAmelCase : List[Any] = { "hidden_size": "emb_dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", "n_words": "vocab_size", # For backward compatibility } def __init__( self : Dict , lowercase : Optional[int]=30_145 , lowercase : List[str]=2_048 , lowercase : Union[str, Any]=12 , lowercase : Optional[int]=16 , lowercase : Any=0.1 , lowercase : List[Any]=0.1 , lowercase : str=True , lowercase : Tuple=False , lowercase : Any=False , lowercase : List[str]=False , lowercase : Tuple=1 , lowercase : Union[str, Any]=True , lowercase : Dict=512 , lowercase : Any=2_048**-0.5 , lowercase : Tuple=1E-12 , lowercase : Optional[int]=0.02 , lowercase : Optional[Any]=0 , lowercase : int=1 , lowercase : Any=2 , lowercase : List[str]=3 , lowercase : Any=5 , lowercase : List[str]=True , lowercase : Any="first" , lowercase : Tuple=True , lowercase : Dict=None , lowercase : Optional[Any]=True , lowercase : List[Any]=0.1 , lowercase : Dict=5 , lowercase : Any=5 , lowercase : Optional[Any]=0 , lowercase : int=0 , lowercase : str=2 , lowercase : int=0 , **lowercase : List[str] , ): '''simple docstring''' _snake_case = vocab_size _snake_case = emb_dim _snake_case = n_layers _snake_case = n_heads _snake_case = dropout _snake_case = attention_dropout _snake_case = gelu_activation _snake_case = sinusoidal_embeddings _snake_case = causal _snake_case = asm _snake_case = n_langs _snake_case = use_lang_emb _snake_case = layer_norm_eps _snake_case = bos_index _snake_case = eos_index _snake_case = pad_index _snake_case = unk_index _snake_case = mask_index _snake_case = is_encoder _snake_case = max_position_embeddings _snake_case = embed_init_std _snake_case = init_std _snake_case = summary_type _snake_case = summary_use_proj _snake_case = summary_activation _snake_case = summary_proj_to_labels _snake_case = summary_first_dropout _snake_case = start_n_top _snake_case = end_n_top _snake_case = mask_token_id _snake_case = lang_id if "n_words" in kwargs: _snake_case = kwargs['n_words'] super().__init__(pad_token_id=lowercase , bos_token_id=lowercase , **lowercase ) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' @property def A ( self : Dict ): '''simple docstring''' if self.task == "multiple-choice": _snake_case = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _snake_case = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('token_type_ids', dynamic_axis), ] )
282
from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpec from ...tokenization_utils_base import PreTrainedTokenizerBase _lowerCamelCase : List[Any] = logging.get_logger(__name__) _lowerCamelCase : Union[str, Any] = { '''allenai/longformer-base-4096''': '''https://huggingface.co/allenai/longformer-base-4096/resolve/main/config.json''', '''allenai/longformer-large-4096''': '''https://huggingface.co/allenai/longformer-large-4096/resolve/main/config.json''', '''allenai/longformer-large-4096-finetuned-triviaqa''': ( '''https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/config.json''' ), '''allenai/longformer-base-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/config.json''' ), '''allenai/longformer-large-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/config.json''' ), } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _UpperCAmelCase : Dict = "longformer" def __init__( self : Optional[Any] , lowercase : Union[List[int], int] = 512 , lowercase : int = 2 , lowercase : int = 1 , lowercase : int = 0 , lowercase : int = 2 , lowercase : int = 30_522 , lowercase : int = 768 , lowercase : int = 12 , lowercase : int = 12 , lowercase : int = 3_072 , lowercase : str = "gelu" , lowercase : float = 0.1 , lowercase : float = 0.1 , lowercase : int = 512 , lowercase : int = 2 , lowercase : float = 0.02 , lowercase : float = 1E-12 , lowercase : bool = False , **lowercase : Optional[Any] , ): '''simple docstring''' super().__init__(pad_token_id=lowercase , **lowercase ) _snake_case = attention_window _snake_case = sep_token_id _snake_case = bos_token_id _snake_case = eos_token_id _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = hidden_act _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = onnx_export class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def __init__( self : int , lowercase : "PretrainedConfig" , lowercase : str = "default" , lowercase : "List[PatchingSpec]" = None ): '''simple docstring''' super().__init__(lowercase , lowercase , lowercase ) _snake_case = True @property def A ( self : Union[str, Any] ): '''simple docstring''' if self.task == "multiple-choice": _snake_case = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _snake_case = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('global_attention_mask', dynamic_axis), ] ) @property def A ( self : int ): '''simple docstring''' _snake_case = super().outputs if self.task == "default": _snake_case = {0: 'batch'} return outputs @property def A ( self : List[Any] ): '''simple docstring''' return 1E-4 @property def A ( self : List[str] ): '''simple docstring''' return max(super().default_onnx_opset , 14 ) def A ( self : str , lowercase : "PreTrainedTokenizerBase" , lowercase : int = -1 , lowercase : int = -1 , lowercase : bool = False , lowercase : Optional[TensorType] = None , ): '''simple docstring''' _snake_case = super().generate_dummy_inputs( preprocessor=lowercase , batch_size=lowercase , seq_length=lowercase , is_pair=lowercase , framework=lowercase ) import torch # for some reason, replacing this code by inputs["global_attention_mask"] = torch.randint(2, inputs["input_ids"].shape, dtype=torch.int64) # makes the export fail randomly _snake_case = torch.zeros_like(inputs['input_ids'] ) # make every second token global _snake_case = 1 return inputs
282
1
from __future__ import annotations from collections import deque from collections.abc import Iterator from dataclasses import dataclass @dataclass class SCREAMING_SNAKE_CASE__ : '''simple docstring''' _UpperCAmelCase : int _UpperCAmelCase : int class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : Union[str, Any] , lowercase : int ): '''simple docstring''' _snake_case = [[] for _ in range(lowercase )] _snake_case = size def __getitem__( self : Any , lowercase : int ): '''simple docstring''' return iter(self._graph[vertex] ) @property def A ( self : int ): '''simple docstring''' return self._size def A ( self : Any , lowercase : int , lowercase : int , lowercase : int ): '''simple docstring''' if weight not in (0, 1): raise ValueError('Edge weight must be either 0 or 1.' ) if to_vertex < 0 or to_vertex >= self.size: raise ValueError('Vertex indexes must be in [0; size).' ) self._graph[from_vertex].append(Edge(lowercase , lowercase ) ) def A ( self : int , lowercase : int , lowercase : int ): '''simple docstring''' _snake_case = deque([start_vertex] ) _snake_case = [None] * self.size _snake_case = 0 while queue: _snake_case = queue.popleft() _snake_case = distances[current_vertex] if current_distance is None: continue for edge in self[current_vertex]: _snake_case = current_distance + edge.weight _snake_case = distances[edge.destination_vertex] if ( isinstance(lowercase , lowercase ) and new_distance >= dest_vertex_distance ): continue _snake_case = new_distance if edge.weight == 0: queue.appendleft(edge.destination_vertex ) else: queue.append(edge.destination_vertex ) if distances[finish_vertex] is None: raise ValueError('No path from start_vertex to finish_vertex.' ) return distances[finish_vertex] if __name__ == "__main__": import doctest doctest.testmod()
282
import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging _lowerCamelCase : List[str] = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def __init__( self : Dict , lowercase : Union[List[ControlNetModel], Tuple[ControlNetModel]] ): '''simple docstring''' super().__init__() _snake_case = nn.ModuleList(lowercase ) def A ( self : Optional[int] , lowercase : torch.FloatTensor , lowercase : Union[torch.Tensor, float, int] , lowercase : torch.Tensor , lowercase : List[torch.tensor] , lowercase : List[float] , lowercase : Optional[torch.Tensor] = None , lowercase : Optional[torch.Tensor] = None , lowercase : Optional[torch.Tensor] = None , lowercase : Optional[Dict[str, Any]] = None , lowercase : bool = False , lowercase : bool = True , ): '''simple docstring''' for i, (image, scale, controlnet) in enumerate(zip(lowercase , lowercase , self.nets ) ): _snake_case , _snake_case = controlnet( lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) # merge samples if i == 0: _snake_case , _snake_case = down_samples, mid_sample else: _snake_case = [ samples_prev + samples_curr for samples_prev, samples_curr in zip(lowercase , lowercase ) ] mid_block_res_sample += mid_sample return down_block_res_samples, mid_block_res_sample def A ( self : Dict , lowercase : Union[str, os.PathLike] , lowercase : bool = True , lowercase : Callable = None , lowercase : bool = False , lowercase : Optional[str] = None , ): '''simple docstring''' _snake_case = 0 _snake_case = save_directory for controlnet in self.nets: controlnet.save_pretrained( lowercase , is_main_process=lowercase , save_function=lowercase , safe_serialization=lowercase , variant=lowercase , ) idx += 1 _snake_case = model_path_to_save + f'''_{idx}''' @classmethod def A ( cls : Any , lowercase : Optional[Union[str, os.PathLike]] , **lowercase : List[str] ): '''simple docstring''' _snake_case = 0 _snake_case = [] # load controlnet and append to list until no controlnet directory exists anymore # first controlnet has to be saved under `./mydirectory/controlnet` to be compliant with `DiffusionPipeline.from_prertained` # second, third, ... controlnets have to be saved under `./mydirectory/controlnet_1`, `./mydirectory/controlnet_2`, ... _snake_case = pretrained_model_path while os.path.isdir(lowercase ): _snake_case = ControlNetModel.from_pretrained(lowercase , **lowercase ) controlnets.append(lowercase ) idx += 1 _snake_case = pretrained_model_path + f'''_{idx}''' logger.info(f'''{len(lowercase )} controlnets loaded from {pretrained_model_path}.''' ) if len(lowercase ) == 0: raise ValueError( f'''No ControlNets found under {os.path.dirname(lowercase )}. Expected at least {pretrained_model_path + '_0'}.''' ) return cls(lowercase )
282
1
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=UpperCAmelCase ) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _UpperCAmelCase : str = field(default="language-modeling" ,metadata={"include_in_asdict_even_if_is_default": True} ) _UpperCAmelCase : ClassVar[Features] = Features({"text": Value("string" )} ) _UpperCAmelCase : ClassVar[Features] = Features({} ) _UpperCAmelCase : str = "text" @property def A ( self : str ): '''simple docstring''' return {self.text_column: "text"}
282
class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : List[str] , lowercase : list[int] ): '''simple docstring''' _snake_case = len(lowercase ) _snake_case = [0] * len_array if len_array > 0: _snake_case = array[0] for i in range(1 , lowercase ): _snake_case = self.prefix_sum[i - 1] + array[i] def A ( self : Optional[Any] , lowercase : int , lowercase : int ): '''simple docstring''' if start == 0: return self.prefix_sum[end] return self.prefix_sum[end] - self.prefix_sum[start - 1] def A ( self : Union[str, Any] , lowercase : int ): '''simple docstring''' _snake_case = {0} for sum_item in self.prefix_sum: if sum_item - target_sum in sums: return True sums.add(lowercase ) return False if __name__ == "__main__": import doctest doctest.testmod()
282
1
import unittest from transformers import LiltConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, ) from transformers.models.lilt.modeling_lilt import LILT_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : Dict , lowercase : str , lowercase : List[str]=13 , lowercase : Any=7 , lowercase : Dict=True , lowercase : str=True , lowercase : List[Any]=True , lowercase : Any=True , lowercase : Tuple=99 , lowercase : str=24 , lowercase : str=2 , lowercase : Any=6 , lowercase : Dict=37 , lowercase : List[str]="gelu" , lowercase : Dict=0.1 , lowercase : Tuple=0.1 , lowercase : Optional[Any]=512 , lowercase : List[Any]=16 , lowercase : str=2 , lowercase : int=0.02 , lowercase : List[Any]=3 , lowercase : List[Any]=None , lowercase : int=1_000 , ): '''simple docstring''' _snake_case = parent _snake_case = batch_size _snake_case = seq_length _snake_case = is_training _snake_case = use_input_mask _snake_case = use_token_type_ids _snake_case = use_labels _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = type_sequence_label_size _snake_case = initializer_range _snake_case = num_labels _snake_case = scope _snake_case = range_bbox def A ( self : List[Any] ): '''simple docstring''' _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: _snake_case = bbox[i, j, 3] _snake_case = bbox[i, j, 1] _snake_case = t if bbox[i, j, 2] < bbox[i, j, 0]: _snake_case = bbox[i, j, 2] _snake_case = bbox[i, j, 0] _snake_case = t _snake_case = None if self.use_input_mask: _snake_case = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _snake_case = None if self.use_token_type_ids: _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _snake_case = None _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _snake_case = self.get_config() return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels def A ( self : List[str] ): '''simple docstring''' return LiltConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def A ( self : str , lowercase : Tuple , lowercase : Tuple , lowercase : str , lowercase : Any , lowercase : Union[str, Any] , lowercase : List[str] , lowercase : str , ): '''simple docstring''' _snake_case = LiltModel(config=lowercase ) model.to(lowercase ) model.eval() _snake_case = model(lowercase , bbox=lowercase , attention_mask=lowercase , token_type_ids=lowercase ) _snake_case = model(lowercase , bbox=lowercase , token_type_ids=lowercase ) _snake_case = model(lowercase , bbox=lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A ( self : List[Any] , lowercase : int , lowercase : int , lowercase : Any , lowercase : Optional[int] , lowercase : Union[str, Any] , lowercase : Optional[Any] , lowercase : Optional[int] , ): '''simple docstring''' _snake_case = self.num_labels _snake_case = LiltForTokenClassification(config=lowercase ) model.to(lowercase ) model.eval() _snake_case = model( lowercase , bbox=lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : List[str] , lowercase : Union[str, Any] , lowercase : str , lowercase : Dict , lowercase : Optional[int] , lowercase : List[str] , lowercase : int , lowercase : int , ): '''simple docstring''' _snake_case = LiltForQuestionAnswering(config=lowercase ) model.to(lowercase ) model.eval() _snake_case = model( lowercase , bbox=lowercase , attention_mask=lowercase , token_type_ids=lowercase , start_positions=lowercase , end_positions=lowercase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = self.prepare_config_and_inputs() ( ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ) = config_and_inputs _snake_case = { 'input_ids': input_ids, 'bbox': bbox, 'token_type_ids': token_type_ids, 'attention_mask': input_mask, } return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ,UpperCAmelCase ,UpperCAmelCase ,unittest.TestCase ): '''simple docstring''' _UpperCAmelCase : List[Any] = ( ( LiltModel, LiltForSequenceClassification, LiltForTokenClassification, LiltForQuestionAnswering, ) if is_torch_available() else () ) _UpperCAmelCase : List[str] = ( { "feature-extraction": LiltModel, "question-answering": LiltForQuestionAnswering, "text-classification": LiltForSequenceClassification, "token-classification": LiltForTokenClassification, "zero-shot": LiltForSequenceClassification, } if is_torch_available() else {} ) _UpperCAmelCase : Optional[Any] = False _UpperCAmelCase : Union[str, Any] = False def A ( self : Dict , lowercase : Dict , lowercase : Optional[int] , lowercase : Optional[int] , lowercase : List[str] , lowercase : Tuple ): '''simple docstring''' return True def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = LiltModelTester(self ) _snake_case = ConfigTester(self , config_class=lowercase , hidden_size=37 ) def A ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def A ( self : Dict ): '''simple docstring''' _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def A ( self : List[Any] ): '''simple docstring''' _snake_case = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _snake_case = type self.model_tester.create_and_check_model(*lowercase ) def A ( self : Any ): '''simple docstring''' _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase ) def A ( self : Any ): '''simple docstring''' _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase ) @slow def A ( self : Union[str, Any] ): '''simple docstring''' for model_name in LILT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = LiltModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) @require_torch @slow class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' def A ( self : Tuple ): '''simple docstring''' _snake_case = LiltModel.from_pretrained('SCUT-DLVCLab/lilt-roberta-en-base' ).to(lowercase ) _snake_case = torch.tensor([[1, 2]] , device=lowercase ) _snake_case = torch.tensor([[[1, 2, 3, 4], [5, 6, 7, 8]]] , device=lowercase ) # forward pass with torch.no_grad(): _snake_case = model(input_ids=lowercase , bbox=lowercase ) _snake_case = torch.Size([1, 2, 768] ) _snake_case = torch.tensor( [[-0.0653, 0.0950, -0.0061], [-0.0545, 0.0926, -0.0324]] , device=lowercase , ) self.assertTrue(outputs.last_hidden_state.shape , lowercase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :, :3] , lowercase , atol=1E-3 ) )
282
from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class SCREAMING_SNAKE_CASE__ ( nn.Module ): '''simple docstring''' def __init__( self : Optional[int] , lowercase : int = 16 , lowercase : int = 88 , lowercase : Optional[int] = None , lowercase : int = 1 , lowercase : float = 0.0 , lowercase : int = 32 , lowercase : Optional[int] = None , lowercase : bool = False , lowercase : Optional[int] = None , lowercase : Optional[int] = None , lowercase : str = "geglu" , lowercase : Optional[int] = None , ): '''simple docstring''' super().__init__() _snake_case = nn.ModuleList( [ TransformeraDModel( num_attention_heads=lowercase , attention_head_dim=lowercase , in_channels=lowercase , num_layers=lowercase , dropout=lowercase , norm_num_groups=lowercase , cross_attention_dim=lowercase , attention_bias=lowercase , sample_size=lowercase , num_vector_embeds=lowercase , activation_fn=lowercase , num_embeds_ada_norm=lowercase , ) for _ in range(2 ) ] ) # Variables that can be set by a pipeline: # The ratio of transformer1 to transformer2's output states to be combined during inference _snake_case = 0.5 # The shape of `encoder_hidden_states` is expected to be # `(batch_size, condition_lengths[0]+condition_lengths[1], num_features)` _snake_case = [77, 257] # Which transformer to use to encode which condition. # E.g. `(1, 0)` means that we'll use `transformers[1](conditions[0])` and `transformers[0](conditions[1])` _snake_case = [1, 0] def A ( self : Optional[int] , lowercase : Optional[int] , lowercase : List[Any] , lowercase : List[str]=None , lowercase : Tuple=None , lowercase : Dict=None , lowercase : bool = True , ): '''simple docstring''' _snake_case = hidden_states _snake_case = [] _snake_case = 0 # attention_mask is not used yet for i in range(2 ): # for each of the two transformers, pass the corresponding condition tokens _snake_case = encoder_hidden_states[:, tokens_start : tokens_start + self.condition_lengths[i]] _snake_case = self.transformer_index_for_condition[i] _snake_case = self.transformers[transformer_index]( lowercase , encoder_hidden_states=lowercase , timestep=lowercase , cross_attention_kwargs=lowercase , return_dict=lowercase , )[0] encoded_states.append(encoded_state - input_states ) tokens_start += self.condition_lengths[i] _snake_case = encoded_states[0] * self.mix_ratio + encoded_states[1] * (1 - self.mix_ratio) _snake_case = output_states + input_states if not return_dict: return (output_states,) return TransformeraDModelOutput(sample=lowercase )
282
1
class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : List[str] , lowercase : list[int] ): '''simple docstring''' _snake_case = len(lowercase ) _snake_case = [0] * len_array if len_array > 0: _snake_case = array[0] for i in range(1 , lowercase ): _snake_case = self.prefix_sum[i - 1] + array[i] def A ( self : Optional[Any] , lowercase : int , lowercase : int ): '''simple docstring''' if start == 0: return self.prefix_sum[end] return self.prefix_sum[end] - self.prefix_sum[start - 1] def A ( self : Union[str, Any] , lowercase : int ): '''simple docstring''' _snake_case = {0} for sum_item in self.prefix_sum: if sum_item - target_sum in sums: return True sums.add(lowercase ) return False if __name__ == "__main__": import doctest doctest.testmod()
282
import tempfile import unittest from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from transformers.testing_utils import ( is_torch_available, require_optimum, require_torch, slow, ) if is_torch_available(): import torch @require_torch @require_optimum @slow class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' def A ( self : Optional[int] ): '''simple docstring''' _snake_case = 'hf-internal-testing/tiny-random-t5' _snake_case = AutoTokenizer.from_pretrained(lowercase ) _snake_case = AutoModelForSeqaSeqLM.from_pretrained(lowercase ) _snake_case = tokenizer('This is me' , return_tensors='pt' ) _snake_case = model.to_bettertransformer() self.assertTrue(any('BetterTransformer' in mod.__class__.__name__ for _, mod in model.named_modules() ) ) _snake_case = model.generate(**lowercase ) _snake_case = model.reverse_bettertransformer() self.assertFalse(any('BetterTransformer' in mod.__class__.__name__ for _, mod in model.named_modules() ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowercase ) _snake_case = AutoModelForSeqaSeqLM.from_pretrained(lowercase ) self.assertFalse( any('BetterTransformer' in mod.__class__.__name__ for _, mod in model_reloaded.named_modules() ) ) _snake_case = model_reloaded.generate(**lowercase ) self.assertTrue(torch.allclose(lowercase , lowercase ) ) def A ( self : List[Any] ): '''simple docstring''' _snake_case = 'hf-internal-testing/tiny-random-t5' _snake_case = AutoModelForSeqaSeqLM.from_pretrained(lowercase ) _snake_case = model.to_bettertransformer() with tempfile.TemporaryDirectory() as tmpdirname: with self.assertRaises(lowercase ): model.save_pretrained(lowercase ) _snake_case = model.reverse_bettertransformer() model.save_pretrained(lowercase )
282
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase : Optional[int] = logging.get_logger(__name__) _lowerCamelCase : int = { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/config.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/config.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json''' ), '''distilbert-base-uncased-finetuned-sst-2-english''': ( '''https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json''' ), } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _UpperCAmelCase : List[Any] = "distilbert" _UpperCAmelCase : str = { "hidden_size": "dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", } def __init__( self : Union[str, Any] , lowercase : List[Any]=30_522 , lowercase : Dict=512 , lowercase : int=False , lowercase : int=6 , lowercase : Union[str, Any]=12 , lowercase : Union[str, Any]=768 , lowercase : Union[str, Any]=4 * 768 , lowercase : Dict=0.1 , lowercase : List[Any]=0.1 , lowercase : Dict="gelu" , lowercase : Tuple=0.02 , lowercase : str=0.1 , lowercase : str=0.2 , lowercase : List[str]=0 , **lowercase : Tuple , ): '''simple docstring''' _snake_case = vocab_size _snake_case = max_position_embeddings _snake_case = sinusoidal_pos_embds _snake_case = n_layers _snake_case = n_heads _snake_case = dim _snake_case = hidden_dim _snake_case = dropout _snake_case = attention_dropout _snake_case = activation _snake_case = initializer_range _snake_case = qa_dropout _snake_case = seq_classif_dropout super().__init__(**lowercase , pad_token_id=lowercase ) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' @property def A ( self : Tuple ): '''simple docstring''' if self.task == "multiple-choice": _snake_case = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _snake_case = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
282
import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel _lowerCamelCase : List[Any] = HfApi() _lowerCamelCase : Dict = {} # fmt: off _lowerCamelCase : List[Any] = torch.tensor([ -0.7_5_1_5, -1.6_8_8_3, 0.2_4_2_0, 0.0_3_0_0, 0.6_3_4_7, 1.3_4_3_3, -1.1_7_4_3, -3.7_4_6_7, 1.2_3_4_2, -2.2_4_8_5, 0.4_6_3_6, 0.8_0_7_6, -0.7_9_9_1, 0.3_9_6_9, 0.8_4_9_8, 0.9_1_8_9, -1.8_8_8_7, -3.3_5_2_2, 0.7_6_3_9, 0.2_0_4_0, 0.6_2_7_1, -2.7_1_4_8, -1.6_3_1_6, 3.0_8_3_9, 0.3_1_8_6, 0.2_7_2_1, -0.9_7_5_9, -1.2_4_6_1, 2.6_2_5_7, 1.3_5_5_7 ]) _lowerCamelCase : int = torch.tensor([ -2.3_6_3_9, -2.5_3_4_4, 0.0_0_5_4, -0.6_6_7_4, 1.5_9_9_0, 1.0_1_5_8, 0.3_1_2_4, -2.1_4_3_6, 1.8_7_9_5, -2.5_4_2_9, -0.1_5_6_6, -0.3_9_7_3, 1.2_4_9_0, 2.6_4_4_7, 1.2_2_8_3, -0.5_2_0_8, -2.8_1_5_4, -3.5_1_1_9, 2.3_8_3_8, 1.2_0_3_3, 1.7_2_0_1, -2.1_2_5_6, -1.4_5_7_6, 2.7_9_4_8, 2.4_2_0_4, -0.9_7_5_2, -1.2_5_4_6, 0.8_0_2_7, 3.2_7_5_8, 3.1_3_6_5 ]) _lowerCamelCase : Optional[int] = torch.tensor([ -0.6_5_3_1, -0.6_8_9_1, -0.3_1_7_2, -0.5_3_7_5, -0.9_1_4_0, -0.5_3_6_7, -0.1_1_7_5, -0.7_8_6_9, -0.3_8_0_8, -0.4_5_1_3, -0.2_0_9_8, -0.0_0_8_3, 0.3_1_8_3, 0.5_1_4_0, 0.2_2_4_7, -0.1_3_0_4, -0.1_3_0_2, -0.2_8_0_2, -0.2_0_8_4, -0.2_0_2_5, -0.4_9_6_7, -0.4_8_7_3, -0.0_8_6_1, 0.6_9_2_5, 0.0_2_5_0, 0.1_2_9_0, -0.1_5_4_3, 0.6_3_1_6, 1.0_4_6_0, 1.4_9_4_3 ]) _lowerCamelCase : Dict = torch.tensor([ 0.0_9_1_1, 0.1_1_0_7, 0.0_1_8_2, 0.0_4_3_5, -0.0_8_0_5, -0.0_6_0_8, 0.0_3_8_1, 0.2_1_7_2, -0.0_2_8_0, 0.1_3_2_7, -0.0_2_9_9, -0.0_2_5_5, -0.0_0_5_0, -0.1_1_7_0, -0.1_0_4_6, 0.0_3_0_9, 0.1_3_6_7, 0.1_7_2_8, -0.0_5_3_3, -0.0_7_4_8, -0.0_5_3_4, 0.1_6_2_4, 0.0_3_8_4, -0.1_8_0_5, -0.0_7_0_7, 0.0_6_4_2, 0.0_2_2_0, -0.0_1_3_4, -0.1_3_3_3, -0.1_5_0_5 ]) _lowerCamelCase : Dict = torch.tensor([ 0.1_3_2_1, 0.1_3_3_7, 0.0_4_4_0, 0.0_6_2_2, -0.0_5_9_1, -0.0_3_7_0, 0.0_5_0_3, 0.2_1_3_3, -0.0_1_7_7, 0.1_4_1_5, -0.0_1_1_6, -0.0_1_1_2, 0.0_0_4_4, -0.0_9_8_0, -0.0_7_8_9, 0.0_3_9_5, 0.1_5_0_2, 0.1_7_8_5, -0.0_4_8_8, -0.0_5_1_4, -0.0_4_0_4, 0.1_5_3_9, 0.0_4_5_4, -0.1_5_5_9, -0.0_6_6_5, 0.0_6_5_9, 0.0_3_8_3, -0.0_0_0_5, -0.1_2_6_6, -0.1_3_8_6 ]) _lowerCamelCase : List[Any] = torch.tensor([ 0.1_1_5_4, 0.1_2_1_8, 0.0_3_0_7, 0.0_5_2_6, -0.0_7_1_1, -0.0_5_4_1, 0.0_3_6_6, 0.2_0_7_8, -0.0_2_6_7, 0.1_3_1_7, -0.0_2_2_6, -0.0_1_9_3, -0.0_0_1_4, -0.1_0_5_5, -0.0_9_0_2, 0.0_3_3_0, 0.1_3_9_1, 0.1_7_0_9, -0.0_5_6_2, -0.0_6_9_3, -0.0_5_6_0, 0.1_4_8_2, 0.0_3_8_1, -0.1_6_8_3, -0.0_6_8_1, 0.0_6_6_1, 0.0_3_3_1, -0.0_0_4_6, -0.1_2_6_8, -0.1_4_3_1 ]) _lowerCamelCase : Dict = torch.tensor([ 0.1_1_9_2, 0.1_2_4_0, 0.0_4_1_4, 0.0_6_0_6, -0.0_5_5_7, -0.0_4_1_2, 0.0_4_3_0, 0.2_0_4_2, -0.0_2_0_0, 0.1_3_8_5, -0.0_1_1_5, -0.0_1_3_2, 0.0_0_1_7, -0.0_9_6_5, -0.0_8_0_2, 0.0_3_9_8, 0.1_4_3_3, 0.1_7_4_7, -0.0_4_5_8, -0.0_5_3_3, -0.0_4_0_7, 0.1_5_4_5, 0.0_4_1_9, -0.1_5_7_4, -0.0_6_4_5, 0.0_6_2_6, 0.0_3_4_1, -0.0_0_1_0, -0.1_1_9_9, -0.1_3_9_0 ]) _lowerCamelCase : int = torch.tensor([ 0.1_0_7_5, 0.1_0_7_4, 0.0_2_0_5, 0.0_4_3_1, -0.0_7_7_4, -0.0_6_0_7, 0.0_2_9_8, 0.2_0_4_2, -0.0_3_2_0, 0.1_2_6_7, -0.0_2_8_1, -0.0_2_5_0, -0.0_0_6_4, -0.1_0_9_1, -0.0_9_4_6, 0.0_2_9_0, 0.1_3_2_8, 0.1_6_5_0, -0.0_5_8_0, -0.0_7_3_8, -0.0_5_8_6, 0.1_4_4_0, 0.0_3_3_7, -0.1_7_4_6, -0.0_7_1_2, 0.0_6_0_5, 0.0_2_5_0, -0.0_0_9_9, -0.1_3_1_6, -0.1_4_7_3 ]) _lowerCamelCase : int = torch.tensor([ -1.4_5_7_2, -2.0_4_8_1, -0.0_4_1_4, -0.6_0_0_5, 1.4_1_3_6, 0.5_8_4_8, 0.4_0_2_8, -2.7_3_3_0, 1.2_2_1_2, -2.1_2_2_8, 0.2_1_5_5, 0.4_0_3_9, 0.7_6_6_2, 2.0_5_3_5, 0.7_4_7_7, -0.3_2_4_3, -2.1_7_5_8, -2.7_6_4_8, 1.6_9_4_7, 0.7_0_2_6, 1.2_3_3_8, -1.6_0_7_8, -0.8_6_8_2, 2.2_8_1_0, 1.8_5_7_4, -0.5_7_1_8, -0.5_5_8_6, -0.0_1_8_6, 2.3_4_1_5, 2.1_2_5_1]) _lowerCamelCase : Tuple = torch.tensor([ -1.3_6_9_0, -1.9_7_2_0, -0.4_0_9_0, -0.6_9_6_6, 1.4_6_6_0, 0.9_9_3_8, -0.1_3_8_5, -2.7_3_2_4, 0.7_7_3_6, -1.8_9_1_7, 0.2_9_2_3, 0.4_2_9_3, 0.1_6_9_3, 1.4_1_1_2, 1.1_8_8_7, -0.3_1_8_1, -2.2_1_6_0, -2.6_3_8_1, 1.3_1_7_0, 0.8_1_6_3, 0.9_2_4_0, -1.6_5_4_4, -0.6_0_9_9, 2.5_2_5_9, 1.6_4_3_0, -0.9_0_9_0, -0.9_3_9_2, -0.0_1_2_6, 2.4_2_6_8, 2.3_2_6_6 ]) _lowerCamelCase : List[str] = torch.tensor([ -1.3_5_2_5, -1.9_6_2_8, -0.3_9_5_6, -0.6_8_6_0, 1.4_6_6_4, 1.0_0_1_4, -0.1_2_5_9, -2.7_2_1_2, 0.7_7_7_2, -1.8_8_1_1, 0.2_9_9_6, 0.4_3_8_8, 0.1_7_0_4, 1.4_0_2_9, 1.1_7_0_1, -0.3_0_2_7, -2.2_0_5_3, -2.6_2_8_7, 1.3_3_5_0, 0.8_1_3_1, 0.9_2_7_4, -1.6_2_9_2, -0.6_0_9_8, 2.5_1_3_1, 1.6_5_0_5, -0.8_9_5_8, -0.9_2_9_8, -0.0_1_5_1, 2.4_2_5_7, 2.3_3_5_5 ]) _lowerCamelCase : int = torch.tensor([ -2.0_5_8_5, -2.7_8_9_7, -0.2_8_5_0, -0.8_9_4_0, 1.9_0_5_2, 0.5_7_0_2, 0.6_3_4_5, -3.8_9_5_9, 1.5_9_3_2, -3.2_3_1_9, 0.1_9_7_4, 0.0_2_8_7, 1.7_5_6_6, 2.6_5_4_3, 0.8_3_8_7, -0.5_3_5_1, -3.2_7_3_6, -4.3_3_7_5, 2.9_0_2_9, 1.6_3_9_0, 1.4_6_4_0, -2.1_7_0_1, -1.9_0_1_3, 2.9_3_4_1, 3.4_9_8_1, -0.6_2_5_5, -1.1_6_4_4, -0.1_5_9_1, 3.7_0_9_7, 3.2_0_6_6 ]) _lowerCamelCase : Tuple = torch.tensor([ -2.3_1_3_9, -2.5_5_9_4, -0.0_1_9_7, -0.6_7_8_5, 1.7_0_0_1, 1.1_6_0_6, 0.3_0_7_5, -2.1_7_4_0, 1.8_0_7_1, -2.5_6_3_0, -0.0_9_2_6, -0.3_8_1_1, 1.2_1_1_6, 2.6_2_4_6, 1.2_7_3_1, -0.5_3_9_8, -2.8_1_5_3, -3.6_1_4_0, 2.3_8_9_3, 1.3_2_6_2, 1.6_2_5_8, -2.1_8_5_6, -1.3_2_6_7, 2.8_3_9_5, 2.3_7_7_9, -1.0_6_2_3, -1.2_4_6_8, 0.8_9_5_9, 3.3_3_6_7, 3.2_2_4_3 ]) _lowerCamelCase : int = torch.tensor([ -2.0_6_2_8, -2.7_6_6_7, -0.2_0_8_9, -0.8_2_6_3, 2.0_5_3_9, 0.5_9_9_2, 0.6_4_9_5, -3.8_3_3_6, 1.6_0_2_5, -3.2_8_1_7, 0.1_7_2_1, -0.0_6_3_3, 1.7_5_1_6, 2.7_0_3_9, 0.8_1_0_0, -0.5_9_0_8, -3.2_1_1_3, -4.4_3_4_3, 2.9_2_5_7, 1.3_6_3_2, 1.5_5_6_2, -2.1_4_8_9, -1.9_8_9_4, 3.0_5_6_0, 3.3_3_9_6, -0.7_3_2_8, -1.0_4_1_7, 0.0_3_8_3, 3.7_0_9_3, 3.2_3_4_3 ]) _lowerCamelCase : List[Any] = torch.tensor([ -1.4_5_7_4, -2.0_5_6_9, -0.0_4_7_3, -0.6_1_1_7, 1.4_0_1_8, 0.5_7_6_9, 0.4_1_2_9, -2.7_3_4_4, 1.2_2_4_1, -2.1_3_9_7, 0.2_0_0_0, 0.3_9_3_7, 0.7_6_1_6, 2.0_4_5_3, 0.7_3_2_4, -0.3_3_9_1, -2.1_7_4_6, -2.7_7_4_4, 1.6_9_6_3, 0.6_9_2_1, 1.2_1_8_7, -1.6_1_7_2, -0.8_8_7_7, 2.2_4_3_9, 1.8_4_7_1, -0.5_8_3_9, -0.5_6_0_5, -0.0_4_6_4, 2.3_2_5_0, 2.1_2_1_9 ]) # fmt: on _lowerCamelCase : List[str] = api.list_models(filter='''diffusers''') for mod in models: if "google" in mod.author or mod.modelId == "CompVis/ldm-celebahq-256": _lowerCamelCase : Any = '''/home/patrick/google_checkpoints/''' + mod.modelId.split('''/''')[-1] print(F'Started running {mod.modelId}!!!') if mod.modelId.startswith('''CompVis'''): _lowerCamelCase : Optional[Any] = UNetaDModel.from_pretrained(local_checkpoint, subfolder='''unet''') else: _lowerCamelCase : int = UNetaDModel.from_pretrained(local_checkpoint) torch.manual_seed(0) random.seed(0) _lowerCamelCase : Union[str, Any] = torch.randn(1, model.config.in_channels, model.config.sample_size, model.config.sample_size) _lowerCamelCase : int = torch.tensor([10] * noise.shape[0]) with torch.no_grad(): _lowerCamelCase : int = model(noise, time_step).sample assert torch.allclose( logits[0, 0, 0, :30], results['''_'''.join('''_'''.join(mod.modelId.split('''/''')).split('''-'''))], atol=1E-3 ) print(F'{mod.modelId} has passed successfully!!!')
282
1
from __future__ import annotations def a_ ( __lowercase : str ) -> list[int]: return [ord(__lowercase ) - 96 for elem in plain] def a_ ( __lowercase : list[int] ) -> str: return "".join(chr(elem + 96 ) for elem in encoded ) def a_ ( ) -> None: _snake_case = encode(input('-> ' ).strip().lower() ) print('Encoded: ' , __lowercase ) print('Decoded:' , decode(__lowercase ) ) if __name__ == "__main__": main()
282
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def A ( self : List[str] ): '''simple docstring''' _snake_case = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase , 'tf_padding' ) ) self.parent.assertTrue(hasattr(lowercase , 'depth_multiplier' ) ) class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : Dict , lowercase : List[str] , lowercase : Dict=13 , lowercase : Optional[int]=3 , lowercase : Any=32 , lowercase : Any=0.25 , lowercase : Union[str, Any]=8 , lowercase : List[Any]=8 , lowercase : List[Any]=6 , lowercase : Dict=32 , lowercase : Dict=True , lowercase : Optional[Any]=True , lowercase : Tuple=True , lowercase : Tuple="relu6" , lowercase : List[Any]=1_280 , lowercase : Optional[Any]=0.1 , lowercase : int=0.02 , lowercase : Optional[Any]=True , lowercase : List[str]=True , lowercase : List[str]=10 , lowercase : Optional[Any]=None , ): '''simple docstring''' _snake_case = parent _snake_case = batch_size _snake_case = num_channels _snake_case = image_size _snake_case = depth_multiplier _snake_case = depth_divisible_by _snake_case = min_depth _snake_case = expand_ratio _snake_case = tf_padding _snake_case = output_stride _snake_case = first_layer_is_expansion _snake_case = finegrained_output _snake_case = hidden_act _snake_case = last_hidden_size if finegrained_output else int(last_hidden_size * depth_multiplier ) _snake_case = classifier_dropout_prob _snake_case = use_labels _snake_case = is_training _snake_case = num_labels _snake_case = initializer_range _snake_case = scope def A ( self : Union[str, Any] ): '''simple docstring''' _snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case = None _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.num_labels ) _snake_case = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) _snake_case = self.get_config() return config, pixel_values, labels, pixel_labels def A ( self : str ): '''simple docstring''' return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , depth_divisible_by=self.depth_divisible_by , min_depth=self.min_depth , expand_ratio=self.expand_ratio , output_stride=self.output_stride , first_layer_is_expansion=self.first_layer_is_expansion , finegrained_output=self.finegrained_output , hidden_act=self.hidden_act , tf_padding=self.tf_padding , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def A ( self : Optional[Any] , lowercase : str , lowercase : List[str] , lowercase : str , lowercase : Dict ): '''simple docstring''' _snake_case = MobileNetVaModel(config=lowercase ) model.to(lowercase ) model.eval() _snake_case = model(lowercase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) self.parent.assertEqual( result.pooler_output.shape , (self.batch_size, self.last_hidden_size) , ) def A ( self : List[Any] , lowercase : Optional[int] , lowercase : Optional[int] , lowercase : Optional[Any] , lowercase : List[Any] ): '''simple docstring''' _snake_case = self.num_labels _snake_case = MobileNetVaForImageClassification(lowercase ) model.to(lowercase ) model.eval() _snake_case = model(lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Any , lowercase : int , lowercase : Dict , lowercase : int , lowercase : List[Any] ): '''simple docstring''' _snake_case = self.num_labels _snake_case = MobileNetVaForSemanticSegmentation(lowercase ) model.to(lowercase ) model.eval() _snake_case = model(lowercase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) _snake_case = model(lowercase , labels=lowercase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def A ( self : str ): '''simple docstring''' _snake_case = self.prepare_config_and_inputs() _snake_case , _snake_case , _snake_case , _snake_case = config_and_inputs _snake_case = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ,UpperCAmelCase ,unittest.TestCase ): '''simple docstring''' _UpperCAmelCase : str = ( (MobileNetVaModel, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation) if is_torch_available() else () ) _UpperCAmelCase : str = ( { "feature-extraction": MobileNetVaModel, "image-classification": MobileNetVaForImageClassification, "image-segmentation": MobileNetVaForSemanticSegmentation, } if is_torch_available() else {} ) _UpperCAmelCase : Optional[int] = False _UpperCAmelCase : Dict = False _UpperCAmelCase : Dict = False _UpperCAmelCase : Union[str, Any] = False def A ( self : Any ): '''simple docstring''' _snake_case = MobileNetVaModelTester(self ) _snake_case = MobileNetVaConfigTester(self , config_class=lowercase , has_text_modality=lowercase ) def A ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='MobileNetV2 does not use inputs_embeds' ) def A ( self : List[str] ): '''simple docstring''' pass @unittest.skip(reason='MobileNetV2 does not support input and output embeddings' ) def A ( self : int ): '''simple docstring''' pass @unittest.skip(reason='MobileNetV2 does not output attentions' ) def A ( self : Any ): '''simple docstring''' pass def A ( self : Optional[int] ): '''simple docstring''' _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(lowercase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ['pixel_values'] self.assertListEqual(arg_names[:1] , lowercase ) def A ( self : List[str] ): '''simple docstring''' _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def A ( self : List[Any] ): '''simple docstring''' def check_hidden_states_output(lowercase : List[Any] , lowercase : Union[str, Any] , lowercase : str ): _snake_case = model_class(lowercase ) model.to(lowercase ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(lowercase , lowercase ) ) _snake_case = outputs.hidden_states _snake_case = 16 self.assertEqual(len(lowercase ) , lowercase ) _snake_case , _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = True check_hidden_states_output(lowercase , lowercase , lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case = True check_hidden_states_output(lowercase , lowercase , lowercase ) def A ( self : Tuple ): '''simple docstring''' _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase ) def A ( self : Dict ): '''simple docstring''' _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowercase ) @slow def A ( self : List[Any] ): '''simple docstring''' for model_name in MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = MobileNetVaModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) def a_ ( ) -> Union[str, Any]: _snake_case = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' @cached_property def A ( self : Optional[Any] ): '''simple docstring''' return ( MobileNetVaImageProcessor.from_pretrained('google/mobilenet_v2_1.0_224' ) if is_vision_available() else None ) @slow def A ( self : List[Any] ): '''simple docstring''' _snake_case = MobileNetVaForImageClassification.from_pretrained('google/mobilenet_v2_1.0_224' ).to(lowercase ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(images=lowercase , return_tensors='pt' ).to(lowercase ) # forward pass with torch.no_grad(): _snake_case = model(**lowercase ) # verify the logits _snake_case = torch.Size((1, 1_001) ) self.assertEqual(outputs.logits.shape , lowercase ) _snake_case = torch.tensor([0.2445, -1.1993, 0.1905] ).to(lowercase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase , atol=1E-4 ) ) @slow def A ( self : Dict ): '''simple docstring''' _snake_case = MobileNetVaForSemanticSegmentation.from_pretrained('google/deeplabv3_mobilenet_v2_1.0_513' ) _snake_case = model.to(lowercase ) _snake_case = MobileNetVaImageProcessor.from_pretrained('google/deeplabv3_mobilenet_v2_1.0_513' ) _snake_case = prepare_img() _snake_case = image_processor(images=lowercase , return_tensors='pt' ).to(lowercase ) # forward pass with torch.no_grad(): _snake_case = model(**lowercase ) _snake_case = outputs.logits # verify the logits _snake_case = torch.Size((1, 21, 65, 65) ) self.assertEqual(logits.shape , lowercase ) _snake_case = torch.tensor( [ [[17.5790, 17.7581, 18.3355], [18.3257, 18.4230, 18.8973], [18.6169, 18.8650, 19.2187]], [[-2.1595, -2.0977, -2.3741], [-2.4226, -2.3028, -2.6835], [-2.7819, -2.5991, -2.7706]], [[4.2058, 4.8317, 4.7638], [4.4136, 5.0361, 4.9383], [4.5028, 4.9644, 4.8734]], ] , device=lowercase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowercase , atol=1E-4 ) )
282
1
from __future__ import annotations from functools import lru_cache from math import ceil _lowerCamelCase : str = 100 _lowerCamelCase : List[Any] = set(range(3, NUM_PRIMES, 2)) primes.add(2) _lowerCamelCase : int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=100 ) def a_ ( __lowercase : int ) -> set[int]: if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} _snake_case = set() _snake_case = 42 _snake_case = 42 for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def a_ ( __lowercase : int = 5_000 ) -> int | None: for number_to_partition in range(1 , __lowercase ): if len(partition(__lowercase ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(F'{solution() = }')
282
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def a_ ( __lowercase : Dict , __lowercase : int , __lowercase : Optional[Any]=None ) -> Any: # set parameter of one layer assert torch_layer.weight.shape == weight.shape, f'''{torch_layer} layer.weight does not match''' _snake_case = nn.Parameter(__lowercase ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f'''{torch_layer} layer.bias does not match''' _snake_case = nn.Parameter(__lowercase ) def a_ ( __lowercase : Any , __lowercase : Dict , __lowercase : Union[str, Any] ) -> Optional[Any]: # set torch weights for 1-to-1 comparison _snake_case = np.asarray(weights[0] ) _snake_case = np.asarray(weights[1] ) _snake_case = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(__lowercase ).transpose(1 , 2 ).contiguous().view(-1 , __lowercase ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__lowercase ).transpose(1 , 2 ).contiguous().view(-1 , __lowercase ) , ) set_param( torch_layer.output.dense , torch.tensor(__lowercase ).view(-1 , __lowercase ).contiguous().transpose(0 , 1 ) , ) def a_ ( __lowercase : str , __lowercase : Tuple , __lowercase : Any ) -> Optional[Any]: # set torch weights for 1-to-1 comparison _snake_case = np.asarray(weights[0] ) _snake_case = np.asarray(weights[1] ) _snake_case = np.asarray(weights[2] ) _snake_case = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(__lowercase ).transpose(1 , 2 ).contiguous().view(-1 , __lowercase ) , ) set_param( torch_layer.self_attention.key , torch.tensor(__lowercase ).transpose(1 , 2 ).contiguous().view(-1 , __lowercase ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__lowercase ).transpose(1 , 2 ).contiguous().view(-1 , __lowercase ) , ) set_param( torch_layer.output.dense , torch.tensor(__lowercase ).view(-1 , __lowercase ).contiguous().transpose(0 , 1 ) , ) def a_ ( __lowercase : Dict , __lowercase : List[str] , __lowercase : Union[str, Any] ) -> Optional[Any]: # layernorm 1 _snake_case = weights[0][0][0] _snake_case = np.asarray(layer_norm_a[0] ) _snake_case = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(__lowercase ) , torch.tensor(__lowercase ) , ) # lsh weights + output _snake_case = weights[0][1] if len(__lowercase ) < 4: set_layer_weights_in_torch_lsh(__lowercase , torch_block.attention , __lowercase ) else: set_layer_weights_in_torch_local(__lowercase , torch_block.attention , __lowercase ) # intermediate weighs _snake_case = weights[2][0][1][2] # Chunked Feed Forward if len(__lowercase ) == 4: _snake_case = intermediate_weights[2] # layernorm 2 _snake_case = np.asarray(intermediate_weights[0][0] ) _snake_case = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(__lowercase ) , torch.tensor(__lowercase ) , ) # intermediate dense _snake_case = np.asarray(intermediate_weights[1][0] ) _snake_case = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(__lowercase ).transpose(0 , 1 ).contiguous() , torch.tensor(__lowercase ) , ) # intermediate out _snake_case = np.asarray(intermediate_weights[4][0] ) _snake_case = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(__lowercase ).transpose(0 , 1 ).contiguous() , torch.tensor(__lowercase ) , ) def a_ ( __lowercase : Tuple , __lowercase : Tuple , __lowercase : Dict ) -> Optional[int]: # reformer model _snake_case = torch_model.reformer # word embeds _snake_case = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(__lowercase ) , ) if isinstance(weights[3] , __lowercase ): _snake_case = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _snake_case = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f'''{position_embeddings[emb_idx]} emb does not match''' _snake_case = nn.Parameter(torch.tensor(__lowercase ) ) _snake_case = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __lowercase ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _snake_case = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__lowercase , __lowercase , __lowercase ) # output layer norm _snake_case = np.asarray(weights[7][0] ) _snake_case = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(__lowercase ) , torch.tensor(__lowercase ) , ) # output embeddings _snake_case = np.asarray(weights[9][0] ) _snake_case = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(__lowercase ).transpose(0 , 1 ).contiguous() , torch.tensor(__lowercase ) , ) def a_ ( __lowercase : Optional[Any] , __lowercase : Dict , __lowercase : List[Any] ) -> Optional[int]: # Initialise PyTorch model _snake_case = ReformerConfig.from_json_file(__lowercase ) print(f'''Building PyTorch model from configuration: {config}''' ) _snake_case = ReformerModelWithLMHead(__lowercase ) with open(__lowercase , 'rb' ) as f: _snake_case = pickle.load(__lowercase )['weights'] set_model_weights_in_torch(__lowercase , __lowercase , config.hidden_size ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , __lowercase ) if __name__ == "__main__": _lowerCamelCase : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--trax_model_pkl_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained Reformer model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) _lowerCamelCase : List[Any] = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
282
1
import argparse import torch from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_from_original_stable_diffusion_ckpt if __name__ == "__main__": _lowerCamelCase : List[str] = argparse.ArgumentParser() parser.add_argument( '''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the checkpoint to convert.''' ) # !wget https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml parser.add_argument( '''--original_config_file''', default=None, type=str, help='''The YAML config file corresponding to the original architecture.''', ) parser.add_argument( '''--num_in_channels''', default=None, type=int, help='''The number of input channels. If `None` number of input channels will be automatically inferred.''', ) parser.add_argument( '''--scheduler_type''', default='''pndm''', type=str, help='''Type of scheduler to use. Should be one of [\'pndm\', \'lms\', \'ddim\', \'euler\', \'euler-ancestral\', \'dpm\']''', ) parser.add_argument( '''--pipeline_type''', default=None, type=str, help=( '''The pipeline type. One of \'FrozenOpenCLIPEmbedder\', \'FrozenCLIPEmbedder\', \'PaintByExample\'''' '''. If `None` pipeline will be automatically inferred.''' ), ) parser.add_argument( '''--image_size''', default=None, type=int, help=( '''The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2''' ''' Base. Use 768 for Stable Diffusion v2.''' ), ) parser.add_argument( '''--prediction_type''', default=None, type=str, help=( '''The prediction type that the model was trained on. Use \'epsilon\' for Stable Diffusion v1.X and Stable''' ''' Diffusion v2 Base. Use \'v_prediction\' for Stable Diffusion v2.''' ), ) parser.add_argument( '''--extract_ema''', action='''store_true''', help=( '''Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights''' ''' or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield''' ''' higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.''' ), ) parser.add_argument( '''--upcast_attention''', action='''store_true''', help=( '''Whether the attention computation should always be upcasted. This is necessary when running stable''' ''' diffusion 2.1.''' ), ) parser.add_argument( '''--from_safetensors''', action='''store_true''', help='''If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.''', ) parser.add_argument( '''--to_safetensors''', action='''store_true''', help='''Whether to store pipeline in safetensors format or not.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument('''--device''', type=str, help='''Device to use (e.g. cpu, cuda:0, cuda:1, etc.)''') parser.add_argument( '''--stable_unclip''', type=str, default=None, required=False, help='''Set if this is a stable unCLIP model. One of \'txt2img\' or \'img2img\'.''', ) parser.add_argument( '''--stable_unclip_prior''', type=str, default=None, required=False, help='''Set if this is a stable unCLIP txt2img model. Selects which prior to use. If `--stable_unclip` is set to `txt2img`, the karlo prior (https://huggingface.co/kakaobrain/karlo-v1-alpha/tree/main/prior) is selected by default.''', ) parser.add_argument( '''--clip_stats_path''', type=str, help='''Path to the clip stats file. Only required if the stable unclip model\'s config specifies `model.params.noise_aug_config.params.clip_stats_path`.''', required=False, ) parser.add_argument( '''--controlnet''', action='''store_true''', default=None, help='''Set flag if this is a controlnet checkpoint.''' ) parser.add_argument('''--half''', action='''store_true''', help='''Save weights in half precision.''') parser.add_argument( '''--vae_path''', type=str, default=None, required=False, help='''Set to a path, hub id to an already converted vae to not convert it again.''', ) _lowerCamelCase : Dict = parser.parse_args() _lowerCamelCase : Optional[int] = download_from_original_stable_diffusion_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, prediction_type=args.prediction_type, model_type=args.pipeline_type, extract_ema=args.extract_ema, scheduler_type=args.scheduler_type, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, stable_unclip=args.stable_unclip, stable_unclip_prior=args.stable_unclip_prior, clip_stats_path=args.clip_stats_path, controlnet=args.controlnet, vae_path=args.vae_path, ) if args.half: pipe.to(torch_dtype=torch.floataa) if args.controlnet: # only save the controlnet model pipe.controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors) else: pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
282
import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def a_ ( __lowercase : Dict ) -> List[Any]: _snake_case = args.pruning_method _snake_case = args.threshold _snake_case = args.model_name_or_path.rstrip('/' ) _snake_case = args.target_model_path print(f'''Load fine-pruned model from {model_name_or_path}''' ) _snake_case = torch.load(os.path.join(__lowercase , 'pytorch_model.bin' ) ) _snake_case = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: _snake_case = tensor print(f'''Copied layer {name}''' ) elif "classifier" in name or "qa_output" in name: _snake_case = tensor print(f'''Copied layer {name}''' ) elif "bias" in name: _snake_case = tensor print(f'''Copied layer {name}''' ) else: if pruning_method == "magnitude": _snake_case = MagnitudeBinarizer.apply(inputs=__lowercase , threshold=__lowercase ) _snake_case = tensor * mask print(f'''Pruned layer {name}''' ) elif pruning_method == "topK": if "mask_scores" in name: continue _snake_case = name[:-6] _snake_case = model[f'''{prefix_}mask_scores'''] _snake_case = TopKBinarizer.apply(__lowercase , __lowercase ) _snake_case = tensor * mask print(f'''Pruned layer {name}''' ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue _snake_case = name[:-6] _snake_case = model[f'''{prefix_}mask_scores'''] _snake_case = ThresholdBinarizer.apply(__lowercase , __lowercase , __lowercase ) _snake_case = tensor * mask print(f'''Pruned layer {name}''' ) elif pruning_method == "l0": if "mask_scores" in name: continue _snake_case = name[:-6] _snake_case = model[f'''{prefix_}mask_scores'''] _snake_case , _snake_case = -0.1, 1.1 _snake_case = torch.sigmoid(__lowercase ) _snake_case = s * (r - l) + l _snake_case = s_bar.clamp(min=0.0 , max=1.0 ) _snake_case = tensor * mask print(f'''Pruned layer {name}''' ) else: raise ValueError('Unknown pruning method' ) if target_model_path is None: _snake_case = os.path.join( os.path.dirname(__lowercase ) , f'''bertarized_{os.path.basename(__lowercase )}''' ) if not os.path.isdir(__lowercase ): shutil.copytree(__lowercase , __lowercase ) print(f'''\nCreated folder {target_model_path}''' ) torch.save(__lowercase , os.path.join(__lowercase , 'pytorch_model.bin' ) ) print('\nPruned model saved! See you later!' ) if __name__ == "__main__": _lowerCamelCase : Dict = argparse.ArgumentParser() parser.add_argument( '''--pruning_method''', choices=['''l0''', '''magnitude''', '''topK''', '''sigmoied_threshold'''], type=str, required=True, help=( '''Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,''' ''' sigmoied_threshold = Soft movement pruning)''' ), ) parser.add_argument( '''--threshold''', type=float, required=False, help=( '''For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.''' '''For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.''' '''Not needed for `l0`''' ), ) parser.add_argument( '''--model_name_or_path''', type=str, required=True, help='''Folder containing the model that was previously fine-pruned''', ) parser.add_argument( '''--target_model_path''', default=None, type=str, required=False, help='''Folder containing the model that was previously fine-pruned''', ) _lowerCamelCase : int = parser.parse_args() main(args)
282
1
from typing import List from .keymap import KEYMAP, get_character def a_ ( __lowercase : str ) -> Tuple: def decorator(__lowercase : Optional[Any] ): _snake_case = getattr(__lowercase , 'handle_key' , [] ) handle += [key] setattr(__lowercase , 'handle_key' , __lowercase ) return func return decorator def a_ ( *__lowercase : List[str] ) -> str: def decorator(__lowercase : Union[str, Any] ): _snake_case = getattr(__lowercase , 'handle_key' , [] ) handle += keys setattr(__lowercase , 'handle_key' , __lowercase ) return func return decorator class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def __new__( cls : Union[str, Any] , lowercase : str , lowercase : List[str] , lowercase : Any ): '''simple docstring''' _snake_case = super().__new__(cls , lowercase , lowercase , lowercase ) if not hasattr(lowercase , 'key_handler' ): setattr(lowercase , 'key_handler' , {} ) setattr(lowercase , 'handle_input' , KeyHandler.handle_input ) for value in attrs.values(): _snake_case = getattr(lowercase , 'handle_key' , [] ) for key in handled_keys: _snake_case = value return new_cls @staticmethod def A ( cls : List[Any] ): '''simple docstring''' _snake_case = get_character() if char != KEYMAP["undefined"]: _snake_case = ord(lowercase ) _snake_case = cls.key_handler.get(lowercase ) if handler: _snake_case = char return handler(cls ) else: return None def a_ ( cls : str ) -> List[Any]: return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
282
import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class SCREAMING_SNAKE_CASE__ : '''simple docstring''' @property def A ( self : List[str] ): '''simple docstring''' return self.get_dummy_input() @property def A ( self : Any ): '''simple docstring''' if self.block_type == "down": return (4, 32, 16, 16) elif self.block_type == "mid": return (4, 32, 32, 32) elif self.block_type == "up": return (4, 32, 64, 64) raise ValueError(f'''\'{self.block_type}\' is not a supported block_type. Set it to \'up\', \'mid\', or \'down\'.''' ) def A ( self : Union[str, Any] , lowercase : Any=True , lowercase : List[Any]=False , lowercase : List[str]=False , lowercase : Dict=False , ): '''simple docstring''' _snake_case = 4 _snake_case = 32 _snake_case = (32, 32) _snake_case = torch.manual_seed(0 ) _snake_case = torch.device(lowercase ) _snake_case = (batch_size, num_channels) + sizes _snake_case = randn_tensor(lowercase , generator=lowercase , device=lowercase ) _snake_case = {'hidden_states': hidden_states} if include_temb: _snake_case = 128 _snake_case = randn_tensor((batch_size, temb_channels) , generator=lowercase , device=lowercase ) if include_res_hidden_states_tuple: _snake_case = torch.manual_seed(1 ) _snake_case = (randn_tensor(lowercase , generator=lowercase , device=lowercase ),) if include_encoder_hidden_states: _snake_case = floats_tensor((batch_size, 32, 32) ).to(lowercase ) if include_skip_sample: _snake_case = randn_tensor(((batch_size, 3) + sizes) , generator=lowercase , device=lowercase ) return dummy_input def A ( self : Any ): '''simple docstring''' _snake_case = { 'in_channels': 32, 'out_channels': 32, 'temb_channels': 128, } if self.block_type == "up": _snake_case = 32 if self.block_type == "mid": init_dict.pop('out_channels' ) _snake_case = self.dummy_input return init_dict, inputs_dict def A ( self : Dict , lowercase : Optional[int] ): '''simple docstring''' _snake_case , _snake_case = self.prepare_init_args_and_inputs_for_common() _snake_case = self.block_class(**lowercase ) unet_block.to(lowercase ) unet_block.eval() with torch.no_grad(): _snake_case = unet_block(**lowercase ) if isinstance(lowercase , lowercase ): _snake_case = output[0] self.assertEqual(output.shape , self.output_shape ) _snake_case = output[0, -1, -3:, -3:] _snake_case = torch.tensor(lowercase ).to(lowercase ) assert torch_all_close(output_slice.flatten() , lowercase , atol=5E-3 ) @unittest.skipIf(torch_device == 'mps' , 'Training is not supported in mps' ) def A ( self : Dict ): '''simple docstring''' _snake_case , _snake_case = self.prepare_init_args_and_inputs_for_common() _snake_case = self.block_class(**lowercase ) model.to(lowercase ) model.train() _snake_case = model(**lowercase ) if isinstance(lowercase , lowercase ): _snake_case = output[0] _snake_case = torch.device(lowercase ) _snake_case = randn_tensor(output.shape , device=lowercase ) _snake_case = torch.nn.functional.mse_loss(lowercase , lowercase ) loss.backward()
282
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase : Optional[int] = logging.get_logger(__name__) _lowerCamelCase : int = { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json''' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _UpperCAmelCase : Optional[int] = "roformer" def __init__( self : List[str] , lowercase : List[Any]=50_000 , lowercase : Dict=None , lowercase : Tuple=768 , lowercase : Optional[Any]=12 , lowercase : Tuple=12 , lowercase : Dict=3_072 , lowercase : List[Any]="gelu" , lowercase : Optional[Any]=0.1 , lowercase : str=0.1 , lowercase : int=1_536 , lowercase : Optional[int]=2 , lowercase : List[str]=0.02 , lowercase : List[str]=1E-12 , lowercase : Optional[int]=0 , lowercase : List[Any]=False , lowercase : List[str]=True , **lowercase : Dict , ): '''simple docstring''' super().__init__(pad_token_id=lowercase , **lowercase ) _snake_case = vocab_size _snake_case = hidden_size if embedding_size is None else embedding_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = hidden_act _snake_case = intermediate_size _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = rotary_value _snake_case = use_cache class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' @property def A ( self : Tuple ): '''simple docstring''' if self.task == "multiple-choice": _snake_case = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _snake_case = {0: 'batch', 1: 'sequence'} _snake_case = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('token_type_ids', dynamic_axis), ] )
282
_lowerCamelCase : int = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _lowerCamelCase : str = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _lowerCamelCase : List[str] = { 0: '''Sunday''', 1: '''Monday''', 2: '''Tuesday''', 3: '''Wednesday''', 4: '''Thursday''', 5: '''Friday''', 6: '''Saturday''', } def a_ ( __lowercase : int , __lowercase : int , __lowercase : int ) -> str: assert len(str(__lowercase ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: _snake_case = year // 100 _snake_case = (5 * (century % 4) + 2) % 7 _snake_case = year % 100 _snake_case = centurian % 12 _snake_case = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 _snake_case = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 400) == 0) else DOOMSDAY_LEAP[month - 1] ) _snake_case = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
282
1
import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging _lowerCamelCase : Tuple = '''\ ''' _lowerCamelCase : List[str] = ''' Perplexity (PPL) is one of the most common metrics for evaluating language models. It is defined as the exponentiated average negative log-likelihood of a sequence. For more information, see https://huggingface.co/docs/transformers/perplexity ''' _lowerCamelCase : Any = ''' Args: model_id (str): model used for calculating Perplexity NOTE: Perplexity can only be calculated for causal language models. This includes models such as gpt2, causal variations of bert, causal versions of t5, and more (the full list can be found in the AutoModelForCausalLM documentation here: https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM ) input_texts (list of str): input text, each separate text snippet is one list entry. batch_size (int): the batch size to run texts through the model. Defaults to 16. add_start_token (bool): whether to add the start token to the texts, so the perplexity can include the probability of the first word. Defaults to True. device (str): device to run on, defaults to \'cuda\' when available Returns: perplexity: dictionary containing the perplexity scores for the texts in the input list, as well as the mean perplexity. If one of the input texts is longer than the max input length of the model, then it is truncated to the max length for the perplexity computation. Examples: Example 1: >>> perplexity = datasets.load_metric("perplexity") >>> input_texts = ["lorem ipsum", "Happy Birthday!", "Bienvenue"] >>> results = perplexity.compute(model_id=\'gpt2\', ... add_start_token=False, ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) [\'perplexities\', \'mean_perplexity\'] >>> print(round(results["mean_perplexity"], 2)) 78.22 >>> print(round(results["perplexities"][0], 2)) 11.11 Example 2: >>> perplexity = datasets.load_metric("perplexity") >>> input_texts = datasets.load_dataset("wikitext", ... "wikitext-2-raw-v1", ... split="test")["text"][:50] # doctest:+ELLIPSIS [...] >>> input_texts = [s for s in input_texts if s!=\'\'] >>> results = perplexity.compute(model_id=\'gpt2\', ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) [\'perplexities\', \'mean_perplexity\'] >>> print(round(results["mean_perplexity"], 2)) 60.35 >>> print(round(results["perplexities"][0], 2)) 81.12 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE__ ( datasets.Metric ): '''simple docstring''' def A ( self : Union[str, Any] ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'input_texts': datasets.Value('string' ), } ) , reference_urls=['https://huggingface.co/docs/transformers/perplexity'] , ) def A ( self : Optional[int] , lowercase : Optional[Any] , lowercase : Any , lowercase : int = 16 , lowercase : bool = True , lowercase : int=None ): '''simple docstring''' if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": _snake_case = 'cuda' else: _snake_case = 'cuda' if torch.cuda.is_available() else 'cpu' _snake_case = AutoModelForCausalLM.from_pretrained(lowercase ) _snake_case = model.to(lowercase ) _snake_case = AutoTokenizer.from_pretrained(lowercase ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: _snake_case = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(lowercase ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({'pad_token': existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" _snake_case = model.config.max_length - 1 else: _snake_case = model.config.max_length _snake_case = tokenizer( lowercase , add_special_tokens=lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , return_tensors='pt' , return_attention_mask=lowercase , ).to(lowercase ) _snake_case = encodings['input_ids'] _snake_case = encodings['attention_mask'] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." _snake_case = [] _snake_case = CrossEntropyLoss(reduction='none' ) for start_index in logging.tqdm(range(0 , len(lowercase ) , lowercase ) ): _snake_case = min(start_index + batch_size , len(lowercase ) ) _snake_case = encoded_texts[start_index:end_index] _snake_case = attn_masks[start_index:end_index] if add_start_token: _snake_case = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(lowercase ) _snake_case = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) _snake_case = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(lowercase ), attn_mask] , dim=1 ) _snake_case = encoded_batch with torch.no_grad(): _snake_case = model(lowercase , attention_mask=lowercase ).logits _snake_case = out_logits[..., :-1, :].contiguous() _snake_case = labels[..., 1:].contiguous() _snake_case = attn_mask[..., 1:].contiguous() _snake_case = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , lowercase ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(lowercase )}
282
import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow _lowerCamelCase : int = False class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' def A ( self : Union[str, Any] , lowercase : Optional[int]=32 ): '''simple docstring''' set_seed(0 ) _snake_case = UNetaDModel(sample_size=lowercase , in_channels=3 , out_channels=3 ) _snake_case = torch.optim.SGD(model.parameters() , lr=0.0001 ) return model, optimizer @slow def A ( self : List[str] ): '''simple docstring''' _snake_case = 'cpu' # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable _snake_case = DDPMScheduler( num_train_timesteps=1_000 , beta_start=0.0001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=lowercase , ) _snake_case = DDIMScheduler( num_train_timesteps=1_000 , beta_start=0.0001 , beta_end=0.02 , beta_schedule='linear' , clip_sample=lowercase , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) _snake_case = [torch.randn((4, 3, 32, 32) ).clip(-1 , 1 ).to(lowercase ) for _ in range(4 )] _snake_case = [torch.randn((4, 3, 32, 32) ).to(lowercase ) for _ in range(4 )] _snake_case = [torch.randint(0 , 1_000 , (4,) ).long().to(lowercase ) for _ in range(4 )] # train with a DDPM scheduler _snake_case , _snake_case = self.get_model_optimizer(resolution=32 ) model.train().to(lowercase ) for i in range(4 ): optimizer.zero_grad() _snake_case = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) _snake_case = model(lowercase , timesteps[i] ).sample _snake_case = torch.nn.functional.mse_loss(lowercase , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM _snake_case , _snake_case = self.get_model_optimizer(resolution=32 ) model.train().to(lowercase ) for i in range(4 ): optimizer.zero_grad() _snake_case = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) _snake_case = model(lowercase , timesteps[i] ).sample _snake_case = torch.nn.functional.mse_loss(lowercase , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(lowercase , lowercase , atol=1E-5 ) ) self.assertTrue(torch.allclose(lowercase , lowercase , atol=1E-5 ) )
282
1
import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel _lowerCamelCase : List[Any] = HfApi() _lowerCamelCase : Dict = {} # fmt: off _lowerCamelCase : List[Any] = torch.tensor([ -0.7_5_1_5, -1.6_8_8_3, 0.2_4_2_0, 0.0_3_0_0, 0.6_3_4_7, 1.3_4_3_3, -1.1_7_4_3, -3.7_4_6_7, 1.2_3_4_2, -2.2_4_8_5, 0.4_6_3_6, 0.8_0_7_6, -0.7_9_9_1, 0.3_9_6_9, 0.8_4_9_8, 0.9_1_8_9, -1.8_8_8_7, -3.3_5_2_2, 0.7_6_3_9, 0.2_0_4_0, 0.6_2_7_1, -2.7_1_4_8, -1.6_3_1_6, 3.0_8_3_9, 0.3_1_8_6, 0.2_7_2_1, -0.9_7_5_9, -1.2_4_6_1, 2.6_2_5_7, 1.3_5_5_7 ]) _lowerCamelCase : int = torch.tensor([ -2.3_6_3_9, -2.5_3_4_4, 0.0_0_5_4, -0.6_6_7_4, 1.5_9_9_0, 1.0_1_5_8, 0.3_1_2_4, -2.1_4_3_6, 1.8_7_9_5, -2.5_4_2_9, -0.1_5_6_6, -0.3_9_7_3, 1.2_4_9_0, 2.6_4_4_7, 1.2_2_8_3, -0.5_2_0_8, -2.8_1_5_4, -3.5_1_1_9, 2.3_8_3_8, 1.2_0_3_3, 1.7_2_0_1, -2.1_2_5_6, -1.4_5_7_6, 2.7_9_4_8, 2.4_2_0_4, -0.9_7_5_2, -1.2_5_4_6, 0.8_0_2_7, 3.2_7_5_8, 3.1_3_6_5 ]) _lowerCamelCase : Optional[int] = torch.tensor([ -0.6_5_3_1, -0.6_8_9_1, -0.3_1_7_2, -0.5_3_7_5, -0.9_1_4_0, -0.5_3_6_7, -0.1_1_7_5, -0.7_8_6_9, -0.3_8_0_8, -0.4_5_1_3, -0.2_0_9_8, -0.0_0_8_3, 0.3_1_8_3, 0.5_1_4_0, 0.2_2_4_7, -0.1_3_0_4, -0.1_3_0_2, -0.2_8_0_2, -0.2_0_8_4, -0.2_0_2_5, -0.4_9_6_7, -0.4_8_7_3, -0.0_8_6_1, 0.6_9_2_5, 0.0_2_5_0, 0.1_2_9_0, -0.1_5_4_3, 0.6_3_1_6, 1.0_4_6_0, 1.4_9_4_3 ]) _lowerCamelCase : Dict = torch.tensor([ 0.0_9_1_1, 0.1_1_0_7, 0.0_1_8_2, 0.0_4_3_5, -0.0_8_0_5, -0.0_6_0_8, 0.0_3_8_1, 0.2_1_7_2, -0.0_2_8_0, 0.1_3_2_7, -0.0_2_9_9, -0.0_2_5_5, -0.0_0_5_0, -0.1_1_7_0, -0.1_0_4_6, 0.0_3_0_9, 0.1_3_6_7, 0.1_7_2_8, -0.0_5_3_3, -0.0_7_4_8, -0.0_5_3_4, 0.1_6_2_4, 0.0_3_8_4, -0.1_8_0_5, -0.0_7_0_7, 0.0_6_4_2, 0.0_2_2_0, -0.0_1_3_4, -0.1_3_3_3, -0.1_5_0_5 ]) _lowerCamelCase : Dict = torch.tensor([ 0.1_3_2_1, 0.1_3_3_7, 0.0_4_4_0, 0.0_6_2_2, -0.0_5_9_1, -0.0_3_7_0, 0.0_5_0_3, 0.2_1_3_3, -0.0_1_7_7, 0.1_4_1_5, -0.0_1_1_6, -0.0_1_1_2, 0.0_0_4_4, -0.0_9_8_0, -0.0_7_8_9, 0.0_3_9_5, 0.1_5_0_2, 0.1_7_8_5, -0.0_4_8_8, -0.0_5_1_4, -0.0_4_0_4, 0.1_5_3_9, 0.0_4_5_4, -0.1_5_5_9, -0.0_6_6_5, 0.0_6_5_9, 0.0_3_8_3, -0.0_0_0_5, -0.1_2_6_6, -0.1_3_8_6 ]) _lowerCamelCase : List[Any] = torch.tensor([ 0.1_1_5_4, 0.1_2_1_8, 0.0_3_0_7, 0.0_5_2_6, -0.0_7_1_1, -0.0_5_4_1, 0.0_3_6_6, 0.2_0_7_8, -0.0_2_6_7, 0.1_3_1_7, -0.0_2_2_6, -0.0_1_9_3, -0.0_0_1_4, -0.1_0_5_5, -0.0_9_0_2, 0.0_3_3_0, 0.1_3_9_1, 0.1_7_0_9, -0.0_5_6_2, -0.0_6_9_3, -0.0_5_6_0, 0.1_4_8_2, 0.0_3_8_1, -0.1_6_8_3, -0.0_6_8_1, 0.0_6_6_1, 0.0_3_3_1, -0.0_0_4_6, -0.1_2_6_8, -0.1_4_3_1 ]) _lowerCamelCase : Dict = torch.tensor([ 0.1_1_9_2, 0.1_2_4_0, 0.0_4_1_4, 0.0_6_0_6, -0.0_5_5_7, -0.0_4_1_2, 0.0_4_3_0, 0.2_0_4_2, -0.0_2_0_0, 0.1_3_8_5, -0.0_1_1_5, -0.0_1_3_2, 0.0_0_1_7, -0.0_9_6_5, -0.0_8_0_2, 0.0_3_9_8, 0.1_4_3_3, 0.1_7_4_7, -0.0_4_5_8, -0.0_5_3_3, -0.0_4_0_7, 0.1_5_4_5, 0.0_4_1_9, -0.1_5_7_4, -0.0_6_4_5, 0.0_6_2_6, 0.0_3_4_1, -0.0_0_1_0, -0.1_1_9_9, -0.1_3_9_0 ]) _lowerCamelCase : int = torch.tensor([ 0.1_0_7_5, 0.1_0_7_4, 0.0_2_0_5, 0.0_4_3_1, -0.0_7_7_4, -0.0_6_0_7, 0.0_2_9_8, 0.2_0_4_2, -0.0_3_2_0, 0.1_2_6_7, -0.0_2_8_1, -0.0_2_5_0, -0.0_0_6_4, -0.1_0_9_1, -0.0_9_4_6, 0.0_2_9_0, 0.1_3_2_8, 0.1_6_5_0, -0.0_5_8_0, -0.0_7_3_8, -0.0_5_8_6, 0.1_4_4_0, 0.0_3_3_7, -0.1_7_4_6, -0.0_7_1_2, 0.0_6_0_5, 0.0_2_5_0, -0.0_0_9_9, -0.1_3_1_6, -0.1_4_7_3 ]) _lowerCamelCase : int = torch.tensor([ -1.4_5_7_2, -2.0_4_8_1, -0.0_4_1_4, -0.6_0_0_5, 1.4_1_3_6, 0.5_8_4_8, 0.4_0_2_8, -2.7_3_3_0, 1.2_2_1_2, -2.1_2_2_8, 0.2_1_5_5, 0.4_0_3_9, 0.7_6_6_2, 2.0_5_3_5, 0.7_4_7_7, -0.3_2_4_3, -2.1_7_5_8, -2.7_6_4_8, 1.6_9_4_7, 0.7_0_2_6, 1.2_3_3_8, -1.6_0_7_8, -0.8_6_8_2, 2.2_8_1_0, 1.8_5_7_4, -0.5_7_1_8, -0.5_5_8_6, -0.0_1_8_6, 2.3_4_1_5, 2.1_2_5_1]) _lowerCamelCase : Tuple = torch.tensor([ -1.3_6_9_0, -1.9_7_2_0, -0.4_0_9_0, -0.6_9_6_6, 1.4_6_6_0, 0.9_9_3_8, -0.1_3_8_5, -2.7_3_2_4, 0.7_7_3_6, -1.8_9_1_7, 0.2_9_2_3, 0.4_2_9_3, 0.1_6_9_3, 1.4_1_1_2, 1.1_8_8_7, -0.3_1_8_1, -2.2_1_6_0, -2.6_3_8_1, 1.3_1_7_0, 0.8_1_6_3, 0.9_2_4_0, -1.6_5_4_4, -0.6_0_9_9, 2.5_2_5_9, 1.6_4_3_0, -0.9_0_9_0, -0.9_3_9_2, -0.0_1_2_6, 2.4_2_6_8, 2.3_2_6_6 ]) _lowerCamelCase : List[str] = torch.tensor([ -1.3_5_2_5, -1.9_6_2_8, -0.3_9_5_6, -0.6_8_6_0, 1.4_6_6_4, 1.0_0_1_4, -0.1_2_5_9, -2.7_2_1_2, 0.7_7_7_2, -1.8_8_1_1, 0.2_9_9_6, 0.4_3_8_8, 0.1_7_0_4, 1.4_0_2_9, 1.1_7_0_1, -0.3_0_2_7, -2.2_0_5_3, -2.6_2_8_7, 1.3_3_5_0, 0.8_1_3_1, 0.9_2_7_4, -1.6_2_9_2, -0.6_0_9_8, 2.5_1_3_1, 1.6_5_0_5, -0.8_9_5_8, -0.9_2_9_8, -0.0_1_5_1, 2.4_2_5_7, 2.3_3_5_5 ]) _lowerCamelCase : int = torch.tensor([ -2.0_5_8_5, -2.7_8_9_7, -0.2_8_5_0, -0.8_9_4_0, 1.9_0_5_2, 0.5_7_0_2, 0.6_3_4_5, -3.8_9_5_9, 1.5_9_3_2, -3.2_3_1_9, 0.1_9_7_4, 0.0_2_8_7, 1.7_5_6_6, 2.6_5_4_3, 0.8_3_8_7, -0.5_3_5_1, -3.2_7_3_6, -4.3_3_7_5, 2.9_0_2_9, 1.6_3_9_0, 1.4_6_4_0, -2.1_7_0_1, -1.9_0_1_3, 2.9_3_4_1, 3.4_9_8_1, -0.6_2_5_5, -1.1_6_4_4, -0.1_5_9_1, 3.7_0_9_7, 3.2_0_6_6 ]) _lowerCamelCase : Tuple = torch.tensor([ -2.3_1_3_9, -2.5_5_9_4, -0.0_1_9_7, -0.6_7_8_5, 1.7_0_0_1, 1.1_6_0_6, 0.3_0_7_5, -2.1_7_4_0, 1.8_0_7_1, -2.5_6_3_0, -0.0_9_2_6, -0.3_8_1_1, 1.2_1_1_6, 2.6_2_4_6, 1.2_7_3_1, -0.5_3_9_8, -2.8_1_5_3, -3.6_1_4_0, 2.3_8_9_3, 1.3_2_6_2, 1.6_2_5_8, -2.1_8_5_6, -1.3_2_6_7, 2.8_3_9_5, 2.3_7_7_9, -1.0_6_2_3, -1.2_4_6_8, 0.8_9_5_9, 3.3_3_6_7, 3.2_2_4_3 ]) _lowerCamelCase : int = torch.tensor([ -2.0_6_2_8, -2.7_6_6_7, -0.2_0_8_9, -0.8_2_6_3, 2.0_5_3_9, 0.5_9_9_2, 0.6_4_9_5, -3.8_3_3_6, 1.6_0_2_5, -3.2_8_1_7, 0.1_7_2_1, -0.0_6_3_3, 1.7_5_1_6, 2.7_0_3_9, 0.8_1_0_0, -0.5_9_0_8, -3.2_1_1_3, -4.4_3_4_3, 2.9_2_5_7, 1.3_6_3_2, 1.5_5_6_2, -2.1_4_8_9, -1.9_8_9_4, 3.0_5_6_0, 3.3_3_9_6, -0.7_3_2_8, -1.0_4_1_7, 0.0_3_8_3, 3.7_0_9_3, 3.2_3_4_3 ]) _lowerCamelCase : List[Any] = torch.tensor([ -1.4_5_7_4, -2.0_5_6_9, -0.0_4_7_3, -0.6_1_1_7, 1.4_0_1_8, 0.5_7_6_9, 0.4_1_2_9, -2.7_3_4_4, 1.2_2_4_1, -2.1_3_9_7, 0.2_0_0_0, 0.3_9_3_7, 0.7_6_1_6, 2.0_4_5_3, 0.7_3_2_4, -0.3_3_9_1, -2.1_7_4_6, -2.7_7_4_4, 1.6_9_6_3, 0.6_9_2_1, 1.2_1_8_7, -1.6_1_7_2, -0.8_8_7_7, 2.2_4_3_9, 1.8_4_7_1, -0.5_8_3_9, -0.5_6_0_5, -0.0_4_6_4, 2.3_2_5_0, 2.1_2_1_9 ]) # fmt: on _lowerCamelCase : List[str] = api.list_models(filter='''diffusers''') for mod in models: if "google" in mod.author or mod.modelId == "CompVis/ldm-celebahq-256": _lowerCamelCase : Any = '''/home/patrick/google_checkpoints/''' + mod.modelId.split('''/''')[-1] print(F'Started running {mod.modelId}!!!') if mod.modelId.startswith('''CompVis'''): _lowerCamelCase : Optional[Any] = UNetaDModel.from_pretrained(local_checkpoint, subfolder='''unet''') else: _lowerCamelCase : int = UNetaDModel.from_pretrained(local_checkpoint) torch.manual_seed(0) random.seed(0) _lowerCamelCase : Union[str, Any] = torch.randn(1, model.config.in_channels, model.config.sample_size, model.config.sample_size) _lowerCamelCase : int = torch.tensor([10] * noise.shape[0]) with torch.no_grad(): _lowerCamelCase : int = model(noise, time_step).sample assert torch.allclose( logits[0, 0, 0, :30], results['''_'''.join('''_'''.join(mod.modelId.split('''/''')).split('''-'''))], atol=1E-3 ) print(F'{mod.modelId} has passed successfully!!!')
282
import numpy as np def a_ ( __lowercase : np.array ) -> np.array: return 1 / (1 + np.exp(-vector )) if __name__ == "__main__": import doctest doctest.testmod()
282
1
# XXX: we want transformers master here - in the absense of conftest manipulating sys.path: # hack it in for now: import sys from pathlib import Path _lowerCamelCase : List[str] = Path(__file__).resolve().parents[3] / '''src''' sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa import io # noqa import itertools # noqa import json # noqa import os # noqa import unittest # noqa from copy import deepcopy # noqa from parameterized import parameterized # noqa from transformers import TrainingArguments, is_torch_available # noqa from transformers.deepspeed import is_deepspeed_available # noqa from transformers.file_utils import WEIGHTS_NAME # noqa from transformers.testing_utils import ( # noqa CaptureLogger, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, mockenv_context, require_deepspeed, require_torch_gpu, require_torch_multi_gpu, slow, ) from transformers.trainer_utils import set_seed # noqa set_seed(42) _lowerCamelCase : Tuple = {'''base''': '''patrickvonplaten/wav2vec2_tiny_random''', '''robust''': '''patrickvonplaten/wav2vec2_tiny_random_robust'''} _lowerCamelCase : Optional[Any] = '''zero2''' _lowerCamelCase : str = '''zero3''' _lowerCamelCase : Dict = [ZEROa, ZEROa] def a_ ( __lowercase : List[Any] , __lowercase : Optional[Any] , __lowercase : Optional[int] ) -> int: # customize the test name generator function as we want both params to appear in the sub-test # name, as by default it shows only the first param _snake_case = parameterized.to_safe_name('_'.join(str(__lowercase ) for x in param.args ) ) return f'''{func.__name__}_{param_based_name}''' # Cartesian-product of zero stages with models to test _lowerCamelCase : str = list(itertools.product(stages, models.keys())) @slow @require_deepspeed @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' @parameterized.expand(lowercase , name_func=lowercase ) def A ( self : List[str] , lowercase : List[Any] , lowercase : List[Any] ): '''simple docstring''' self.run_and_check( stage=lowercase , model=lowercase , distributed=lowercase , fpaa=lowercase , ) @require_torch_multi_gpu @parameterized.expand(lowercase , name_func=lowercase ) def A ( self : Optional[int] , lowercase : int , lowercase : List[str] ): '''simple docstring''' self.run_and_check( stage=lowercase , model=lowercase , distributed=lowercase , fpaa=lowercase , ) @parameterized.expand(lowercase , name_func=lowercase ) def A ( self : str , lowercase : int , lowercase : Dict ): '''simple docstring''' self.run_and_check( stage=lowercase , model=lowercase , distributed=lowercase , fpaa=lowercase , ) @require_torch_multi_gpu @parameterized.expand(lowercase , name_func=lowercase ) def A ( self : int , lowercase : str , lowercase : int ): '''simple docstring''' self.run_and_check( stage=lowercase , model=lowercase , distributed=lowercase , fpaa=lowercase , ) def A ( self : Union[str, Any] , lowercase : Optional[Any] ): '''simple docstring''' pass def A ( self : Dict , lowercase : str , lowercase : str , lowercase : int = 10 , lowercase : bool = True , lowercase : bool = True , lowercase : bool = True , ): '''simple docstring''' _snake_case = models[model] _snake_case = self.run_trainer( stage=lowercase , model_name=lowercase , eval_steps=lowercase , num_train_epochs=1 , distributed=lowercase , fpaa=lowercase , ) self.do_checks(lowercase ) return output_dir def A ( self : str , lowercase : str , lowercase : str , lowercase : int = 10 , lowercase : int = 1 , lowercase : bool = True , lowercase : bool = True , ): '''simple docstring''' _snake_case = self.get_auto_remove_tmp_dir('./xxx' , after=lowercase ) _snake_case = f''' --model_name_or_path {model_name} --dataset_name hf-internal-testing/librispeech_asr_dummy --dataset_config_name clean --train_split_name validation --validation_split_name validation --output_dir {output_dir} --num_train_epochs {str(lowercase )} --per_device_train_batch_size 2 --per_device_eval_batch_size 2 --evaluation_strategy steps --learning_rate 5e-4 --warmup_steps 8 --orthography timit --preprocessing_num_workers 1 --group_by_length --freeze_feature_extractor --report_to none --save_steps 0 --eval_steps {eval_steps} --report_to none '''.split() if fpaa: args.extend(['--fp16'] ) # currently ds_config_wav2vec2_zero.json requires "zero_optimization.find_unused_parameters": true, # hence the separate config files _snake_case = f'''--deepspeed {self.test_file_dir_str}/ds_config_wav2vec2_{stage}.json'''.split() _snake_case = [f'''{self.examples_dir_str}/research_projects/wav2vec2/run_asr.py'''] _snake_case = self.get_launcher(lowercase ) _snake_case = launcher + script + args + ds_args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(lowercase , env=self.get_env() ) return output_dir def A ( self : List[Any] , lowercase : str=False ): '''simple docstring''' _snake_case = min(2 , get_gpu_count() ) if distributed else 1 return f'''deepspeed --num_nodes 1 --num_gpus {num_gpus}'''.split()
282
import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sentencepiece @require_tokenizers @require_flax class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' @slow def A ( self : int ): '''simple docstring''' _snake_case = FlaxXLMRobertaModel.from_pretrained('xlm-roberta-base' ) _snake_case = AutoTokenizer.from_pretrained('xlm-roberta-base' ) _snake_case = 'The dog is cute and lives in the garden house' _snake_case = jnp.array([tokenizer.encode(lowercase )] ) _snake_case = (1, 12, 768) # batch_size, sequence_length, embedding_vector_dim _snake_case = jnp.array( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) _snake_case = model(lowercase )['last_hidden_state'] self.assertEqual(output.shape , lowercase ) # compare the actual values for a slice of last dim self.assertTrue(jnp.allclose(output[:, :, -1] , lowercase , atol=1E-3 ) )
282
1