code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' from importlib import import_module from .logging import get_logger a_ = get_logger(__name__) class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase=None ): '''simple docstring''' __lowerCamelCase = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith('''__''' ): setattr(self , __A , getattr(__A , __A ) ) __lowerCamelCase = module._original_module if isinstance(__A , _PatchedModuleObj ) else module class __lowerCAmelCase : lowerCAmelCase__ = [] def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None ): '''simple docstring''' __lowerCamelCase = obj __lowerCamelCase = target __lowerCamelCase = new __lowerCamelCase = target.split('''.''' )[0] __lowerCamelCase = {} __lowerCamelCase = attrs or [] def __enter__( self ): '''simple docstring''' __lowerCamelCase = self.target.split('''.''' ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(__A ) ): try: __lowerCamelCase = import_module('''.'''.join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): __lowerCamelCase = getattr(self.obj , __A ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(__A , _PatchedModuleObj ) and obj_attr._original_module is submodule) ): __lowerCamelCase = obj_attr # patch at top level setattr(self.obj , __A , _PatchedModuleObj(__A , attrs=self.attrs ) ) __lowerCamelCase = getattr(self.obj , __A ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(__A , __A , _PatchedModuleObj(getattr(__A , __A , __A ) , attrs=self.attrs ) ) __lowerCamelCase = getattr(__A , __A ) # finally set the target attribute setattr(__A , __A , self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: __lowerCamelCase = getattr(import_module('''.'''.join(__A ) ) , __A ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , __A ) is attr_value: __lowerCamelCase = getattr(self.obj , __A ) setattr(self.obj , __A , self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" __lowerCamelCase = globals()["__builtins__"][target_attr] setattr(self.obj , __A , self.new ) else: raise RuntimeError(F"""Tried to patch attribute {target_attr} instead of a submodule.""" ) def __exit__( self , *__UpperCAmelCase ): '''simple docstring''' for attr in list(self.original ): setattr(self.obj , __A , self.original.pop(__A ) ) def lowerCamelCase ( self ): '''simple docstring''' self.__enter__() self._active_patches.append(self ) def lowerCamelCase ( self ): '''simple docstring''' try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
711
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def a__ ( _UpperCamelCase : Optional[int] ,_UpperCamelCase : List[str] ,_UpperCamelCase : List[Any]=None ,_UpperCamelCase : Any=None ): if attention_mask is None: __lowerCamelCase = tf.cast(tf.math.not_equal(_UpperCamelCase ,config.pad_token_id ) ,tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class __lowerCAmelCase : lowerCAmelCase__ = OPTConfig lowerCAmelCase__ = {} lowerCAmelCase__ = """gelu""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=99 , __UpperCAmelCase=16 , __UpperCAmelCase=2 , __UpperCAmelCase=4 , __UpperCAmelCase=4 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=20 , __UpperCAmelCase=2 , __UpperCAmelCase=1 , __UpperCAmelCase=0 , __UpperCAmelCase=16 , __UpperCAmelCase=16 , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = eos_token_id __lowerCamelCase = pad_token_id __lowerCamelCase = bos_token_id __lowerCamelCase = embed_dim __lowerCamelCase = word_embed_proj_dim __lowerCamelCase = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowerCamelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowerCamelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowerCamelCase = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=__UpperCAmelCase , **self.config_updates , ) __lowerCamelCase = prepare_opt_inputs_dict(__UpperCAmelCase , __UpperCAmelCase ) return config, inputs_dict def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = TFOPTModel(config=__UpperCAmelCase ) __lowerCamelCase = inputs_dict['''input_ids'''] __lowerCamelCase = input_ids[:1, :] __lowerCamelCase = inputs_dict['''attention_mask'''][:1, :] __lowerCamelCase = 1 # first forward pass __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , use_cache=__UpperCAmelCase ) __lowerCamelCase ,__lowerCamelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowerCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowerCamelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowerCamelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowerCamelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase )[0] __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , past_key_values=__UpperCAmelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowerCamelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowerCamelCase = output_from_no_past[:, -3:, random_slice_idx] __lowerCamelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__UpperCAmelCase , __UpperCAmelCase , rtol=1E-3 ) @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () lowerCAmelCase__ = (TFOPTForCausalLM,) if is_tf_available() else () lowerCAmelCase__ = ( {"""feature-extraction""": TFOPTModel, """text-generation""": TFOPTForCausalLM} if is_tf_available() else {} ) lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = 1_0 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(__UpperCAmelCase , __UpperCAmelCase ): if hasattr(__UpperCAmelCase , '''weight''' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(__UpperCAmelCase , '''weight''' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 10, config.vocab_size + 10]: # build the embeddings __lowerCamelCase = model_class(config=__UpperCAmelCase ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_input_embeddings() ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(__UpperCAmelCase ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_input_embeddings() ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. __lowerCamelCase = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , __UpperCAmelCase ) # check that weights remain the same after resizing __lowerCamelCase = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCamelCase = False self.assertTrue(__UpperCAmelCase ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , __UpperCAmelCase ) __lowerCamelCase = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCamelCase = False self.assertTrue(__UpperCAmelCase ) def a__ ( _UpperCamelCase : Optional[Any] ): return tf.constant(_UpperCamelCase ,dtype=tf.intaa ) @require_tf class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = 9_9 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = tf.ones((4, 1) , dtype=tf.intaa ) * 2 __lowerCamelCase = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) __lowerCamelCase = input_ids.shape[0] __lowerCamelCase = OPTConfig( vocab_size=self.vocab_size , hidden_size=24 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class __lowerCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTModel.from_pretrained('''facebook/opt-350m''' ) __lowerCamelCase = _long_tensor([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] ) __lowerCamelCase = tf.not_equal(__UpperCAmelCase , model.config.pad_token_id ) with tf.GradientTape(): __lowerCamelCase = model(input_ids=__UpperCAmelCase , attention_mask=__UpperCAmelCase ).last_hidden_state __lowerCamelCase = (1, 11, 512) self.assertEqual(output.shape , __UpperCAmelCase ) __lowerCamelCase = tf.constant( [[-0.2_873, -1.9_218, -0.3_033], [-1.2_710, -0.1_338, -0.1_902], [0.4_095, 0.1_214, -1.3_121]] ) self.assertTrue(np.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=4E-3 ) ) __lowerCamelCase = tf.function(__UpperCAmelCase , jit_compile=__UpperCAmelCase ) __lowerCamelCase = xla_generate(__UpperCAmelCase , __UpperCAmelCase )[0] self.assertTrue(np.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=4E-2 ) ) @require_tf @slow class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' super().setUp() __lowerCamelCase = '''facebook/opt-350m''' def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTForCausalLM.from_pretrained(self.path_model ) __lowerCamelCase = GPTaTokenizer.from_pretrained(self.path_model ) __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of''', '''Paris is the capital of France and''', '''Computers and mobile phones have taken''', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' , padding=__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) __lowerCamelCase = tf.constant( [ [1.3_851, -13.8_923, -10.5_229, -10.7_533, -0.2_309, -10.2_384, -0.5_365, -9.0_947, -5.1_670], [-4.7_073, -10.6_276, -3.9_415, -21.5_242, -0.2_822, -0.2_822, -0.2_822, -0.2_822, -0.2_822], [0.6_247, -3.4_229, -8.9_179, -1.4_297, -14.1_650, 1.4_146, -9.0_218, -0.2_703, -0.2_703], [6.4_783, -1.9_913, -10.7_926, -2.3_336, 1.5_092, -0.9_974, -6.8_213, 1.3_477, 1.3_477], ] ) self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-4 ) ) __lowerCamelCase = tf.function(__UpperCAmelCase , jit_compile=__UpperCAmelCase ) __lowerCamelCase = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-4 ) ) @require_tf @slow class __lowerCAmelCase ( unittest.TestCase ): @property def lowerCamelCase ( self ): '''simple docstring''' return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-125m''' __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of New York, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] __lowerCamelCase = [] __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) for prompt in self.prompts: __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(__UpperCAmelCase , max_length=10 ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) predicted_outputs += generated_string self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-350m''' __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = '''left''' # use different length sentences to test batching __lowerCamelCase = [ '''Hello, my dog is a little''', '''Today, I''', ] __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' , padding=__UpperCAmelCase ) __lowerCamelCase = inputs['''input_ids'''] __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase , attention_mask=inputs['''attention_mask'''] ) __lowerCamelCase = tokenizer(sentences[0] , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase ) __lowerCamelCase = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['''attention_mask'''][-1] , tf.intaa ) ) __lowerCamelCase = tokenizer(sentences[1] , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase , max_length=model.config.max_length - num_paddings ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = [ '''Hello, my dog is a little bit of a dork.\nI\'m a little bit''', '''Today, I was in the middle of a conversation with a friend about the''', ] self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , [non_padded_sentence, padded_sentence] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-350m''' __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of San Francisco, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] __lowerCamelCase = [] __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) for prompt in self.prompts: __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(__UpperCAmelCase , max_length=10 ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) predicted_outputs += generated_string self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase )
622
0
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def a__ ( _UpperCamelCase : Tuple ): __lowerCamelCase = filter(lambda _UpperCamelCase : p.requires_grad ,model.parameters() ) __lowerCamelCase = sum([np.prod(p.size() ) for p in model_parameters] ) return params a_ = logging.getLogger(__name__) def a__ ( _UpperCamelCase : Optional[Any] ,_UpperCamelCase : Tuple ): if metric == "rouge2": __lowerCamelCase = """{val_avg_rouge2:.4f}-{step_count}""" elif metric == "bleu": __lowerCamelCase = """{val_avg_bleu:.4f}-{step_count}""" elif metric == "em": __lowerCamelCase = """{val_avg_em:.4f}-{step_count}""" elif metric == "loss": __lowerCamelCase = """{val_avg_loss:.4f}-{step_count}""" else: raise NotImplementedError( F"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" ''' function.''' ) __lowerCamelCase = ModelCheckpoint( dirpath=__lowerCAmelCase ,filename=__lowerCAmelCase ,monitor=F"""val_{metric}""" ,mode='''max''' ,save_top_k=1 ,every_n_epochs=1 ,) return checkpoint_callback def a__ ( _UpperCamelCase : List[Any] ,_UpperCamelCase : Dict ): return EarlyStopping( monitor=F"""val_{metric}""" ,mode='''min''' if '''loss''' in metric else '''max''' ,patience=__lowerCAmelCase ,verbose=__lowerCAmelCase ,) class __lowerCAmelCase ( pl.Callback ): def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = {F"""lr_group_{i}""": param["""lr"""] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(_a ) @rank_zero_only def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=True ): '''simple docstring''' logger.info(F"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) __lowerCamelCase = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['''log''', '''progress_bar''', '''preds''']} ) # Log results __lowerCamelCase = Path(pl_module.hparams.output_dir ) if type_path == "test": __lowerCamelCase = od / """test_results.txt""" __lowerCamelCase = od / """test_generations.txt""" else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __lowerCamelCase = od / F"""{type_path}_results/{trainer.global_step:05d}.txt""" __lowerCamelCase = od / F"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=_a ) generations_file.parent.mkdir(exist_ok=_a ) with open(_a , '''a+''' ) as writer: for key in sorted(_a ): if key in ["log", "progress_bar", "preds"]: continue __lowerCamelCase = metrics[key] if isinstance(_a , torch.Tensor ): __lowerCamelCase = val.item() __lowerCamelCase = F"""{key}: {val:.6f}\n""" writer.write(_a ) if not save_generations: return if "preds" in metrics: __lowerCamelCase = """\n""".join(metrics['''preds'''] ) generations_file.open('''w+''' ).write(_a ) @rank_zero_only def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' try: __lowerCamelCase = pl_module.model.model.num_parameters() except AttributeError: __lowerCamelCase = pl_module.model.num_parameters() __lowerCamelCase = count_trainable_parameters(_a ) # mp stands for million parameters trainer.logger.log_metrics({'''n_params''': npars, '''mp''': npars / 1E6, '''grad_mp''': n_trainable_pars / 1E6} ) @rank_zero_only def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(_a , _a , '''test''' ) @rank_zero_only def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
712
import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) a_ = logging.getLogger(__name__) def a__ ( _UpperCamelCase : str ,_UpperCamelCase : List[Any] ): __lowerCamelCase = np.argmax(_UpperCamelCase ,axis=1 ) return np.sum(outputs == labels ) def a__ ( _UpperCamelCase : Optional[int] ): with open(_UpperCamelCase ,encoding='''utf_8''' ) as f: __lowerCamelCase = csv.reader(_UpperCamelCase ) __lowerCamelCase = [] next(_UpperCamelCase ) # skip the first line for line in tqdm(_UpperCamelCase ): output.append((''' '''.join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def a__ ( _UpperCamelCase : Any ,_UpperCamelCase : Dict ,_UpperCamelCase : str ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Any ,_UpperCamelCase : Dict ): __lowerCamelCase = [] for dataset in encoded_datasets: __lowerCamelCase = len(_UpperCamelCase ) __lowerCamelCase = np.zeros((n_batch, 2, input_len) ,dtype=np.intaa ) __lowerCamelCase = np.zeros((n_batch, 2) ,dtype=np.intaa ) __lowerCamelCase = np.full((n_batch, 2, input_len) ,fill_value=-1_00 ,dtype=np.intaa ) __lowerCamelCase = np.zeros((n_batch,) ,dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(_UpperCamelCase ): __lowerCamelCase = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __lowerCamelCase = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __lowerCamelCase = with_conta __lowerCamelCase = with_conta __lowerCamelCase = len(_UpperCamelCase ) - 1 __lowerCamelCase = len(_UpperCamelCase ) - 1 __lowerCamelCase = with_conta __lowerCamelCase = with_conta __lowerCamelCase = mc_label __lowerCamelCase = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(_UpperCamelCase ) for t in all_inputs ) ) return tensor_datasets def a__ ( ): __lowerCamelCase = argparse.ArgumentParser() parser.add_argument('''--model_name''' ,type=_UpperCamelCase ,default='''openai-gpt''' ,help='''pretrained model name''' ) parser.add_argument('''--do_train''' ,action='''store_true''' ,help='''Whether to run training.''' ) parser.add_argument('''--do_eval''' ,action='''store_true''' ,help='''Whether to run eval on the dev set.''' ) parser.add_argument( '''--output_dir''' ,default=_UpperCamelCase ,type=_UpperCamelCase ,required=_UpperCamelCase ,help='''The output directory where the model predictions and checkpoints will be written.''' ,) parser.add_argument('''--train_dataset''' ,type=_UpperCamelCase ,default='''''' ) parser.add_argument('''--eval_dataset''' ,type=_UpperCamelCase ,default='''''' ) parser.add_argument('''--seed''' ,type=_UpperCamelCase ,default=42 ) parser.add_argument('''--num_train_epochs''' ,type=_UpperCamelCase ,default=3 ) parser.add_argument('''--train_batch_size''' ,type=_UpperCamelCase ,default=8 ) parser.add_argument('''--eval_batch_size''' ,type=_UpperCamelCase ,default=16 ) parser.add_argument('''--adam_epsilon''' ,default=1e-8 ,type=_UpperCamelCase ,help='''Epsilon for Adam optimizer.''' ) parser.add_argument('''--max_grad_norm''' ,type=_UpperCamelCase ,default=1 ) parser.add_argument( '''--max_steps''' ,default=-1 ,type=_UpperCamelCase ,help=( '''If > 0: set total number of training steps to perform. Override num_train_epochs.''' ) ,) parser.add_argument( '''--gradient_accumulation_steps''' ,type=_UpperCamelCase ,default=1 ,help='''Number of updates steps to accumulate before performing a backward/update pass.''' ,) parser.add_argument('''--learning_rate''' ,type=_UpperCamelCase ,default=6.25e-5 ) parser.add_argument('''--warmup_steps''' ,default=0 ,type=_UpperCamelCase ,help='''Linear warmup over warmup_steps.''' ) parser.add_argument('''--lr_schedule''' ,type=_UpperCamelCase ,default='''warmup_linear''' ) parser.add_argument('''--weight_decay''' ,type=_UpperCamelCase ,default=0.01 ) parser.add_argument('''--lm_coef''' ,type=_UpperCamelCase ,default=0.9 ) parser.add_argument('''--n_valid''' ,type=_UpperCamelCase ,default=3_74 ) parser.add_argument('''--server_ip''' ,type=_UpperCamelCase ,default='''''' ,help='''Can be used for distant debugging.''' ) parser.add_argument('''--server_port''' ,type=_UpperCamelCase ,default='''''' ,help='''Can be used for distant debugging.''' ) __lowerCamelCase = parser.parse_args() print(_UpperCamelCase ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('''Waiting for debugger attach''' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) ,redirect_output=_UpperCamelCase ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) __lowerCamelCase = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) __lowerCamelCase = torch.cuda.device_count() logger.info('''device: {}, n_gpu {}'''.format(_UpperCamelCase ,_UpperCamelCase ) ) if not args.do_train and not args.do_eval: raise ValueError('''At least one of `do_train` or `do_eval` must be True.''' ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset __lowerCamelCase = ['''_start_''', '''_delimiter_''', '''_classify_'''] __lowerCamelCase = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(_UpperCamelCase ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(_UpperCamelCase ) __lowerCamelCase = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(_UpperCamelCase ) ) model.to(_UpperCamelCase ) # Load and encode the datasets def tokenize_and_encode(_UpperCamelCase : Dict ): if isinstance(_UpperCamelCase ,_UpperCamelCase ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(_UpperCamelCase ) ) elif isinstance(_UpperCamelCase ,_UpperCamelCase ): return obj return [tokenize_and_encode(_UpperCamelCase ) for o in obj] logger.info('''Encoding dataset...''' ) __lowerCamelCase = load_rocstories_dataset(args.train_dataset ) __lowerCamelCase = load_rocstories_dataset(args.eval_dataset ) __lowerCamelCase = (train_dataset, eval_dataset) __lowerCamelCase = tokenize_and_encode(_UpperCamelCase ) # Compute the max input length for the Transformer __lowerCamelCase = model.config.n_positions // 2 - 2 __lowerCamelCase = max( len(story[:max_length] ) + max(len(conta[:max_length] ) ,len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) __lowerCamelCase = min(_UpperCamelCase ,model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders __lowerCamelCase = pre_process_datasets(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ,*_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = tensor_datasets[0], tensor_datasets[1] __lowerCamelCase = TensorDataset(*_UpperCamelCase ) __lowerCamelCase = RandomSampler(_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,sampler=_UpperCamelCase ,batch_size=args.train_batch_size ) __lowerCamelCase = TensorDataset(*_UpperCamelCase ) __lowerCamelCase = SequentialSampler(_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,sampler=_UpperCamelCase ,batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: __lowerCamelCase = args.max_steps __lowerCamelCase = args.max_steps // (len(_UpperCamelCase ) // args.gradient_accumulation_steps) + 1 else: __lowerCamelCase = len(_UpperCamelCase ) // args.gradient_accumulation_steps * args.num_train_epochs __lowerCamelCase = list(model.named_parameters() ) __lowerCamelCase = ['''bias''', '''LayerNorm.bias''', '''LayerNorm.weight'''] __lowerCamelCase = [ { '''params''': [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], '''weight_decay''': args.weight_decay, }, {'''params''': [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0}, ] __lowerCamelCase = AdamW(_UpperCamelCase ,lr=args.learning_rate ,eps=args.adam_epsilon ) __lowerCamelCase = get_linear_schedule_with_warmup( _UpperCamelCase ,num_warmup_steps=args.warmup_steps ,num_training_steps=_UpperCamelCase ) if args.do_train: __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) ,desc='''Epoch''' ): __lowerCamelCase = 0 __lowerCamelCase = 0 __lowerCamelCase = tqdm(_UpperCamelCase ,desc='''Training''' ) for step, batch in enumerate(_UpperCamelCase ): __lowerCamelCase = tuple(t.to(_UpperCamelCase ) for t in batch ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = batch __lowerCamelCase = model(_UpperCamelCase ,mc_token_ids=_UpperCamelCase ,lm_labels=_UpperCamelCase ,mc_labels=_UpperCamelCase ) __lowerCamelCase = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() __lowerCamelCase = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 __lowerCamelCase = '''Training loss: {:.2e} lr: {:.2e}'''.format(_UpperCamelCase ,scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer __lowerCamelCase = model.module if hasattr(_UpperCamelCase ,'''module''' ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` __lowerCamelCase = os.path.join(args.output_dir ,_UpperCamelCase ) __lowerCamelCase = os.path.join(args.output_dir ,_UpperCamelCase ) torch.save(model_to_save.state_dict() ,_UpperCamelCase ) model_to_save.config.to_json_file(_UpperCamelCase ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned __lowerCamelCase = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) __lowerCamelCase = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(_UpperCamelCase ) if args.do_eval: model.eval() __lowerCamelCase ,__lowerCamelCase = 0, 0 __lowerCamelCase ,__lowerCamelCase = 0, 0 for batch in tqdm(_UpperCamelCase ,desc='''Evaluating''' ): __lowerCamelCase = tuple(t.to(_UpperCamelCase ) for t in batch ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = batch with torch.no_grad(): __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = model( _UpperCamelCase ,mc_token_ids=_UpperCamelCase ,lm_labels=_UpperCamelCase ,mc_labels=_UpperCamelCase ) __lowerCamelCase = mc_logits.detach().cpu().numpy() __lowerCamelCase = mc_labels.to('''cpu''' ).numpy() __lowerCamelCase = accuracy(_UpperCamelCase ,_UpperCamelCase ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 __lowerCamelCase = eval_loss / nb_eval_steps __lowerCamelCase = eval_accuracy / nb_eval_examples __lowerCamelCase = tr_loss / nb_tr_steps if args.do_train else None __lowerCamelCase = {'''eval_loss''': eval_loss, '''eval_accuracy''': eval_accuracy, '''train_loss''': train_loss} __lowerCamelCase = os.path.join(args.output_dir ,'''eval_results.txt''' ) with open(_UpperCamelCase ,'''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key in sorted(result.keys() ): logger.info(''' %s = %s''' ,_UpperCamelCase ,str(result[key] ) ) writer.write('''%s = %s\n''' % (key, str(result[key] )) ) if __name__ == "__main__": main()
622
0
import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Dict=0.999 ,_UpperCamelCase : List[Any]="cosine" ,): if alpha_transform_type == "cosine": def alpha_bar_fn(_UpperCamelCase : Dict ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(_UpperCamelCase : Union[str, Any] ): return math.exp(t * -12.0 ) else: raise ValueError(F"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) __lowerCamelCase = [] for i in range(__UpperCAmelCase ): __lowerCamelCase = i / num_diffusion_timesteps __lowerCamelCase = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__UpperCAmelCase ) / alpha_bar_fn(__UpperCAmelCase ) ,__UpperCAmelCase ) ) return torch.tensor(__UpperCAmelCase ,dtype=torch.floataa ) class __lowerCAmelCase ( _snake_case , _snake_case ): lowerCAmelCase__ = [e.name for e in KarrasDiffusionSchedulers] lowerCAmelCase__ = 2 @register_to_config def __init__( self , __UpperCAmelCase = 1000 , __UpperCAmelCase = 0.00_085 , __UpperCAmelCase = 0.012 , __UpperCAmelCase = "linear" , __UpperCAmelCase = None , __UpperCAmelCase = "epsilon" , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = 1.0 , __UpperCAmelCase = "linspace" , __UpperCAmelCase = 0 , ): '''simple docstring''' if trained_betas is not None: __lowerCamelCase = torch.tensor(__UpperCAmelCase , dtype=torch.floataa ) elif beta_schedule == "linear": __lowerCamelCase = torch.linspace(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. __lowerCamelCase = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , __UpperCAmelCase , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule __lowerCamelCase = betas_for_alpha_bar(__UpperCAmelCase , alpha_transform_type='''cosine''' ) elif beta_schedule == "exp": __lowerCamelCase = betas_for_alpha_bar(__UpperCAmelCase , alpha_transform_type='''exp''' ) else: raise NotImplementedError(F"""{beta_schedule} does is not implemented for {self.__class__}""" ) __lowerCamelCase = 1.0 - self.betas __lowerCamelCase = torch.cumprod(self.alphas , dim=0 ) # set all values self.set_timesteps(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = use_karras_sigmas def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase=None ): '''simple docstring''' if schedule_timesteps is None: __lowerCamelCase = self.timesteps __lowerCamelCase = (schedule_timesteps == timestep).nonzero() # The sigma index that is taken for the **very** first `step` # is always the second index (or the last index if there is only 1) # This way we can ensure we don't accidentally skip a sigma in # case we start in the middle of the denoising schedule (e.g. for image-to-image) if len(self._index_counter ) == 0: __lowerCamelCase = 1 if len(__UpperCAmelCase ) > 1 else 0 else: __lowerCamelCase = timestep.cpu().item() if torch.is_tensor(__UpperCAmelCase ) else timestep __lowerCamelCase = self._index_counter[timestep_int] return indices[pos].item() @property def lowerCamelCase ( self ): '''simple docstring''' if self.config.timestep_spacing in ["linspace", "trailing"]: return self.sigmas.max() return (self.sigmas.max() ** 2 + 1) ** 0.5 def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.index_for_timestep(__UpperCAmelCase ) __lowerCamelCase = self.sigmas[step_index] __lowerCamelCase = sample / ((sigma**2 + 1) ** 0.5) return sample def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , ): '''simple docstring''' __lowerCamelCase = num_inference_steps __lowerCamelCase = num_train_timesteps or self.config.num_train_timesteps # "linspace", "leading", "trailing" corresponds to annotation of Table 2. of https://arxiv.org/abs/2305.08891 if self.config.timestep_spacing == "linspace": __lowerCamelCase = np.linspace(0 , num_train_timesteps - 1 , __UpperCAmelCase , dtype=__UpperCAmelCase )[::-1].copy() elif self.config.timestep_spacing == "leading": __lowerCamelCase = num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 __lowerCamelCase = (np.arange(0 , __UpperCAmelCase ) * step_ratio).round()[::-1].copy().astype(__UpperCAmelCase ) timesteps += self.config.steps_offset elif self.config.timestep_spacing == "trailing": __lowerCamelCase = num_train_timesteps / self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 __lowerCamelCase = (np.arange(__UpperCAmelCase , 0 , -step_ratio )).round().copy().astype(__UpperCAmelCase ) timesteps -= 1 else: raise ValueError( F"""{self.config.timestep_spacing} is not supported. Please make sure to choose one of 'linspace', 'leading' or 'trailing'.""" ) __lowerCamelCase = np.array(((1 - self.alphas_cumprod) / self.alphas_cumprod) ** 0.5 ) __lowerCamelCase = np.log(__UpperCAmelCase ) __lowerCamelCase = np.interp(__UpperCAmelCase , np.arange(0 , len(__UpperCAmelCase ) ) , __UpperCAmelCase ) if self.config.use_karras_sigmas: __lowerCamelCase = self._convert_to_karras(in_sigmas=__UpperCAmelCase , num_inference_steps=self.num_inference_steps ) __lowerCamelCase = np.array([self._sigma_to_t(__UpperCAmelCase , __UpperCAmelCase ) for sigma in sigmas] ) __lowerCamelCase = np.concatenate([sigmas, [0.0]] ).astype(np.floataa ) __lowerCamelCase = torch.from_numpy(__UpperCAmelCase ).to(device=__UpperCAmelCase ) __lowerCamelCase = torch.cat([sigmas[:1], sigmas[1:-1].repeat_interleave(2 ), sigmas[-1:]] ) __lowerCamelCase = torch.from_numpy(__UpperCAmelCase ) __lowerCamelCase = torch.cat([timesteps[:1], timesteps[1:].repeat_interleave(2 )] ) if str(__UpperCAmelCase ).startswith('''mps''' ): # mps does not support float64 __lowerCamelCase = timesteps.to(__UpperCAmelCase , dtype=torch.floataa ) else: __lowerCamelCase = timesteps.to(device=__UpperCAmelCase ) # empty dt and derivative __lowerCamelCase = None __lowerCamelCase = None # for exp beta schedules, such as the one for `pipeline_shap_e.py` # we need an index counter __lowerCamelCase = defaultdict(__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = np.log(__UpperCAmelCase ) # get distribution __lowerCamelCase = log_sigma - log_sigmas[:, np.newaxis] # get sigmas range __lowerCamelCase = np.cumsum((dists >= 0) , axis=0 ).argmax(axis=0 ).clip(max=log_sigmas.shape[0] - 2 ) __lowerCamelCase = low_idx + 1 __lowerCamelCase = log_sigmas[low_idx] __lowerCamelCase = log_sigmas[high_idx] # interpolate sigmas __lowerCamelCase = (low - log_sigma) / (low - high) __lowerCamelCase = np.clip(__UpperCAmelCase , 0 , 1 ) # transform interpolation to time range __lowerCamelCase = (1 - w) * low_idx + w * high_idx __lowerCamelCase = t.reshape(sigma.shape ) return t def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = in_sigmas[-1].item() __lowerCamelCase = in_sigmas[0].item() __lowerCamelCase = 7.0 # 7.0 is the value used in the paper __lowerCamelCase = np.linspace(0 , 1 , __UpperCAmelCase ) __lowerCamelCase = sigma_min ** (1 / rho) __lowerCamelCase = sigma_max ** (1 / rho) __lowerCamelCase = (max_inv_rho + ramp * (min_inv_rho - max_inv_rho)) ** rho return sigmas @property def lowerCamelCase ( self ): '''simple docstring''' return self.dt is None def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = True , ): '''simple docstring''' __lowerCamelCase = self.index_for_timestep(__UpperCAmelCase ) # advance index counter by 1 __lowerCamelCase = timestep.cpu().item() if torch.is_tensor(__UpperCAmelCase ) else timestep self._index_counter[timestep_int] += 1 if self.state_in_first_order: __lowerCamelCase = self.sigmas[step_index] __lowerCamelCase = self.sigmas[step_index + 1] else: # 2nd order / Heun's method __lowerCamelCase = self.sigmas[step_index - 1] __lowerCamelCase = self.sigmas[step_index] # currently only gamma=0 is supported. This usually works best anyways. # We can support gamma in the future but then need to scale the timestep before # passing it to the model which requires a change in API __lowerCamelCase = 0 __lowerCamelCase = sigma * (gamma + 1) # Note: sigma_hat == sigma for now # 1. compute predicted original sample (x_0) from sigma-scaled predicted noise if self.config.prediction_type == "epsilon": __lowerCamelCase = sigma_hat if self.state_in_first_order else sigma_next __lowerCamelCase = sample - sigma_input * model_output elif self.config.prediction_type == "v_prediction": __lowerCamelCase = sigma_hat if self.state_in_first_order else sigma_next __lowerCamelCase = model_output * (-sigma_input / (sigma_input**2 + 1) ** 0.5) + ( sample / (sigma_input**2 + 1) ) elif self.config.prediction_type == "sample": __lowerCamelCase = model_output else: raise ValueError( F"""prediction_type given as {self.config.prediction_type} must be one of `epsilon`, or `v_prediction`""" ) if self.config.clip_sample: __lowerCamelCase = pred_original_sample.clamp( -self.config.clip_sample_range , self.config.clip_sample_range ) if self.state_in_first_order: # 2. Convert to an ODE derivative for 1st order __lowerCamelCase = (sample - pred_original_sample) / sigma_hat # 3. delta timestep __lowerCamelCase = sigma_next - sigma_hat # store for 2nd order step __lowerCamelCase = derivative __lowerCamelCase = dt __lowerCamelCase = sample else: # 2. 2nd order / Heun's method __lowerCamelCase = (sample - pred_original_sample) / sigma_next __lowerCamelCase = (self.prev_derivative + derivative) / 2 # 3. take prev timestep & sample __lowerCamelCase = self.dt __lowerCamelCase = self.sample # free dt and derivative # Note, this puts the scheduler in "first order mode" __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = sample + derivative * dt if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.sigmas.to(device=original_samples.device , dtype=original_samples.dtype ) if original_samples.device.type == "mps" and torch.is_floating_point(__UpperCAmelCase ): # mps does not support float64 __lowerCamelCase = self.timesteps.to(original_samples.device , dtype=torch.floataa ) __lowerCamelCase = timesteps.to(original_samples.device , dtype=torch.floataa ) else: __lowerCamelCase = self.timesteps.to(original_samples.device ) __lowerCamelCase = timesteps.to(original_samples.device ) __lowerCamelCase = [self.index_for_timestep(__UpperCAmelCase , __UpperCAmelCase ) for t in timesteps] __lowerCamelCase = sigmas[step_indices].flatten() while len(sigma.shape ) < len(original_samples.shape ): __lowerCamelCase = sigma.unsqueeze(-1 ) __lowerCamelCase = original_samples + noise * sigma return noisy_samples def __len__( self ): '''simple docstring''' return self.config.num_train_timesteps
713
import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class __lowerCAmelCase ( lowerCAmelCase__ ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=1024 , __UpperCAmelCase=1024 , __UpperCAmelCase=3.6 ): '''simple docstring''' __lowerCamelCase = tokenizer __lowerCamelCase = tokenizer.bos_token_id __lowerCamelCase = dataset __lowerCamelCase = seq_length __lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self ): '''simple docstring''' __lowerCamelCase = iter(self.dataset ) __lowerCamelCase = True while more_examples: __lowerCamelCase ,__lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(__UpperCAmelCase )['''content'''] ) buffer_len += len(buffer[-1] ) except StopIteration: __lowerCamelCase = False break __lowerCamelCase = tokenizer(__UpperCAmelCase , truncation=__UpperCAmelCase )['''input_ids'''] __lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(__UpperCAmelCase ) , self.seq_length ): __lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(__UpperCAmelCase ) == self.seq_length: yield torch.tensor(__UpperCAmelCase ) def a__ ( _UpperCamelCase : List[Any] ): __lowerCamelCase = {'''streaming''': True} __lowerCamelCase = load_dataset(args.dataset_name ,split='''train''' ,**_UpperCamelCase ) __lowerCamelCase = ConstantLengthDataset(_UpperCamelCase ,_UpperCamelCase ,seq_length=args.seq_length ) __lowerCamelCase = DataLoader(_UpperCamelCase ,batch_size=args.batch_size ) return eval_dataloader def a__ ( _UpperCamelCase : str ): model.eval() __lowerCamelCase = [] for step, batch in enumerate(_UpperCamelCase ): with torch.no_grad(): __lowerCamelCase = model(_UpperCamelCase ,labels=_UpperCamelCase ) __lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(_UpperCamelCase ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break __lowerCamelCase = torch.mean(torch.cat(_UpperCamelCase ) ) try: __lowerCamelCase = torch.exp(_UpperCamelCase ) except OverflowError: __lowerCamelCase = float('''inf''' ) return loss.item(), perplexity.item() # Setup Accelerator a_ = Accelerator() # Parse configuration a_ = HfArgumentParser(EvaluationArguments) a_ = parser.parse_args() set_seed(args.seed) # Logging a_ = logging.getLogger(__name__) logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) # Load model and tokenizer a_ = AutoModelForCausalLM.from_pretrained(args.model_ckpt) a_ = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader a_ = create_dataloader(args) # Prepare everything with our `accelerator`. a_ , a_ = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info("""Evaluating and saving model after training""") a_ , a_ = evaluate(args) logger.info(f"loss/eval: {eval_loss}, perplexity: {perplexity}")
622
0
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """xlnet-base-cased""": """https://huggingface.co/xlnet-base-cased/resolve/main/config.json""", """xlnet-large-cased""": """https://huggingface.co/xlnet-large-cased/resolve/main/config.json""", } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """xlnet""" lowerCAmelCase__ = ["""mems"""] lowerCAmelCase__ = { """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , __UpperCAmelCase=32000 , __UpperCAmelCase=1024 , __UpperCAmelCase=24 , __UpperCAmelCase=16 , __UpperCAmelCase=4096 , __UpperCAmelCase="gelu" , __UpperCAmelCase=True , __UpperCAmelCase="bi" , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=None , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=False , __UpperCAmelCase=-1 , __UpperCAmelCase=False , __UpperCAmelCase="last" , __UpperCAmelCase=True , __UpperCAmelCase="tanh" , __UpperCAmelCase=0.1 , __UpperCAmelCase=5 , __UpperCAmelCase=5 , __UpperCAmelCase=5 , __UpperCAmelCase=1 , __UpperCAmelCase=2 , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = vocab_size __lowerCamelCase = d_model __lowerCamelCase = n_layer __lowerCamelCase = n_head if d_model % n_head != 0: raise ValueError(F"""'d_model % n_head' ({d_model % n_head}) should be equal to 0""" ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( F"""`d_head` ({kwargs["d_head"]}) should be equal to `d_model // n_head` ({d_model // n_head})""" ) __lowerCamelCase = d_model // n_head __lowerCamelCase = ff_activation __lowerCamelCase = d_inner __lowerCamelCase = untie_r __lowerCamelCase = attn_type __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = dropout __lowerCamelCase = mem_len __lowerCamelCase = reuse_len __lowerCamelCase = bi_data __lowerCamelCase = clamp_len __lowerCamelCase = same_length __lowerCamelCase = summary_type __lowerCamelCase = summary_use_proj __lowerCamelCase = summary_activation __lowerCamelCase = summary_last_dropout __lowerCamelCase = start_n_top __lowerCamelCase = end_n_top __lowerCamelCase = bos_token_id __lowerCamelCase = pad_token_id __lowerCamelCase = eos_token_id if "use_cache" in kwargs: warnings.warn( '''The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`''' ''' instead.''' , _UpperCAmelCase , ) __lowerCamelCase = kwargs['''use_cache'''] __lowerCamelCase = use_mems_eval __lowerCamelCase = use_mems_train super().__init__(pad_token_id=_UpperCAmelCase , bos_token_id=_UpperCAmelCase , eos_token_id=_UpperCAmelCase , **_UpperCAmelCase ) @property def lowerCamelCase ( self ): '''simple docstring''' logger.info(F"""The model {self.model_type} is one of the few models that has no sequence length limit.""" ) return -1 @max_position_embeddings.setter def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' raise NotImplementedError( F"""The model {self.model_type} is one of the few models that has no sequence length limit.""" )
714
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """lxmert""" lowerCAmelCase__ = {} def __init__( self , __UpperCAmelCase=30522 , __UpperCAmelCase=768 , __UpperCAmelCase=12 , __UpperCAmelCase=9500 , __UpperCAmelCase=1600 , __UpperCAmelCase=400 , __UpperCAmelCase=3072 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=9 , __UpperCAmelCase=5 , __UpperCAmelCase=5 , __UpperCAmelCase=2048 , __UpperCAmelCase=4 , __UpperCAmelCase=6.67 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = num_qa_labels __lowerCamelCase = num_object_labels __lowerCamelCase = num_attr_labels __lowerCamelCase = l_layers __lowerCamelCase = x_layers __lowerCamelCase = r_layers __lowerCamelCase = visual_feat_dim __lowerCamelCase = visual_pos_dim __lowerCamelCase = visual_loss_normalizer __lowerCamelCase = task_matched __lowerCamelCase = task_mask_lm __lowerCamelCase = task_obj_predict __lowerCamelCase = task_qa __lowerCamelCase = visual_obj_loss __lowerCamelCase = visual_attr_loss __lowerCamelCase = visual_feat_loss __lowerCamelCase = {'''vision''': r_layers, '''cross_encoder''': x_layers, '''language''': l_layers} super().__init__(**__UpperCAmelCase )
622
0
'''simple docstring''' from __future__ import annotations from collections import deque from collections.abc import Iterator from dataclasses import dataclass @dataclass class __lowerCAmelCase : lowerCAmelCase__ = 4_2 lowerCAmelCase__ = 4_2 class __lowerCAmelCase : def __init__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = [[] for _ in range(_UpperCamelCase )] __lowerCamelCase = size def __getitem__( self , __UpperCAmelCase ): '''simple docstring''' return iter(self._graph[vertex] ) @property def lowerCamelCase ( self ): '''simple docstring''' return self._size def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' if weight not in (0, 1): raise ValueError('''Edge weight must be either 0 or 1.''' ) if to_vertex < 0 or to_vertex >= self.size: raise ValueError('''Vertex indexes must be in [0; size).''' ) self._graph[from_vertex].append(Edge(_UpperCamelCase , _UpperCamelCase ) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = deque([start_vertex] ) __lowerCamelCase = [None] * self.size __lowerCamelCase = 0 while queue: __lowerCamelCase = queue.popleft() __lowerCamelCase = distances[current_vertex] if current_distance is None: continue for edge in self[current_vertex]: __lowerCamelCase = current_distance + edge.weight __lowerCamelCase = distances[edge.destination_vertex] if ( isinstance(_UpperCamelCase , _UpperCamelCase ) and new_distance >= dest_vertex_distance ): continue __lowerCamelCase = new_distance if edge.weight == 0: queue.appendleft(edge.destination_vertex ) else: queue.append(edge.destination_vertex ) if distances[finish_vertex] is None: raise ValueError('''No path from start_vertex to finish_vertex.''' ) return distances[finish_vertex] if __name__ == "__main__": import doctest doctest.testmod()
715
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Any ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Any ): if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = np.full((len(_UpperCamelCase ), sequence_length, 2) ,_UpperCamelCase ) else: __lowerCamelCase = np.full((len(_UpperCamelCase ), sequence_length) ,_UpperCamelCase ) for i, tensor in enumerate(_UpperCamelCase ): if padding_side == "right": if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = tensor[:sequence_length] else: __lowerCamelCase = tensor[:sequence_length] else: if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = tensor[:sequence_length] else: __lowerCamelCase = tensor[:sequence_length] return out_tensor.tolist() def a__ ( _UpperCamelCase : Dict ): __lowerCamelCase = ord(_UpperCamelCase ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 1_23 and cp <= 1_26): return True __lowerCamelCase = unicodedata.category(_UpperCamelCase ) if cat.startswith('''P''' ): return True return False @dataclass class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = 42 lowerCAmelCase__ = True lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = -1_0_0 lowerCAmelCase__ = "pt" def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' import torch __lowerCamelCase = '''label''' if '''label''' in features[0].keys() else '''labels''' __lowerCamelCase = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __lowerCamelCase = self.tokenizer.pad( __UpperCAmelCase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch __lowerCamelCase = torch.tensor(batch['''entity_ids'''] ).shape[1] __lowerCamelCase = self.tokenizer.padding_side if padding_side == "right": __lowerCamelCase = [ list(__UpperCAmelCase ) + [self.label_pad_token_id] * (sequence_length - len(__UpperCAmelCase )) for label in labels ] else: __lowerCamelCase = [ [self.label_pad_token_id] * (sequence_length - len(__UpperCAmelCase )) + list(__UpperCAmelCase ) for label in labels ] __lowerCamelCase = [feature['''ner_tags'''] for feature in features] __lowerCamelCase = padding_tensor(__UpperCAmelCase , -1 , __UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = [feature['''original_entity_spans'''] for feature in features] __lowerCamelCase = padding_tensor(__UpperCAmelCase , (-1, -1) , __UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = {k: torch.tensor(__UpperCAmelCase , dtype=torch.intaa ) for k, v in batch.items()} return batch
622
0
from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig a_ = logging.get_logger(__name__) # General docstring a_ = """MobileNetV1Config""" # Base docstring a_ = """google/mobilenet_v1_1.0_224""" a_ = [1, 1_024, 7, 7] # Image classification docstring a_ = """google/mobilenet_v1_1.0_224""" a_ = """tabby, tabby cat""" a_ = [ """google/mobilenet_v1_1.0_224""", """google/mobilenet_v1_0.75_192""", # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def a__ ( _UpperCamelCase : Any ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : List[Any]=None ): __lowerCamelCase = {} if isinstance(_lowerCamelCase ,_lowerCamelCase ): __lowerCamelCase = model.mobilenet_va else: __lowerCamelCase = model __lowerCamelCase = 'MobilenetV1/Conv2d_0/' __lowerCamelCase = backbone.conv_stem.convolution.weight __lowerCamelCase = backbone.conv_stem.normalization.bias __lowerCamelCase = backbone.conv_stem.normalization.weight __lowerCamelCase = backbone.conv_stem.normalization.running_mean __lowerCamelCase = backbone.conv_stem.normalization.running_var for i in range(13 ): __lowerCamelCase = i + 1 __lowerCamelCase = i * 2 __lowerCamelCase = backbone.layer[pt_index] __lowerCamelCase = F"""MobilenetV1/Conv2d_{tf_index}_depthwise/""" __lowerCamelCase = pointer.convolution.weight __lowerCamelCase = pointer.normalization.bias __lowerCamelCase = pointer.normalization.weight __lowerCamelCase = pointer.normalization.running_mean __lowerCamelCase = pointer.normalization.running_var __lowerCamelCase = backbone.layer[pt_index + 1] __lowerCamelCase = F"""MobilenetV1/Conv2d_{tf_index}_pointwise/""" __lowerCamelCase = pointer.convolution.weight __lowerCamelCase = pointer.normalization.bias __lowerCamelCase = pointer.normalization.weight __lowerCamelCase = pointer.normalization.running_mean __lowerCamelCase = pointer.normalization.running_var if isinstance(_lowerCamelCase ,_lowerCamelCase ): __lowerCamelCase = 'MobilenetV1/Logits/Conv2d_1c_1x1/' __lowerCamelCase = model.classifier.weight __lowerCamelCase = model.classifier.bias return tf_to_pt_map def a__ ( _UpperCamelCase : List[str] ,_UpperCamelCase : str ,_UpperCamelCase : Optional[Any] ): try: import numpy as np import tensorflow as tf except ImportError: logger.error( '''Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see ''' '''https://www.tensorflow.org/install/ for installation instructions.''' ) raise # Load weights from TF model __lowerCamelCase = tf.train.list_variables(_lowerCamelCase ) __lowerCamelCase = {} for name, shape in init_vars: logger.info(F"""Loading TF weight {name} with shape {shape}""" ) __lowerCamelCase = tf.train.load_variable(_lowerCamelCase ,_lowerCamelCase ) __lowerCamelCase = array # Build TF to PyTorch weights loading map __lowerCamelCase = _build_tf_to_pytorch_map(_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ) for name, pointer in tf_to_pt_map.items(): logger.info(F"""Importing {name}""" ) if name not in tf_weights: logger.info(F"""{name} not in tf pre-trained weights, skipping""" ) continue __lowerCamelCase = tf_weights[name] if "depthwise_weights" in name: logger.info('''Transposing depthwise''' ) __lowerCamelCase = np.transpose(_lowerCamelCase ,(2, 3, 0, 1) ) elif "weights" in name: logger.info('''Transposing''' ) if len(pointer.shape ) == 2: # copying into linear layer __lowerCamelCase = array.squeeze().transpose() else: __lowerCamelCase = np.transpose(_lowerCamelCase ,(3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F"""Pointer shape {pointer.shape} and array shape {array.shape} mismatched""" ) logger.info(F"""Initialize PyTorch weight {name} {array.shape}""" ) __lowerCamelCase = torch.from_numpy(_lowerCamelCase ) tf_weights.pop(_lowerCamelCase ,_lowerCamelCase ) tf_weights.pop(name + '''/RMSProp''' ,_lowerCamelCase ) tf_weights.pop(name + '''/RMSProp_1''' ,_lowerCamelCase ) tf_weights.pop(name + '''/ExponentialMovingAverage''' ,_lowerCamelCase ) logger.info(F"""Weights not copied to PyTorch model: {", ".join(tf_weights.keys() )}""" ) return model def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : Union[str, Any] ): __lowerCamelCase = features.shape[-2:] __lowerCamelCase = conv_layer.stride __lowerCamelCase = conv_layer.kernel_size if in_height % stride_height == 0: __lowerCamelCase = max(kernel_height - stride_height ,0 ) else: __lowerCamelCase = max(kernel_height - (in_height % stride_height) ,0 ) if in_width % stride_width == 0: __lowerCamelCase = max(kernel_width - stride_width ,0 ) else: __lowerCamelCase = max(kernel_width - (in_width % stride_width) ,0 ) __lowerCamelCase = pad_along_width // 2 __lowerCamelCase = pad_along_width - pad_left __lowerCamelCase = pad_along_height // 2 __lowerCamelCase = pad_along_height - pad_top __lowerCamelCase = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(_lowerCamelCase ,_lowerCamelCase ,'''constant''' ,0.0 ) class __lowerCAmelCase ( nn.Module ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = 1 , __UpperCAmelCase = 1 , __UpperCAmelCase = False , __UpperCAmelCase = True , __UpperCAmelCase = True , ): '''simple docstring''' super().__init__() __lowerCamelCase = config if in_channels % groups != 0: raise ValueError(F"""Input channels ({in_channels}) are not divisible by {groups} groups.""" ) if out_channels % groups != 0: raise ValueError(F"""Output channels ({out_channels}) are not divisible by {groups} groups.""" ) __lowerCamelCase = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) __lowerCamelCase = nn.Convad( in_channels=_A , out_channels=_A , kernel_size=_A , stride=_A , padding=_A , groups=_A , bias=_A , padding_mode='''zeros''' , ) if use_normalization: __lowerCamelCase = nn.BatchNormad( num_features=_A , eps=config.layer_norm_eps , momentum=0.9_997 , affine=_A , track_running_stats=_A , ) else: __lowerCamelCase = None if use_activation: if isinstance(_A , _A ): __lowerCamelCase = ACTaFN[use_activation] elif isinstance(config.hidden_act , _A ): __lowerCamelCase = ACTaFN[config.hidden_act] else: __lowerCamelCase = config.hidden_act else: __lowerCamelCase = None def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' if self.config.tf_padding: __lowerCamelCase = apply_tf_padding(_A , self.convolution ) __lowerCamelCase = self.convolution(_A ) if self.normalization is not None: __lowerCamelCase = self.normalization(_A ) if self.activation is not None: __lowerCamelCase = self.activation(_A ) return features class __lowerCAmelCase ( a__ ): lowerCAmelCase__ = MobileNetVaConfig lowerCAmelCase__ = load_tf_weights_in_mobilenet_va lowerCAmelCase__ = """mobilenet_v1""" lowerCAmelCase__ = """pixel_values""" lowerCAmelCase__ = False def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' if isinstance(_A , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(_A , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) a_ = R""" This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. """ a_ = R""" Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. """ @add_start_docstrings( """The bare MobileNetV1 model outputting raw hidden-states without any specific head on top.""" , a__ , ) class __lowerCAmelCase ( a__ ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase = True ): '''simple docstring''' super().__init__(_A ) __lowerCamelCase = config __lowerCamelCase = 32 __lowerCamelCase = max(int(depth * config.depth_multiplier ) , config.min_depth ) __lowerCamelCase = MobileNetVaConvLayer( _A , in_channels=config.num_channels , out_channels=_A , kernel_size=3 , stride=2 , ) __lowerCamelCase = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] __lowerCamelCase = nn.ModuleList() for i in range(13 ): __lowerCamelCase = out_channels if strides[i] == 2 or i == 0: depth *= 2 __lowerCamelCase = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( _A , in_channels=_A , out_channels=_A , kernel_size=3 , stride=strides[i] , groups=_A , ) ) self.layer.append( MobileNetVaConvLayer( _A , in_channels=_A , out_channels=_A , kernel_size=1 , ) ) __lowerCamelCase = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' raise NotImplementedError @add_start_docstrings_to_model_forward(_A ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=_A , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowerCamelCase ( self , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , ): '''simple docstring''' __lowerCamelCase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __lowerCamelCase = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('''You have to specify pixel_values''' ) __lowerCamelCase = self.conv_stem(_A ) __lowerCamelCase = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): __lowerCamelCase = layer_module(_A ) if output_hidden_states: __lowerCamelCase = all_hidden_states + (hidden_states,) __lowerCamelCase = hidden_states if self.pooler is not None: __lowerCamelCase = torch.flatten(self.pooler(_A ) , start_dim=1 ) else: __lowerCamelCase = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=_A , pooler_output=_A , hidden_states=_A , ) @add_start_docstrings( """\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n """ , a__ , ) class __lowerCAmelCase ( a__ ): def __init__( self , __UpperCAmelCase ): '''simple docstring''' super().__init__(_A ) __lowerCamelCase = config.num_labels __lowerCamelCase = MobileNetVaModel(_A ) __lowerCamelCase = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head __lowerCamelCase = nn.Dropout(config.classifier_dropout_prob , inplace=_A ) __lowerCamelCase = nn.Linear(_A , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_A ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=_A , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowerCamelCase ( self , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , ): '''simple docstring''' __lowerCamelCase = return_dict if return_dict is not None else self.config.use_return_dict __lowerCamelCase = self.mobilenet_va(_A , output_hidden_states=_A , return_dict=_A ) __lowerCamelCase = outputs.pooler_output if return_dict else outputs[1] __lowerCamelCase = self.classifier(self.dropout(_A ) ) __lowerCamelCase = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: __lowerCamelCase = 'regression' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): __lowerCamelCase = 'single_label_classification' else: __lowerCamelCase = 'multi_label_classification' if self.config.problem_type == "regression": __lowerCamelCase = MSELoss() if self.num_labels == 1: __lowerCamelCase = loss_fct(logits.squeeze() , labels.squeeze() ) else: __lowerCamelCase = loss_fct(_A , _A ) elif self.config.problem_type == "single_label_classification": __lowerCamelCase = CrossEntropyLoss() __lowerCamelCase = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": __lowerCamelCase = BCEWithLogitsLoss() __lowerCamelCase = loss_fct(_A , _A ) if not return_dict: __lowerCamelCase = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=_A , logits=_A , hidden_states=outputs.hidden_states , )
716
from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=99 , __UpperCAmelCase=[1, 1, 2] , __UpperCAmelCase=1 , __UpperCAmelCase=32 , __UpperCAmelCase=4 , __UpperCAmelCase=8 , __UpperCAmelCase=37 , __UpperCAmelCase="gelu_new" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.0 , __UpperCAmelCase=512 , __UpperCAmelCase=3 , __UpperCAmelCase=0.02 , __UpperCAmelCase=3 , __UpperCAmelCase=4 , __UpperCAmelCase=None , __UpperCAmelCase=False , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_input_mask __lowerCamelCase = use_token_type_ids __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = block_sizes __lowerCamelCase = num_decoder_layers __lowerCamelCase = d_model __lowerCamelCase = n_head __lowerCamelCase = d_head __lowerCamelCase = d_inner __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout __lowerCamelCase = attention_dropout __lowerCamelCase = activation_dropout __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = 2 __lowerCamelCase = num_labels __lowerCamelCase = num_choices __lowerCamelCase = scope __lowerCamelCase = initializer_std # Used in the tests to check the size of the first attention layer __lowerCamelCase = n_head # Used in the tests to check the size of the first hidden state __lowerCamelCase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowerCamelCase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowerCamelCase = self.num_hidden_layers + 2 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCamelCase = None if self.use_token_type_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCamelCase = FunnelConfig( vocab_size=self.vocab_size , block_sizes=self.block_sizes , num_decoder_layers=self.num_decoder_layers , d_model=self.d_model , n_head=self.n_head , d_head=self.d_head , d_inner=self.d_inner , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , activation_dropout=self.activation_dropout , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_std=self.initializer_std , ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = [input_ids, input_mask] __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = [input_ids, input_mask] __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 3, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForPreTraining(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForMaskedLM(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = TFFunnelForSequenceClassification(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_choices __lowerCamelCase = TFFunnelForMultipleChoice(config=__UpperCAmelCase ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = TFFunnelForTokenClassification(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForQuestionAnswering(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.prepare_config_and_inputs() ( ( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) , ) = config_and_inputs __lowerCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) lowerCAmelCase__ = ( { """feature-extraction""": (TFFunnelBaseModel, TFFunnelModel), """fill-mask""": TFFunnelForMaskedLM, """question-answering""": TFFunnelForQuestionAnswering, """text-classification""": TFFunnelForSequenceClassification, """token-classification""": TFFunnelForTokenClassification, """zero-shot""": TFFunnelForSequenceClassification, } if is_tf_available() else {} ) lowerCAmelCase__ = False lowerCAmelCase__ = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFFunnelModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCAmelCase ) @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) lowerCAmelCase__ = False lowerCAmelCase__ = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFFunnelModelTester(self , base=__UpperCAmelCase ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCAmelCase )
622
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available a_ = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ["""BartphoTokenizer"""] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
717
from collections import namedtuple import requests from lxml import html # type: ignore a_ = namedtuple("""covid_data""", """cases deaths recovered""") def a__ ( _UpperCamelCase : str = "https://www.worldometers.info/coronavirus/" ): __lowerCamelCase = '''//div[@class = "maincounter-number"]/span/text()''' return covid_data(*html.fromstring(requests.get(_UpperCamelCase ).content ).xpath(_UpperCamelCase ) ) a_ = """Total COVID-19 cases in the world: {} Total deaths due to COVID-19 in the world: {} Total COVID-19 patients recovered in the world: {}""" print(fmt.format(*covid_stats()))
622
0
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser a_ = logging.getLogger(__name__) torch.set_grad_enabled(False) a_ = "cuda" if torch.cuda.is_available() else "cpu" def a__ ( _UpperCamelCase : str ,_UpperCamelCase : Optional[Any]=1_00 ,_UpperCamelCase : Dict=" " ): __lowerCamelCase = text.split(a_ ) return [character.join(text[i : i + n] ).strip() for i in range(0 ,len(a_ ) ,a_ )] def a__ ( _UpperCamelCase : dict ): __lowerCamelCase = [], [] for title, text in zip(documents['''title'''] ,documents['''text'''] ): if text is not None: for passage in split_text(a_ ): titles.append(title if title is not None else '''''' ) texts.append(a_ ) return {"title": titles, "text": texts} def a__ ( _UpperCamelCase : dict ,_UpperCamelCase : DPRContextEncoder ,_UpperCamelCase : DPRContextEncoderTokenizerFast ): __lowerCamelCase = ctx_tokenizer( documents['''title'''] ,documents['''text'''] ,truncation=a_ ,padding='''longest''' ,return_tensors='''pt''' )['''input_ids'''] __lowerCamelCase = ctx_encoder(input_ids.to(device=a_ ) ,return_dict=a_ ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def a__ ( _UpperCamelCase : "RagExampleArguments" ,_UpperCamelCase : "ProcessingArguments" ,_UpperCamelCase : "IndexHnswArguments" ,): logger.info('''Step 1 - Create the dataset''' ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way __lowerCamelCase = load_dataset( '''csv''' ,data_files=[rag_example_args.csv_path] ,split='''train''' ,delimiter='''\t''' ,column_names=['''title''', '''text'''] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words __lowerCamelCase = dataset.map(a_ ,batched=a_ ,num_proc=processing_args.num_proc ) # And compute the embeddings __lowerCamelCase = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=a_ ) __lowerCamelCase = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) __lowerCamelCase = Features( {'''text''': Value('''string''' ), '''title''': Value('''string''' ), '''embeddings''': Sequence(Value('''float32''' ) )} ) # optional, save as float32 instead of float64 to save space __lowerCamelCase = dataset.map( partial(a_ ,ctx_encoder=a_ ,ctx_tokenizer=a_ ) ,batched=a_ ,batch_size=processing_args.batch_size ,features=a_ ,) # And finally save your dataset __lowerCamelCase = os.path.join(rag_example_args.output_dir ,'''my_knowledge_dataset''' ) dataset.save_to_disk(a_ ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info('''Step 2 - Index the dataset''' ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search __lowerCamelCase = faiss.IndexHNSWFlat(index_hnsw_args.d ,index_hnsw_args.m ,faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index('''embeddings''' ,custom_index=a_ ) # And save the index __lowerCamelCase = os.path.join(rag_example_args.output_dir ,'''my_knowledge_dataset_hnsw_index.faiss''' ) dataset.get_index('''embeddings''' ).save(a_ ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class __lowerCAmelCase : lowerCAmelCase__ = field( default=str(Path(_UpperCAmelCase ).parent / """test_run""" / """dummy-kb""" / """my_knowledge_dataset.csv""" ) , metadata={"""help""": """Path to a tab-separated csv file with columns 'title' and 'text'"""} , ) lowerCAmelCase__ = field( default=_UpperCAmelCase , metadata={"""help""": """Question that is passed as input to RAG. Default is 'What does Moses' rod turn into ?'."""} , ) lowerCAmelCase__ = field( default="""facebook/rag-sequence-nq""" , metadata={"""help""": """The RAG model to use. Either 'facebook/rag-sequence-nq' or 'facebook/rag-token-nq'"""} , ) lowerCAmelCase__ = field( default="""facebook/dpr-ctx_encoder-multiset-base""" , metadata={ """help""": ( """The DPR context encoder model to use. Either 'facebook/dpr-ctx_encoder-single-nq-base' or""" """ 'facebook/dpr-ctx_encoder-multiset-base'""" ) } , ) lowerCAmelCase__ = field( default=str(Path(_UpperCAmelCase ).parent / """test_run""" / """dummy-kb""" ) , metadata={"""help""": """Path to a directory where the dataset passages and the index will be saved"""} , ) @dataclass class __lowerCAmelCase : lowerCAmelCase__ = field( default=_UpperCAmelCase , metadata={ """help""": """The number of processes to use to split the documents into passages. Default is single process.""" } , ) lowerCAmelCase__ = field( default=1_6 , metadata={ """help""": """The batch size to use when computing the passages embeddings using the DPR context encoder.""" } , ) @dataclass class __lowerCAmelCase : lowerCAmelCase__ = field( default=7_6_8 , metadata={"""help""": """The dimension of the embeddings to pass to the HNSW Faiss index."""} , ) lowerCAmelCase__ = field( default=1_2_8 , metadata={ """help""": ( """The number of bi-directional links created for every new element during the HNSW index construction.""" ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) a_ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) a_ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: a_ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
718
def a__ ( _UpperCamelCase : str ,_UpperCamelCase : str = " " ): __lowerCamelCase = [] __lowerCamelCase = 0 for index, char in enumerate(_UpperCamelCase ): if char == separator: split_words.append(string[last_index:index] ) __lowerCamelCase = index + 1 elif index + 1 == len(_UpperCamelCase ): split_words.append(string[last_index : index + 1] ) return split_words if __name__ == "__main__": from doctest import testmod testmod()
622
0
import argparse import os import transformers from .convert_slow_tokenizer import SLOW_TO_FAST_CONVERTERS from .utils import logging logging.set_verbosity_info() a_ = logging.get_logger(__name__) a_ = {name: getattr(transformers, name + """Fast""") for name in SLOW_TO_FAST_CONVERTERS} def a__ ( _UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Dict ,_UpperCamelCase : Optional[Any] ): if tokenizer_name is not None and tokenizer_name not in TOKENIZER_CLASSES: raise ValueError(F"""Unrecognized tokenizer name, should be one of {list(TOKENIZER_CLASSES.keys() )}.""" ) if tokenizer_name is None: __lowerCamelCase = TOKENIZER_CLASSES else: __lowerCamelCase = {tokenizer_name: getattr(_A ,tokenizer_name + '''Fast''' )} logger.info(F"""Loading tokenizer classes: {tokenizer_names}""" ) for tokenizer_name in tokenizer_names: __lowerCamelCase = TOKENIZER_CLASSES[tokenizer_name] __lowerCamelCase = True if checkpoint_name is None: __lowerCamelCase = list(tokenizer_class.max_model_input_sizes.keys() ) else: __lowerCamelCase = [checkpoint_name] logger.info(F"""For tokenizer {tokenizer_class.__class__.__name__} loading checkpoints: {checkpoint_names}""" ) for checkpoint in checkpoint_names: logger.info(F"""Loading {tokenizer_class.__class__.__name__} {checkpoint}""" ) # Load tokenizer __lowerCamelCase = tokenizer_class.from_pretrained(_A ,force_download=_A ) # Save fast tokenizer logger.info(F"""Save fast tokenizer to {dump_path} with prefix {checkpoint} add_prefix {add_prefix}""" ) # For organization names we create sub-directories if "/" in checkpoint: __lowerCamelCase ,__lowerCamelCase = checkpoint.split('''/''' ) __lowerCamelCase = os.path.join(_A ,_A ) elif add_prefix: __lowerCamelCase = checkpoint __lowerCamelCase = dump_path else: __lowerCamelCase = None __lowerCamelCase = dump_path logger.info(F"""=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}""" ) if checkpoint in list(tokenizer.pretrained_vocab_files_map.values() )[0]: __lowerCamelCase = list(tokenizer.pretrained_vocab_files_map.values() )[0][checkpoint] __lowerCamelCase = file_path.split(_A )[-1][0] if next_char == "/": __lowerCamelCase = os.path.join(_A ,_A ) __lowerCamelCase = None logger.info(F"""=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}""" ) __lowerCamelCase = tokenizer.save_pretrained( _A ,legacy_format=_A ,filename_prefix=_A ) logger.info(F"""=> File names {file_names}""" ) for file_name in file_names: if not file_name.endswith('''tokenizer.json''' ): os.remove(_A ) logger.info(F"""=> removing {file_name}""" ) if __name__ == "__main__": a_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--dump_path""", default=None, type=str, required=True, help="""Path to output generated fast tokenizer files.""" ) parser.add_argument( """--tokenizer_name""", default=None, type=str, help=( f"Optional tokenizer type selected in the list of {list(TOKENIZER_CLASSES.keys())}. If not given, will " """download and convert all the checkpoints from AWS.""" ), ) parser.add_argument( """--checkpoint_name""", default=None, type=str, help="""Optional checkpoint name. If not given, will download and convert the canonical checkpoints from AWS.""", ) parser.add_argument( """--force_download""", action="""store_true""", help="""Re-download checkpoints.""", ) a_ = parser.parse_args() convert_slow_checkpoint_to_fast(args.tokenizer_name, args.checkpoint_name, args.dump_path, args.force_download)
719
import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class __lowerCAmelCase ( lowerCAmelCase__ ): def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = tempfile.mkdtemp() __lowerCamelCase = 8 # DPR tok __lowerCamelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] __lowerCamelCase = os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) __lowerCamelCase = os.path.join(__UpperCAmelCase , DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok __lowerCamelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] __lowerCamelCase = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) __lowerCamelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] __lowerCamelCase = {'''unk_token''': '''<unk>'''} __lowerCamelCase = os.path.join(self.tmpdirname , '''bart_tokenizer''' ) os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) __lowerCamelCase = os.path.join(__UpperCAmelCase , BART_VOCAB_FILES_NAMES['''vocab_file'''] ) __lowerCamelCase = os.path.join(__UpperCAmelCase , BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(__UpperCAmelCase ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(__UpperCAmelCase ) ) def lowerCamelCase ( self ): '''simple docstring''' return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''bart_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_dataset() __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: __lowerCamelCase = dataset __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.get_dummy_dataset() __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''custom''' , ) if from_disk: __lowerCamelCase = os.path.join(self.tmpdirname , '''dataset''' ) __lowerCamelCase = os.path.join(self.tmpdirname , '''index.faiss''' ) dataset.get_index('''embeddings''' ).save(os.path.join(self.tmpdirname , '''index.faiss''' ) ) dataset.drop_index('''embeddings''' ) dataset.save_to_disk(os.path.join(self.tmpdirname , '''dataset''' ) ) del dataset __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , __UpperCAmelCase ) , ) return retriever def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) __lowerCamelCase = os.path.join(self.tmpdirname , '''hf_bert_base.hnswSQ8_correct_phi_128.c_index''' ) dataset.save_faiss_index('''embeddings''' , index_file_name + '''.index.dpr''' ) pickle.dump(dataset['''id'''] , open(index_file_name + '''.index_meta.dpr''' , '''wb''' ) ) __lowerCamelCase = os.path.join(self.tmpdirname , '''psgs_w100.tsv.pkl''' ) __lowerCamelCase = {sample['''id''']: [sample['''text'''], sample['''title''']] for sample in dataset} pickle.dump(__UpperCAmelCase , open(__UpperCAmelCase , '''wb''' ) ) __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''legacy''' , index_path=self.tmpdirname , ) __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: __lowerCamelCase = self.get_dummy_dataset() retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_legacy_index_retriever() __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''text'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''text'''][0] , '''bar''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''text'''][0] , '''foo''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def lowerCamelCase ( self ): '''simple docstring''' import torch __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() __lowerCamelCase = [[5, 7], [10, 11]] __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever(__UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = ( out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , np.ndarray ) __lowerCamelCase = retriever( __UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase , return_tensors='''pt''' , ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = ( # noqa: F841 out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], out['''doc_ids'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dpr_ctx_encoder_tokenizer() __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) retriever.set_ctx_encoder_tokenizer(__UpperCAmelCase ) __lowerCamelCase = [[5, 7], [10, 11]] __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever(__UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase ) self.assertEqual( len(__UpperCAmelCase ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('''tokenized_doc_ids''', '''tokenized_doc_attention_mask''') ) , __UpperCAmelCase ) # check for doc token related keys in dictionary.
622
0
import tempfile import unittest import numpy as np import transformers from transformers import GPTaTokenizer, GPTJConfig, is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax, tooslow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax import jax.numpy as jnp from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) from transformers.models.gptj.modeling_flax_gptj import FlaxGPTJForCausalLM, FlaxGPTJModel if is_torch_available(): import torch class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase=14 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase=99 , __UpperCAmelCase=32 , __UpperCAmelCase=4 , __UpperCAmelCase=4 , __UpperCAmelCase=4 , __UpperCAmelCase=37 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=0.02 , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_input_mask __lowerCamelCase = use_token_type_ids __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = rotary_dim __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = initializer_range __lowerCamelCase = None __lowerCamelCase = vocab_size - 1 __lowerCamelCase = vocab_size - 1 __lowerCamelCase = vocab_size - 1 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCamelCase = GPTJConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , use_cache=snake_case_ , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , rotary_dim=self.rotary_dim , ) return (config, input_ids, input_mask) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.prepare_config_and_inputs() __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = config_and_inputs __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = 20 __lowerCamelCase = model_class_name(snake_case_ ) __lowerCamelCase = model.init_cache(input_ids.shape[0] , snake_case_ ) __lowerCamelCase = jnp.ones((input_ids.shape[0], max_decoder_length) , dtype='''i4''' ) __lowerCamelCase = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) __lowerCamelCase = model( input_ids[:, :-1] , attention_mask=snake_case_ , past_key_values=snake_case_ , position_ids=snake_case_ , ) __lowerCamelCase = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype='''i4''' ) __lowerCamelCase = model( input_ids[:, -1:] , attention_mask=snake_case_ , past_key_values=outputs_cache.past_key_values , position_ids=snake_case_ , ) __lowerCamelCase = model(snake_case_ ) __lowerCamelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F"""Max diff is {diff}""" ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = 20 __lowerCamelCase = model_class_name(snake_case_ ) __lowerCamelCase = jnp.concatenate( [attention_mask, jnp.zeros((attention_mask.shape[0], max_decoder_length - attention_mask.shape[1]) )] , axis=-1 , ) __lowerCamelCase = model.init_cache(input_ids.shape[0] , snake_case_ ) __lowerCamelCase = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) __lowerCamelCase = model( input_ids[:, :-1] , attention_mask=snake_case_ , past_key_values=snake_case_ , position_ids=snake_case_ , ) __lowerCamelCase = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype='''i4''' ) __lowerCamelCase = model( input_ids[:, -1:] , past_key_values=outputs_cache.past_key_values , attention_mask=snake_case_ , position_ids=snake_case_ , ) __lowerCamelCase = model(snake_case_ , attention_mask=snake_case_ ) __lowerCamelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F"""Max diff is {diff}""" ) @require_flax class __lowerCAmelCase ( _a , _a , unittest.TestCase ): lowerCAmelCase__ = (FlaxGPTJModel, FlaxGPTJForCausalLM) if is_flax_available() else () lowerCAmelCase__ = (FlaxGPTJForCausalLM,) if is_flax_available() else () def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = FlaxGPTJModelTester(self ) def lowerCamelCase ( self ): '''simple docstring''' for model_class_name in self.all_model_classes: __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) def lowerCamelCase ( self ): '''simple docstring''' for model_class_name in self.all_model_classes: __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward_with_attn_mask( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) @tooslow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = GPTaTokenizer.from_pretrained('''gpt2''' , pad_token='''<|endoftext|>''' , padding_side='''left''' ) __lowerCamelCase = tokenizer(['''Hello this is a long string''', '''Hey'''] , return_tensors='''np''' , padding=snake_case_ , truncation=snake_case_ ) __lowerCamelCase = FlaxGPTJForCausalLM.from_pretrained('''EleutherAI/gpt-j-6B''' ) __lowerCamelCase = False __lowerCamelCase = model.config.eos_token_id __lowerCamelCase = jax.jit(model.generate ) __lowerCamelCase = jit_generate( inputs['''input_ids'''] , attention_mask=inputs['''attention_mask'''] , pad_token_id=tokenizer.pad_token_id ).sequences __lowerCamelCase = tokenizer.batch_decode(snake_case_ , skip_special_tokens=snake_case_ ) __lowerCamelCase = [ '''Hello this is a long string of text.\n\nI\'m trying to get the text of the''', '''Hey, I\'m a little late to the party. I\'m going to''', ] self.assertListEqual(snake_case_ , snake_case_ ) @is_pt_flax_cross_test def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs __lowerCamelCase = self._prepare_for_class(snake_case_ , snake_case_ ) __lowerCamelCase = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class __lowerCamelCase = model_class.__name__[4:] # Skip the "Flax" at the beginning __lowerCamelCase = getattr(snake_case_ , snake_case_ ) __lowerCamelCase ,__lowerCamelCase = pt_inputs['''input_ids'''].shape __lowerCamelCase = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(snake_case_ ): __lowerCamelCase = 0 __lowerCamelCase = 1 __lowerCamelCase = 0 __lowerCamelCase = 1 __lowerCamelCase = pt_model_class(snake_case_ ).eval() __lowerCamelCase = model_class(snake_case_ , dtype=jnp.floataa ) __lowerCamelCase = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , snake_case_ ) __lowerCamelCase = fx_state with torch.no_grad(): __lowerCamelCase = pt_model(**snake_case_ ).to_tuple() __lowerCamelCase = fx_model(**snake_case_ ).to_tuple() self.assertEqual(len(snake_case_ ) , len(snake_case_ ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output, pt_output in zip(snake_case_ , snake_case_ ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(snake_case_ ) __lowerCamelCase = model_class.from_pretrained(snake_case_ , from_pt=snake_case_ ) __lowerCamelCase = fx_model_loaded(**snake_case_ ).to_tuple() self.assertEqual( len(snake_case_ ) , len(snake_case_ ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output_loaded, pt_output in zip(snake_case_ , snake_case_ ): self.assert_almost_equals(fx_output_loaded[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @is_pt_flax_cross_test def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs __lowerCamelCase = self._prepare_for_class(snake_case_ , snake_case_ ) __lowerCamelCase = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class __lowerCamelCase = model_class.__name__[4:] # Skip the "Flax" at the beginning __lowerCamelCase = getattr(snake_case_ , snake_case_ ) __lowerCamelCase = pt_model_class(snake_case_ ).eval() __lowerCamelCase = model_class(snake_case_ , dtype=jnp.floataa ) __lowerCamelCase = load_flax_weights_in_pytorch_model(snake_case_ , fx_model.params ) __lowerCamelCase ,__lowerCamelCase = pt_inputs['''input_ids'''].shape __lowerCamelCase = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(snake_case_ ): __lowerCamelCase = 0 __lowerCamelCase = 1 __lowerCamelCase = 0 __lowerCamelCase = 1 # make sure weights are tied in PyTorch pt_model.tie_weights() with torch.no_grad(): __lowerCamelCase = pt_model(**snake_case_ ).to_tuple() __lowerCamelCase = fx_model(**snake_case_ ).to_tuple() self.assertEqual(len(snake_case_ ) , len(snake_case_ ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output, pt_output in zip(snake_case_ , snake_case_ ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(snake_case_ ) __lowerCamelCase = pt_model_class.from_pretrained(snake_case_ , from_flax=snake_case_ ) with torch.no_grad(): __lowerCamelCase = pt_model_loaded(**snake_case_ ).to_tuple() self.assertEqual( len(snake_case_ ) , len(snake_case_ ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output, pt_output in zip(snake_case_ , snake_case_ ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @tooslow def lowerCamelCase ( self ): '''simple docstring''' for model_class_name in self.all_model_classes: __lowerCamelCase = model_class_name.from_pretrained('''EleutherAI/gpt-j-6B''' ) __lowerCamelCase = model(np.ones((1, 1) ) ) self.assertIsNotNone(snake_case_ )
720
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """sail/poolformer_s12""": """https://huggingface.co/sail/poolformer_s12/resolve/main/config.json""", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """poolformer""" def __init__( self , __UpperCAmelCase=3 , __UpperCAmelCase=16 , __UpperCAmelCase=16 , __UpperCAmelCase=3 , __UpperCAmelCase=4.0 , __UpperCAmelCase=[2, 2, 6, 2] , __UpperCAmelCase=[64, 128, 320, 512] , __UpperCAmelCase=[7, 3, 3, 3] , __UpperCAmelCase=[4, 2, 2, 2] , __UpperCAmelCase=[2, 1, 1, 1] , __UpperCAmelCase=4 , __UpperCAmelCase=0.0 , __UpperCAmelCase="gelu" , __UpperCAmelCase=True , __UpperCAmelCase=1E-5 , __UpperCAmelCase=0.02 , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = num_channels __lowerCamelCase = patch_size __lowerCamelCase = stride __lowerCamelCase = padding __lowerCamelCase = pool_size __lowerCamelCase = hidden_sizes __lowerCamelCase = mlp_ratio __lowerCamelCase = depths __lowerCamelCase = patch_sizes __lowerCamelCase = strides __lowerCamelCase = num_encoder_blocks __lowerCamelCase = drop_path_rate __lowerCamelCase = hidden_act __lowerCamelCase = use_layer_scale __lowerCamelCase = layer_scale_init_value __lowerCamelCase = initializer_range super().__init__(**__UpperCAmelCase ) class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = version.parse("""1.11""" ) @property def lowerCamelCase ( self ): '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def lowerCamelCase ( self ): '''simple docstring''' return 2E-3
622
0
a_ = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' a_ = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] a_ = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
721
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """uclanlp/visualbert-vqa""": """https://huggingface.co/uclanlp/visualbert-vqa/resolve/main/config.json""", """uclanlp/visualbert-vqa-pre""": """https://huggingface.co/uclanlp/visualbert-vqa-pre/resolve/main/config.json""", """uclanlp/visualbert-vqa-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-vqa-coco-pre/resolve/main/config.json""" ), """uclanlp/visualbert-vcr""": """https://huggingface.co/uclanlp/visualbert-vcr/resolve/main/config.json""", """uclanlp/visualbert-vcr-pre""": """https://huggingface.co/uclanlp/visualbert-vcr-pre/resolve/main/config.json""", """uclanlp/visualbert-vcr-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-vcr-coco-pre/resolve/main/config.json""" ), """uclanlp/visualbert-nlvr2""": """https://huggingface.co/uclanlp/visualbert-nlvr2/resolve/main/config.json""", """uclanlp/visualbert-nlvr2-pre""": """https://huggingface.co/uclanlp/visualbert-nlvr2-pre/resolve/main/config.json""", """uclanlp/visualbert-nlvr2-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-nlvr2-coco-pre/resolve/main/config.json""" ) # See all VisualBERT models at https://huggingface.co/models?filter=visual_bert } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """visual_bert""" def __init__( self , __UpperCAmelCase=30522 , __UpperCAmelCase=768 , __UpperCAmelCase=512 , __UpperCAmelCase=12 , __UpperCAmelCase=12 , __UpperCAmelCase=3072 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase=1 , __UpperCAmelCase=0 , __UpperCAmelCase=2 , **__UpperCAmelCase , ): '''simple docstring''' super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) __lowerCamelCase = vocab_size __lowerCamelCase = max_position_embeddings __lowerCamelCase = hidden_size __lowerCamelCase = visual_embedding_dim __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = initializer_range __lowerCamelCase = type_vocab_size __lowerCamelCase = layer_norm_eps __lowerCamelCase = bypass_transformer __lowerCamelCase = special_visual_initialize
622
0
import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __lowerCAmelCase ( _lowercase , unittest.TestCase ): lowerCAmelCase__ = '''hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline''' def lowerCamelCase ( self , __UpperCAmelCase=0 ): '''simple docstring''' __lowerCamelCase = floats_tensor((1, 3, 128, 128) , rng=random.Random(A_ ) ) __lowerCamelCase = np.random.RandomState(A_ ) __lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 3, '''strength''': 0.75, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) pipe.set_progress_bar_config(disable=A_ ) __lowerCamelCase = self.get_dummy_inputs() __lowerCamelCase = pipe(**A_ ).images __lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 128, 128, 3) __lowerCamelCase = np.array([0.69_643, 0.58_484, 0.50_314, 0.58_760, 0.55_368, 0.59_643, 0.51_529, 0.41_217, 0.49_087] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) __lowerCamelCase = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=A_ ) pipe.set_progress_bar_config(disable=A_ ) __lowerCamelCase = self.get_dummy_inputs() __lowerCamelCase = pipe(**A_ ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __lowerCamelCase = np.array([0.61_737, 0.54_642, 0.53_183, 0.54_465, 0.52_742, 0.60_525, 0.49_969, 0.40_655, 0.48_154] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) __lowerCamelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=A_ ) # warmup pass to apply optimizations __lowerCamelCase = pipe(**self.get_dummy_inputs() ) __lowerCamelCase = self.get_dummy_inputs() __lowerCamelCase = pipe(**A_ ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __lowerCamelCase = np.array([0.52_761, 0.59_977, 0.49_033, 0.49_619, 0.54_282, 0.50_311, 0.47_600, 0.40_918, 0.45_203] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) __lowerCamelCase = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=A_ ) __lowerCamelCase = self.get_dummy_inputs() __lowerCamelCase = pipe(**A_ ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __lowerCamelCase = np.array([0.52_911, 0.60_004, 0.49_229, 0.49_805, 0.54_502, 0.50_680, 0.47_777, 0.41_028, 0.45_304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) __lowerCamelCase = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=A_ ) __lowerCamelCase = self.get_dummy_inputs() __lowerCamelCase = pipe(**A_ ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __lowerCamelCase = np.array([0.52_911, 0.60_004, 0.49_229, 0.49_805, 0.54_502, 0.50_680, 0.47_777, 0.41_028, 0.45_304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) __lowerCamelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=A_ ) __lowerCamelCase = self.get_dummy_inputs() __lowerCamelCase = pipe(**A_ ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __lowerCamelCase = np.array([0.65_331, 0.58_277, 0.48_204, 0.56_059, 0.53_665, 0.56_235, 0.50_969, 0.40_009, 0.46_552] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): @property def lowerCamelCase ( self ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ort.SessionOptions() __lowerCamelCase = False return options def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) __lowerCamelCase = init_image.resize((768, 512) ) # using the PNDM scheduler by default __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''onnx''' , safety_checker=A_ , feature_extractor=A_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=A_ ) __lowerCamelCase = '''A fantasy landscape, trending on artstation''' __lowerCamelCase = np.random.RandomState(0 ) __lowerCamelCase = pipe( prompt=A_ , image=A_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=A_ , output_type='''np''' , ) __lowerCamelCase = output.images __lowerCamelCase = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __lowerCamelCase = np.array([0.4_909, 0.5_059, 0.5_372, 0.4_623, 0.4_876, 0.5_049, 0.4_820, 0.4_956, 0.5_019] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) __lowerCamelCase = init_image.resize((768, 512) ) __lowerCamelCase = LMSDiscreteScheduler.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , subfolder='''scheduler''' , revision='''onnx''' ) __lowerCamelCase = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , revision='''onnx''' , scheduler=A_ , safety_checker=A_ , feature_extractor=A_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=A_ ) __lowerCamelCase = '''A fantasy landscape, trending on artstation''' __lowerCamelCase = np.random.RandomState(0 ) __lowerCamelCase = pipe( prompt=A_ , image=A_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=A_ , output_type='''np''' , ) __lowerCamelCase = output.images __lowerCamelCase = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __lowerCamelCase = np.array([0.8_043, 0.926, 0.9_581, 0.8_119, 0.8_954, 0.913, 0.7_209, 0.7_463, 0.7_431] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
700
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging a_ = logging.get_logger(__name__) a_ = {"""vocab_file""": """spiece.model"""} a_ = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", } } a_ = { """albert-base-v1""": 512, """albert-large-v1""": 512, """albert-xlarge-v1""": 512, """albert-xxlarge-v1""": 512, """albert-base-v2""": 512, """albert-large-v2""": 512, """albert-xlarge-v2""": 512, """albert-xxlarge-v2""": 512, } a_ = """▁""" class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , __UpperCAmelCase , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="<unk>" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="<pad>" , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[MASK]" , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. __lowerCamelCase = ( AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase , normalized=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token ) __lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__UpperCAmelCase , remove_space=__UpperCAmelCase , keep_accents=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) __lowerCamelCase = do_lower_case __lowerCamelCase = remove_space __lowerCamelCase = keep_accents __lowerCamelCase = vocab_file __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCAmelCase ) @property def lowerCamelCase ( self ): '''simple docstring''' return len(self.sp_model ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): '''simple docstring''' __lowerCamelCase = self.__dict__.copy() __lowerCamelCase = None return state def __setstate__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCamelCase = {} __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' if self.remove_space: __lowerCamelCase = ''' '''.join(inputs.strip().split() ) else: __lowerCamelCase = inputs __lowerCamelCase = outputs.replace('''``''' , '''"''' ).replace('''\'\'''' , '''"''' ) if not self.keep_accents: __lowerCamelCase = unicodedata.normalize('''NFKD''' , __UpperCAmelCase ) __lowerCamelCase = ''''''.join([c for c in outputs if not unicodedata.combining(__UpperCAmelCase )] ) if self.do_lower_case: __lowerCamelCase = outputs.lower() return outputs def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.preprocess_text(__UpperCAmelCase ) __lowerCamelCase = self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) __lowerCamelCase = [] for piece in pieces: if len(__UpperCAmelCase ) > 1 and piece[-1] == str(''',''' ) and piece[-2].isdigit(): __lowerCamelCase = self.sp_model.EncodeAsPieces(piece[:-1].replace(__UpperCAmelCase , '''''' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __lowerCamelCase = cur_pieces[1:] else: __lowerCamelCase = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__UpperCAmelCase ) else: new_pieces.append(__UpperCAmelCase ) return new_pieces def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.PieceToId(__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.IdToPiece(__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = '''''' __lowerCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__UpperCAmelCase ) + token __lowerCamelCase = True __lowerCamelCase = [] else: current_sub_tokens.append(__UpperCAmelCase ) __lowerCamelCase = False out_string += self.sp_model.decode(__UpperCAmelCase ) return out_string.strip() def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase ) if token_ids_a is not None: return [1] + ([0] * len(__UpperCAmelCase )) + [1] + ([0] * len(__UpperCAmelCase )) + [1] return [1] + ([0] * len(__UpperCAmelCase )) + [1] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' if not os.path.isdir(__UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCamelCase = os.path.join( __UpperCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , '''wb''' ) as fi: __lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,)
622
0
a_ = """\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n""" a_ = [{"""type""": """code""", """content""": INSTALL_CONTENT}] a_ = { """{processor_class}""": """FakeProcessorClass""", """{model_class}""": """FakeModelClass""", """{object_class}""": """FakeObjectClass""", }
701
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed a_ = """true""" def a__ ( _UpperCamelCase : Union[str, Any] ,_UpperCamelCase : List[str]=82 ,_UpperCamelCase : Optional[Any]=16 ): set_seed(42 ) __lowerCamelCase = RegressionModel() __lowerCamelCase = deepcopy(_UpperCamelCase ) __lowerCamelCase = RegressionDataset(length=_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,batch_size=_UpperCamelCase ) model.to(accelerator.device ) __lowerCamelCase ,__lowerCamelCase = accelerator.prepare(_UpperCamelCase ,_UpperCamelCase ) return model, ddp_model, dataloader def a__ ( _UpperCamelCase : Accelerator ,_UpperCamelCase : str=False ): __lowerCamelCase = AutoTokenizer.from_pretrained('''hf-internal-testing/mrpc-bert-base-cased''' ) __lowerCamelCase = load_dataset('''glue''' ,'''mrpc''' ,split='''validation''' ) def tokenize_function(_UpperCamelCase : int ): __lowerCamelCase = tokenizer(examples['''sentence1'''] ,examples['''sentence2'''] ,truncation=_UpperCamelCase ,max_length=_UpperCamelCase ) return outputs with accelerator.main_process_first(): __lowerCamelCase = dataset.map( _UpperCamelCase ,batched=_UpperCamelCase ,remove_columns=['''idx''', '''sentence1''', '''sentence2'''] ,) __lowerCamelCase = tokenized_datasets.rename_column('''label''' ,'''labels''' ) def collate_fn(_UpperCamelCase : Any ): if use_longest: return tokenizer.pad(_UpperCamelCase ,padding='''longest''' ,return_tensors='''pt''' ) return tokenizer.pad(_UpperCamelCase ,padding='''max_length''' ,max_length=1_28 ,return_tensors='''pt''' ) return DataLoader(_UpperCamelCase ,shuffle=_UpperCamelCase ,collate_fn=_UpperCamelCase ,batch_size=16 ) def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : List[str] ): __lowerCamelCase = Accelerator(dispatch_batches=_UpperCamelCase ,split_batches=_UpperCamelCase ) __lowerCamelCase = get_dataloader(_UpperCamelCase ,not dispatch_batches ) __lowerCamelCase = AutoModelForSequenceClassification.from_pretrained( '''hf-internal-testing/mrpc-bert-base-cased''' ,return_dict=_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = accelerator.prepare(_UpperCamelCase ,_UpperCamelCase ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : Optional[Any] ,_UpperCamelCase : Union[str, Any] ): __lowerCamelCase = [] for batch in dataloader: __lowerCamelCase ,__lowerCamelCase = batch.values() with torch.no_grad(): __lowerCamelCase = model(_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) __lowerCamelCase ,__lowerCamelCase = [], [] for logit, targ in logits_and_targets: logits.append(_UpperCamelCase ) targs.append(_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = torch.cat(_UpperCamelCase ), torch.cat(_UpperCamelCase ) return logits, targs def a__ ( _UpperCamelCase : Accelerator ,_UpperCamelCase : List[Any]=82 ,_UpperCamelCase : str=False ,_UpperCamelCase : List[str]=False ,_UpperCamelCase : Optional[int]=16 ): __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = get_basic_setup(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = generate_predictions(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ) assert ( len(_UpperCamelCase ) == num_samples ), F"""Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(_UpperCamelCase )}""" def a__ ( _UpperCamelCase : bool = False ,_UpperCamelCase : bool = False ): __lowerCamelCase = evaluate.load('''glue''' ,'''mrpc''' ) __lowerCamelCase ,__lowerCamelCase = get_mrpc_setup(_UpperCamelCase ,_UpperCamelCase ) # First do baseline __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = setup['''no'''] model.to(_UpperCamelCase ) model.eval() for batch in dataloader: batch.to(_UpperCamelCase ) with torch.inference_mode(): __lowerCamelCase = model(**_UpperCamelCase ) __lowerCamelCase = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=_UpperCamelCase ,references=batch['''labels'''] ) __lowerCamelCase = metric.compute() # Then do distributed __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = setup['''ddp'''] model.eval() for batch in dataloader: with torch.inference_mode(): __lowerCamelCase = model(**_UpperCamelCase ) __lowerCamelCase = outputs.logits.argmax(dim=-1 ) __lowerCamelCase = batch['''labels'''] __lowerCamelCase ,__lowerCamelCase = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=_UpperCamelCase ,references=_UpperCamelCase ) __lowerCamelCase = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] ,distributed[key] ), F"""Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n""" def a__ ( ): __lowerCamelCase = Accelerator(split_batches=_UpperCamelCase ,dispatch_batches=_UpperCamelCase ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print('''**Testing gather_for_metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(F"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`""" ) test_mrpc(_UpperCamelCase ,_UpperCamelCase ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test torch metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: __lowerCamelCase = Accelerator(split_batches=_UpperCamelCase ,dispatch_batches=_UpperCamelCase ) if accelerator.is_local_main_process: print(F"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99""" ) test_torch_metrics(_UpperCamelCase ,99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test last batch is not dropped when perfectly divisible**''' ) __lowerCamelCase = Accelerator() test_torch_metrics(_UpperCamelCase ,5_12 ) accelerator.state._reset_state() def a__ ( _UpperCamelCase : Optional[int] ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
622
0
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Sequence, Value from .base import TaskTemplate @dataclass(frozen=lowerCAmelCase__ ) class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = field(default="""question-answering-extractive""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) lowerCAmelCase__ = Features({"""question""": Value("""string""" ), """context""": Value("""string""" )} ) lowerCAmelCase__ = Features( { """answers""": Sequence( { """text""": Value("""string""" ), """answer_start""": Value("""int32""" ), } ) } ) lowerCAmelCase__ = """question""" lowerCAmelCase__ = """context""" lowerCAmelCase__ = """answers""" @property def lowerCamelCase ( self ): '''simple docstring''' return {self.question_column: "question", self.context_column: "context", self.answers_column: "answers"}
702
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = StableDiffusionXLImgaImgPipeline lowerCAmelCase__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} lowerCAmelCase__ = PipelineTesterMixin.required_optional_params - {"""latents"""} lowerCAmelCase__ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowerCAmelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS lowerCAmelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowerCamelCase ( self ): '''simple docstring''' torch.manual_seed(0 ) __lowerCamelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=__UpperCAmelCase , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) __lowerCamelCase = EulerDiscreteScheduler( beta_start=0.00_085 , beta_end=0.012 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) __lowerCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='''gelu''' , projection_dim=32 , ) __lowerCamelCase = CLIPTextModel(__UpperCAmelCase ) __lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__UpperCAmelCase ) __lowerCamelCase = CLIPTextModelWithProjection(__UpperCAmelCase ) __lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__UpperCAmelCase ) __lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase=0 ): '''simple docstring''' __lowerCamelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__UpperCAmelCase ) ).to(__UpperCAmelCase ) __lowerCamelCase = image / 2 + 0.5 if str(__UpperCAmelCase ).startswith('''mps''' ): __lowerCamelCase = torch.manual_seed(__UpperCAmelCase ) else: __lowerCamelCase = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) __lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.75, } return inputs def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = StableDiffusionXLImgaImgPipeline(**__UpperCAmelCase ) __lowerCamelCase = sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase = sd_pipe(**__UpperCAmelCase ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCamelCase = np.array([0.4_656, 0.4_840, 0.4_439, 0.6_698, 0.5_574, 0.4_524, 0.5_799, 0.5_943, 0.5_165] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase ( self ): '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def lowerCamelCase ( self ): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = StableDiffusionXLImgaImgPipeline(**__UpperCAmelCase ) __lowerCamelCase = sd_pipe.to(__UpperCAmelCase ) __lowerCamelCase = sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) # forward without prompt embeds __lowerCamelCase = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase = 3 * ['''this is a negative prompt'''] __lowerCamelCase = negative_prompt __lowerCamelCase = 3 * [inputs['''prompt''']] __lowerCamelCase = sd_pipe(**__UpperCAmelCase ) __lowerCamelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds __lowerCamelCase = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase = 3 * ['''this is a negative prompt'''] __lowerCamelCase = 3 * [inputs.pop('''prompt''' )] ( ( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) , ) = sd_pipe.encode_prompt(__UpperCAmelCase , negative_prompt=__UpperCAmelCase ) __lowerCamelCase = sd_pipe( **__UpperCAmelCase , prompt_embeds=__UpperCAmelCase , negative_prompt_embeds=__UpperCAmelCase , pooled_prompt_embeds=__UpperCAmelCase , negative_pooled_prompt_embeds=__UpperCAmelCase , ) __lowerCamelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase="cpu" , __UpperCAmelCase=torch.floataa , __UpperCAmelCase=0 ): '''simple docstring''' __lowerCamelCase = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) __lowerCamelCase = np.random.RandomState(__UpperCAmelCase ).standard_normal((1, 4, 64, 64) ) __lowerCamelCase = torch.from_numpy(__UpperCAmelCase ).to(device=__UpperCAmelCase , dtype=__UpperCAmelCase ) __lowerCamelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = self.get_inputs(__UpperCAmelCase ) __lowerCamelCase = pipe(**__UpperCAmelCase ).images __lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.49_493, 0.47_896, 0.40_798, 0.54_214, 0.53_212, 0.48_202, 0.47_656, 0.46_329, 0.48_506] ) assert np.abs(image_slice - expected_slice ).max() < 7E-3
622
0
from math import pi, sqrt, tan def a__ ( _UpperCamelCase : float ): if side_length < 0: raise ValueError('''surface_area_cube() only accepts non-negative values''' ) return 6 * side_length**2 def a__ ( _UpperCamelCase : float ,_UpperCamelCase : float ,_UpperCamelCase : float ): if length < 0 or breadth < 0 or height < 0: raise ValueError('''surface_area_cuboid() only accepts non-negative values''' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def a__ ( _UpperCamelCase : float ): if radius < 0: raise ValueError('''surface_area_sphere() only accepts non-negative values''' ) return 4 * pi * radius**2 def a__ ( _UpperCamelCase : float ): if radius < 0: raise ValueError('''surface_area_hemisphere() only accepts non-negative values''' ) return 3 * pi * radius**2 def a__ ( _UpperCamelCase : float ,_UpperCamelCase : float ): if radius < 0 or height < 0: raise ValueError('''surface_area_cone() only accepts non-negative values''' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def a__ ( _UpperCamelCase : float ,_UpperCamelCase : float ,_UpperCamelCase : float ): if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( '''surface_area_conical_frustum() only accepts non-negative values''' ) __lowerCamelCase = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def a__ ( _UpperCamelCase : float ,_UpperCamelCase : float ): if radius < 0 or height < 0: raise ValueError('''surface_area_cylinder() only accepts non-negative values''' ) return 2 * pi * radius * (height + radius) def a__ ( _UpperCamelCase : float ,_UpperCamelCase : float ): if torus_radius < 0 or tube_radius < 0: raise ValueError('''surface_area_torus() only accepts non-negative values''' ) if torus_radius < tube_radius: raise ValueError( '''surface_area_torus() does not support spindle or self intersecting tori''' ) return 4 * pow(snake_case_ ,2 ) * torus_radius * tube_radius def a__ ( _UpperCamelCase : float ,_UpperCamelCase : float ): if length < 0 or width < 0: raise ValueError('''area_rectangle() only accepts non-negative values''' ) return length * width def a__ ( _UpperCamelCase : float ): if side_length < 0: raise ValueError('''area_square() only accepts non-negative values''' ) return side_length**2 def a__ ( _UpperCamelCase : float ,_UpperCamelCase : float ): if base < 0 or height < 0: raise ValueError('''area_triangle() only accepts non-negative values''' ) return (base * height) / 2 def a__ ( _UpperCamelCase : float ,_UpperCamelCase : float ,_UpperCamelCase : float ): if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('''area_triangle_three_sides() only accepts non-negative values''' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('''Given three sides do not form a triangle''' ) __lowerCamelCase = (sidea + sidea + sidea) / 2 __lowerCamelCase = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def a__ ( _UpperCamelCase : float ,_UpperCamelCase : float ): if base < 0 or height < 0: raise ValueError('''area_parallelogram() only accepts non-negative values''' ) return base * height def a__ ( _UpperCamelCase : float ,_UpperCamelCase : float ,_UpperCamelCase : float ): if basea < 0 or basea < 0 or height < 0: raise ValueError('''area_trapezium() only accepts non-negative values''' ) return 1 / 2 * (basea + basea) * height def a__ ( _UpperCamelCase : float ): if radius < 0: raise ValueError('''area_circle() only accepts non-negative values''' ) return pi * radius**2 def a__ ( _UpperCamelCase : float ,_UpperCamelCase : float ): if radius_x < 0 or radius_y < 0: raise ValueError('''area_ellipse() only accepts non-negative values''' ) return pi * radius_x * radius_y def a__ ( _UpperCamelCase : float ,_UpperCamelCase : float ): if diagonal_a < 0 or diagonal_a < 0: raise ValueError('''area_rhombus() only accepts non-negative values''' ) return 1 / 2 * diagonal_a * diagonal_a def a__ ( _UpperCamelCase : int ,_UpperCamelCase : float ): if not isinstance(snake_case_ ,snake_case_ ) or sides < 3: raise ValueError( '''area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides''' ) elif length < 0: raise ValueError( '''area_reg_polygon() only accepts non-negative values as \ length of a side''' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print("""[DEMO] Areas of various geometric shapes: \n""") print(f"Rectangle: {area_rectangle(10, 20) = }") print(f"Square: {area_square(10) = }") print(f"Triangle: {area_triangle(10, 10) = }") print(f"Triangle: {area_triangle_three_sides(5, 12, 13) = }") print(f"Parallelogram: {area_parallelogram(10, 20) = }") print(f"Rhombus: {area_rhombus(10, 20) = }") print(f"Trapezium: {area_trapezium(10, 20, 30) = }") print(f"Circle: {area_circle(20) = }") print(f"Ellipse: {area_ellipse(10, 20) = }") print("""\nSurface Areas of various geometric shapes: \n""") print(f"Cube: {surface_area_cube(20) = }") print(f"Cuboid: {surface_area_cuboid(10, 20, 30) = }") print(f"Sphere: {surface_area_sphere(20) = }") print(f"Hemisphere: {surface_area_hemisphere(20) = }") print(f"Cone: {surface_area_cone(10, 20) = }") print(f"Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }") print(f"Cylinder: {surface_area_cylinder(10, 20) = }") print(f"Torus: {surface_area_torus(20, 10) = }") print(f"Equilateral Triangle: {area_reg_polygon(3, 10) = }") print(f"Square: {area_reg_polygon(4, 10) = }") print(f"Reqular Pentagon: {area_reg_polygon(5, 10) = }")
703
import torch from diffusers import StableDiffusionPipeline a_ = """path-to-your-trained-model""" a_ = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to("""cuda""") a_ = """A photo of sks dog in a bucket""" a_ = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save("""dog-bucket.png""")
622
0
import unittest from pathlib import Path from shutil import copyfile from transformers import SPIECE_UNDERLINE, is_sentencepiece_available from transformers.models.speech_to_text import SpeechaTextTokenizer from transformers.models.speech_to_text.tokenization_speech_to_text import VOCAB_FILES_NAMES, save_json from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin a_ = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_sentencepiece_available(): import sentencepiece as sp a_ = 5 a_ = 10 @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( lowercase__ , unittest.TestCase ): lowerCAmelCase__ = SpeechaTextTokenizer lowerCAmelCase__ = False lowerCAmelCase__ = True def lowerCamelCase ( self ): '''simple docstring''' super().setUp() __lowerCamelCase = sp.SentencePieceProcessor() spm_model.Load(__UpperCAmelCase ) __lowerCamelCase = ["<s>", "<pad>", "</s>", "<unk>"] vocab += [spm_model.IdToPiece(id_ ) for id_ in range(len(__UpperCAmelCase ) )] __lowerCamelCase = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) __lowerCamelCase = Path(self.tmpdirname ) save_json(__UpperCAmelCase , save_dir / VOCAB_FILES_NAMES['''vocab_file'''] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(__UpperCAmelCase , save_dir / VOCAB_FILES_NAMES['''spm_file'''] ) __lowerCamelCase = SpeechaTextTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = "<pad>" __lowerCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCAmelCase ) , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(__UpperCAmelCase ) , 1001 ) def lowerCamelCase ( self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1001 ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = SpeechaTextTokenizer.from_pretrained(self.tmpdirname ) __lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__UpperCAmelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [289, 50, 14, 174, 386] , ) __lowerCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __UpperCAmelCase , [SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.'''] , ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , [12, 25, 88, 59, 28, 23, 11, 4, 606, 351, 351, 351, 7, 16, 70, 50, 76, 84, 10, 4, 8] ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.'''] , ) @slow def lowerCamelCase ( self ): '''simple docstring''' # fmt: off __lowerCamelCase = {"input_ids": [[3791, 797, 31, 11, 64, 797, 31, 2429, 433, 12, 1176, 12, 20, 786, 915, 142, 2413, 240, 37, 3238, 797, 31, 11, 35, 93, 915, 142, 2413, 240, 37, 5540, 567, 1276, 93, 37, 610, 40, 62, 455, 657, 1042, 123, 780, 177, 37, 309, 241, 1298, 514, 20, 292, 2737, 114, 2469, 241, 85, 64, 302, 548, 528, 423, 4, 509, 406, 423, 37, 601, 4, 777, 302, 548, 528, 423, 284, 4, 3388, 511, 459, 4, 3555, 40, 321, 302, 705, 4, 3388, 511, 583, 326, 5, 5, 5, 62, 3310, 560, 177, 2680, 217, 1508, 32, 31, 853, 418, 64, 583, 511, 1605, 62, 35, 93, 560, 177, 2680, 217, 1508, 1521, 64, 583, 511, 519, 62, 20, 1515, 764, 20, 149, 261, 5625, 7972, 20, 5540, 567, 1276, 93, 3925, 1675, 11, 15, 802, 7972, 576, 217, 1508, 11, 35, 93, 1253, 2441, 15, 289, 652, 31, 416, 321, 3842, 115, 40, 911, 8, 476, 619, 4, 380, 142, 423, 335, 240, 35, 93, 264, 8, 11, 335, 569, 420, 163, 5, 2], [260, 548, 528, 423, 20, 451, 20, 2681, 1153, 3434, 20, 5540, 37, 567, 126, 1253, 2441, 3376, 449, 210, 431, 1563, 177, 767, 5540, 11, 1203, 472, 11, 2953, 685, 285, 364, 706, 1153, 20, 6799, 20, 2869, 20, 4464, 126, 40, 2429, 20, 1040, 866, 2664, 418, 20, 318, 20, 1726, 186, 20, 265, 522, 35, 93, 2191, 4634, 20, 1040, 12, 6799, 15, 228, 2356, 142, 31, 11, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [2575, 2666, 684, 1582, 1176, 12, 627, 149, 619, 20, 4902, 563, 11, 20, 149, 261, 3420, 2356, 174, 142, 4714, 131, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__UpperCAmelCase , model_name='''facebook/s2t-small-mustc-en-de-st''' , revision='''a14f04cf0776c02f62a8cb800cf7909e15ea23ad''' , ) @require_sentencepiece class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = """valhalla/s2t_mustc_multilinguial_medium""" lowerCAmelCase__ = """C'est trop cool""" lowerCAmelCase__ = """Esto es genial""" @classmethod def lowerCamelCase ( cls ): '''simple docstring''' __lowerCamelCase = SpeechaTextTokenizer.from_pretrained(cls.checkpoint_name ) return cls def lowerCamelCase ( self ): '''simple docstring''' self.assertEqual(self.tokenizer.lang_code_to_id['''pt'''] , 4 ) self.assertEqual(self.tokenizer.lang_code_to_id['''ru'''] , 6 ) self.assertEqual(self.tokenizer.lang_code_to_id['''it'''] , 9 ) self.assertEqual(self.tokenizer.lang_code_to_id['''de'''] , 11 ) def lowerCamelCase ( self ): '''simple docstring''' self.assertEqual(self.tokenizer.vocab_size , 10000 ) def lowerCamelCase ( self ): '''simple docstring''' self.assertIn(__UpperCAmelCase , self.tokenizer.all_special_ids ) __lowerCamelCase = [ES_CODE, 4, 1601, 47, 7647, 2] __lowerCamelCase = self.tokenizer.decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=__UpperCAmelCase ) self.assertEqual(__UpperCAmelCase , __UpperCAmelCase ) self.assertNotIn(self.tokenizer.eos_token , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = "fr" __lowerCamelCase = self.tokenizer(self.french_text ).input_ids self.assertEqual(encoded[0] , __UpperCAmelCase ) self.assertEqual(encoded[-1] , self.tokenizer.eos_token_id ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = "fr" self.assertListEqual(self.tokenizer.prefix_tokens , [FR_CODE] ) __lowerCamelCase = "es" self.assertListEqual(self.tokenizer.prefix_tokens , [ES_CODE] )
704
import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class __lowerCAmelCase : @staticmethod def lowerCamelCase ( *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' pass def a__ ( _UpperCamelCase : List[str] ): return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. a_ = ( """https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png""" ) @is_pipeline_test @require_torch @require_vision class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model=__UpperCAmelCase , tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '''''' ) ) ) __lowerCamelCase = '''What is the placebo?''' __lowerCamelCase = [ { '''image''': load_image(__UpperCAmelCase ), '''question''': question, }, { '''image''': image, '''question''': question, }, { '''image''': image, '''question''': question, '''word_boxes''': word_boxes, }, ] return dqa_pipeline, examples def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = dqa_pipeline(__UpperCAmelCase , top_k=2 ) self.assertEqual( __UpperCAmelCase , [ [ {'''score''': ANY(__UpperCAmelCase ), '''answer''': ANY(__UpperCAmelCase ), '''start''': ANY(__UpperCAmelCase ), '''end''': ANY(__UpperCAmelCase )}, {'''score''': ANY(__UpperCAmelCase ), '''answer''': ANY(__UpperCAmelCase ), '''start''': ANY(__UpperCAmelCase ), '''end''': ANY(__UpperCAmelCase )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline('''document-question-answering''' , model='''hf-internal-testing/tiny-random-layoutlmv2''' ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''How many cats are there?''' __lowerCamelCase = [ {'''score''': 0.0_001, '''answer''': '''oy 2312/2019''', '''start''': 38, '''end''': 39}, {'''score''': 0.0_001, '''answer''': '''oy 2312/2019 DUE''', '''start''': 38, '''end''': 40}, ] __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , __UpperCAmelCase ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , __UpperCAmelCase ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably __lowerCamelCase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(__UpperCAmelCase , [] ) # We can optionnally pass directly the words and bounding boxes __lowerCamelCase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' __lowerCamelCase = [] __lowerCamelCase = [] __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , words=__UpperCAmelCase , boxes=__UpperCAmelCase , top_k=2 ) self.assertEqual(__UpperCAmelCase , [] ) @slow @require_torch @require_detectrona @require_pytesseract def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_944, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_009, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_944, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_009, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.9_944, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_009, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , max_seq_len=50 , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_974, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_948, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_974, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_948, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.9_974, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_948, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__UpperCAmelCase ) __lowerCamelCase = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__UpperCAmelCase , revision='''3dc6de3''' , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] ] * 2 , ) __lowerCamelCase = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '''''' ) ) ) # This model should also work if `image` is set to None __lowerCamelCase = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__UpperCAmelCase ) __lowerCamelCase = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__UpperCAmelCase , revision='''3dc6de3''' , max_seq_len=50 , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_999, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_998, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.9_999, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_998, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] ] * 2 , ) __lowerCamelCase = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '''''' ) ) ) # This model should also work if `image` is set to None __lowerCamelCase = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_999, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_998, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) @slow @require_torch def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model='''naver-clova-ix/donut-base-finetuned-docvqa''' , tokenizer=AutoTokenizer.from_pretrained('''naver-clova-ix/donut-base-finetuned-docvqa''' ) , feature_extractor='''naver-clova-ix/donut-base-finetuned-docvqa''' , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , [{'''answer''': '''us-001'''}] ) @require_tf @unittest.skip('''Document question answering not implemented in TF''' ) def lowerCamelCase ( self ): '''simple docstring''' pass
622
0
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): lowerCAmelCase__ = ["""image_processor""", """tokenizer"""] lowerCAmelCase__ = """Pix2StructImageProcessor""" lowerCAmelCase__ = ("""T5Tokenizer""", """T5TokenizerFast""") def __init__( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = False super().__init__(_a , _a ) def __call__( self , __UpperCAmelCase=None , __UpperCAmelCase = None , __UpperCAmelCase = True , __UpperCAmelCase = False , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = 2048 , __UpperCAmelCase = 0 , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = True , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' if images is None and text is None: raise ValueError('''You have to specify either images or text.''' ) # Get only text if images is None and not self.image_processor.is_vqa: __lowerCamelCase = self.tokenizer __lowerCamelCase = self.tokenizer( text=_a , add_special_tokens=_a , padding=_a , truncation=_a , max_length=_a , stride=_a , pad_to_multiple_of=_a , return_attention_mask=_a , return_overflowing_tokens=_a , return_special_tokens_mask=_a , return_offsets_mapping=_a , return_token_type_ids=_a , return_length=_a , verbose=_a , return_tensors=_a , **_a , ) return text_encoding if not self.image_processor.is_vqa: # add pixel_values __lowerCamelCase = self.image_processor( _a , return_tensors=_a , max_patches=_a , **_a ) else: # add pixel_values and bbox __lowerCamelCase = self.image_processor( _a , return_tensors=_a , max_patches=_a , header_text=_a , **_a ) if text is not None and not self.image_processor.is_vqa: __lowerCamelCase = self.tokenizer( text=_a , add_special_tokens=_a , padding=_a , truncation=_a , max_length=_a , stride=_a , pad_to_multiple_of=_a , return_attention_mask=_a , return_overflowing_tokens=_a , return_special_tokens_mask=_a , return_offsets_mapping=_a , return_token_type_ids=_a , return_length=_a , verbose=_a , return_tensors=_a , **_a , ) if "attention_mask" in text_encoding: __lowerCamelCase = text_encoding.pop('''attention_mask''' ) if "input_ids" in text_encoding: __lowerCamelCase = text_encoding.pop('''input_ids''' ) else: __lowerCamelCase = None if text_encoding is not None: encoding_image_processor.update(_a ) return encoding_image_processor def lowerCamelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' return self.tokenizer.batch_decode(*_a , **_a ) def lowerCamelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' return self.tokenizer.decode(*_a , **_a ) @property def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.tokenizer.model_input_names __lowerCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
705
import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin a_ = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = XLMProphetNetTokenizer lowerCAmelCase__ = False lowerCAmelCase__ = True def lowerCamelCase ( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __lowerCamelCase = XLMProphetNetTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''[PAD]''' __lowerCamelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCAmelCase ) , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''[PAD]''' ) self.assertEqual(vocab_keys[1] , '''[CLS]''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(__UpperCAmelCase ) , 1012 ) def lowerCamelCase ( self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1012 ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = XLMProphetNetTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) __lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__UpperCAmelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __lowerCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] , ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''[UNK]''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''[UNK]''', '''.''', ] , ) @cached_property def lowerCamelCase ( self ): '''simple docstring''' return XLMProphetNetTokenizer.from_pretrained('''microsoft/xprophetnet-large-wiki100-cased''' ) @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''Hello World!''' __lowerCamelCase = [35389, 6672, 49, 2] self.assertListEqual(__UpperCAmelCase , self.big_tokenizer.encode(__UpperCAmelCase ) ) @slow def lowerCamelCase ( self ): '''simple docstring''' # fmt: off __lowerCamelCase = {'''input_ids''': [[11073, 82783, 18, 26, 82783, 549, 51540, 248, 17209, 1301, 217, 20, 215186, 1325, 147, 17209, 1301, 217, 20, 56370, 53, 122020, 20, 16477, 27, 87355, 4548, 20, 4728, 78392, 17, 159969, 18, 26, 24491, 629, 15, 538, 22704, 5439, 15, 2788, 24491, 9885, 15, 43534, 605, 15, 814, 18403, 33200, 29, 15, 43534, 24458, 12410, 111, 24966, 83669, 9637, 144068, 26, 850, 22346, 27, 147, 24966, 83669, 83490, 26, 39113, 735, 27, 689, 656, 2800, 1339, 4600, 53, 122020, 115785, 34, 816, 1339, 46887, 18, 147, 53905, 1951, 42238, 41170, 17732, 834, 436, 15, 27523, 98733, 217, 147, 5542, 4981, 930, 17347, 16, 2], [20091, 629, 94, 82786, 58, 490, 20, 1528, 84, 53905, 344, 80592, 110128, 18822, 5267, 1306, 62, 152537, 308, 7997, 401, 124427, 549, 35442, 225, 109, 15055, 25748, 147, 7119, 43712, 34, 767, 135366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 63784, 119466, 17, 147808, 88214, 18, 656, 81, 32, 3296, 10280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__UpperCAmelCase , model_name='''microsoft/xprophetnet-large-wiki100-cased''' , revision='''1acad1643ddd54a44df6a1b797ada8373685d90e''' , )
622
0
from datasets.utils.patching import _PatchedModuleObj, patch_submodule from . import _test_patching def a__ ( ): import os as original_os from os import path as original_path from os import rename as original_rename from os.path import dirname as original_dirname from os.path import join as original_join assert _test_patching.os is original_os assert _test_patching.path is original_path assert _test_patching.join is original_join assert _test_patching.renamed_os is original_os assert _test_patching.renamed_path is original_path assert _test_patching.renamed_join is original_join __lowerCamelCase = """__test_patch_submodule_mock__""" with patch_submodule(_test_patching ,'''os.path.join''' ,snake_case__ ): # Every way to access os.path.join must be patched, and the rest must stay untouched # check os.path.join assert isinstance(_test_patching.os ,_PatchedModuleObj ) assert isinstance(_test_patching.os.path ,_PatchedModuleObj ) assert _test_patching.os.path.join is mock # check path.join assert isinstance(_test_patching.path ,_PatchedModuleObj ) assert _test_patching.path.join is mock # check join assert _test_patching.join is mock # check that the other attributes are untouched assert _test_patching.os.rename is original_rename assert _test_patching.path.dirname is original_dirname assert _test_patching.os.path.dirname is original_dirname # Even renamed modules or objects must be patched # check renamed_os.path.join assert isinstance(_test_patching.renamed_os ,_PatchedModuleObj ) assert isinstance(_test_patching.renamed_os.path ,_PatchedModuleObj ) assert _test_patching.renamed_os.path.join is mock # check renamed_path.join assert isinstance(_test_patching.renamed_path ,_PatchedModuleObj ) assert _test_patching.renamed_path.join is mock # check renamed_join assert _test_patching.renamed_join is mock # check that the other attributes are untouched assert _test_patching.renamed_os.rename is original_rename assert _test_patching.renamed_path.dirname is original_dirname assert _test_patching.renamed_os.path.dirname is original_dirname # check that everthing is back to normal when the patch is over assert _test_patching.os is original_os assert _test_patching.path is original_path assert _test_patching.join is original_join assert _test_patching.renamed_os is original_os assert _test_patching.renamed_path is original_path assert _test_patching.renamed_join is original_join def a__ ( ): assert _test_patching.open is open __lowerCamelCase = """__test_patch_submodule_builtin_mock__""" # _test_patching has "open" in its globals assert _test_patching.open is open with patch_submodule(_test_patching ,'''open''' ,snake_case__ ): assert _test_patching.open is mock # check that everthing is back to normal when the patch is over assert _test_patching.open is open def a__ ( ): __lowerCamelCase = """__test_patch_submodule_missing_mock__""" with patch_submodule(_test_patching ,'''pandas.read_csv''' ,snake_case__ ): pass def a__ ( ): __lowerCamelCase = """__test_patch_submodule_missing_builtin_mock__""" # _test_patching doesn't have "len" in its globals assert getattr(_test_patching ,'''len''' ,snake_case__ ) is None with patch_submodule(_test_patching ,'''len''' ,snake_case__ ): assert _test_patching.len is mock assert _test_patching.len is len def a__ ( ): __lowerCamelCase = """__test_patch_submodule_start_and_stop_mock__""" __lowerCamelCase = patch_submodule(_test_patching ,'''open''' ,snake_case__ ) assert _test_patching.open is open patch.start() assert _test_patching.open is mock patch.stop() assert _test_patching.open is open def a__ ( ): from os import rename as original_rename from os.path import dirname as original_dirname from os.path import join as original_join __lowerCamelCase = """__test_patch_submodule_successive_join__""" __lowerCamelCase = """__test_patch_submodule_successive_dirname__""" __lowerCamelCase = """__test_patch_submodule_successive_rename__""" assert _test_patching.os.path.join is original_join assert _test_patching.os.path.dirname is original_dirname assert _test_patching.os.rename is original_rename with patch_submodule(_test_patching ,'''os.path.join''' ,snake_case__ ): with patch_submodule(_test_patching ,'''os.rename''' ,snake_case__ ): with patch_submodule(_test_patching ,'''os.path.dirname''' ,snake_case__ ): assert _test_patching.os.path.join is mock_join assert _test_patching.os.path.dirname is mock_dirname assert _test_patching.os.rename is mock_rename # try another order with patch_submodule(_test_patching ,'''os.rename''' ,snake_case__ ): with patch_submodule(_test_patching ,'''os.path.join''' ,snake_case__ ): with patch_submodule(_test_patching ,'''os.path.dirname''' ,snake_case__ ): assert _test_patching.os.path.join is mock_join assert _test_patching.os.path.dirname is mock_dirname assert _test_patching.os.rename is mock_rename assert _test_patching.os.path.join is original_join assert _test_patching.os.path.dirname is original_dirname assert _test_patching.os.rename is original_rename def a__ ( ): __lowerCamelCase = """__test_patch_submodule_doesnt_exist_mock__""" with patch_submodule(_test_patching ,'''__module_that_doesn_exist__.__attribute_that_doesn_exist__''' ,snake_case__ ): pass with patch_submodule(_test_patching ,'''os.__attribute_that_doesn_exist__''' ,snake_case__ ): pass
706
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py a_ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. a_ = direct_transformers_import(PATH_TO_TRANSFORMERS) a_ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` a_ = re.compile(R"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") a_ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def a__ ( _UpperCamelCase : Union[str, Any] ): __lowerCamelCase = None # source code of `config_class` __lowerCamelCase = inspect.getsource(_UpperCamelCase ) __lowerCamelCase = _re_checkpoint.findall(_UpperCamelCase ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('''/''' ): __lowerCamelCase = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link __lowerCamelCase = F"""https://huggingface.co/{ckpt_name}""" if ckpt_link == ckpt_link_from_name: __lowerCamelCase = ckpt_name break return checkpoint def a__ ( ): __lowerCamelCase = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue __lowerCamelCase = get_checkpoint_from_config_class(_UpperCamelCase ) __lowerCamelCase = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(_UpperCamelCase ) if len(_UpperCamelCase ) > 0: __lowerCamelCase = '''\n'''.join(sorted(_UpperCamelCase ) ) raise ValueError(F"""The following configurations don't contain any valid checkpoint:\n{message}""" ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
622
0
import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''ZinengTang/tvlt-base''' __lowerCamelCase = tempfile.mkdtemp() def lowerCamelCase ( self , **__UpperCAmelCase ): '''simple docstring''' return TvltImageProcessor.from_pretrained(self.checkpoint , **__lowerCAmelCase ) def lowerCamelCase ( self , **__UpperCAmelCase ): '''simple docstring''' return TvltFeatureExtractor.from_pretrained(self.checkpoint , **__lowerCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_image_processor() __lowerCamelCase = self.get_feature_extractor() __lowerCamelCase = TvltProcessor(image_processor=__lowerCAmelCase , feature_extractor=__lowerCAmelCase ) processor.save_pretrained(self.tmpdirname ) __lowerCamelCase = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , __lowerCAmelCase ) self.assertIsInstance(processor.image_processor , __lowerCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_image_processor() __lowerCamelCase = self.get_feature_extractor() __lowerCamelCase = TvltProcessor(image_processor=__lowerCAmelCase , feature_extractor=__lowerCAmelCase ) __lowerCamelCase = np.ones([12000] ) __lowerCamelCase = feature_extractor(__lowerCAmelCase , return_tensors='''np''' ) __lowerCamelCase = processor(audio=__lowerCAmelCase , return_tensors='''np''' ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1E-2 ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_image_processor() __lowerCamelCase = self.get_feature_extractor() __lowerCamelCase = TvltProcessor(image_processor=__lowerCAmelCase , feature_extractor=__lowerCAmelCase ) __lowerCamelCase = np.ones([3, 224, 224] ) __lowerCamelCase = image_processor(__lowerCAmelCase , return_tensors='''np''' ) __lowerCamelCase = processor(images=__lowerCAmelCase , return_tensors='''np''' ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1E-2 ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_image_processor() __lowerCamelCase = self.get_feature_extractor() __lowerCamelCase = TvltProcessor(image_processor=__lowerCAmelCase , feature_extractor=__lowerCAmelCase ) __lowerCamelCase = np.ones([12000] ) __lowerCamelCase = np.ones([3, 224, 224] ) __lowerCamelCase = processor(audio=__lowerCAmelCase , images=__lowerCAmelCase ) self.assertListEqual(list(inputs.keys() ) , ['''audio_values''', '''audio_mask''', '''pixel_values''', '''pixel_mask'''] ) # test if it raises when no input is passed with pytest.raises(__lowerCAmelCase ): processor() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_image_processor() __lowerCamelCase = self.get_feature_extractor() __lowerCamelCase = TvltProcessor(image_processor=__lowerCAmelCase , feature_extractor=__lowerCAmelCase ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='''`processor` and `image_processor`+`feature_extractor` model input names do not match''' , )
707
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a_ = { """configuration_clipseg""": [ """CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP""", """CLIPSegConfig""", """CLIPSegTextConfig""", """CLIPSegVisionConfig""", ], """processing_clipseg""": ["""CLIPSegProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ """CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST""", """CLIPSegModel""", """CLIPSegPreTrainedModel""", """CLIPSegTextModel""", """CLIPSegVisionModel""", """CLIPSegForImageSegmentation""", ] if TYPE_CHECKING: from .configuration_clipseg import ( CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPSegConfig, CLIPSegTextConfig, CLIPSegVisionConfig, ) from .processing_clipseg import CLIPSegProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clipseg import ( CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPSegForImageSegmentation, CLIPSegModel, CLIPSegPreTrainedModel, CLIPSegTextModel, CLIPSegVisionModel, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
622
0
import argparse from pathlib import Path import torch from packaging import version from torch.onnx import export from diffusers import AutoencoderKL a_ = version.parse(version.parse(torch.__version__).base_version) < version.parse("""1.11""") def a__ ( _UpperCamelCase : List[str] ,_UpperCamelCase : tuple ,_UpperCamelCase : Path ,_UpperCamelCase : Tuple ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : Optional[Any] ,_UpperCamelCase : List[str]=False ,): output_path.parent.mkdir(parents=UpperCAmelCase__ ,exist_ok=UpperCAmelCase__ ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( UpperCAmelCase__ ,UpperCAmelCase__ ,f=output_path.as_posix() ,input_names=UpperCAmelCase__ ,output_names=UpperCAmelCase__ ,dynamic_axes=UpperCAmelCase__ ,do_constant_folding=UpperCAmelCase__ ,use_external_data_format=UpperCAmelCase__ ,enable_onnx_checker=UpperCAmelCase__ ,opset_version=UpperCAmelCase__ ,) else: export( UpperCAmelCase__ ,UpperCAmelCase__ ,f=output_path.as_posix() ,input_names=UpperCAmelCase__ ,output_names=UpperCAmelCase__ ,dynamic_axes=UpperCAmelCase__ ,do_constant_folding=UpperCAmelCase__ ,opset_version=UpperCAmelCase__ ,) @torch.no_grad() def a__ ( _UpperCamelCase : str ,_UpperCamelCase : str ,_UpperCamelCase : int ,_UpperCamelCase : bool = False ): __lowerCamelCase = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): __lowerCamelCase = """cuda""" elif fpaa and not torch.cuda.is_available(): raise ValueError('''`float16` model export is only supported on GPUs with CUDA''' ) else: __lowerCamelCase = """cpu""" __lowerCamelCase = Path(UpperCAmelCase__ ) # VAE DECODER __lowerCamelCase = AutoencoderKL.from_pretrained(model_path + '''/vae''' ) __lowerCamelCase = vae_decoder.config.latent_channels # forward only through the decoder part __lowerCamelCase = vae_decoder.decode onnx_export( UpperCAmelCase__ ,model_args=( torch.randn(1 ,UpperCAmelCase__ ,25 ,25 ).to(device=UpperCAmelCase__ ,dtype=UpperCAmelCase__ ), False, ) ,output_path=output_path / '''vae_decoder''' / '''model.onnx''' ,ordered_input_names=['''latent_sample''', '''return_dict'''] ,output_names=['''sample'''] ,dynamic_axes={ '''latent_sample''': {0: '''batch''', 1: '''channels''', 2: '''height''', 3: '''width'''}, } ,opset=UpperCAmelCase__ ,) del vae_decoder if __name__ == "__main__": a_ = argparse.ArgumentParser() parser.add_argument( """--model_path""", type=str, required=True, help="""Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).""", ) parser.add_argument("""--output_path""", type=str, required=True, help="""Path to the output model.""") parser.add_argument( """--opset""", default=14, type=int, help="""The version of the ONNX operator set to use.""", ) parser.add_argument("""--fp16""", action="""store_true""", default=False, help="""Export the models in `float16` mode""") a_ = parser.parse_args() print(args.output_path) convert_models(args.model_path, args.output_path, args.opset, args.fpaa) print("""SD: Done: ONNX""")
708
import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = RoFormerTokenizer lowerCAmelCase__ = RoFormerTokenizerFast lowerCAmelCase__ = True lowerCAmelCase__ = True def lowerCamelCase ( self ): '''simple docstring''' super().setUp() def lowerCamelCase ( self , **__UpperCAmelCase ): '''simple docstring''' return self.tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **__UpperCAmelCase ) def lowerCamelCase ( self , **__UpperCAmelCase ): '''simple docstring''' return self.rust_tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''永和服装饰品有限公司,今天天气非常好''' __lowerCamelCase = '''永和 服装 饰品 有限公司 , 今 天 天 气 非常 好''' return input_text, output_text def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_tokenizer() __lowerCamelCase ,__lowerCamelCase = self.get_chinese_input_output_texts() __lowerCamelCase = tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , output_text.split() ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [22943, 21332, 34431, 45904, 117, 306, 1231, 1231, 2653, 33994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_rust_tokenizer() __lowerCamelCase ,__lowerCamelCase = self.get_chinese_input_output_texts() __lowerCamelCase = tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , output_text.split() ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [22943, 21332, 34431, 45904, 117, 306, 1231, 1231, 2653, 33994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' pass
622
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) a_ = { '''configuration_longformer''': [ '''LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LongformerConfig''', '''LongformerOnnxConfig''', ], '''tokenization_longformer''': ['''LongformerTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ['''LongformerTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ '''LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LongformerForMaskedLM''', '''LongformerForMultipleChoice''', '''LongformerForQuestionAnswering''', '''LongformerForSequenceClassification''', '''LongformerForTokenClassification''', '''LongformerModel''', '''LongformerPreTrainedModel''', '''LongformerSelfAttention''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ '''TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFLongformerForMaskedLM''', '''TFLongformerForMultipleChoice''', '''TFLongformerForQuestionAnswering''', '''TFLongformerForSequenceClassification''', '''TFLongformerForTokenClassification''', '''TFLongformerModel''', '''TFLongformerPreTrainedModel''', '''TFLongformerSelfAttention''', ] if TYPE_CHECKING: from .configuration_longformer import ( LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, LongformerConfig, LongformerOnnxConfig, ) from .tokenization_longformer import LongformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_longformer_fast import LongformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longformer import ( LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, LongformerForMaskedLM, LongformerForMultipleChoice, LongformerForQuestionAnswering, LongformerForSequenceClassification, LongformerForTokenClassification, LongformerModel, LongformerPreTrainedModel, LongformerSelfAttention, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_longformer import ( TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFLongformerForMaskedLM, TFLongformerForMultipleChoice, TFLongformerForQuestionAnswering, TFLongformerForSequenceClassification, TFLongformerForTokenClassification, TFLongformerModel, TFLongformerPreTrainedModel, TFLongformerSelfAttention, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
709
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device a_ = False class __lowerCAmelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained('''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt='''first prompt''' , image=__UpperCAmelCase , text_to_image_strength=0.75 , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained(__UpperCAmelCase , torch_dtype=torch.floataa ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = generator.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt='''first prompt''' , image=__UpperCAmelCase , text_to_image_strength=0.75 , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained('''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = '''cyberpunk 2077''' __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt=__UpperCAmelCase , image=__UpperCAmelCase , text_to_image_strength=0.75 , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' , ).images __lowerCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __lowerCamelCase = '''A painting of a squirrel eating a burger ''' __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.text_to_image( prompt=__UpperCAmelCase , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' ).images __lowerCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __lowerCamelCase = pipe.image_variation(__UpperCAmelCase , generator=__UpperCAmelCase , output_type='''numpy''' ).images __lowerCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
622
0
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger a_ = get_logger(__name__) class __lowerCAmelCase : def __init__( self , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = ( os.path.join(__lowerCAmelCase , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) __lowerCamelCase = Extractor def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" __lowerCamelCase = os.path.abspath(__lowerCAmelCase ) return os.path.join(self.extract_dir , hash_url_to_filename(__lowerCAmelCase ) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' return force_extract or ( not os.path.isfile(__lowerCAmelCase ) and not (os.path.isdir(__lowerCAmelCase ) and os.listdir(__lowerCAmelCase )) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = False ): '''simple docstring''' __lowerCamelCase = self.extractor.infer_extractor_format(__lowerCAmelCase ) if not extractor_format: return input_path __lowerCamelCase = self._get_output_path(__lowerCAmelCase ) if self._do_extract(__lowerCAmelCase , __lowerCAmelCase ): self.extractor.extract(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) return output_path class __lowerCAmelCase ( UpperCAmelCase__ ): @classmethod @abstractmethod def lowerCamelCase ( cls , __UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' ... @staticmethod @abstractmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' ... class __lowerCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ ): lowerCAmelCase__ = [] @staticmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' with open(__lowerCAmelCase , '''rb''' ) as f: return f.read(__lowerCAmelCase ) @classmethod def lowerCamelCase ( cls , __UpperCAmelCase , __UpperCAmelCase = b"" ): '''simple docstring''' if not magic_number: __lowerCamelCase = max(len(__lowerCAmelCase ) for cls_magic_number in cls.magic_numbers ) try: __lowerCamelCase = cls.read_magic_number(__lowerCAmelCase , __lowerCAmelCase ) except OSError: return False return any(magic_number.startswith(__lowerCAmelCase ) for cls_magic_number in cls.magic_numbers ) class __lowerCAmelCase ( UpperCAmelCase__ ): @classmethod def lowerCamelCase ( cls , __UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' return tarfile.is_tarfile(__lowerCAmelCase ) @staticmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' def resolved(__UpperCAmelCase ) -> str: return os.path.realpath(os.path.abspath(__lowerCAmelCase ) ) def badpath(__UpperCAmelCase , __UpperCAmelCase ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(__lowerCAmelCase , __lowerCAmelCase ) ).startswith(__lowerCAmelCase ) def badlink(__UpperCAmelCase , __UpperCAmelCase ) -> bool: # Links are interpreted relative to the directory containing the link __lowerCamelCase = resolved(os.path.join(__lowerCAmelCase , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=__lowerCAmelCase ) __lowerCamelCase = resolved(__lowerCAmelCase ) for finfo in members: if badpath(finfo.name , __lowerCAmelCase ): logger.error(F"""Extraction of {finfo.name} is blocked (illegal path)""" ) elif finfo.issym() and badlink(__lowerCAmelCase , __lowerCAmelCase ): logger.error(F"""Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}""" ) elif finfo.islnk() and badlink(__lowerCAmelCase , __lowerCAmelCase ): logger.error(F"""Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}""" ) else: yield finfo @staticmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) __lowerCamelCase = tarfile.open(__lowerCAmelCase ) tar_file.extractall(__lowerCAmelCase , members=TarExtractor.safemembers(__lowerCAmelCase , __lowerCAmelCase ) ) tar_file.close() class __lowerCAmelCase ( UpperCAmelCase__ ): lowerCAmelCase__ = [b"""\x1F\x8B"""] @staticmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' with gzip.open(__lowerCAmelCase , '''rb''' ) as gzip_file: with open(__lowerCAmelCase , '''wb''' ) as extracted_file: shutil.copyfileobj(__lowerCAmelCase , __lowerCAmelCase ) class __lowerCAmelCase ( UpperCAmelCase__ ): lowerCAmelCase__ = [ b"""PK\x03\x04""", b"""PK\x05\x06""", # empty archive b"""PK\x07\x08""", # spanned archive ] @classmethod def lowerCamelCase ( cls , __UpperCAmelCase , __UpperCAmelCase = b"" ): '''simple docstring''' if super().is_extractable(__lowerCAmelCase , magic_number=__lowerCAmelCase ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(__lowerCAmelCase , '''rb''' ) as fp: __lowerCamelCase = _EndRecData(__lowerCAmelCase ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: __lowerCamelCase = fp.read(__lowerCAmelCase ) # CD is where we expect it to be if len(__lowerCAmelCase ) == sizeCentralDir: __lowerCamelCase = struct.unpack(__lowerCAmelCase , __lowerCAmelCase ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) with zipfile.ZipFile(__lowerCAmelCase , '''r''' ) as zip_file: zip_file.extractall(__lowerCAmelCase ) zip_file.close() class __lowerCAmelCase ( UpperCAmelCase__ ): lowerCAmelCase__ = [b"""\xFD\x37\x7A\x58\x5A\x00"""] @staticmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' with lzma.open(__lowerCAmelCase ) as compressed_file: with open(__lowerCAmelCase , '''wb''' ) as extracted_file: shutil.copyfileobj(__lowerCAmelCase , __lowerCAmelCase ) class __lowerCAmelCase ( UpperCAmelCase__ ): lowerCAmelCase__ = [b"""Rar!\x1a\x07\x00""", b"""Rar!\x1a\x07\x01\x00"""] # RAR_ID # RAR5_ID @staticmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' if not config.RARFILE_AVAILABLE: raise ImportError('''Please pip install rarfile''' ) import rarfile os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) __lowerCamelCase = rarfile.RarFile(__lowerCAmelCase ) rf.extractall(__lowerCAmelCase ) rf.close() class __lowerCAmelCase ( UpperCAmelCase__ ): lowerCAmelCase__ = [b"""\x28\xb5\x2F\xFD"""] @staticmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' if not config.ZSTANDARD_AVAILABLE: raise ImportError('''Please pip install zstandard''' ) import zstandard as zstd __lowerCamelCase = zstd.ZstdDecompressor() with open(__lowerCAmelCase , '''rb''' ) as ifh, open(__lowerCAmelCase , '''wb''' ) as ofh: dctx.copy_stream(__lowerCAmelCase , __lowerCAmelCase ) class __lowerCAmelCase ( UpperCAmelCase__ ): lowerCAmelCase__ = [b"""\x42\x5A\x68"""] @staticmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' with bza.open(__lowerCAmelCase , '''rb''' ) as compressed_file: with open(__lowerCAmelCase , '''wb''' ) as extracted_file: shutil.copyfileobj(__lowerCAmelCase , __lowerCAmelCase ) class __lowerCAmelCase ( UpperCAmelCase__ ): lowerCAmelCase__ = [b"""\x37\x7A\xBC\xAF\x27\x1C"""] @staticmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' if not config.PY7ZR_AVAILABLE: raise ImportError('''Please pip install py7zr''' ) import pyazr os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) with pyazr.SevenZipFile(__lowerCAmelCase , '''r''' ) as archive: archive.extractall(__lowerCAmelCase ) class __lowerCAmelCase ( UpperCAmelCase__ ): lowerCAmelCase__ = [b"""\x04\x22\x4D\x18"""] @staticmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' if not config.LZ4_AVAILABLE: raise ImportError('''Please pip install lz4''' ) import lza.frame with lza.frame.open(__lowerCAmelCase , '''rb''' ) as compressed_file: with open(__lowerCAmelCase , '''wb''' ) as extracted_file: shutil.copyfileobj(__lowerCAmelCase , __lowerCAmelCase ) class __lowerCAmelCase : # Put zip file to the last, b/c it is possible wrongly detected as zip (I guess it means: as tar or gzip) lowerCAmelCase__ = { """tar""": TarExtractor, """gzip""": GzipExtractor, """zip""": ZipExtractor, """xz""": XzExtractor, """rar""": RarExtractor, """zstd""": ZstdExtractor, """bz2""": BzipaExtractor, """7z""": SevenZipExtractor, # <Added version="2.4.0"/> """lz4""": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def lowerCamelCase ( cls ): '''simple docstring''' return max( len(__lowerCAmelCase ) for extractor in cls.extractors.values() if issubclass(__lowerCAmelCase , __lowerCAmelCase ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' try: return MagicNumberBaseExtractor.read_magic_number(__lowerCAmelCase , magic_number_length=__lowerCAmelCase ) except OSError: return b"" @classmethod def lowerCamelCase ( cls , __UpperCAmelCase , __UpperCAmelCase = False ): '''simple docstring''' warnings.warn( '''Method \'is_extractable\' was deprecated in version 2.4.0 and will be removed in 3.0.0. ''' '''Use \'infer_extractor_format\' instead.''' , category=__lowerCAmelCase , ) __lowerCamelCase = cls.infer_extractor_format(__lowerCAmelCase ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def lowerCamelCase ( cls , __UpperCAmelCase ): # <Added version="2.4.0"/> '''simple docstring''' __lowerCamelCase = cls._get_magic_number_max_length() __lowerCamelCase = cls._read_magic_number(__lowerCAmelCase , __lowerCAmelCase ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(__lowerCAmelCase , magic_number=__lowerCAmelCase ): return extractor_format @classmethod def lowerCamelCase ( cls , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = "deprecated" , ): '''simple docstring''' os.makedirs(os.path.dirname(__lowerCAmelCase ) , exist_ok=__lowerCAmelCase ) # Prevent parallel extractions __lowerCamelCase = str(Path(__lowerCAmelCase ).with_suffix('''.lock''' ) ) with FileLock(__lowerCAmelCase ): shutil.rmtree(__lowerCAmelCase , ignore_errors=__lowerCAmelCase ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(__lowerCAmelCase , __lowerCAmelCase ): # passed as positional arg warnings.warn( '''Parameter \'extractor\' was deprecated in version 2.4.0 and will be removed in 3.0.0. ''' '''Use \'extractor_format\' instead.''' , category=__lowerCAmelCase , ) __lowerCamelCase = extractor if extractor != '''deprecated''' else extractor_format else: __lowerCamelCase = cls.extractors[extractor_format] return extractor.extract(__lowerCAmelCase , __lowerCAmelCase ) else: warnings.warn( '''Parameter \'extractor_format\' was made required in version 2.4.0 and not passing it will raise an ''' '''exception in 3.0.0.''' , category=__lowerCAmelCase , ) for extractor in cls.extractors.values(): if extractor.is_extractable(__lowerCAmelCase ): return extractor.extract(__lowerCAmelCase , __lowerCAmelCase )
710
import argparse import datetime import json import time import warnings from logging import getLogger from pathlib import Path from typing import Dict, List import torch from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import calculate_bleu, calculate_rouge, chunks, parse_numeric_n_bool_cl_kwargs, use_task_specific_params a_ = getLogger(__name__) a_ = """cuda""" if torch.cuda.is_available() else """cpu""" def a__ ( _UpperCamelCase : List[str] ,_UpperCamelCase : str ,_UpperCamelCase : str ,_UpperCamelCase : int = 8 ,_UpperCamelCase : str = DEFAULT_DEVICE ,_UpperCamelCase : Dict=False ,_UpperCamelCase : Dict="summarization" ,_UpperCamelCase : Optional[int]=None ,**_UpperCamelCase : Dict ,): __lowerCamelCase = Path(_UpperCamelCase ).open('''w''' ,encoding='''utf-8''' ) __lowerCamelCase = str(_UpperCamelCase ) __lowerCamelCase = AutoModelForSeqaSeqLM.from_pretrained(_UpperCamelCase ).to(_UpperCamelCase ) if fpaa: __lowerCamelCase = model.half() __lowerCamelCase = AutoTokenizer.from_pretrained(_UpperCamelCase ) logger.info(F"""Inferred tokenizer type: {tokenizer.__class__}""" ) # if this is wrong, check config.model_type. __lowerCamelCase = time.time() # update config with task specific params use_task_specific_params(_UpperCamelCase ,_UpperCamelCase ) if prefix is None: __lowerCamelCase = prefix or getattr(model.config ,'''prefix''' ,'''''' ) or '''''' for examples_chunk in tqdm(list(chunks(_UpperCamelCase ,_UpperCamelCase ) ) ): __lowerCamelCase = [prefix + text for text in examples_chunk] __lowerCamelCase = tokenizer(_UpperCamelCase ,return_tensors='''pt''' ,truncation=_UpperCamelCase ,padding='''longest''' ).to(_UpperCamelCase ) __lowerCamelCase = model.generate( input_ids=batch.input_ids ,attention_mask=batch.attention_mask ,**_UpperCamelCase ,) __lowerCamelCase = tokenizer.batch_decode(_UpperCamelCase ,skip_special_tokens=_UpperCamelCase ,clean_up_tokenization_spaces=_UpperCamelCase ) for hypothesis in dec: fout.write(hypothesis + '''\n''' ) fout.flush() fout.close() __lowerCamelCase = int(time.time() - start_time ) # seconds __lowerCamelCase = len(_UpperCamelCase ) return {"n_obs": n_obs, "runtime": runtime, "seconds_per_sample": round(runtime / n_obs ,4 )} def a__ ( ): return datetime.datetime.now().strftime('''%Y-%m-%d %H:%M:%S''' ) def a__ ( _UpperCamelCase : Union[str, Any]=True ): __lowerCamelCase = argparse.ArgumentParser() parser.add_argument('''model_name''' ,type=_UpperCamelCase ,help='''like facebook/bart-large-cnn,t5-base, etc.''' ) parser.add_argument('''input_path''' ,type=_UpperCamelCase ,help='''like cnn_dm/test.source''' ) parser.add_argument('''save_path''' ,type=_UpperCamelCase ,help='''where to save summaries''' ) parser.add_argument('''--reference_path''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,help='''like cnn_dm/test.target''' ) parser.add_argument('''--score_path''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,default='''metrics.json''' ,help='''where to save metrics''' ) parser.add_argument('''--device''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,default=_UpperCamelCase ,help='''cuda, cuda:1, cpu etc.''' ) parser.add_argument( '''--prefix''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,default=_UpperCamelCase ,help='''will be added to the begininng of src examples''' ) parser.add_argument('''--task''' ,type=_UpperCamelCase ,default='''summarization''' ,help='''used for task_specific_params + metrics''' ) parser.add_argument('''--bs''' ,type=_UpperCamelCase ,default=8 ,required=_UpperCamelCase ,help='''batch size''' ) parser.add_argument( '''--n_obs''' ,type=_UpperCamelCase ,default=-1 ,required=_UpperCamelCase ,help='''How many observations. Defaults to all.''' ) parser.add_argument('''--fp16''' ,action='''store_true''' ) parser.add_argument('''--dump-args''' ,action='''store_true''' ,help='''print the custom hparams with the results''' ) parser.add_argument( '''--info''' ,nargs='''?''' ,type=_UpperCamelCase ,const=datetime_now() ,help=( '''use in conjunction w/ --dump-args to print with the results whatever other info you\'d like, e.g.''' ''' lang=en-ru. If no value is passed, the current datetime string will be used.''' ) ,) # Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate __lowerCamelCase ,__lowerCamelCase = parser.parse_known_args() __lowerCamelCase = parse_numeric_n_bool_cl_kwargs(_UpperCamelCase ) if parsed_args and verbose: print(F"""parsed the following generate kwargs: {parsed_args}""" ) __lowerCamelCase = [''' ''' + x.rstrip() if '''t5''' in args.model_name else x.rstrip() for x in open(args.input_path ).readlines()] if args.n_obs > 0: __lowerCamelCase = examples[: args.n_obs] Path(args.save_path ).parent.mkdir(exist_ok=_UpperCamelCase ) if args.reference_path is None and Path(args.score_path ).exists(): warnings.warn(F"""score_path {args.score_path} will be overwritten unless you type ctrl-c.""" ) if args.device == "cpu" and args.fpaa: # this mix leads to RuntimeError: "threshold_cpu" not implemented for 'Half' raise ValueError('''Can\'t mix --fp16 and --device cpu''' ) __lowerCamelCase = generate_summaries_or_translations( _UpperCamelCase ,args.save_path ,args.model_name ,batch_size=args.bs ,device=args.device ,fpaa=args.fpaa ,task=args.task ,prefix=args.prefix ,**_UpperCamelCase ,) if args.reference_path is None: return {} # Compute scores __lowerCamelCase = calculate_bleu if '''translation''' in args.task else calculate_rouge __lowerCamelCase = [x.rstrip() for x in open(args.save_path ).readlines()] __lowerCamelCase = [x.rstrip() for x in open(args.reference_path ).readlines()][: len(_UpperCamelCase )] __lowerCamelCase = score_fn(_UpperCamelCase ,_UpperCamelCase ) scores.update(_UpperCamelCase ) if args.dump_args: scores.update(_UpperCamelCase ) if args.info: __lowerCamelCase = args.info if verbose: print(_UpperCamelCase ) if args.score_path is not None: json.dump(_UpperCamelCase ,open(args.score_path ,'''w''' ) ) return scores if __name__ == "__main__": # Usage for MT: # python run_eval.py MODEL_NAME $DATA_DIR/test.source $save_dir/test_translations.txt --reference_path $DATA_DIR/test.target --score_path $save_dir/test_bleu.json --task translation $@ run_generate(verbose=True)
622
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class __lowerCAmelCase : lowerCAmelCase__ = BlenderbotConfig lowerCAmelCase__ = {} lowerCAmelCase__ = """gelu""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=99 , __UpperCAmelCase=32 , __UpperCAmelCase=2 , __UpperCAmelCase=4 , __UpperCAmelCase=37 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=20 , __UpperCAmelCase=2 , __UpperCAmelCase=1 , __UpperCAmelCase=0 , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = eos_token_id __lowerCamelCase = pad_token_id __lowerCamelCase = bos_token_id def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowerCamelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowerCamelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) __lowerCamelCase = prepare_blenderbot_inputs_dict(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) return config, inputs_dict def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = TFBlenderbotModel(config=UpperCAmelCase_ ).get_decoder() __lowerCamelCase = inputs_dict['''input_ids'''] __lowerCamelCase = input_ids[:1, :] __lowerCamelCase = inputs_dict['''attention_mask'''][:1, :] __lowerCamelCase = inputs_dict['''head_mask'''] __lowerCamelCase = 1 # first forward pass __lowerCamelCase = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , head_mask=UpperCAmelCase_ , use_cache=UpperCAmelCase_ ) __lowerCamelCase ,__lowerCamelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowerCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowerCamelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowerCamelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowerCamelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowerCamelCase = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ )[0] __lowerCamelCase = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , past_key_values=UpperCAmelCase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowerCamelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowerCamelCase = output_from_no_past[:, -3:, random_slice_idx] __lowerCamelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(UpperCAmelCase_ , UpperCAmelCase_ , rtol=1E-3 ) def a__ ( _UpperCamelCase : Optional[Any] ,_UpperCamelCase : str ,_UpperCamelCase : List[Any] ,_UpperCamelCase : Optional[Any]=None ,_UpperCamelCase : List[Any]=None ,_UpperCamelCase : Optional[int]=None ,_UpperCamelCase : Any=None ,_UpperCamelCase : int=None ,): if attention_mask is None: __lowerCamelCase = tf.cast(tf.math.not_equal(_snake_case ,config.pad_token_id ) ,tf.inta ) if decoder_attention_mask is None: __lowerCamelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape ,dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] ,config.pad_token_id ) ,tf.inta ), ] ,axis=-1 ,) if head_mask is None: __lowerCamelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: __lowerCamelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: __lowerCamelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __lowerCAmelCase ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): lowerCAmelCase__ = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () lowerCAmelCase__ = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () lowerCAmelCase__ = ( { """conversational""": TFBlenderbotForConditionalGeneration, """feature-extraction""": TFBlenderbotModel, """summarization""": TFBlenderbotForConditionalGeneration, """text2text-generation""": TFBlenderbotForConditionalGeneration, """translation""": TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) lowerCAmelCase__ = True lowerCAmelCase__ = False lowerCAmelCase__ = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFBlenderbotModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=UpperCAmelCase_ ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*UpperCAmelCase_ ) @require_tokenizers @require_tf class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = ["""My friends are cool but they eat too many carbs."""] lowerCAmelCase__ = """facebook/blenderbot-400M-distill""" @cached_property def lowerCamelCase ( self ): '''simple docstring''' return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.tokenizer(self.src_text , return_tensors='''tf''' ) __lowerCamelCase = self.model.generate( model_inputs.input_ids , ) __lowerCamelCase = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=UpperCAmelCase_ )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
711
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def a__ ( _UpperCamelCase : Optional[int] ,_UpperCamelCase : List[str] ,_UpperCamelCase : List[Any]=None ,_UpperCamelCase : Any=None ): if attention_mask is None: __lowerCamelCase = tf.cast(tf.math.not_equal(_UpperCamelCase ,config.pad_token_id ) ,tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class __lowerCAmelCase : lowerCAmelCase__ = OPTConfig lowerCAmelCase__ = {} lowerCAmelCase__ = """gelu""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=99 , __UpperCAmelCase=16 , __UpperCAmelCase=2 , __UpperCAmelCase=4 , __UpperCAmelCase=4 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=20 , __UpperCAmelCase=2 , __UpperCAmelCase=1 , __UpperCAmelCase=0 , __UpperCAmelCase=16 , __UpperCAmelCase=16 , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = eos_token_id __lowerCamelCase = pad_token_id __lowerCamelCase = bos_token_id __lowerCamelCase = embed_dim __lowerCamelCase = word_embed_proj_dim __lowerCamelCase = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowerCamelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowerCamelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowerCamelCase = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=__UpperCAmelCase , **self.config_updates , ) __lowerCamelCase = prepare_opt_inputs_dict(__UpperCAmelCase , __UpperCAmelCase ) return config, inputs_dict def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = TFOPTModel(config=__UpperCAmelCase ) __lowerCamelCase = inputs_dict['''input_ids'''] __lowerCamelCase = input_ids[:1, :] __lowerCamelCase = inputs_dict['''attention_mask'''][:1, :] __lowerCamelCase = 1 # first forward pass __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , use_cache=__UpperCAmelCase ) __lowerCamelCase ,__lowerCamelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowerCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowerCamelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowerCamelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowerCamelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase )[0] __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , past_key_values=__UpperCAmelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowerCamelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowerCamelCase = output_from_no_past[:, -3:, random_slice_idx] __lowerCamelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__UpperCAmelCase , __UpperCAmelCase , rtol=1E-3 ) @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () lowerCAmelCase__ = (TFOPTForCausalLM,) if is_tf_available() else () lowerCAmelCase__ = ( {"""feature-extraction""": TFOPTModel, """text-generation""": TFOPTForCausalLM} if is_tf_available() else {} ) lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = 1_0 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(__UpperCAmelCase , __UpperCAmelCase ): if hasattr(__UpperCAmelCase , '''weight''' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(__UpperCAmelCase , '''weight''' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 10, config.vocab_size + 10]: # build the embeddings __lowerCamelCase = model_class(config=__UpperCAmelCase ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_input_embeddings() ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(__UpperCAmelCase ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_input_embeddings() ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. __lowerCamelCase = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , __UpperCAmelCase ) # check that weights remain the same after resizing __lowerCamelCase = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCamelCase = False self.assertTrue(__UpperCAmelCase ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , __UpperCAmelCase ) __lowerCamelCase = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCamelCase = False self.assertTrue(__UpperCAmelCase ) def a__ ( _UpperCamelCase : Optional[Any] ): return tf.constant(_UpperCamelCase ,dtype=tf.intaa ) @require_tf class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = 9_9 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = tf.ones((4, 1) , dtype=tf.intaa ) * 2 __lowerCamelCase = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) __lowerCamelCase = input_ids.shape[0] __lowerCamelCase = OPTConfig( vocab_size=self.vocab_size , hidden_size=24 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class __lowerCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTModel.from_pretrained('''facebook/opt-350m''' ) __lowerCamelCase = _long_tensor([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] ) __lowerCamelCase = tf.not_equal(__UpperCAmelCase , model.config.pad_token_id ) with tf.GradientTape(): __lowerCamelCase = model(input_ids=__UpperCAmelCase , attention_mask=__UpperCAmelCase ).last_hidden_state __lowerCamelCase = (1, 11, 512) self.assertEqual(output.shape , __UpperCAmelCase ) __lowerCamelCase = tf.constant( [[-0.2_873, -1.9_218, -0.3_033], [-1.2_710, -0.1_338, -0.1_902], [0.4_095, 0.1_214, -1.3_121]] ) self.assertTrue(np.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=4E-3 ) ) __lowerCamelCase = tf.function(__UpperCAmelCase , jit_compile=__UpperCAmelCase ) __lowerCamelCase = xla_generate(__UpperCAmelCase , __UpperCAmelCase )[0] self.assertTrue(np.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=4E-2 ) ) @require_tf @slow class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' super().setUp() __lowerCamelCase = '''facebook/opt-350m''' def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTForCausalLM.from_pretrained(self.path_model ) __lowerCamelCase = GPTaTokenizer.from_pretrained(self.path_model ) __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of''', '''Paris is the capital of France and''', '''Computers and mobile phones have taken''', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' , padding=__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) __lowerCamelCase = tf.constant( [ [1.3_851, -13.8_923, -10.5_229, -10.7_533, -0.2_309, -10.2_384, -0.5_365, -9.0_947, -5.1_670], [-4.7_073, -10.6_276, -3.9_415, -21.5_242, -0.2_822, -0.2_822, -0.2_822, -0.2_822, -0.2_822], [0.6_247, -3.4_229, -8.9_179, -1.4_297, -14.1_650, 1.4_146, -9.0_218, -0.2_703, -0.2_703], [6.4_783, -1.9_913, -10.7_926, -2.3_336, 1.5_092, -0.9_974, -6.8_213, 1.3_477, 1.3_477], ] ) self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-4 ) ) __lowerCamelCase = tf.function(__UpperCAmelCase , jit_compile=__UpperCAmelCase ) __lowerCamelCase = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-4 ) ) @require_tf @slow class __lowerCAmelCase ( unittest.TestCase ): @property def lowerCamelCase ( self ): '''simple docstring''' return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-125m''' __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of New York, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] __lowerCamelCase = [] __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) for prompt in self.prompts: __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(__UpperCAmelCase , max_length=10 ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) predicted_outputs += generated_string self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-350m''' __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = '''left''' # use different length sentences to test batching __lowerCamelCase = [ '''Hello, my dog is a little''', '''Today, I''', ] __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' , padding=__UpperCAmelCase ) __lowerCamelCase = inputs['''input_ids'''] __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase , attention_mask=inputs['''attention_mask'''] ) __lowerCamelCase = tokenizer(sentences[0] , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase ) __lowerCamelCase = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['''attention_mask'''][-1] , tf.intaa ) ) __lowerCamelCase = tokenizer(sentences[1] , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase , max_length=model.config.max_length - num_paddings ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = [ '''Hello, my dog is a little bit of a dork.\nI\'m a little bit''', '''Today, I was in the middle of a conversation with a friend about the''', ] self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , [non_padded_sentence, padded_sentence] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-350m''' __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of San Francisco, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] __lowerCamelCase = [] __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) for prompt in self.prompts: __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(__UpperCAmelCase , max_length=10 ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) predicted_outputs += generated_string self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase )
622
0
from typing import Dict, List, Optional from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'nielsr/canine-s': 2_048, } # Unicode defines 1,114,112 total “codepoints” a_ = 1_114_112 # Below: Constants defining canonical codepoints for special, pseudo-characters. # Copied from https://github.com/google-research/language/blob/master/language/canine/special_codepoints.py a_ = 0 a_ = 0xe_0_0_0 a_ = 0xe_0_0_1 a_ = 0xe_0_0_2 a_ = 0xe_0_0_3 a_ = 0xe_0_0_4 # Maps special codepoints to human-readable names. a_ = { # Special symbols are represented using codepoints values that are valid, # but designated as "Private Use", meaning that they will never be assigned # characters by the Unicode Consortium, and are thus safe for use here. # # NOTE: Do *NOT* add any sort of [UNK_CHAR] here. They are explicitly # excluded and should fail with a hard error. CLS: "[CLS]", SEP: "[SEP]", BOS: "[BOS]", MASK: "[MASK]", PAD: "[PAD]", RESERVED: "[RESERVED]", } # Maps special codepoint human-readable names to their codepoint values. a_ = {name: codepoint for codepoint, name in SPECIAL_CODEPOINTS.items()} class __lowerCAmelCase ( UpperCamelCase_ ): lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , __UpperCAmelCase=chr(UpperCamelCase__ ) , __UpperCAmelCase=chr(UpperCamelCase__ ) , __UpperCAmelCase=chr(UpperCamelCase__ ) , __UpperCAmelCase=chr(UpperCamelCase__ ) , __UpperCAmelCase=chr(UpperCamelCase__ ) , __UpperCAmelCase=chr(UpperCamelCase__ ) , __UpperCAmelCase=False , __UpperCAmelCase=2048 , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else bos_token __lowerCamelCase = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else eos_token __lowerCamelCase = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else sep_token __lowerCamelCase = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else cls_token __lowerCamelCase = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it __lowerCamelCase = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token super().__init__( bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , model_max_length=UpperCamelCase__ , **UpperCamelCase__ , ) # Creates a mapping for looking up the IDs of special symbols. __lowerCamelCase = {} for codepoint, name in SPECIAL_CODEPOINTS.items(): __lowerCamelCase = codepoint # Creates a mapping for looking up the string forms of special symbol IDs. __lowerCamelCase = { codepoint: name for name, codepoint in self._special_codepoints.items() } __lowerCamelCase = UNICODE_VOCAB_SIZE __lowerCamelCase = len(self._special_codepoints ) @property def lowerCamelCase ( self ): '''simple docstring''' return self._unicode_vocab_size def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return list(UpperCamelCase__ ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' try: return ord(UpperCamelCase__ ) except TypeError: raise ValueError(F"""invalid token: '{token}'""" ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' try: if index in SPECIAL_CODEPOINTS: return SPECIAL_CODEPOINTS[index] return chr(UpperCamelCase__ ) except TypeError: raise ValueError(F"""invalid id: {index}""" ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return "".join(UpperCamelCase__ ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] __lowerCamelCase = cls + token_ids_a + sep if token_ids_a is not None: result += token_ids_a + sep return result def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase__ , token_ids_a=UpperCamelCase__ , already_has_special_tokens=UpperCamelCase__ ) __lowerCamelCase = [1] + ([0] * len(UpperCamelCase__ )) + [1] if token_ids_a is not None: result += ([0] * len(UpperCamelCase__ )) + [1] return result def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] __lowerCamelCase = len(cls + token_ids_a + sep ) * [0] if token_ids_a is not None: result += len(token_ids_a + sep ) * [1] return result def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' return ()
712
import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) a_ = logging.getLogger(__name__) def a__ ( _UpperCamelCase : str ,_UpperCamelCase : List[Any] ): __lowerCamelCase = np.argmax(_UpperCamelCase ,axis=1 ) return np.sum(outputs == labels ) def a__ ( _UpperCamelCase : Optional[int] ): with open(_UpperCamelCase ,encoding='''utf_8''' ) as f: __lowerCamelCase = csv.reader(_UpperCamelCase ) __lowerCamelCase = [] next(_UpperCamelCase ) # skip the first line for line in tqdm(_UpperCamelCase ): output.append((''' '''.join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def a__ ( _UpperCamelCase : Any ,_UpperCamelCase : Dict ,_UpperCamelCase : str ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Any ,_UpperCamelCase : Dict ): __lowerCamelCase = [] for dataset in encoded_datasets: __lowerCamelCase = len(_UpperCamelCase ) __lowerCamelCase = np.zeros((n_batch, 2, input_len) ,dtype=np.intaa ) __lowerCamelCase = np.zeros((n_batch, 2) ,dtype=np.intaa ) __lowerCamelCase = np.full((n_batch, 2, input_len) ,fill_value=-1_00 ,dtype=np.intaa ) __lowerCamelCase = np.zeros((n_batch,) ,dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(_UpperCamelCase ): __lowerCamelCase = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __lowerCamelCase = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __lowerCamelCase = with_conta __lowerCamelCase = with_conta __lowerCamelCase = len(_UpperCamelCase ) - 1 __lowerCamelCase = len(_UpperCamelCase ) - 1 __lowerCamelCase = with_conta __lowerCamelCase = with_conta __lowerCamelCase = mc_label __lowerCamelCase = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(_UpperCamelCase ) for t in all_inputs ) ) return tensor_datasets def a__ ( ): __lowerCamelCase = argparse.ArgumentParser() parser.add_argument('''--model_name''' ,type=_UpperCamelCase ,default='''openai-gpt''' ,help='''pretrained model name''' ) parser.add_argument('''--do_train''' ,action='''store_true''' ,help='''Whether to run training.''' ) parser.add_argument('''--do_eval''' ,action='''store_true''' ,help='''Whether to run eval on the dev set.''' ) parser.add_argument( '''--output_dir''' ,default=_UpperCamelCase ,type=_UpperCamelCase ,required=_UpperCamelCase ,help='''The output directory where the model predictions and checkpoints will be written.''' ,) parser.add_argument('''--train_dataset''' ,type=_UpperCamelCase ,default='''''' ) parser.add_argument('''--eval_dataset''' ,type=_UpperCamelCase ,default='''''' ) parser.add_argument('''--seed''' ,type=_UpperCamelCase ,default=42 ) parser.add_argument('''--num_train_epochs''' ,type=_UpperCamelCase ,default=3 ) parser.add_argument('''--train_batch_size''' ,type=_UpperCamelCase ,default=8 ) parser.add_argument('''--eval_batch_size''' ,type=_UpperCamelCase ,default=16 ) parser.add_argument('''--adam_epsilon''' ,default=1e-8 ,type=_UpperCamelCase ,help='''Epsilon for Adam optimizer.''' ) parser.add_argument('''--max_grad_norm''' ,type=_UpperCamelCase ,default=1 ) parser.add_argument( '''--max_steps''' ,default=-1 ,type=_UpperCamelCase ,help=( '''If > 0: set total number of training steps to perform. Override num_train_epochs.''' ) ,) parser.add_argument( '''--gradient_accumulation_steps''' ,type=_UpperCamelCase ,default=1 ,help='''Number of updates steps to accumulate before performing a backward/update pass.''' ,) parser.add_argument('''--learning_rate''' ,type=_UpperCamelCase ,default=6.25e-5 ) parser.add_argument('''--warmup_steps''' ,default=0 ,type=_UpperCamelCase ,help='''Linear warmup over warmup_steps.''' ) parser.add_argument('''--lr_schedule''' ,type=_UpperCamelCase ,default='''warmup_linear''' ) parser.add_argument('''--weight_decay''' ,type=_UpperCamelCase ,default=0.01 ) parser.add_argument('''--lm_coef''' ,type=_UpperCamelCase ,default=0.9 ) parser.add_argument('''--n_valid''' ,type=_UpperCamelCase ,default=3_74 ) parser.add_argument('''--server_ip''' ,type=_UpperCamelCase ,default='''''' ,help='''Can be used for distant debugging.''' ) parser.add_argument('''--server_port''' ,type=_UpperCamelCase ,default='''''' ,help='''Can be used for distant debugging.''' ) __lowerCamelCase = parser.parse_args() print(_UpperCamelCase ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('''Waiting for debugger attach''' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) ,redirect_output=_UpperCamelCase ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) __lowerCamelCase = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) __lowerCamelCase = torch.cuda.device_count() logger.info('''device: {}, n_gpu {}'''.format(_UpperCamelCase ,_UpperCamelCase ) ) if not args.do_train and not args.do_eval: raise ValueError('''At least one of `do_train` or `do_eval` must be True.''' ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset __lowerCamelCase = ['''_start_''', '''_delimiter_''', '''_classify_'''] __lowerCamelCase = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(_UpperCamelCase ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(_UpperCamelCase ) __lowerCamelCase = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(_UpperCamelCase ) ) model.to(_UpperCamelCase ) # Load and encode the datasets def tokenize_and_encode(_UpperCamelCase : Dict ): if isinstance(_UpperCamelCase ,_UpperCamelCase ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(_UpperCamelCase ) ) elif isinstance(_UpperCamelCase ,_UpperCamelCase ): return obj return [tokenize_and_encode(_UpperCamelCase ) for o in obj] logger.info('''Encoding dataset...''' ) __lowerCamelCase = load_rocstories_dataset(args.train_dataset ) __lowerCamelCase = load_rocstories_dataset(args.eval_dataset ) __lowerCamelCase = (train_dataset, eval_dataset) __lowerCamelCase = tokenize_and_encode(_UpperCamelCase ) # Compute the max input length for the Transformer __lowerCamelCase = model.config.n_positions // 2 - 2 __lowerCamelCase = max( len(story[:max_length] ) + max(len(conta[:max_length] ) ,len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) __lowerCamelCase = min(_UpperCamelCase ,model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders __lowerCamelCase = pre_process_datasets(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ,*_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = tensor_datasets[0], tensor_datasets[1] __lowerCamelCase = TensorDataset(*_UpperCamelCase ) __lowerCamelCase = RandomSampler(_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,sampler=_UpperCamelCase ,batch_size=args.train_batch_size ) __lowerCamelCase = TensorDataset(*_UpperCamelCase ) __lowerCamelCase = SequentialSampler(_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,sampler=_UpperCamelCase ,batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: __lowerCamelCase = args.max_steps __lowerCamelCase = args.max_steps // (len(_UpperCamelCase ) // args.gradient_accumulation_steps) + 1 else: __lowerCamelCase = len(_UpperCamelCase ) // args.gradient_accumulation_steps * args.num_train_epochs __lowerCamelCase = list(model.named_parameters() ) __lowerCamelCase = ['''bias''', '''LayerNorm.bias''', '''LayerNorm.weight'''] __lowerCamelCase = [ { '''params''': [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], '''weight_decay''': args.weight_decay, }, {'''params''': [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0}, ] __lowerCamelCase = AdamW(_UpperCamelCase ,lr=args.learning_rate ,eps=args.adam_epsilon ) __lowerCamelCase = get_linear_schedule_with_warmup( _UpperCamelCase ,num_warmup_steps=args.warmup_steps ,num_training_steps=_UpperCamelCase ) if args.do_train: __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) ,desc='''Epoch''' ): __lowerCamelCase = 0 __lowerCamelCase = 0 __lowerCamelCase = tqdm(_UpperCamelCase ,desc='''Training''' ) for step, batch in enumerate(_UpperCamelCase ): __lowerCamelCase = tuple(t.to(_UpperCamelCase ) for t in batch ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = batch __lowerCamelCase = model(_UpperCamelCase ,mc_token_ids=_UpperCamelCase ,lm_labels=_UpperCamelCase ,mc_labels=_UpperCamelCase ) __lowerCamelCase = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() __lowerCamelCase = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 __lowerCamelCase = '''Training loss: {:.2e} lr: {:.2e}'''.format(_UpperCamelCase ,scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer __lowerCamelCase = model.module if hasattr(_UpperCamelCase ,'''module''' ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` __lowerCamelCase = os.path.join(args.output_dir ,_UpperCamelCase ) __lowerCamelCase = os.path.join(args.output_dir ,_UpperCamelCase ) torch.save(model_to_save.state_dict() ,_UpperCamelCase ) model_to_save.config.to_json_file(_UpperCamelCase ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned __lowerCamelCase = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) __lowerCamelCase = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(_UpperCamelCase ) if args.do_eval: model.eval() __lowerCamelCase ,__lowerCamelCase = 0, 0 __lowerCamelCase ,__lowerCamelCase = 0, 0 for batch in tqdm(_UpperCamelCase ,desc='''Evaluating''' ): __lowerCamelCase = tuple(t.to(_UpperCamelCase ) for t in batch ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = batch with torch.no_grad(): __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = model( _UpperCamelCase ,mc_token_ids=_UpperCamelCase ,lm_labels=_UpperCamelCase ,mc_labels=_UpperCamelCase ) __lowerCamelCase = mc_logits.detach().cpu().numpy() __lowerCamelCase = mc_labels.to('''cpu''' ).numpy() __lowerCamelCase = accuracy(_UpperCamelCase ,_UpperCamelCase ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 __lowerCamelCase = eval_loss / nb_eval_steps __lowerCamelCase = eval_accuracy / nb_eval_examples __lowerCamelCase = tr_loss / nb_tr_steps if args.do_train else None __lowerCamelCase = {'''eval_loss''': eval_loss, '''eval_accuracy''': eval_accuracy, '''train_loss''': train_loss} __lowerCamelCase = os.path.join(args.output_dir ,'''eval_results.txt''' ) with open(_UpperCamelCase ,'''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key in sorted(result.keys() ): logger.info(''' %s = %s''' ,_UpperCamelCase ,str(result[key] ) ) writer.write('''%s = %s\n''' % (key, str(result[key] )) ) if __name__ == "__main__": main()
622
0
import argparse import gc import json import os import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler a_ = 16 a_ = 32 def a__ ( _UpperCamelCase : Dict ): return int(x / 2**20 ) class __lowerCAmelCase : def __enter__( self ): '''simple docstring''' gc.collect() torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() # reset the peak gauge to zero __lowerCamelCase = torch.cuda.memory_allocated() return self def __exit__( self , *__UpperCAmelCase ): '''simple docstring''' gc.collect() torch.cuda.empty_cache() __lowerCamelCase = torch.cuda.memory_allocated() __lowerCamelCase = torch.cuda.max_memory_allocated() __lowerCamelCase = bamb(self.end - self.begin ) __lowerCamelCase = bamb(self.peak - self.begin ) # print(f"delta used/peak {self.used:4d}/{self.peaked:4d}") def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Dict = 16 ,_UpperCamelCase : Union[str, Any] = "bert-base-cased" ,_UpperCamelCase : Dict = 3_20 ,_UpperCamelCase : int = 1_60 ,): __lowerCamelCase = AutoTokenizer.from_pretrained(__UpperCamelCase ) __lowerCamelCase = load_dataset( '''glue''' ,'''mrpc''' ,split={'''train''': F"""train[:{n_train}]""", '''validation''': F"""validation[:{n_val}]"""} ) def tokenize_function(_UpperCamelCase : List[str] ): # max_length=None => use the model max length (it's actually the default) __lowerCamelCase = tokenizer(examples['''sentence1'''] ,examples['''sentence2'''] ,truncation=__UpperCamelCase ,max_length=__UpperCamelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset __lowerCamelCase = datasets.map( __UpperCamelCase ,batched=__UpperCamelCase ,remove_columns=['''idx''', '''sentence1''', '''sentence2'''] ,load_from_cache_file=__UpperCamelCase ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __lowerCamelCase = tokenized_datasets.rename_column('''label''' ,'''labels''' ) def collate_fn(_UpperCamelCase : int ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(__UpperCamelCase ,padding='''max_length''' ,max_length=1_28 ,return_tensors='''pt''' ) return tokenizer.pad(__UpperCamelCase ,padding='''longest''' ,return_tensors='''pt''' ) # Instantiate dataloaders. __lowerCamelCase = DataLoader( tokenized_datasets['''train'''] ,shuffle=__UpperCamelCase ,collate_fn=__UpperCamelCase ,batch_size=__UpperCamelCase ) __lowerCamelCase = DataLoader( tokenized_datasets['''validation'''] ,shuffle=__UpperCamelCase ,collate_fn=__UpperCamelCase ,batch_size=__UpperCamelCase ) return train_dataloader, eval_dataloader def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Optional[Any] ): # Initialize accelerator __lowerCamelCase = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __lowerCamelCase = config['''lr'''] __lowerCamelCase = int(config['''num_epochs'''] ) __lowerCamelCase = int(config['''seed'''] ) __lowerCamelCase = int(config['''batch_size'''] ) __lowerCamelCase = args.model_name_or_path set_seed(__UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = get_dataloaders(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,args.n_train ,args.n_val ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __lowerCamelCase = AutoModelForSequenceClassification.from_pretrained(__UpperCamelCase ,return_dict=__UpperCamelCase ) # Instantiate optimizer __lowerCamelCase = ( AdamW if accelerator.state.deepspeed_plugin is None or '''optimizer''' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) __lowerCamelCase = optimizer_cls(params=model.parameters() ,lr=__UpperCamelCase ) if accelerator.state.deepspeed_plugin is not None: __lowerCamelCase = accelerator.state.deepspeed_plugin.deepspeed_config[ '''gradient_accumulation_steps''' ] else: __lowerCamelCase = 1 __lowerCamelCase = (len(__UpperCamelCase ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): __lowerCamelCase = get_linear_schedule_with_warmup( optimizer=__UpperCamelCase ,num_warmup_steps=0 ,num_training_steps=__UpperCamelCase ,) else: __lowerCamelCase = DummyScheduler(__UpperCamelCase ,total_num_steps=__UpperCamelCase ,warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = accelerator.prepare( __UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) # We need to keep track of how many total steps we have iterated over __lowerCamelCase = 0 # We also need to keep track of the stating epoch so files are named properly __lowerCamelCase = 0 # Now we train the model __lowerCamelCase = {} for epoch in range(__UpperCamelCase ,__UpperCamelCase ): with TorchTracemalloc() as tracemalloc: model.train() for step, batch in enumerate(__UpperCamelCase ): __lowerCamelCase = model(**__UpperCamelCase ) __lowerCamelCase = outputs.loss __lowerCamelCase = loss / gradient_accumulation_steps accelerator.backward(__UpperCamelCase ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 # Printing the GPU memory usage details such as allocated memory, peak memory, and total memory usage accelerator.print('''Memory before entering the train : {}'''.format(bamb(tracemalloc.begin ) ) ) accelerator.print('''Memory consumed at the end of the train (end-begin): {}'''.format(tracemalloc.used ) ) accelerator.print('''Peak Memory consumed during the train (max-begin): {}'''.format(tracemalloc.peaked ) ) accelerator.print( '''Total Peak Memory consumed during the train (max): {}'''.format( tracemalloc.peaked + bamb(tracemalloc.begin ) ) ) __lowerCamelCase = tracemalloc.peaked + bamb(tracemalloc.begin ) if args.peak_memory_upper_bound is not None: assert ( train_total_peak_memory[F"""epoch-{epoch}"""] <= args.peak_memory_upper_bound ), "Peak memory usage exceeded the upper bound" accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir ,'''peak_memory_utilization.json''' ) ,'''w''' ) as f: json.dump(__UpperCamelCase ,__UpperCamelCase ) def a__ ( ): __lowerCamelCase = argparse.ArgumentParser(description='''Simple example of training script tracking peak GPU memory usage.''' ) parser.add_argument( '''--model_name_or_path''' ,type=__UpperCamelCase ,default='''bert-base-cased''' ,help='''Path to pretrained model or model identifier from huggingface.co/models.''' ,required=__UpperCamelCase ,) parser.add_argument( '''--output_dir''' ,type=__UpperCamelCase ,default='''.''' ,help='''Optional save directory where all checkpoint folders will be stored. Default is the current working directory.''' ,) parser.add_argument( '''--peak_memory_upper_bound''' ,type=__UpperCamelCase ,default=__UpperCamelCase ,help='''The upper bound of peak memory usage in MB. If set, the training will throw an error if the peak memory usage exceeds this value.''' ,) parser.add_argument( '''--n_train''' ,type=__UpperCamelCase ,default=3_20 ,help='''Number of training examples to use.''' ,) parser.add_argument( '''--n_val''' ,type=__UpperCamelCase ,default=1_60 ,help='''Number of validation examples to use.''' ,) parser.add_argument( '''--num_epochs''' ,type=__UpperCamelCase ,default=1 ,help='''Number of train epochs.''' ,) __lowerCamelCase = parser.parse_args() __lowerCamelCase = {'''lr''': 2e-5, '''num_epochs''': args.num_epochs, '''seed''': 42, '''batch_size''': 16} training_function(__UpperCamelCase ,__UpperCamelCase ) if __name__ == "__main__": main()
713
import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class __lowerCAmelCase ( lowerCAmelCase__ ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=1024 , __UpperCAmelCase=1024 , __UpperCAmelCase=3.6 ): '''simple docstring''' __lowerCamelCase = tokenizer __lowerCamelCase = tokenizer.bos_token_id __lowerCamelCase = dataset __lowerCamelCase = seq_length __lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self ): '''simple docstring''' __lowerCamelCase = iter(self.dataset ) __lowerCamelCase = True while more_examples: __lowerCamelCase ,__lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(__UpperCAmelCase )['''content'''] ) buffer_len += len(buffer[-1] ) except StopIteration: __lowerCamelCase = False break __lowerCamelCase = tokenizer(__UpperCAmelCase , truncation=__UpperCAmelCase )['''input_ids'''] __lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(__UpperCAmelCase ) , self.seq_length ): __lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(__UpperCAmelCase ) == self.seq_length: yield torch.tensor(__UpperCAmelCase ) def a__ ( _UpperCamelCase : List[Any] ): __lowerCamelCase = {'''streaming''': True} __lowerCamelCase = load_dataset(args.dataset_name ,split='''train''' ,**_UpperCamelCase ) __lowerCamelCase = ConstantLengthDataset(_UpperCamelCase ,_UpperCamelCase ,seq_length=args.seq_length ) __lowerCamelCase = DataLoader(_UpperCamelCase ,batch_size=args.batch_size ) return eval_dataloader def a__ ( _UpperCamelCase : str ): model.eval() __lowerCamelCase = [] for step, batch in enumerate(_UpperCamelCase ): with torch.no_grad(): __lowerCamelCase = model(_UpperCamelCase ,labels=_UpperCamelCase ) __lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(_UpperCamelCase ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break __lowerCamelCase = torch.mean(torch.cat(_UpperCamelCase ) ) try: __lowerCamelCase = torch.exp(_UpperCamelCase ) except OverflowError: __lowerCamelCase = float('''inf''' ) return loss.item(), perplexity.item() # Setup Accelerator a_ = Accelerator() # Parse configuration a_ = HfArgumentParser(EvaluationArguments) a_ = parser.parse_args() set_seed(args.seed) # Logging a_ = logging.getLogger(__name__) logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) # Load model and tokenizer a_ = AutoModelForCausalLM.from_pretrained(args.model_ckpt) a_ = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader a_ = create_dataloader(args) # Prepare everything with our `accelerator`. a_ , a_ = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info("""Evaluating and saving model after training""") a_ , a_ = evaluate(args) logger.info(f"loss/eval: {eval_loss}, perplexity: {perplexity}")
622
0
import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand a_ = ( """4S 3H 2C 7S 5H""", """9D 8H 2C 6S 7H""", """2D 6D 9D TH 7D""", """TC 8C 2S JH 6C""", """JH 8S TH AH QH""", """TS KS 5S 9S AC""", """KD 6S 9D TH AD""", """KS 8D 4D 9S 4S""", # pair """8C 4S KH JS 4D""", # pair """QH 8H KD JH 8S""", # pair """KC 4H KS 2H 8D""", # pair """KD 4S KC 3H 8S""", # pair """AH 8S AS KC JH""", # pair """3H 4C 4H 3S 2H""", # 2 pairs """5S 5D 2C KH KH""", # 2 pairs """3C KH 5D 5S KH""", # 2 pairs """AS 3C KH AD KH""", # 2 pairs """7C 7S 3S 7H 5S""", # 3 of a kind """7C 7S KH 2H 7H""", # 3 of a kind """AC KH QH AH AS""", # 3 of a kind """2H 4D 3C AS 5S""", # straight (low ace) """3C 5C 4C 2C 6H""", # straight """6S 8S 7S 5H 9H""", # straight """JS QS 9H TS KH""", # straight """QC KH TS JS AH""", # straight (high ace) """8C 9C 5C 3C TC""", # flush """3S 8S 9S 5S KS""", # flush """4C 5C 9C 8C KC""", # flush """JH 8H AH KH QH""", # flush """3D 2H 3H 2C 2D""", # full house """2H 2C 3S 3H 3D""", # full house """KH KC 3S 3H 3D""", # full house """JC 6H JS JD JH""", # 4 of a kind """JC 7H JS JD JH""", # 4 of a kind """JC KH JS JD JH""", # 4 of a kind """2S AS 4S 5S 3S""", # straight flush (low ace) """2D 6D 3D 4D 5D""", # straight flush """5C 6C 3C 7C 4C""", # straight flush """JH 9H TH KH QH""", # straight flush """JH AH TH KH QH""", # royal flush (high ace straight flush) ) a_ = ( ("""2H 3H 4H 5H 6H""", """KS AS TS QS JS""", """Loss"""), ("""2H 3H 4H 5H 6H""", """AS AD AC AH JD""", """Win"""), ("""AS AH 2H AD AC""", """JS JD JC JH 3D""", """Win"""), ("""2S AH 2H AS AC""", """JS JD JC JH AD""", """Loss"""), ("""2S AH 2H AS AC""", """2H 3H 5H 6H 7H""", """Win"""), ("""AS 3S 4S 8S 2S""", """2H 3H 5H 6H 7H""", """Win"""), ("""2H 3H 5H 6H 7H""", """2S 3H 4H 5S 6C""", """Win"""), ("""2S 3H 4H 5S 6C""", """3D 4C 5H 6H 2S""", """Tie"""), ("""2S 3H 4H 5S 6C""", """AH AC 5H 6H AS""", """Win"""), ("""2S 2H 4H 5S 4C""", """AH AC 5H 6H AS""", """Loss"""), ("""2S 2H 4H 5S 4C""", """AH AC 5H 6H 7S""", """Win"""), ("""6S AD 7H 4S AS""", """AH AC 5H 6H 7S""", """Loss"""), ("""2S AH 4H 5S KC""", """AH AC 5H 6H 7S""", """Loss"""), ("""2S 3H 6H 7S 9C""", """7H 3C TH 6H 9S""", """Loss"""), ("""4S 5H 6H TS AC""", """3S 5H 6H TS AC""", """Win"""), ("""2S AH 4H 5S 6C""", """AD 4C 5H 6H 2C""", """Tie"""), ("""AS AH 3H AD AC""", """AS AH 2H AD AC""", """Win"""), ("""AH AC 5H 5C QS""", """AH AC 5H 5C KS""", """Loss"""), ("""AH AC 5H 5C QS""", """KH KC 5H 5C QS""", """Win"""), ("""7C 7S KH 2H 7H""", """3C 3S AH 2H 3H""", """Win"""), ("""3C 3S AH 2H 3H""", """7C 7S KH 2H 7H""", """Loss"""), ("""6H 5H 4H 3H 2H""", """5H 4H 3H 2H AH""", """Win"""), ("""5H 4H 3H 2H AH""", """5H 4H 3H 2H AH""", """Tie"""), ("""5H 4H 3H 2H AH""", """6H 5H 4H 3H 2H""", """Loss"""), ("""AH AD KS KC AC""", """AH KD KH AC KC""", """Win"""), ("""2H 4D 3C AS 5S""", """2H 4D 3C 6S 5S""", """Loss"""), ("""2H 3S 3C 3H 2S""", """3S 3C 2S 2H 2D""", """Win"""), ("""4D 6D 5D 2D JH""", """3S 8S 3H TC KH""", """Loss"""), ("""4S 6C 8S 3S 7S""", """AD KS 2D 7D 7C""", """Loss"""), ("""6S 4C 7H 8C 3H""", """5H JC AH 9D 9C""", """Loss"""), ("""9D 9H JH TC QH""", """3C 2S JS 5C 7H""", """Win"""), ("""2H TC 8S AD 9S""", """4H TS 7H 2C 5C""", """Win"""), ("""9D 3S 2C 7S 7C""", """JC TD 3C TC 9H""", """Loss"""), ) a_ = ( ("""2H 3H 4H 5H 6H""", True), ("""AS AH 2H AD AC""", False), ("""2H 3H 5H 6H 7H""", True), ("""KS AS TS QS JS""", True), ("""8H 9H QS JS TH""", False), ("""AS 3S 4S 8S 2S""", True), ) a_ = ( ("""2H 3H 4H 5H 6H""", True), ("""AS AH 2H AD AC""", False), ("""2H 3H 5H 6H 7H""", False), ("""KS AS TS QS JS""", True), ("""8H 9H QS JS TH""", True), ) a_ = ( ("""2H 4D 3C AS 5S""", True, [5, 4, 3, 2, 14]), ("""2H 5D 3C AS 5S""", False, [14, 5, 5, 3, 2]), ("""JH QD KC AS TS""", False, [14, 13, 12, 11, 10]), ("""9D 3S 2C 7S 7C""", False, [9, 7, 7, 3, 2]), ) a_ = ( ("""JH AH TH KH QH""", 0), ("""JH 9H TH KH QH""", 0), ("""JC KH JS JD JH""", 7), ("""KH KC 3S 3H 3D""", 6), ("""8C 9C 5C 3C TC""", 0), ("""JS QS 9H TS KH""", 0), ("""7C 7S KH 2H 7H""", 3), ("""3C KH 5D 5S KH""", 2), ("""QH 8H KD JH 8S""", 1), ("""2D 6D 9D TH 7D""", 0), ) a_ = ( ("""JH AH TH KH QH""", 23), ("""JH 9H TH KH QH""", 22), ("""JC KH JS JD JH""", 21), ("""KH KC 3S 3H 3D""", 20), ("""8C 9C 5C 3C TC""", 19), ("""JS QS 9H TS KH""", 18), ("""7C 7S KH 2H 7H""", 17), ("""3C KH 5D 5S KH""", 16), ("""QH 8H KD JH 8S""", 15), ("""2D 6D 9D TH 7D""", 14), ) def a__ ( ): __lowerCamelCase = randrange(len(_UpperCamelCase ) ), randrange(len(_UpperCamelCase ) ) __lowerCamelCase = ["""Loss""", """Tie""", """Win"""][(play >= oppo) + (play > oppo)] __lowerCamelCase = SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def a__ ( _UpperCamelCase : List[Any] = 1_00 ): return (generate_random_hand() for _ in range(_UpperCamelCase )) @pytest.mark.parametrize('''hand, expected''' ,_UpperCamelCase ) def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Tuple ): assert PokerHand(_UpperCamelCase )._is_flush() == expected @pytest.mark.parametrize('''hand, expected''' ,_UpperCamelCase ) def a__ ( _UpperCamelCase : Any ,_UpperCamelCase : Union[str, Any] ): assert PokerHand(_UpperCamelCase )._is_straight() == expected @pytest.mark.parametrize('''hand, expected, card_values''' ,_UpperCamelCase ) def a__ ( _UpperCamelCase : str ,_UpperCamelCase : List[str] ,_UpperCamelCase : int ): __lowerCamelCase = PokerHand(_UpperCamelCase ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize('''hand, expected''' ,_UpperCamelCase ) def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : int ): assert PokerHand(_UpperCamelCase )._is_same_kind() == expected @pytest.mark.parametrize('''hand, expected''' ,_UpperCamelCase ) def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : List[str] ): assert PokerHand(_UpperCamelCase )._hand_type == expected @pytest.mark.parametrize('''hand, other, expected''' ,_UpperCamelCase ) def a__ ( _UpperCamelCase : Optional[int] ,_UpperCamelCase : List[str] ,_UpperCamelCase : List[str] ): assert PokerHand(_UpperCamelCase ).compare_with(PokerHand(_UpperCamelCase ) ) == expected @pytest.mark.parametrize('''hand, other, expected''' ,generate_random_hands() ) def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : Tuple ): assert PokerHand(_UpperCamelCase ).compare_with(PokerHand(_UpperCamelCase ) ) == expected def a__ ( ): __lowerCamelCase = [PokerHand(_UpperCamelCase ) for hand in SORTED_HANDS] __lowerCamelCase = poker_hands.copy() shuffle(_UpperCamelCase ) __lowerCamelCase = chain(sorted(_UpperCamelCase ) ) for index, hand in enumerate(_UpperCamelCase ): assert hand == poker_hands[index] def a__ ( ): # Test that five high straights are compared correctly. __lowerCamelCase = [PokerHand('''2D AC 3H 4H 5S''' ), PokerHand('''2S 3H 4H 5S 6C''' )] pokerhands.sort(reverse=_UpperCamelCase ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def a__ ( ): # Multiple calls to five_high_straight function should still return True # and shouldn't mutate the list in every call other than the first. __lowerCamelCase = PokerHand('''2C 4S AS 3D 5C''' ) __lowerCamelCase = True __lowerCamelCase = [5, 4, 3, 2, 14] for _ in range(10 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def a__ ( ): # Problem number 54 from Project Euler # Testing from poker_hands.txt file __lowerCamelCase = 0 __lowerCamelCase = os.path.abspath(os.path.dirname(_UpperCamelCase ) ) __lowerCamelCase = os.path.join(_UpperCamelCase ,'''poker_hands.txt''' ) with open(_UpperCamelCase ) as file_hand: for line in file_hand: __lowerCamelCase = line[:14].strip() __lowerCamelCase = line[15:].strip() __lowerCamelCase = PokerHand(_UpperCamelCase ), PokerHand(_UpperCamelCase ) __lowerCamelCase = player.compare_with(_UpperCamelCase ) if output == "Win": answer += 1 assert answer == 3_76
714
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """lxmert""" lowerCAmelCase__ = {} def __init__( self , __UpperCAmelCase=30522 , __UpperCAmelCase=768 , __UpperCAmelCase=12 , __UpperCAmelCase=9500 , __UpperCAmelCase=1600 , __UpperCAmelCase=400 , __UpperCAmelCase=3072 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=9 , __UpperCAmelCase=5 , __UpperCAmelCase=5 , __UpperCAmelCase=2048 , __UpperCAmelCase=4 , __UpperCAmelCase=6.67 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = num_qa_labels __lowerCamelCase = num_object_labels __lowerCamelCase = num_attr_labels __lowerCamelCase = l_layers __lowerCamelCase = x_layers __lowerCamelCase = r_layers __lowerCamelCase = visual_feat_dim __lowerCamelCase = visual_pos_dim __lowerCamelCase = visual_loss_normalizer __lowerCamelCase = task_matched __lowerCamelCase = task_mask_lm __lowerCamelCase = task_obj_predict __lowerCamelCase = task_qa __lowerCamelCase = visual_obj_loss __lowerCamelCase = visual_attr_loss __lowerCamelCase = visual_feat_loss __lowerCamelCase = {'''vision''': r_layers, '''cross_encoder''': x_layers, '''language''': l_layers} super().__init__(**__UpperCAmelCase )
622
0
'''simple docstring''' from typing import List, Union import numpy as np from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_DEPTH_ESTIMATION_MAPPING a_ = logging.get_logger(__name__) @add_end_docstrings(lowercase__ ) class __lowerCAmelCase ( lowercase__ ): def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' super().__init__(*__lowercase , **__lowercase ) requires_backends(self , '''vision''' ) self.check_model_type(__lowercase ) def __call__( self , __UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' return super().__call__(__lowercase , **__lowercase ) def lowerCamelCase ( self , **__UpperCAmelCase ): '''simple docstring''' return {}, {}, {} def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = load_image(__lowercase ) __lowerCamelCase = image.size __lowerCamelCase = self.image_processor(images=__lowercase , return_tensors=self.framework ) return model_inputs def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.model(**__lowercase ) return model_outputs def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = model_outputs.predicted_depth __lowerCamelCase = torch.nn.functional.interpolate( predicted_depth.unsqueeze(1 ) , size=self.image_size[::-1] , mode='''bicubic''' , align_corners=__lowercase ) __lowerCamelCase = prediction.squeeze().cpu().numpy() __lowerCamelCase = (output * 255 / np.max(__lowercase )).astype('''uint8''' ) __lowerCamelCase = Image.fromarray(__lowercase ) __lowerCamelCase = {} __lowerCamelCase = predicted_depth __lowerCamelCase = depth return output_dict
715
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Any ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Any ): if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = np.full((len(_UpperCamelCase ), sequence_length, 2) ,_UpperCamelCase ) else: __lowerCamelCase = np.full((len(_UpperCamelCase ), sequence_length) ,_UpperCamelCase ) for i, tensor in enumerate(_UpperCamelCase ): if padding_side == "right": if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = tensor[:sequence_length] else: __lowerCamelCase = tensor[:sequence_length] else: if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = tensor[:sequence_length] else: __lowerCamelCase = tensor[:sequence_length] return out_tensor.tolist() def a__ ( _UpperCamelCase : Dict ): __lowerCamelCase = ord(_UpperCamelCase ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 1_23 and cp <= 1_26): return True __lowerCamelCase = unicodedata.category(_UpperCamelCase ) if cat.startswith('''P''' ): return True return False @dataclass class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = 42 lowerCAmelCase__ = True lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = -1_0_0 lowerCAmelCase__ = "pt" def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' import torch __lowerCamelCase = '''label''' if '''label''' in features[0].keys() else '''labels''' __lowerCamelCase = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __lowerCamelCase = self.tokenizer.pad( __UpperCAmelCase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch __lowerCamelCase = torch.tensor(batch['''entity_ids'''] ).shape[1] __lowerCamelCase = self.tokenizer.padding_side if padding_side == "right": __lowerCamelCase = [ list(__UpperCAmelCase ) + [self.label_pad_token_id] * (sequence_length - len(__UpperCAmelCase )) for label in labels ] else: __lowerCamelCase = [ [self.label_pad_token_id] * (sequence_length - len(__UpperCAmelCase )) + list(__UpperCAmelCase ) for label in labels ] __lowerCamelCase = [feature['''ner_tags'''] for feature in features] __lowerCamelCase = padding_tensor(__UpperCAmelCase , -1 , __UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = [feature['''original_entity_spans'''] for feature in features] __lowerCamelCase = padding_tensor(__UpperCAmelCase , (-1, -1) , __UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = {k: torch.tensor(__UpperCAmelCase , dtype=torch.intaa ) for k, v in batch.items()} return batch
622
0
import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin a_ = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right a_ = 250_004 a_ = 250_020 @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( __snake_case , unittest.TestCase ): lowerCAmelCase__ = MBartTokenizer lowerCAmelCase__ = MBartTokenizerFast lowerCAmelCase__ = True lowerCAmelCase__ = True def lowerCamelCase ( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __lowerCamelCase = MBartTokenizer(A_ , keep_accents=A_ ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = MBartTokenizer(A_ , keep_accents=A_ ) __lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(A_ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(A_ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __lowerCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( A_ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(A_ ) self.assertListEqual( A_ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(A_ ) self.assertListEqual( A_ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def lowerCamelCase ( self ): '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __lowerCamelCase = (self.rust_tokenizer_class, "hf-internal-testing/tiny-random-mbart", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __lowerCamelCase = self.rust_tokenizer_class.from_pretrained(A_ , **A_ ) __lowerCamelCase = self.tokenizer_class.from_pretrained(A_ , **A_ ) __lowerCamelCase = tempfile.mkdtemp() __lowerCamelCase = tokenizer_r.save_pretrained(A_ ) __lowerCamelCase = tokenizer_p.save_pretrained(A_ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) __lowerCamelCase = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(A_ , A_ ) # Checks everything loads correctly in the same way __lowerCamelCase = tokenizer_r.from_pretrained(A_ ) __lowerCamelCase = tokenizer_p.from_pretrained(A_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A_ , A_ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(A_ ) # Save tokenizer rust, legacy_format=True __lowerCamelCase = tempfile.mkdtemp() __lowerCamelCase = tokenizer_r.save_pretrained(A_ , legacy_format=A_ ) __lowerCamelCase = tokenizer_p.save_pretrained(A_ ) # Checks it save with the same files self.assertSequenceEqual(A_ , A_ ) # Checks everything loads correctly in the same way __lowerCamelCase = tokenizer_r.from_pretrained(A_ ) __lowerCamelCase = tokenizer_p.from_pretrained(A_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A_ , A_ ) ) shutil.rmtree(A_ ) # Save tokenizer rust, legacy_format=False __lowerCamelCase = tempfile.mkdtemp() __lowerCamelCase = tokenizer_r.save_pretrained(A_ , legacy_format=A_ ) __lowerCamelCase = tokenizer_p.save_pretrained(A_ ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way __lowerCamelCase = tokenizer_r.from_pretrained(A_ ) __lowerCamelCase = tokenizer_p.from_pretrained(A_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A_ , A_ ) ) shutil.rmtree(A_ ) @require_torch @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = 'facebook/mbart-large-en-ro' lowerCAmelCase__ = [ ' UN Chief Says There Is No Military Solution in Syria', ' Secretary-General Ban Ki-moon says his response to Russia\'s stepped up military support for Syria is that "there is no military solution" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.', ] lowerCAmelCase__ = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', 'Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei' ' pentru Siria este că "nu există o soluţie militară" la conflictul de aproape cinci ani şi că noi arme nu vor' ' face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.', ] lowerCAmelCase__ = [8_2_7_4, 1_2_7_8_7_3, 2_5_9_1_6, 7, 8_6_2_2, 2_0_7_1, 4_3_8, 6_7_4_8_5, 5_3, 1_8_7_8_9_5, 2_3, 5_1_7_1_2, 2, EN_CODE] @classmethod def lowerCamelCase ( cls ): '''simple docstring''' __lowerCamelCase = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang='''en_XX''' , tgt_lang='''ro_RO''' ) __lowerCamelCase = 1 return cls def lowerCamelCase ( self ): '''simple docstring''' self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ar_AR'''] , 250001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''en_EN'''] , 250004 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ro_RO'''] , 250020 ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , A_ ) def lowerCamelCase ( self ): '''simple docstring''' self.assertIn(A_ , self.tokenizer.all_special_ids ) __lowerCamelCase = [RO_CODE, 884, 9019, 96, 9, 916, 86792, 36, 18743, 15596, 5, 2] __lowerCamelCase = self.tokenizer.decode(A_ , skip_special_tokens=A_ ) __lowerCamelCase = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=A_ ) self.assertEqual(A_ , A_ ) self.assertNotIn(self.tokenizer.eos_token , A_ ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ["this is gunna be a long sentence " * 20] assert isinstance(src_text[0] , A_ ) __lowerCamelCase = 10 __lowerCamelCase = self.tokenizer(A_ , max_length=A_ , truncation=A_ ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , A_ ) self.assertEqual(len(A_ ) , A_ ) def lowerCamelCase ( self ): '''simple docstring''' self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['''<mask>''', '''ar_AR'''] ) , [250026, 250001] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = tempfile.mkdtemp() __lowerCamelCase = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(A_ ) __lowerCamelCase = MBartTokenizer.from_pretrained(A_ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , A_ ) @require_torch def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=A_ , return_tensors='''pt''' ) __lowerCamelCase = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=A_ , truncation=A_ , max_length=len(self.expected_src_tokens ) , return_tensors='''pt''' , ) __lowerCamelCase = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) self.assertIsInstance(A_ , A_ ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) __lowerCamelCase = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , A_ ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.tokenizer(self.src_text , padding=A_ , truncation=A_ , max_length=3 , return_tensors='''pt''' ) __lowerCamelCase = self.tokenizer( text_target=self.tgt_text , padding=A_ , truncation=A_ , max_length=10 , return_tensors='''pt''' ) __lowerCamelCase = targets["input_ids"] __lowerCamelCase = shift_tokens_right(A_ , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.tokenizer._build_translation_inputs( '''A test''' , return_tensors='''pt''' , src_lang='''en_XX''' , tgt_lang='''ar_AR''' ) self.assertEqual( nested_simplify(A_ ) , { # A, test, EOS, en_XX '''input_ids''': [[62, 3034, 2, 250004]], '''attention_mask''': [[1, 1, 1, 1]], # ar_AR '''forced_bos_token_id''': 250001, } , )
716
from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=99 , __UpperCAmelCase=[1, 1, 2] , __UpperCAmelCase=1 , __UpperCAmelCase=32 , __UpperCAmelCase=4 , __UpperCAmelCase=8 , __UpperCAmelCase=37 , __UpperCAmelCase="gelu_new" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.0 , __UpperCAmelCase=512 , __UpperCAmelCase=3 , __UpperCAmelCase=0.02 , __UpperCAmelCase=3 , __UpperCAmelCase=4 , __UpperCAmelCase=None , __UpperCAmelCase=False , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_input_mask __lowerCamelCase = use_token_type_ids __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = block_sizes __lowerCamelCase = num_decoder_layers __lowerCamelCase = d_model __lowerCamelCase = n_head __lowerCamelCase = d_head __lowerCamelCase = d_inner __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout __lowerCamelCase = attention_dropout __lowerCamelCase = activation_dropout __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = 2 __lowerCamelCase = num_labels __lowerCamelCase = num_choices __lowerCamelCase = scope __lowerCamelCase = initializer_std # Used in the tests to check the size of the first attention layer __lowerCamelCase = n_head # Used in the tests to check the size of the first hidden state __lowerCamelCase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowerCamelCase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowerCamelCase = self.num_hidden_layers + 2 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCamelCase = None if self.use_token_type_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCamelCase = FunnelConfig( vocab_size=self.vocab_size , block_sizes=self.block_sizes , num_decoder_layers=self.num_decoder_layers , d_model=self.d_model , n_head=self.n_head , d_head=self.d_head , d_inner=self.d_inner , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , activation_dropout=self.activation_dropout , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_std=self.initializer_std , ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = [input_ids, input_mask] __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = [input_ids, input_mask] __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 3, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForPreTraining(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForMaskedLM(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = TFFunnelForSequenceClassification(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_choices __lowerCamelCase = TFFunnelForMultipleChoice(config=__UpperCAmelCase ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = TFFunnelForTokenClassification(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForQuestionAnswering(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.prepare_config_and_inputs() ( ( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) , ) = config_and_inputs __lowerCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) lowerCAmelCase__ = ( { """feature-extraction""": (TFFunnelBaseModel, TFFunnelModel), """fill-mask""": TFFunnelForMaskedLM, """question-answering""": TFFunnelForQuestionAnswering, """text-classification""": TFFunnelForSequenceClassification, """token-classification""": TFFunnelForTokenClassification, """zero-shot""": TFFunnelForSequenceClassification, } if is_tf_available() else {} ) lowerCAmelCase__ = False lowerCAmelCase__ = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFFunnelModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCAmelCase ) @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) lowerCAmelCase__ = False lowerCAmelCase__ = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFFunnelModelTester(self , base=__UpperCAmelCase ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCAmelCase )
622
0
from __future__ import annotations import math def a__ ( _UpperCamelCase : int ): if num <= 0: __lowerCamelCase = F"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(_UpperCamelCase ) __lowerCamelCase = [True] * (num + 1) __lowerCamelCase = [] __lowerCamelCase = 2 __lowerCamelCase = int(math.sqrt(_UpperCamelCase ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(_UpperCamelCase ) # Set multiples of start be False for i in range(start * start ,num + 1 ,_UpperCamelCase ): if sieve[i] is True: __lowerCamelCase = False start += 1 for j in range(end + 1 ,num + 1 ): if sieve[j] is True: prime.append(_UpperCamelCase ) return prime if __name__ == "__main__": print(prime_sieve(int(input("""Enter a positive integer: """).strip())))
717
from collections import namedtuple import requests from lxml import html # type: ignore a_ = namedtuple("""covid_data""", """cases deaths recovered""") def a__ ( _UpperCamelCase : str = "https://www.worldometers.info/coronavirus/" ): __lowerCamelCase = '''//div[@class = "maincounter-number"]/span/text()''' return covid_data(*html.fromstring(requests.get(_UpperCamelCase ).content ).xpath(_UpperCamelCase ) ) a_ = """Total COVID-19 cases in the world: {} Total deaths due to COVID-19 in the world: {} Total COVID-19 patients recovered in the world: {}""" print(fmt.format(*covid_stats()))
622
0
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class __lowerCAmelCase ( UpperCAmelCase__ ): lowerCAmelCase__ = 42 class __lowerCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ ): @register_to_config def __init__( self , __UpperCAmelCase = 65536 , __UpperCAmelCase = None , __UpperCAmelCase = 2 , __UpperCAmelCase = 2 , __UpperCAmelCase = 0 , __UpperCAmelCase = "fourier" , __UpperCAmelCase = True , __UpperCAmelCase = False , __UpperCAmelCase = 0.0 , __UpperCAmelCase = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , __UpperCAmelCase = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , __UpperCAmelCase = "UNetMidBlock1D" , __UpperCAmelCase = None , __UpperCAmelCase = (32, 32, 64) , __UpperCAmelCase = None , __UpperCAmelCase = 8 , __UpperCAmelCase = 1 , __UpperCAmelCase = False , ): '''simple docstring''' super().__init__() __lowerCamelCase = sample_size # time if time_embedding_type == "fourier": __lowerCamelCase = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=lowerCamelCase__ , log=lowerCamelCase__ , flip_sin_to_cos=lowerCamelCase__ ) __lowerCamelCase = 2 * block_out_channels[0] elif time_embedding_type == "positional": __lowerCamelCase = Timesteps( block_out_channels[0] , flip_sin_to_cos=lowerCamelCase__ , downscale_freq_shift=lowerCamelCase__ ) __lowerCamelCase = block_out_channels[0] if use_timestep_embedding: __lowerCamelCase = block_out_channels[0] * 4 __lowerCamelCase = TimestepEmbedding( in_channels=lowerCamelCase__ , time_embed_dim=lowerCamelCase__ , act_fn=lowerCamelCase__ , out_dim=block_out_channels[0] , ) __lowerCamelCase = nn.ModuleList([] ) __lowerCamelCase = None __lowerCamelCase = nn.ModuleList([] ) __lowerCamelCase = None # down __lowerCamelCase = in_channels for i, down_block_type in enumerate(lowerCamelCase__ ): __lowerCamelCase = output_channel __lowerCamelCase = block_out_channels[i] if i == 0: input_channel += extra_in_channels __lowerCamelCase = i == len(lowerCamelCase__ ) - 1 __lowerCamelCase = get_down_block( lowerCamelCase__ , num_layers=lowerCamelCase__ , in_channels=lowerCamelCase__ , out_channels=lowerCamelCase__ , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(lowerCamelCase__ ) # mid __lowerCamelCase = get_mid_block( lowerCamelCase__ , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=lowerCamelCase__ , add_downsample=lowerCamelCase__ , ) # up __lowerCamelCase = list(reversed(lowerCamelCase__ ) ) __lowerCamelCase = reversed_block_out_channels[0] if out_block_type is None: __lowerCamelCase = out_channels else: __lowerCamelCase = block_out_channels[0] for i, up_block_type in enumerate(lowerCamelCase__ ): __lowerCamelCase = output_channel __lowerCamelCase = ( reversed_block_out_channels[i + 1] if i < len(lowerCamelCase__ ) - 1 else final_upsample_channels ) __lowerCamelCase = i == len(lowerCamelCase__ ) - 1 __lowerCamelCase = get_up_block( lowerCamelCase__ , num_layers=lowerCamelCase__ , in_channels=lowerCamelCase__ , out_channels=lowerCamelCase__ , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(lowerCamelCase__ ) __lowerCamelCase = output_channel # out __lowerCamelCase = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) __lowerCamelCase = get_out_block( out_block_type=lowerCamelCase__ , num_groups_out=lowerCamelCase__ , embed_dim=block_out_channels[0] , out_channels=lowerCamelCase__ , act_fn=lowerCamelCase__ , fc_dim=block_out_channels[-1] // 4 , ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = True , ): '''simple docstring''' __lowerCamelCase = timestep if not torch.is_tensor(lowerCamelCase__ ): __lowerCamelCase = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(lowerCamelCase__ ) and len(timesteps.shape ) == 0: __lowerCamelCase = timesteps[None].to(sample.device ) __lowerCamelCase = self.time_proj(lowerCamelCase__ ) if self.config.use_timestep_embedding: __lowerCamelCase = self.time_mlp(lowerCamelCase__ ) else: __lowerCamelCase = timestep_embed[..., None] __lowerCamelCase = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) __lowerCamelCase = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down __lowerCamelCase = () for downsample_block in self.down_blocks: __lowerCamelCase = downsample_block(hidden_states=lowerCamelCase__ , temb=lowerCamelCase__ ) down_block_res_samples += res_samples # 3. mid if self.mid_block: __lowerCamelCase = self.mid_block(lowerCamelCase__ , lowerCamelCase__ ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): __lowerCamelCase = down_block_res_samples[-1:] __lowerCamelCase = down_block_res_samples[:-1] __lowerCamelCase = upsample_block(lowerCamelCase__ , res_hidden_states_tuple=lowerCamelCase__ , temb=lowerCamelCase__ ) # 5. post-process if self.out_block: __lowerCamelCase = self.out_block(lowerCamelCase__ , lowerCamelCase__ ) if not return_dict: return (sample,) return UNetaDOutput(sample=lowerCamelCase__ )
718
def a__ ( _UpperCamelCase : str ,_UpperCamelCase : str = " " ): __lowerCamelCase = [] __lowerCamelCase = 0 for index, char in enumerate(_UpperCamelCase ): if char == separator: split_words.append(string[last_index:index] ) __lowerCamelCase = index + 1 elif index + 1 == len(_UpperCamelCase ): split_words.append(string[last_index : index + 1] ) return split_words if __name__ == "__main__": from doctest import testmod testmod()
622
0
import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class __lowerCAmelCase : def __init__( self , __UpperCAmelCase = "cpu" , __UpperCAmelCase = "openai/clip-vit-large-patch14" ): '''simple docstring''' __lowerCamelCase = device __lowerCamelCase = CLIPTokenizerFast.from_pretrained(lowerCAmelCase_ ) __lowerCamelCase = [0.48_145_466, 0.4_578_275, 0.40_821_073] __lowerCamelCase = [0.26_862_954, 0.26_130_258, 0.27_577_711] __lowerCamelCase = torchvision.transforms.Normalize(self.image_mean , self.image_std ) __lowerCamelCase = torchvision.transforms.Resize(224 ) __lowerCamelCase = torchvision.transforms.CenterCrop(224 ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.resize(lowerCAmelCase_ ) __lowerCamelCase = self.center_crop(lowerCAmelCase_ ) __lowerCamelCase = self.normalize(lowerCAmelCase_ ) return images def __call__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , **__UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.tokenizer(text=lowerCAmelCase_ , **lowerCAmelCase_ ) __lowerCamelCase = self.preprocess_img(lowerCAmelCase_ ) __lowerCamelCase = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class __lowerCAmelCase ( nn.Module ): def __init__( self , __UpperCAmelCase=10 , __UpperCAmelCase=0.01 , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase="image" , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=False , __UpperCAmelCase=False , ): '''simple docstring''' super().__init__() __lowerCamelCase = None __lowerCamelCase = device if device else get_device() if vqgan: __lowerCamelCase = vqgan else: __lowerCamelCase = load_vqgan(self.device , conf_path=lowerCAmelCase_ , ckpt_path=lowerCAmelCase_ ) self.vqgan.eval() if clip: __lowerCamelCase = clip else: __lowerCamelCase = CLIPModel.from_pretrained('''openai/clip-vit-base-patch32''' ) self.clip.to(self.device ) __lowerCamelCase = ProcessorGradientFlow(device=self.device ) __lowerCamelCase = iterations __lowerCamelCase = lr __lowerCamelCase = log __lowerCamelCase = make_grid __lowerCamelCase = return_val __lowerCamelCase = quantize __lowerCamelCase = self.vqgan.decoder.z_shape def lowerCamelCase ( self , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=5 , __UpperCAmelCase=True ): '''simple docstring''' __lowerCamelCase = [] if output_path is None: __lowerCamelCase = '''./animation.gif''' if input_path is None: __lowerCamelCase = self.save_path __lowerCamelCase = sorted(glob(input_path + '''/*''' ) ) if not len(lowerCAmelCase_ ): raise ValueError( '''No images found in save path, aborting (did you pass save_intermediate=True to the generate''' ''' function?)''' ) if len(lowerCAmelCase_ ) == 1: print('''Only one image found in save path, (did you pass save_intermediate=True to the generate function?)''' ) __lowerCamelCase = total_duration / len(lowerCAmelCase_ ) __lowerCamelCase = [frame_duration] * len(lowerCAmelCase_ ) if extend_frames: __lowerCamelCase = 1.5 __lowerCamelCase = 3 for file_name in paths: if file_name.endswith('''.png''' ): images.append(imageio.imread(lowerCAmelCase_ ) ) imageio.mimsave(lowerCAmelCase_ , lowerCAmelCase_ , duration=lowerCAmelCase_ ) print(F"""gif saved to {output_path}""" ) def lowerCamelCase ( self , __UpperCAmelCase=None , __UpperCAmelCase=None ): '''simple docstring''' if not (path or img): raise ValueError('''Input either path or tensor''' ) if img is not None: raise NotImplementedError __lowerCamelCase = preprocess(Image.open(lowerCAmelCase_ ) , target_image_size=256 ).to(self.device ) __lowerCamelCase = preprocess_vqgan(lowerCAmelCase_ ) __lowerCamelCase ,*__lowerCamelCase = self.vqgan.encode(lowerCAmelCase_ ) return z def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.latent.detach().requires_grad_() __lowerCamelCase = base_latent + transform_vector if self.quantize: __lowerCamelCase ,*__lowerCamelCase = self.vqgan.quantize(lowerCAmelCase_ ) else: __lowerCamelCase = trans_latent return self.vqgan.decode(lowerCAmelCase_ ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None ): '''simple docstring''' __lowerCamelCase = self.clip_preprocessor(text=lowerCAmelCase_ , images=lowerCAmelCase_ , return_tensors='''pt''' , padding=lowerCAmelCase_ ) __lowerCamelCase = self.clip(**lowerCAmelCase_ ) __lowerCamelCase = clip_outputs.logits_per_image if weights is not None: __lowerCamelCase = similarity_logits * weights return similarity_logits.sum() def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self._get_clip_similarity(pos_prompts['''prompts'''] , lowerCAmelCase_ , weights=(1 / pos_prompts['''weights''']) ) if neg_prompts: __lowerCamelCase = self._get_clip_similarity(neg_prompts['''prompts'''] , lowerCAmelCase_ , weights=neg_prompts['''weights'''] ) else: __lowerCamelCase = torch.tensor([1] , device=self.device ) __lowerCamelCase = -torch.log(lowerCAmelCase_ ) + torch.log(lowerCAmelCase_ ) return loss def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = torch.randn_like(self.latent , requires_grad=lowerCAmelCase_ , device=self.device ) __lowerCamelCase = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() __lowerCamelCase = self._add_vector(lowerCAmelCase_ ) __lowerCamelCase = loop_post_process(lowerCAmelCase_ ) __lowerCamelCase = self._get_CLIP_loss(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) print('''CLIP loss''' , lowerCAmelCase_ ) if self.log: wandb.log({'''CLIP Loss''': clip_loss} ) clip_loss.backward(retain_graph=lowerCAmelCase_ ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' wandb.init(reinit=lowerCAmelCase_ , project='''face-editor''' ) wandb.config.update({'''Positive Prompts''': positive_prompts} ) wandb.config.update({'''Negative Prompts''': negative_prompts} ) wandb.config.update({'''lr''': self.lr, '''iterations''': self.iterations} ) if image_path: __lowerCamelCase = Image.open(lowerCAmelCase_ ) __lowerCamelCase = image.resize((256, 256) ) wandb.log('''Original Image''' , wandb.Image(lowerCAmelCase_ ) ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' if not prompts: return [] __lowerCamelCase = [] __lowerCamelCase = [] if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): __lowerCamelCase = [prompt.strip() for prompt in prompts.split('''|''' )] for prompt in prompts: if isinstance(lowerCAmelCase_ , (tuple, list) ): __lowerCamelCase = prompt[0] __lowerCamelCase = float(prompt[1] ) elif ":" in prompt: __lowerCamelCase ,__lowerCamelCase = prompt.split(''':''' ) __lowerCamelCase = float(lowerCAmelCase_ ) else: __lowerCamelCase = prompt __lowerCamelCase = 1.0 processed_prompts.append(lowerCAmelCase_ ) weights.append(lowerCAmelCase_ ) return { "prompts": processed_prompts, "weights": torch.tensor(lowerCAmelCase_ , device=self.device ), } def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=None , ): '''simple docstring''' if image_path: __lowerCamelCase = self._get_latent(lowerCAmelCase_ ) else: __lowerCamelCase = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) assert pos_prompts, "You must provide at least one positive prompt." __lowerCamelCase = self.process_prompts(lowerCAmelCase_ ) __lowerCamelCase = self.process_prompts(lowerCAmelCase_ ) if save_final and save_path is None: __lowerCamelCase = os.path.join('''./outputs/''' , '''_'''.join(pos_prompts['''prompts'''] ) ) if not os.path.exists(lowerCAmelCase_ ): os.makedirs(lowerCAmelCase_ ) else: __lowerCamelCase = save_path + '''_''' + get_timestamp() os.makedirs(lowerCAmelCase_ ) __lowerCamelCase = save_path __lowerCamelCase = self.vqgan.decode(self.latent )[0] if show_intermediate: print('''Original Image''' ) show_pil(custom_to_pil(lowerCAmelCase_ ) ) __lowerCamelCase = loop_post_process(lowerCAmelCase_ ) for iter, transformed_img in enumerate(self._optimize_CLIP(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ): if show_intermediate: show_pil(lowerCAmelCase_ ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , F"""iter_{iter:03d}.png""" ) ) if self.log: wandb.log({'''Image''': wandb.Image(lowerCAmelCase_ )} ) if show_final: show_pil(lowerCAmelCase_ ) if save_final: transformed_img.save(os.path.join(self.save_path , F"""iter_{iter:03d}_final.png""" ) )
719
import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class __lowerCAmelCase ( lowerCAmelCase__ ): def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = tempfile.mkdtemp() __lowerCamelCase = 8 # DPR tok __lowerCamelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] __lowerCamelCase = os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) __lowerCamelCase = os.path.join(__UpperCAmelCase , DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok __lowerCamelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] __lowerCamelCase = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) __lowerCamelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] __lowerCamelCase = {'''unk_token''': '''<unk>'''} __lowerCamelCase = os.path.join(self.tmpdirname , '''bart_tokenizer''' ) os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) __lowerCamelCase = os.path.join(__UpperCAmelCase , BART_VOCAB_FILES_NAMES['''vocab_file'''] ) __lowerCamelCase = os.path.join(__UpperCAmelCase , BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(__UpperCAmelCase ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(__UpperCAmelCase ) ) def lowerCamelCase ( self ): '''simple docstring''' return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''bart_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_dataset() __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: __lowerCamelCase = dataset __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.get_dummy_dataset() __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''custom''' , ) if from_disk: __lowerCamelCase = os.path.join(self.tmpdirname , '''dataset''' ) __lowerCamelCase = os.path.join(self.tmpdirname , '''index.faiss''' ) dataset.get_index('''embeddings''' ).save(os.path.join(self.tmpdirname , '''index.faiss''' ) ) dataset.drop_index('''embeddings''' ) dataset.save_to_disk(os.path.join(self.tmpdirname , '''dataset''' ) ) del dataset __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , __UpperCAmelCase ) , ) return retriever def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) __lowerCamelCase = os.path.join(self.tmpdirname , '''hf_bert_base.hnswSQ8_correct_phi_128.c_index''' ) dataset.save_faiss_index('''embeddings''' , index_file_name + '''.index.dpr''' ) pickle.dump(dataset['''id'''] , open(index_file_name + '''.index_meta.dpr''' , '''wb''' ) ) __lowerCamelCase = os.path.join(self.tmpdirname , '''psgs_w100.tsv.pkl''' ) __lowerCamelCase = {sample['''id''']: [sample['''text'''], sample['''title''']] for sample in dataset} pickle.dump(__UpperCAmelCase , open(__UpperCAmelCase , '''wb''' ) ) __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''legacy''' , index_path=self.tmpdirname , ) __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: __lowerCamelCase = self.get_dummy_dataset() retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_legacy_index_retriever() __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''text'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''text'''][0] , '''bar''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''text'''][0] , '''foo''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def lowerCamelCase ( self ): '''simple docstring''' import torch __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() __lowerCamelCase = [[5, 7], [10, 11]] __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever(__UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = ( out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , np.ndarray ) __lowerCamelCase = retriever( __UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase , return_tensors='''pt''' , ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = ( # noqa: F841 out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], out['''doc_ids'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dpr_ctx_encoder_tokenizer() __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) retriever.set_ctx_encoder_tokenizer(__UpperCAmelCase ) __lowerCamelCase = [[5, 7], [10, 11]] __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever(__UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase ) self.assertEqual( len(__UpperCAmelCase ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('''tokenized_doc_ids''', '''tokenized_doc_attention_mask''') ) , __UpperCAmelCase ) # check for doc token related keys in dictionary.
622
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) a_ = { 'configuration_encodec': [ 'ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP', 'EncodecConfig', ], 'feature_extraction_encodec': ['EncodecFeatureExtractor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ 'ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST', 'EncodecModel', 'EncodecPreTrainedModel', ] if TYPE_CHECKING: from .configuration_encodec import ( ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP, EncodecConfig, ) from .feature_extraction_encodec import EncodecFeatureExtractor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encodec import ( ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST, EncodecModel, EncodecPreTrainedModel, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
720
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """sail/poolformer_s12""": """https://huggingface.co/sail/poolformer_s12/resolve/main/config.json""", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """poolformer""" def __init__( self , __UpperCAmelCase=3 , __UpperCAmelCase=16 , __UpperCAmelCase=16 , __UpperCAmelCase=3 , __UpperCAmelCase=4.0 , __UpperCAmelCase=[2, 2, 6, 2] , __UpperCAmelCase=[64, 128, 320, 512] , __UpperCAmelCase=[7, 3, 3, 3] , __UpperCAmelCase=[4, 2, 2, 2] , __UpperCAmelCase=[2, 1, 1, 1] , __UpperCAmelCase=4 , __UpperCAmelCase=0.0 , __UpperCAmelCase="gelu" , __UpperCAmelCase=True , __UpperCAmelCase=1E-5 , __UpperCAmelCase=0.02 , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = num_channels __lowerCamelCase = patch_size __lowerCamelCase = stride __lowerCamelCase = padding __lowerCamelCase = pool_size __lowerCamelCase = hidden_sizes __lowerCamelCase = mlp_ratio __lowerCamelCase = depths __lowerCamelCase = patch_sizes __lowerCamelCase = strides __lowerCamelCase = num_encoder_blocks __lowerCamelCase = drop_path_rate __lowerCamelCase = hidden_act __lowerCamelCase = use_layer_scale __lowerCamelCase = layer_scale_init_value __lowerCamelCase = initializer_range super().__init__(**__UpperCAmelCase ) class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = version.parse("""1.11""" ) @property def lowerCamelCase ( self ): '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def lowerCamelCase ( self ): '''simple docstring''' return 2E-3
622
0
import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": a_ = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, required=True, help="""Path to the checkpoint to convert.""" ) parser.add_argument( """--original_config_file""", type=str, required=True, help="""The YAML config file corresponding to the original architecture.""", ) parser.add_argument( """--num_in_channels""", default=None, type=int, help="""The number of input channels. If `None` number of input channels will be automatically inferred.""", ) parser.add_argument( """--image_size""", default=512, type=int, help=( """The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2""" """ Base. Use 768 for Stable Diffusion v2.""" ), ) parser.add_argument( """--extract_ema""", action="""store_true""", help=( """Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights""" """ or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield""" """ higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.""" ), ) parser.add_argument( """--upcast_attention""", action="""store_true""", help=( """Whether the attention computation should always be upcasted. This is necessary when running stable""" """ diffusion 2.1.""" ), ) parser.add_argument( """--from_safetensors""", action="""store_true""", help="""If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.""", ) parser.add_argument( """--to_safetensors""", action="""store_true""", help="""Whether to store pipeline in safetensors format or not.""", ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument("""--device""", type=str, help="""Device to use (e.g. cpu, cuda:0, cuda:1, etc.)""") def a__ ( _UpperCamelCase : Dict ): if string == "True": return True elif string == "False": return False else: raise ValueError(F"""could not parse string as bool {string}""" ) parser.add_argument( """--use_linear_projection""", help="""Override for use linear projection""", required=False, type=parse_bool ) parser.add_argument("""--cross_attention_dim""", help="""Override for cross attention_dim""", required=False, type=int) a_ = parser.parse_args() a_ = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
721
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """uclanlp/visualbert-vqa""": """https://huggingface.co/uclanlp/visualbert-vqa/resolve/main/config.json""", """uclanlp/visualbert-vqa-pre""": """https://huggingface.co/uclanlp/visualbert-vqa-pre/resolve/main/config.json""", """uclanlp/visualbert-vqa-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-vqa-coco-pre/resolve/main/config.json""" ), """uclanlp/visualbert-vcr""": """https://huggingface.co/uclanlp/visualbert-vcr/resolve/main/config.json""", """uclanlp/visualbert-vcr-pre""": """https://huggingface.co/uclanlp/visualbert-vcr-pre/resolve/main/config.json""", """uclanlp/visualbert-vcr-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-vcr-coco-pre/resolve/main/config.json""" ), """uclanlp/visualbert-nlvr2""": """https://huggingface.co/uclanlp/visualbert-nlvr2/resolve/main/config.json""", """uclanlp/visualbert-nlvr2-pre""": """https://huggingface.co/uclanlp/visualbert-nlvr2-pre/resolve/main/config.json""", """uclanlp/visualbert-nlvr2-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-nlvr2-coco-pre/resolve/main/config.json""" ) # See all VisualBERT models at https://huggingface.co/models?filter=visual_bert } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """visual_bert""" def __init__( self , __UpperCAmelCase=30522 , __UpperCAmelCase=768 , __UpperCAmelCase=512 , __UpperCAmelCase=12 , __UpperCAmelCase=12 , __UpperCAmelCase=3072 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase=1 , __UpperCAmelCase=0 , __UpperCAmelCase=2 , **__UpperCAmelCase , ): '''simple docstring''' super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) __lowerCamelCase = vocab_size __lowerCamelCase = max_position_embeddings __lowerCamelCase = hidden_size __lowerCamelCase = visual_embedding_dim __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = initializer_range __lowerCamelCase = type_vocab_size __lowerCamelCase = layer_norm_eps __lowerCamelCase = bypass_transformer __lowerCamelCase = special_visual_initialize
622
0
import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from datasets.iterable_dataset import IterableDataset from datasets.load import dataset_module_factory, import_main_class from datasets.utils.file_utils import cached_path a_ = [ {"""dataset""": """wikipedia""", """config_name""": """20220301.de"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.en"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.fr"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.frr"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.it"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.simple"""}, {"""dataset""": """snli""", """config_name""": """plain_text"""}, {"""dataset""": """eli5""", """config_name""": """LFQA_reddit"""}, {"""dataset""": """wiki40b""", """config_name""": """en"""}, {"""dataset""": """wiki_dpr""", """config_name""": """psgs_w100.nq.compressed"""}, {"""dataset""": """wiki_dpr""", """config_name""": """psgs_w100.nq.no_index"""}, {"""dataset""": """wiki_dpr""", """config_name""": """psgs_w100.multiset.no_index"""}, {"""dataset""": """natural_questions""", """config_name""": """default"""}, ] def a__ ( _UpperCamelCase : Dict=True ): if with_config: return [ { "testcase_name": d["dataset"] + "/" + d["config_name"], "dataset": d["dataset"], "config_name": d["config_name"], } for d in DATASETS_ON_HF_GCP ] else: return [ {"testcase_name": dataset, "dataset": dataset} for dataset in {d["dataset"] for d in DATASETS_ON_HF_GCP} ] @parameterized.named_parameters(list_datasets_on_hf_gcp_parameters(with_config=__a ) ) class __lowerCAmelCase ( __a ): lowerCAmelCase__ = None lowerCAmelCase__ = None def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' with TemporaryDirectory() as tmp_dir: __lowerCamelCase = dataset_module_factory(A__ , cache_dir=A__ ) __lowerCamelCase = import_main_class(dataset_module.module_path , dataset=A__ ) __lowerCamelCase = builder_cls( cache_dir=A__ , config_name=A__ , hash=dataset_module.hash , ) __lowerCamelCase = '''/'''.join( [ HF_GCP_BASE_URL, builder_instance._relative_data_dir(with_hash=A__ ).replace(os.sep , '''/''' ), config.DATASET_INFO_FILENAME, ] ) __lowerCamelCase = cached_path(A__ , cache_dir=A__ ) self.assertTrue(os.path.exists(A__ ) ) @pytest.mark.integration def a__ ( _UpperCamelCase : List[str] ): __lowerCamelCase = tmp_path_factory.mktemp('''test_hf_gcp''' ) / '''test_wikipedia_simple''' __lowerCamelCase = dataset_module_factory('''wikipedia''' ,cache_dir=_UpperCamelCase ) __lowerCamelCase = import_main_class(dataset_module.module_path ) __lowerCamelCase = builder_cls( cache_dir=_UpperCamelCase ,config_name='''20220301.frr''' ,hash=dataset_module.hash ,) # use the HF cloud storage, not the original download_and_prepare that uses apache-beam __lowerCamelCase = None builder_instance.download_and_prepare() __lowerCamelCase = builder_instance.as_dataset() assert ds @pytest.mark.integration def a__ ( _UpperCamelCase : Any ): __lowerCamelCase = dataset_module_factory('''wikipedia''' ,cache_dir=_UpperCamelCase ) __lowerCamelCase = import_main_class(dataset_module.module_path ,dataset=_UpperCamelCase ) __lowerCamelCase = builder_cls( cache_dir=_UpperCamelCase ,config_name='''20220301.frr''' ,hash=dataset_module.hash ,) __lowerCamelCase = builder_instance.as_streaming_dataset() assert ds assert isinstance(_UpperCamelCase ,_UpperCamelCase ) assert "train" in ds assert isinstance(ds['''train'''] ,_UpperCamelCase ) assert next(iter(ds['''train'''] ) )
700
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging a_ = logging.get_logger(__name__) a_ = {"""vocab_file""": """spiece.model"""} a_ = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", } } a_ = { """albert-base-v1""": 512, """albert-large-v1""": 512, """albert-xlarge-v1""": 512, """albert-xxlarge-v1""": 512, """albert-base-v2""": 512, """albert-large-v2""": 512, """albert-xlarge-v2""": 512, """albert-xxlarge-v2""": 512, } a_ = """▁""" class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , __UpperCAmelCase , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="<unk>" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="<pad>" , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[MASK]" , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. __lowerCamelCase = ( AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase , normalized=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token ) __lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__UpperCAmelCase , remove_space=__UpperCAmelCase , keep_accents=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) __lowerCamelCase = do_lower_case __lowerCamelCase = remove_space __lowerCamelCase = keep_accents __lowerCamelCase = vocab_file __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCAmelCase ) @property def lowerCamelCase ( self ): '''simple docstring''' return len(self.sp_model ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): '''simple docstring''' __lowerCamelCase = self.__dict__.copy() __lowerCamelCase = None return state def __setstate__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCamelCase = {} __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' if self.remove_space: __lowerCamelCase = ''' '''.join(inputs.strip().split() ) else: __lowerCamelCase = inputs __lowerCamelCase = outputs.replace('''``''' , '''"''' ).replace('''\'\'''' , '''"''' ) if not self.keep_accents: __lowerCamelCase = unicodedata.normalize('''NFKD''' , __UpperCAmelCase ) __lowerCamelCase = ''''''.join([c for c in outputs if not unicodedata.combining(__UpperCAmelCase )] ) if self.do_lower_case: __lowerCamelCase = outputs.lower() return outputs def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.preprocess_text(__UpperCAmelCase ) __lowerCamelCase = self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) __lowerCamelCase = [] for piece in pieces: if len(__UpperCAmelCase ) > 1 and piece[-1] == str(''',''' ) and piece[-2].isdigit(): __lowerCamelCase = self.sp_model.EncodeAsPieces(piece[:-1].replace(__UpperCAmelCase , '''''' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __lowerCamelCase = cur_pieces[1:] else: __lowerCamelCase = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__UpperCAmelCase ) else: new_pieces.append(__UpperCAmelCase ) return new_pieces def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.PieceToId(__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.IdToPiece(__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = '''''' __lowerCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__UpperCAmelCase ) + token __lowerCamelCase = True __lowerCamelCase = [] else: current_sub_tokens.append(__UpperCAmelCase ) __lowerCamelCase = False out_string += self.sp_model.decode(__UpperCAmelCase ) return out_string.strip() def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase ) if token_ids_a is not None: return [1] + ([0] * len(__UpperCAmelCase )) + [1] + ([0] * len(__UpperCAmelCase )) + [1] return [1] + ([0] * len(__UpperCAmelCase )) + [1] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' if not os.path.isdir(__UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCamelCase = os.path.join( __UpperCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , '''wb''' ) as fi: __lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,)
622
0
# flake8: noqa # Lint as: python3 from typing import Dict, List, Optional, Type from .. import config from ..utils import logging from .formatting import ( ArrowFormatter, CustomFormatter, Formatter, PandasFormatter, PythonFormatter, TensorFormatter, format_table, query_table, ) from .np_formatter import NumpyFormatter a_ = logging.get_logger(__name__) a_ = {} a_ = {} a_ = {} def snake_case_ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Optional[Any] = None ,): __lowerCamelCase = aliases if aliases is not None else [] if format_type in _FORMAT_TYPES: logger.warning( F"""Overwriting format type '{format_type}' ({_FORMAT_TYPES[format_type].__name__} -> {formatter_cls.__name__})""" ) __lowerCamelCase = formatter_cls for alias in set(aliases + [format_type] ): if alias in _FORMAT_TYPES_ALIASES: logger.warning( F"""Overwriting format type alias '{alias}' ({_FORMAT_TYPES_ALIASES[alias]} -> {format_type})""" ) __lowerCamelCase = format_type def snake_case_ ( _UpperCamelCase : List[str] ,_UpperCamelCase : List[Any] ,_UpperCamelCase : Any = None ): __lowerCamelCase = aliases if aliases is not None else [] for alias in set(aliases + [format_type] ): __lowerCamelCase = unavailable_error # Here we define all the available formatting functions that can be used by `Dataset.set_format` _register_formatter(PythonFormatter, None, aliases=["""python"""]) _register_formatter(ArrowFormatter, """arrow""", aliases=["""pa""", """pyarrow"""]) _register_formatter(NumpyFormatter, """numpy""", aliases=["""np"""]) _register_formatter(PandasFormatter, """pandas""", aliases=["""pd"""]) _register_formatter(CustomFormatter, """custom""") if config.TORCH_AVAILABLE: from .torch_formatter import TorchFormatter _register_formatter(TorchFormatter, """torch""", aliases=["""pt""", """pytorch"""]) else: a_ = ValueError("""PyTorch needs to be installed to be able to return PyTorch tensors.""") _register_unavailable_formatter(_torch_error, """torch""", aliases=["""pt""", """pytorch"""]) if config.TF_AVAILABLE: from .tf_formatter import TFFormatter _register_formatter(TFFormatter, """tensorflow""", aliases=["""tf"""]) else: a_ = ValueError("""Tensorflow needs to be installed to be able to return Tensorflow tensors.""") _register_unavailable_formatter(_tf_error, """tensorflow""", aliases=["""tf"""]) if config.JAX_AVAILABLE: from .jax_formatter import JaxFormatter _register_formatter(JaxFormatter, """jax""", aliases=[]) else: a_ = ValueError("""JAX needs to be installed to be able to return JAX arrays.""") _register_unavailable_formatter(_jax_error, """jax""", aliases=[]) def snake_case_ ( _UpperCamelCase : Any ): if format_type in _FORMAT_TYPES_ALIASES: return _FORMAT_TYPES_ALIASES[format_type] else: return format_type def snake_case_ ( _UpperCamelCase : Optional[Any] ,**_UpperCamelCase : List[str] ): __lowerCamelCase = get_format_type_from_alias(_SCREAMING_SNAKE_CASE ) if format_type in _FORMAT_TYPES: return _FORMAT_TYPES[format_type](**_SCREAMING_SNAKE_CASE ) if format_type in _FORMAT_TYPES_ALIASES_UNAVAILABLE: raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type] else: raise ValueError( F"""Return type should be None or selected in {list(type for type in _FORMAT_TYPES.keys() if type != None )}, but got '{format_type}'""" )
701
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed a_ = """true""" def a__ ( _UpperCamelCase : Union[str, Any] ,_UpperCamelCase : List[str]=82 ,_UpperCamelCase : Optional[Any]=16 ): set_seed(42 ) __lowerCamelCase = RegressionModel() __lowerCamelCase = deepcopy(_UpperCamelCase ) __lowerCamelCase = RegressionDataset(length=_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,batch_size=_UpperCamelCase ) model.to(accelerator.device ) __lowerCamelCase ,__lowerCamelCase = accelerator.prepare(_UpperCamelCase ,_UpperCamelCase ) return model, ddp_model, dataloader def a__ ( _UpperCamelCase : Accelerator ,_UpperCamelCase : str=False ): __lowerCamelCase = AutoTokenizer.from_pretrained('''hf-internal-testing/mrpc-bert-base-cased''' ) __lowerCamelCase = load_dataset('''glue''' ,'''mrpc''' ,split='''validation''' ) def tokenize_function(_UpperCamelCase : int ): __lowerCamelCase = tokenizer(examples['''sentence1'''] ,examples['''sentence2'''] ,truncation=_UpperCamelCase ,max_length=_UpperCamelCase ) return outputs with accelerator.main_process_first(): __lowerCamelCase = dataset.map( _UpperCamelCase ,batched=_UpperCamelCase ,remove_columns=['''idx''', '''sentence1''', '''sentence2'''] ,) __lowerCamelCase = tokenized_datasets.rename_column('''label''' ,'''labels''' ) def collate_fn(_UpperCamelCase : Any ): if use_longest: return tokenizer.pad(_UpperCamelCase ,padding='''longest''' ,return_tensors='''pt''' ) return tokenizer.pad(_UpperCamelCase ,padding='''max_length''' ,max_length=1_28 ,return_tensors='''pt''' ) return DataLoader(_UpperCamelCase ,shuffle=_UpperCamelCase ,collate_fn=_UpperCamelCase ,batch_size=16 ) def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : List[str] ): __lowerCamelCase = Accelerator(dispatch_batches=_UpperCamelCase ,split_batches=_UpperCamelCase ) __lowerCamelCase = get_dataloader(_UpperCamelCase ,not dispatch_batches ) __lowerCamelCase = AutoModelForSequenceClassification.from_pretrained( '''hf-internal-testing/mrpc-bert-base-cased''' ,return_dict=_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = accelerator.prepare(_UpperCamelCase ,_UpperCamelCase ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : Optional[Any] ,_UpperCamelCase : Union[str, Any] ): __lowerCamelCase = [] for batch in dataloader: __lowerCamelCase ,__lowerCamelCase = batch.values() with torch.no_grad(): __lowerCamelCase = model(_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) __lowerCamelCase ,__lowerCamelCase = [], [] for logit, targ in logits_and_targets: logits.append(_UpperCamelCase ) targs.append(_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = torch.cat(_UpperCamelCase ), torch.cat(_UpperCamelCase ) return logits, targs def a__ ( _UpperCamelCase : Accelerator ,_UpperCamelCase : List[Any]=82 ,_UpperCamelCase : str=False ,_UpperCamelCase : List[str]=False ,_UpperCamelCase : Optional[int]=16 ): __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = get_basic_setup(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = generate_predictions(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ) assert ( len(_UpperCamelCase ) == num_samples ), F"""Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(_UpperCamelCase )}""" def a__ ( _UpperCamelCase : bool = False ,_UpperCamelCase : bool = False ): __lowerCamelCase = evaluate.load('''glue''' ,'''mrpc''' ) __lowerCamelCase ,__lowerCamelCase = get_mrpc_setup(_UpperCamelCase ,_UpperCamelCase ) # First do baseline __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = setup['''no'''] model.to(_UpperCamelCase ) model.eval() for batch in dataloader: batch.to(_UpperCamelCase ) with torch.inference_mode(): __lowerCamelCase = model(**_UpperCamelCase ) __lowerCamelCase = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=_UpperCamelCase ,references=batch['''labels'''] ) __lowerCamelCase = metric.compute() # Then do distributed __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = setup['''ddp'''] model.eval() for batch in dataloader: with torch.inference_mode(): __lowerCamelCase = model(**_UpperCamelCase ) __lowerCamelCase = outputs.logits.argmax(dim=-1 ) __lowerCamelCase = batch['''labels'''] __lowerCamelCase ,__lowerCamelCase = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=_UpperCamelCase ,references=_UpperCamelCase ) __lowerCamelCase = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] ,distributed[key] ), F"""Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n""" def a__ ( ): __lowerCamelCase = Accelerator(split_batches=_UpperCamelCase ,dispatch_batches=_UpperCamelCase ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print('''**Testing gather_for_metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(F"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`""" ) test_mrpc(_UpperCamelCase ,_UpperCamelCase ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test torch metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: __lowerCamelCase = Accelerator(split_batches=_UpperCamelCase ,dispatch_batches=_UpperCamelCase ) if accelerator.is_local_main_process: print(F"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99""" ) test_torch_metrics(_UpperCamelCase ,99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test last batch is not dropped when perfectly divisible**''' ) __lowerCamelCase = Accelerator() test_torch_metrics(_UpperCamelCase ,5_12 ) accelerator.state._reset_state() def a__ ( _UpperCamelCase : Optional[int] ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
622
0
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __lowerCAmelCase ( __snake_case ): lowerCAmelCase__ = ["""image_processor""", """tokenizer"""] lowerCAmelCase__ = """AutoImageProcessor""" lowerCAmelCase__ = """AutoTokenizer""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' super().__init__(__UpperCamelCase , __UpperCamelCase ) __lowerCamelCase = self.image_processor def __call__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=None , **__UpperCAmelCase ): '''simple docstring''' if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: __lowerCamelCase = self.tokenizer(__UpperCamelCase , return_tensors=__UpperCamelCase , **__UpperCamelCase ) if images is not None: __lowerCamelCase = self.image_processor(__UpperCamelCase , return_tensors=__UpperCamelCase , **__UpperCamelCase ) if text is not None and images is not None: __lowerCamelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__UpperCamelCase ) , tensor_type=__UpperCamelCase ) def lowerCamelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' return self.tokenizer.batch_decode(*__UpperCamelCase , **__UpperCamelCase ) def lowerCamelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' return self.tokenizer.decode(*__UpperCamelCase , **__UpperCamelCase ) @property def lowerCamelCase ( self ): '''simple docstring''' return ["input_ids", "attention_mask", "pixel_values"]
702
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = StableDiffusionXLImgaImgPipeline lowerCAmelCase__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} lowerCAmelCase__ = PipelineTesterMixin.required_optional_params - {"""latents"""} lowerCAmelCase__ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowerCAmelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS lowerCAmelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowerCamelCase ( self ): '''simple docstring''' torch.manual_seed(0 ) __lowerCamelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=__UpperCAmelCase , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) __lowerCamelCase = EulerDiscreteScheduler( beta_start=0.00_085 , beta_end=0.012 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) __lowerCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='''gelu''' , projection_dim=32 , ) __lowerCamelCase = CLIPTextModel(__UpperCAmelCase ) __lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__UpperCAmelCase ) __lowerCamelCase = CLIPTextModelWithProjection(__UpperCAmelCase ) __lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__UpperCAmelCase ) __lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase=0 ): '''simple docstring''' __lowerCamelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__UpperCAmelCase ) ).to(__UpperCAmelCase ) __lowerCamelCase = image / 2 + 0.5 if str(__UpperCAmelCase ).startswith('''mps''' ): __lowerCamelCase = torch.manual_seed(__UpperCAmelCase ) else: __lowerCamelCase = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) __lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.75, } return inputs def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = StableDiffusionXLImgaImgPipeline(**__UpperCAmelCase ) __lowerCamelCase = sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase = sd_pipe(**__UpperCAmelCase ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCamelCase = np.array([0.4_656, 0.4_840, 0.4_439, 0.6_698, 0.5_574, 0.4_524, 0.5_799, 0.5_943, 0.5_165] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase ( self ): '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def lowerCamelCase ( self ): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = StableDiffusionXLImgaImgPipeline(**__UpperCAmelCase ) __lowerCamelCase = sd_pipe.to(__UpperCAmelCase ) __lowerCamelCase = sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) # forward without prompt embeds __lowerCamelCase = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase = 3 * ['''this is a negative prompt'''] __lowerCamelCase = negative_prompt __lowerCamelCase = 3 * [inputs['''prompt''']] __lowerCamelCase = sd_pipe(**__UpperCAmelCase ) __lowerCamelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds __lowerCamelCase = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase = 3 * ['''this is a negative prompt'''] __lowerCamelCase = 3 * [inputs.pop('''prompt''' )] ( ( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) , ) = sd_pipe.encode_prompt(__UpperCAmelCase , negative_prompt=__UpperCAmelCase ) __lowerCamelCase = sd_pipe( **__UpperCAmelCase , prompt_embeds=__UpperCAmelCase , negative_prompt_embeds=__UpperCAmelCase , pooled_prompt_embeds=__UpperCAmelCase , negative_pooled_prompt_embeds=__UpperCAmelCase , ) __lowerCamelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase="cpu" , __UpperCAmelCase=torch.floataa , __UpperCAmelCase=0 ): '''simple docstring''' __lowerCamelCase = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) __lowerCamelCase = np.random.RandomState(__UpperCAmelCase ).standard_normal((1, 4, 64, 64) ) __lowerCamelCase = torch.from_numpy(__UpperCAmelCase ).to(device=__UpperCAmelCase , dtype=__UpperCAmelCase ) __lowerCamelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = self.get_inputs(__UpperCAmelCase ) __lowerCamelCase = pipe(**__UpperCAmelCase ).images __lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.49_493, 0.47_896, 0.40_798, 0.54_214, 0.53_212, 0.48_202, 0.47_656, 0.46_329, 0.48_506] ) assert np.abs(image_slice - expected_slice ).max() < 7E-3
622
0
import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class __lowerCAmelCase ( a__ ): lowerCAmelCase__ = """facebook/bart-large-mnli""" lowerCAmelCase__ = ( """This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which """ """should be the text to classify, and `labels`, which should be the list of labels to use for classification. """ """It returns the most likely label in the list of provided `labels` for the input text.""" ) lowerCAmelCase__ = """text_classifier""" lowerCAmelCase__ = AutoTokenizer lowerCAmelCase__ = AutoModelForSequenceClassification lowerCAmelCase__ = ["""text""", ["""text"""]] lowerCAmelCase__ = ["""text"""] def lowerCamelCase ( self ): '''simple docstring''' super().setup() __lowerCamelCase = self.model.config __lowerCamelCase = -1 for idx, label in config.idalabel.items(): if label.lower().startswith('''entail''' ): __lowerCamelCase = int(lowercase__ ) if self.entailment_id == -1: raise ValueError('''Could not determine the entailment ID from the model config, please pass it at init.''' ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = labels return self.pre_processor( [text] * len(lowercase__ ) , [F"""This example is {label}""" for label in labels] , return_tensors='''pt''' , padding='''max_length''' , ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = outputs.logits __lowerCamelCase = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
703
import torch from diffusers import StableDiffusionPipeline a_ = """path-to-your-trained-model""" a_ = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to("""cuda""") a_ = """A photo of sks dog in a bucket""" a_ = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save("""dog-bucket.png""")
622
0
import argparse import os import torch from transformers import FlavaConfig, FlavaForPreTraining from transformers.models.flava.convert_dalle_to_flava_codebook import convert_dalle_checkpoint def a__ ( _UpperCamelCase : Tuple ): return sum(param.float().sum() if '''encoder.embeddings''' not in key else 0 for key, param in state_dict.items() ) def a__ ( _UpperCamelCase : Any ,_UpperCamelCase : List[str] ): __lowerCamelCase = {} for key, value in state_dict.items(): if "text_encoder.embeddings" in key or "image_encoder.embeddings" in key: continue __lowerCamelCase = key.replace('''heads.cmd.mim_head.cls.predictions''' ,'''mmm_image_head''' ) __lowerCamelCase = key.replace('''heads.cmd.mlm_head.cls.predictions''' ,'''mmm_text_head''' ) __lowerCamelCase = key.replace('''heads.cmd.itm_head.cls''' ,'''itm_head''' ) __lowerCamelCase = key.replace('''heads.cmd.itm_head.pooler''' ,'''itm_head.pooler''' ) __lowerCamelCase = key.replace('''heads.cmd.clip_head.logit_scale''' ,'''flava.logit_scale''' ) __lowerCamelCase = key.replace('''heads.fairseq_mlm.cls.predictions''' ,'''mlm_head''' ) __lowerCamelCase = key.replace('''heads.imagenet.mim_head.cls.predictions''' ,'''mim_head''' ) __lowerCamelCase = key.replace('''mm_text_projection''' ,'''flava.text_to_mm_projection''' ) __lowerCamelCase = key.replace('''mm_image_projection''' ,'''flava.image_to_mm_projection''' ) __lowerCamelCase = key.replace('''image_encoder.module''' ,'''flava.image_model''' ) __lowerCamelCase = key.replace('''text_encoder.module''' ,'''flava.text_model''' ) __lowerCamelCase = key.replace('''mm_encoder.module.encoder.cls_token''' ,'''flava.multimodal_model.cls_token''' ) __lowerCamelCase = key.replace('''mm_encoder.module''' ,'''flava.multimodal_model''' ) __lowerCamelCase = key.replace('''text_projection''' ,'''flava.text_projection''' ) __lowerCamelCase = key.replace('''image_projection''' ,'''flava.image_projection''' ) __lowerCamelCase = value.float() for key, value in codebook_state_dict.items(): __lowerCamelCase = value return upgrade @torch.no_grad() def a__ ( _UpperCamelCase : Any ,_UpperCamelCase : List[str] ,_UpperCamelCase : str ,_UpperCamelCase : Tuple=None ): if config_path is not None: __lowerCamelCase = FlavaConfig.from_pretrained(SCREAMING_SNAKE_CASE__ ) else: __lowerCamelCase = FlavaConfig() __lowerCamelCase = FlavaForPreTraining(SCREAMING_SNAKE_CASE__ ).eval() __lowerCamelCase = convert_dalle_checkpoint(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,save_checkpoint=SCREAMING_SNAKE_CASE__ ) if os.path.exists(SCREAMING_SNAKE_CASE__ ): __lowerCamelCase = torch.load(SCREAMING_SNAKE_CASE__ ,map_location='''cpu''' ) else: __lowerCamelCase = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE__ ,map_location='''cpu''' ) __lowerCamelCase = upgrade_state_dict(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) hf_model.load_state_dict(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = hf_model.state_dict() __lowerCamelCase = count_parameters(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = count_parameters(SCREAMING_SNAKE_CASE__ ) + count_parameters(SCREAMING_SNAKE_CASE__ ) assert torch.allclose(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,atol=1e-3 ) hf_model.save_pretrained(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": a_ = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to flava checkpoint""") parser.add_argument("""--codebook_path""", default=None, type=str, help="""Path to flava codebook checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") a_ = parser.parse_args() convert_flava_checkpoint(args.checkpoint_path, args.codebook_path, args.pytorch_dump_folder_path, args.config_path)
704
import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class __lowerCAmelCase : @staticmethod def lowerCamelCase ( *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' pass def a__ ( _UpperCamelCase : List[str] ): return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. a_ = ( """https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png""" ) @is_pipeline_test @require_torch @require_vision class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model=__UpperCAmelCase , tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '''''' ) ) ) __lowerCamelCase = '''What is the placebo?''' __lowerCamelCase = [ { '''image''': load_image(__UpperCAmelCase ), '''question''': question, }, { '''image''': image, '''question''': question, }, { '''image''': image, '''question''': question, '''word_boxes''': word_boxes, }, ] return dqa_pipeline, examples def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = dqa_pipeline(__UpperCAmelCase , top_k=2 ) self.assertEqual( __UpperCAmelCase , [ [ {'''score''': ANY(__UpperCAmelCase ), '''answer''': ANY(__UpperCAmelCase ), '''start''': ANY(__UpperCAmelCase ), '''end''': ANY(__UpperCAmelCase )}, {'''score''': ANY(__UpperCAmelCase ), '''answer''': ANY(__UpperCAmelCase ), '''start''': ANY(__UpperCAmelCase ), '''end''': ANY(__UpperCAmelCase )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline('''document-question-answering''' , model='''hf-internal-testing/tiny-random-layoutlmv2''' ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''How many cats are there?''' __lowerCamelCase = [ {'''score''': 0.0_001, '''answer''': '''oy 2312/2019''', '''start''': 38, '''end''': 39}, {'''score''': 0.0_001, '''answer''': '''oy 2312/2019 DUE''', '''start''': 38, '''end''': 40}, ] __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , __UpperCAmelCase ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , __UpperCAmelCase ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably __lowerCamelCase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(__UpperCAmelCase , [] ) # We can optionnally pass directly the words and bounding boxes __lowerCamelCase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' __lowerCamelCase = [] __lowerCamelCase = [] __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , words=__UpperCAmelCase , boxes=__UpperCAmelCase , top_k=2 ) self.assertEqual(__UpperCAmelCase , [] ) @slow @require_torch @require_detectrona @require_pytesseract def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_944, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_009, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_944, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_009, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.9_944, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_009, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , max_seq_len=50 , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_974, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_948, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_974, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_948, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.9_974, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_948, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__UpperCAmelCase ) __lowerCamelCase = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__UpperCAmelCase , revision='''3dc6de3''' , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] ] * 2 , ) __lowerCamelCase = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '''''' ) ) ) # This model should also work if `image` is set to None __lowerCamelCase = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__UpperCAmelCase ) __lowerCamelCase = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__UpperCAmelCase , revision='''3dc6de3''' , max_seq_len=50 , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_999, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_998, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.9_999, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_998, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] ] * 2 , ) __lowerCamelCase = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '''''' ) ) ) # This model should also work if `image` is set to None __lowerCamelCase = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_999, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_998, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) @slow @require_torch def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model='''naver-clova-ix/donut-base-finetuned-docvqa''' , tokenizer=AutoTokenizer.from_pretrained('''naver-clova-ix/donut-base-finetuned-docvqa''' ) , feature_extractor='''naver-clova-ix/donut-base-finetuned-docvqa''' , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , [{'''answer''': '''us-001'''}] ) @require_tf @unittest.skip('''Document question answering not implemented in TF''' ) def lowerCamelCase ( self ): '''simple docstring''' pass
622
0
import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin a_ = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): lowerCAmelCase__ = XLMRobertaTokenizer lowerCAmelCase__ = XLMRobertaTokenizerFast lowerCAmelCase__ = True lowerCAmelCase__ = True def lowerCamelCase ( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __lowerCamelCase = XLMRobertaTokenizer(__snake_case , keep_accents=__snake_case ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''<pad>''' __lowerCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__snake_case ) , __snake_case ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__snake_case ) , __snake_case ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(__snake_case ) , 1002 ) def lowerCamelCase ( self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1002 ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = XLMRobertaTokenizer(__snake_case , keep_accents=__snake_case ) __lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__snake_case , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__snake_case ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __lowerCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __snake_case , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(__snake_case ) self.assertListEqual( __snake_case , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(__snake_case ) self.assertListEqual( __snake_case , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def lowerCamelCase ( self ): '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __lowerCamelCase = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-xlm-roberta''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __lowerCamelCase = self.rust_tokenizer_class.from_pretrained(__snake_case , **__snake_case ) __lowerCamelCase = self.tokenizer_class.from_pretrained(__snake_case , **__snake_case ) __lowerCamelCase = tempfile.mkdtemp() __lowerCamelCase = tokenizer_r.save_pretrained(__snake_case ) __lowerCamelCase = tokenizer_p.save_pretrained(__snake_case ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) __lowerCamelCase = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(__snake_case , __snake_case ) # Checks everything loads correctly in the same way __lowerCamelCase = tokenizer_r.from_pretrained(__snake_case ) __lowerCamelCase = tokenizer_p.from_pretrained(__snake_case ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__snake_case , __snake_case ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(__snake_case ) # Save tokenizer rust, legacy_format=True __lowerCamelCase = tempfile.mkdtemp() __lowerCamelCase = tokenizer_r.save_pretrained(__snake_case , legacy_format=__snake_case ) __lowerCamelCase = tokenizer_p.save_pretrained(__snake_case ) # Checks it save with the same files self.assertSequenceEqual(__snake_case , __snake_case ) # Checks everything loads correctly in the same way __lowerCamelCase = tokenizer_r.from_pretrained(__snake_case ) __lowerCamelCase = tokenizer_p.from_pretrained(__snake_case ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__snake_case , __snake_case ) ) shutil.rmtree(__snake_case ) # Save tokenizer rust, legacy_format=False __lowerCamelCase = tempfile.mkdtemp() __lowerCamelCase = tokenizer_r.save_pretrained(__snake_case , legacy_format=__snake_case ) __lowerCamelCase = tokenizer_p.save_pretrained(__snake_case ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way __lowerCamelCase = tokenizer_r.from_pretrained(__snake_case ) __lowerCamelCase = tokenizer_p.from_pretrained(__snake_case ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__snake_case , __snake_case ) ) shutil.rmtree(__snake_case ) @cached_property def lowerCamelCase ( self ): '''simple docstring''' return XLMRobertaTokenizer.from_pretrained('''xlm-roberta-base''' ) def lowerCamelCase ( self ): '''simple docstring''' with tempfile.NamedTemporaryFile() as f: shutil.copyfile(__snake_case , f.name ) __lowerCamelCase = XLMRobertaTokenizer(f.name , keep_accents=__snake_case ) __lowerCamelCase = pickle.dumps(__snake_case ) pickle.loads(__snake_case ) def lowerCamelCase ( self ): '''simple docstring''' if not self.test_rust_tokenizer: return __lowerCamelCase = self.get_tokenizer() __lowerCamelCase = self.get_rust_tokenizer() __lowerCamelCase = '''I was born in 92000, and this is falsé.''' __lowerCamelCase = tokenizer.tokenize(__snake_case ) __lowerCamelCase = rust_tokenizer.tokenize(__snake_case ) self.assertListEqual(__snake_case , __snake_case ) __lowerCamelCase = tokenizer.encode(__snake_case , add_special_tokens=__snake_case ) __lowerCamelCase = rust_tokenizer.encode(__snake_case , add_special_tokens=__snake_case ) self.assertListEqual(__snake_case , __snake_case ) __lowerCamelCase = self.get_rust_tokenizer() __lowerCamelCase = tokenizer.encode(__snake_case ) __lowerCamelCase = rust_tokenizer.encode(__snake_case ) self.assertListEqual(__snake_case , __snake_case ) @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''Hello World!''' __lowerCamelCase = [0, 35378, 6661, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__snake_case , self.big_tokenizer.encode(__snake_case ) ) @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) __lowerCamelCase = [ 0, 3293, 83, 10, 4552, 4989, 7986, 678, 10, 5915, 111, 179459, 124850, 4, 6044, 237, 12, 6, 5, 6, 4, 6780, 705, 15, 1388, 44, 378, 10114, 711, 152, 20, 6, 5, 22376, 642, 1221, 15190, 34153, 450, 5608, 959, 1119, 57702, 136, 186, 47, 1098, 29367, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6044, 237, 6284, 50901, 528, 31, 90, 34, 927, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__snake_case , self.big_tokenizer.encode(__snake_case ) ) @slow def lowerCamelCase ( self ): '''simple docstring''' # fmt: off __lowerCamelCase = {'''input_ids''': [[0, 11062, 82772, 7, 15, 82772, 538, 51529, 237, 17198, 1290, 206, 9, 215175, 1314, 136, 17198, 1290, 206, 9, 56359, 42, 122009, 9, 16466, 16, 87344, 4537, 9, 4717, 78381, 6, 159958, 7, 15, 24480, 618, 4, 527, 22693, 5428, 4, 2777, 24480, 9874, 4, 43523, 594, 4, 803, 18392, 33189, 18, 4, 43523, 24447, 12399, 100, 24955, 83658, 9626, 144057, 15, 839, 22335, 16, 136, 24955, 83658, 83479, 15, 39102, 724, 16, 678, 645, 2789, 1328, 4589, 42, 122009, 115774, 23, 805, 1328, 46876, 7, 136, 53894, 1940, 42227, 41159, 17721, 823, 425, 4, 27512, 98722, 206, 136, 5531, 4970, 919, 17336, 5, 2], [0, 20080, 618, 83, 82775, 47, 479, 9, 1517, 73, 53894, 333, 80581, 110117, 18811, 5256, 1295, 51, 152526, 297, 7986, 390, 124416, 538, 35431, 214, 98, 15044, 25737, 136, 7108, 43701, 23, 756, 135355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 581, 63773, 119455, 6, 147797, 88203, 7, 645, 70, 21, 3285, 10269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__snake_case , model_name='''xlm-roberta-base''' , revision='''d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3''' , )
705
import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin a_ = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = XLMProphetNetTokenizer lowerCAmelCase__ = False lowerCAmelCase__ = True def lowerCamelCase ( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __lowerCamelCase = XLMProphetNetTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''[PAD]''' __lowerCamelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCAmelCase ) , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''[PAD]''' ) self.assertEqual(vocab_keys[1] , '''[CLS]''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(__UpperCAmelCase ) , 1012 ) def lowerCamelCase ( self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1012 ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = XLMProphetNetTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) __lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__UpperCAmelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __lowerCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] , ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''[UNK]''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''[UNK]''', '''.''', ] , ) @cached_property def lowerCamelCase ( self ): '''simple docstring''' return XLMProphetNetTokenizer.from_pretrained('''microsoft/xprophetnet-large-wiki100-cased''' ) @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''Hello World!''' __lowerCamelCase = [35389, 6672, 49, 2] self.assertListEqual(__UpperCAmelCase , self.big_tokenizer.encode(__UpperCAmelCase ) ) @slow def lowerCamelCase ( self ): '''simple docstring''' # fmt: off __lowerCamelCase = {'''input_ids''': [[11073, 82783, 18, 26, 82783, 549, 51540, 248, 17209, 1301, 217, 20, 215186, 1325, 147, 17209, 1301, 217, 20, 56370, 53, 122020, 20, 16477, 27, 87355, 4548, 20, 4728, 78392, 17, 159969, 18, 26, 24491, 629, 15, 538, 22704, 5439, 15, 2788, 24491, 9885, 15, 43534, 605, 15, 814, 18403, 33200, 29, 15, 43534, 24458, 12410, 111, 24966, 83669, 9637, 144068, 26, 850, 22346, 27, 147, 24966, 83669, 83490, 26, 39113, 735, 27, 689, 656, 2800, 1339, 4600, 53, 122020, 115785, 34, 816, 1339, 46887, 18, 147, 53905, 1951, 42238, 41170, 17732, 834, 436, 15, 27523, 98733, 217, 147, 5542, 4981, 930, 17347, 16, 2], [20091, 629, 94, 82786, 58, 490, 20, 1528, 84, 53905, 344, 80592, 110128, 18822, 5267, 1306, 62, 152537, 308, 7997, 401, 124427, 549, 35442, 225, 109, 15055, 25748, 147, 7119, 43712, 34, 767, 135366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 63784, 119466, 17, 147808, 88214, 18, 656, 81, 32, 3296, 10280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__UpperCAmelCase , model_name='''microsoft/xprophetnet-large-wiki100-cased''' , revision='''1acad1643ddd54a44df6a1b797ada8373685d90e''' , )
622
0
import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal a_ = datasets.utils.logging.get_logger(__name__) a_ = ["""names""", """prefix"""] a_ = ["""warn_bad_lines""", """error_bad_lines""", """mangle_dupe_cols"""] a_ = ["""encoding_errors""", """on_bad_lines"""] a_ = ["""date_format"""] @dataclass class __lowerCAmelCase ( datasets.BuilderConfig ): lowerCAmelCase__ = "," lowerCAmelCase__ = None lowerCAmelCase__ = "infer" lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = True lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = False lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = True lowerCAmelCase__ = True lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = None lowerCAmelCase__ = "." lowerCAmelCase__ = None lowerCAmelCase__ = '"' lowerCAmelCase__ = 0 lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = True lowerCAmelCase__ = True lowerCAmelCase__ = 0 lowerCAmelCase__ = True lowerCAmelCase__ = False lowerCAmelCase__ = None lowerCAmelCase__ = 1_0_0_0_0 lowerCAmelCase__ = None lowerCAmelCase__ = "strict" lowerCAmelCase__ = "error" lowerCAmelCase__ = None def lowerCamelCase ( self ): '''simple docstring''' if self.delimiter is not None: __lowerCamelCase = self.delimiter if self.column_names is not None: __lowerCamelCase = self.column_names @property def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = { '''sep''': self.sep, '''header''': self.header, '''names''': self.names, '''index_col''': self.index_col, '''usecols''': self.usecols, '''prefix''': self.prefix, '''mangle_dupe_cols''': self.mangle_dupe_cols, '''engine''': self.engine, '''converters''': self.converters, '''true_values''': self.true_values, '''false_values''': self.false_values, '''skipinitialspace''': self.skipinitialspace, '''skiprows''': self.skiprows, '''nrows''': self.nrows, '''na_values''': self.na_values, '''keep_default_na''': self.keep_default_na, '''na_filter''': self.na_filter, '''verbose''': self.verbose, '''skip_blank_lines''': self.skip_blank_lines, '''thousands''': self.thousands, '''decimal''': self.decimal, '''lineterminator''': self.lineterminator, '''quotechar''': self.quotechar, '''quoting''': self.quoting, '''escapechar''': self.escapechar, '''comment''': self.comment, '''encoding''': self.encoding, '''dialect''': self.dialect, '''error_bad_lines''': self.error_bad_lines, '''warn_bad_lines''': self.warn_bad_lines, '''skipfooter''': self.skipfooter, '''doublequote''': self.doublequote, '''memory_map''': self.memory_map, '''float_precision''': self.float_precision, '''chunksize''': self.chunksize, '''encoding_errors''': self.encoding_errors, '''on_bad_lines''': self.on_bad_lines, '''date_format''': self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , __UpperCAmelCase ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __lowerCAmelCase ( datasets.ArrowBasedBuilder ): lowerCAmelCase__ = CsvConfig def lowerCamelCase ( self ): '''simple docstring''' return datasets.DatasetInfo(features=self.config.features ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) __lowerCamelCase = dl_manager.download_and_extract(self.config.data_files ) if isinstance(__UpperCAmelCase , (str, list, tuple) ): __lowerCamelCase = data_files if isinstance(__UpperCAmelCase , __UpperCAmelCase ): __lowerCamelCase = [files] __lowerCamelCase = [dl_manager.iter_files(__UpperCAmelCase ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''files''': files} )] __lowerCamelCase = [] for split_name, files in data_files.items(): if isinstance(__UpperCAmelCase , __UpperCAmelCase ): __lowerCamelCase = [files] __lowerCamelCase = [dl_manager.iter_files(__UpperCAmelCase ) for file in files] splits.append(datasets.SplitGenerator(name=__UpperCAmelCase , gen_kwargs={'''files''': files} ) ) return splits def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' if self.config.features is not None: __lowerCamelCase = self.config.features.arrow_schema if all(not require_storage_cast(__UpperCAmelCase ) for feature in self.config.features.values() ): # cheaper cast __lowerCamelCase = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=__UpperCAmelCase ) else: # more expensive cast; allows str <-> int/float or str to Audio for example __lowerCamelCase = table_cast(__UpperCAmelCase , __UpperCAmelCase ) return pa_table def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str __lowerCamelCase = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(__UpperCAmelCase ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(__UpperCAmelCase ) ): __lowerCamelCase = pd.read_csv(__UpperCAmelCase , iterator=__UpperCAmelCase , dtype=__UpperCAmelCase , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(__UpperCAmelCase ): __lowerCamelCase = pa.Table.from_pandas(__UpperCAmelCase ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(__UpperCAmelCase ) except ValueError as e: logger.error(F"""Failed to read file '{file}' with error {type(__UpperCAmelCase )}: {e}""" ) raise
706
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py a_ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. a_ = direct_transformers_import(PATH_TO_TRANSFORMERS) a_ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` a_ = re.compile(R"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") a_ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def a__ ( _UpperCamelCase : Union[str, Any] ): __lowerCamelCase = None # source code of `config_class` __lowerCamelCase = inspect.getsource(_UpperCamelCase ) __lowerCamelCase = _re_checkpoint.findall(_UpperCamelCase ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('''/''' ): __lowerCamelCase = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link __lowerCamelCase = F"""https://huggingface.co/{ckpt_name}""" if ckpt_link == ckpt_link_from_name: __lowerCamelCase = ckpt_name break return checkpoint def a__ ( ): __lowerCamelCase = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue __lowerCamelCase = get_checkpoint_from_config_class(_UpperCamelCase ) __lowerCamelCase = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(_UpperCamelCase ) if len(_UpperCamelCase ) > 0: __lowerCamelCase = '''\n'''.join(sorted(_UpperCamelCase ) ) raise ValueError(F"""The following configurations don't contain any valid checkpoint:\n{message}""" ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
622
0
from __future__ import annotations from collections import Counter from random import random class __lowerCAmelCase : def __init__( self ): '''simple docstring''' __lowerCamelCase = {} def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = {} def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' if nodea not in self.connections: self.add_node(__a ) if nodea not in self.connections: self.add_node(__a ) __lowerCamelCase = probability def lowerCamelCase ( self ): '''simple docstring''' return list(self.connections ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = 0 __lowerCamelCase = random() for dest in self.connections[node]: current_probability += self.connections[node][dest] if current_probability > random_value: return dest return "" def a__ ( _UpperCamelCase : List[str] ,_UpperCamelCase : int ,_UpperCamelCase : Union[str, Any] ): __lowerCamelCase = MarkovChainGraphUndirectedUnweighted() for nodea, nodea, probability in transitions: graph.add_transition_probability(lowercase_ ,lowercase_ ,lowercase_ ) __lowerCamelCase = Counter(graph.get_nodes() ) __lowerCamelCase = start for _ in range(lowercase_ ): __lowerCamelCase = graph.transition(lowercase_ ) visited[node] += 1 return visited if __name__ == "__main__": import doctest doctest.testmod()
707
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a_ = { """configuration_clipseg""": [ """CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP""", """CLIPSegConfig""", """CLIPSegTextConfig""", """CLIPSegVisionConfig""", ], """processing_clipseg""": ["""CLIPSegProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ """CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST""", """CLIPSegModel""", """CLIPSegPreTrainedModel""", """CLIPSegTextModel""", """CLIPSegVisionModel""", """CLIPSegForImageSegmentation""", ] if TYPE_CHECKING: from .configuration_clipseg import ( CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPSegConfig, CLIPSegTextConfig, CLIPSegVisionConfig, ) from .processing_clipseg import CLIPSegProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clipseg import ( CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPSegForImageSegmentation, CLIPSegModel, CLIPSegPreTrainedModel, CLIPSegTextModel, CLIPSegVisionModel, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
622
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class __lowerCAmelCase ( UpperCAmelCase_ ): lowerCAmelCase__ = "naver-clova-ix/donut-base-finetuned-docvqa" lowerCAmelCase__ = ( "This is a tool that answers a question about an document (pdf). It takes an input named `document` which " "should be the document containing the information, as well as a `question` that is the question about the " "document. It returns a text that contains the answer to the question." ) lowerCAmelCase__ = "document_qa" lowerCAmelCase__ = AutoProcessor lowerCAmelCase__ = VisionEncoderDecoderModel lowerCAmelCase__ = ["image", "text"] lowerCAmelCase__ = ["text"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' if not is_vision_available(): raise ValueError('''Pillow must be installed to use the DocumentQuestionAnsweringTool.''' ) super().__init__(*_snake_case , **_snake_case ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = '''<s_docvqa><s_question>{user_input}</s_question><s_answer>''' __lowerCamelCase = task_prompt.replace('''{user_input}''' , _snake_case ) __lowerCamelCase = self.pre_processor.tokenizer( _snake_case , add_special_tokens=_snake_case , return_tensors='''pt''' ).input_ids __lowerCamelCase = self.pre_processor(_snake_case , return_tensors='''pt''' ).pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.model.generate( inputs['''pixel_values'''].to(self.device ) , decoder_input_ids=inputs['''decoder_input_ids'''].to(self.device ) , max_length=self.model.decoder.config.max_position_embeddings , early_stopping=_snake_case , pad_token_id=self.pre_processor.tokenizer.pad_token_id , eos_token_id=self.pre_processor.tokenizer.eos_token_id , use_cache=_snake_case , num_beams=1 , bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] , return_dict_in_generate=_snake_case , ).sequences def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.pre_processor.batch_decode(_snake_case )[0] __lowerCamelCase = sequence.replace(self.pre_processor.tokenizer.eos_token , '''''' ) __lowerCamelCase = sequence.replace(self.pre_processor.tokenizer.pad_token , '''''' ) __lowerCamelCase = re.sub(r'''<.*?>''' , '''''' , _snake_case , count=1 ).strip() # remove first task start token __lowerCamelCase = self.pre_processor.tokenajson(_snake_case ) return sequence["answer"]
708
import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = RoFormerTokenizer lowerCAmelCase__ = RoFormerTokenizerFast lowerCAmelCase__ = True lowerCAmelCase__ = True def lowerCamelCase ( self ): '''simple docstring''' super().setUp() def lowerCamelCase ( self , **__UpperCAmelCase ): '''simple docstring''' return self.tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **__UpperCAmelCase ) def lowerCamelCase ( self , **__UpperCAmelCase ): '''simple docstring''' return self.rust_tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''永和服装饰品有限公司,今天天气非常好''' __lowerCamelCase = '''永和 服装 饰品 有限公司 , 今 天 天 气 非常 好''' return input_text, output_text def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_tokenizer() __lowerCamelCase ,__lowerCamelCase = self.get_chinese_input_output_texts() __lowerCamelCase = tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , output_text.split() ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [22943, 21332, 34431, 45904, 117, 306, 1231, 1231, 2653, 33994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_rust_tokenizer() __lowerCamelCase ,__lowerCamelCase = self.get_chinese_input_output_texts() __lowerCamelCase = tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , output_text.split() ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [22943, 21332, 34431, 45904, 117, 306, 1231, 1231, 2653, 33994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' pass
622
0
import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def a__ ( _UpperCamelCase : str=None ,_UpperCamelCase : int=None ): return field(default_factory=lambda: default ,metadata=_UpperCamelCase ) @dataclass class __lowerCAmelCase : lowerCAmelCase__ = field( metadata={"""help""": """The csv file to plot."""} , ) lowerCAmelCase__ = field( default=_lowerCAmelCase , metadata={"""help""": """Whether to plot along batch size or sequence length. Defaults to sequence length."""} , ) lowerCAmelCase__ = field( default=_lowerCAmelCase , metadata={"""help""": """Whether the csv file has time results or memory results. Defaults to memory results."""} , ) lowerCAmelCase__ = field( default=_lowerCAmelCase , metadata={"""help""": """Disable logarithmic scale when plotting"""} , ) lowerCAmelCase__ = field( default=_lowerCAmelCase , metadata={ """help""": """Whether the csv file has training results or inference results. Defaults to inference results.""" } , ) lowerCAmelCase__ = field( default=_lowerCAmelCase , metadata={"""help""": """Filename under which the plot will be saved. If unused no plot is saved."""} , ) lowerCAmelCase__ = list_field( default=_lowerCAmelCase , metadata={"""help""": """List of model names that are used instead of the ones in the csv file."""} ) def a__ ( _UpperCamelCase : Dict ): try: int(_UpperCamelCase ) return True except ValueError: return False def a__ ( _UpperCamelCase : List[str] ): try: float(_UpperCamelCase ) return True except ValueError: return False class __lowerCAmelCase : def __init__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = args __lowerCamelCase = defaultdict(lambda: {"bsz": [], "seq_len": [], "result": {}} ) with open(self.args.csv_file , newline='''''' ) as csv_file: __lowerCamelCase = csv.DictReader(_lowerCAmelCase ) for row in reader: __lowerCamelCase = row['''model'''] self.result_dict[model_name]["bsz"].append(int(row['''batch_size'''] ) ) self.result_dict[model_name]["seq_len"].append(int(row['''sequence_length'''] ) ) if can_convert_to_int(row['''result'''] ): # value is not None __lowerCamelCase = int(row['''result'''] ) elif can_convert_to_float(row['''result'''] ): # value is not None __lowerCamelCase = float(row['''result'''] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase ,__lowerCamelCase = plt.subplots() __lowerCamelCase = '''Time usage''' if self.args.is_time else '''Memory usage''' __lowerCamelCase = title_str + ''' for training''' if self.args.is_train else title_str + ''' for inference''' if not self.args.no_log_scale: # set logarithm scales ax.set_xscale('''log''' ) ax.set_yscale('''log''' ) for axis in [ax.xaxis, ax.yaxis]: axis.set_major_formatter(ScalarFormatter() ) for model_name_idx, model_name in enumerate(self.result_dict.keys() ): __lowerCamelCase = sorted(set(self.result_dict[model_name]['''bsz'''] ) ) __lowerCamelCase = sorted(set(self.result_dict[model_name]['''seq_len'''] ) ) __lowerCamelCase = self.result_dict[model_name]['''result'''] ((__lowerCamelCase) ,(__lowerCamelCase)) = ( (batch_sizes, sequence_lengths) if self.args.plot_along_batch else (sequence_lengths, batch_sizes) ) __lowerCamelCase = ( model_name if self.args.short_model_names is None else self.args.short_model_names[model_name_idx] ) for inner_loop_value in inner_loop_array: if self.args.plot_along_batch: __lowerCamelCase = np.asarray( [results[(x, inner_loop_value)] for x in x_axis_array if (x, inner_loop_value) in results] , dtype=_lowerCAmelCase , ) else: __lowerCamelCase = np.asarray( [results[(inner_loop_value, x)] for x in x_axis_array if (inner_loop_value, x) in results] , dtype=np.floataa , ) ((__lowerCamelCase) ,(__lowerCamelCase)) = ( ('''batch_size''', '''len''') if self.args.plot_along_batch else ('''in #tokens''', '''bsz''') ) __lowerCamelCase = np.asarray(_lowerCAmelCase , _lowerCAmelCase )[: len(_lowerCAmelCase )] plt.scatter( _lowerCAmelCase , _lowerCAmelCase , label=F"""{label_model_name} - {inner_loop_label}: {inner_loop_value}""" ) plt.plot(_lowerCAmelCase , _lowerCAmelCase , '''--''' ) title_str += F""" {label_model_name} vs.""" __lowerCamelCase = title_str[:-4] __lowerCamelCase = '''Time in s''' if self.args.is_time else '''Memory in MB''' # plot plt.title(_lowerCAmelCase ) plt.xlabel(_lowerCAmelCase ) plt.ylabel(_lowerCAmelCase ) plt.legend() if self.args.figure_png_file is not None: plt.savefig(self.args.figure_png_file ) else: plt.show() def a__ ( ): __lowerCamelCase = HfArgumentParser(_UpperCamelCase ) __lowerCamelCase = parser.parse_args_into_dataclasses()[0] __lowerCamelCase = Plot(args=_UpperCamelCase ) plot.plot() if __name__ == "__main__": main()
709
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device a_ = False class __lowerCAmelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained('''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt='''first prompt''' , image=__UpperCAmelCase , text_to_image_strength=0.75 , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained(__UpperCAmelCase , torch_dtype=torch.floataa ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = generator.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt='''first prompt''' , image=__UpperCAmelCase , text_to_image_strength=0.75 , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained('''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = '''cyberpunk 2077''' __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt=__UpperCAmelCase , image=__UpperCAmelCase , text_to_image_strength=0.75 , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' , ).images __lowerCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __lowerCamelCase = '''A painting of a squirrel eating a burger ''' __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.text_to_image( prompt=__UpperCAmelCase , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' ).images __lowerCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __lowerCamelCase = pipe.image_variation(__UpperCAmelCase , generator=__UpperCAmelCase , output_type='''numpy''' ).images __lowerCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
622
0
class __lowerCAmelCase : def __init__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = val __lowerCamelCase = None __lowerCamelCase = None def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' if self.val: if val < self.val: if self.left is None: __lowerCamelCase = Node(_a ) else: self.left.insert(_a ) elif val > self.val: if self.right is None: __lowerCamelCase = Node(_a ) else: self.right.insert(_a ) else: __lowerCamelCase = val def a__ ( _UpperCamelCase : List[Any] ,_UpperCamelCase : List[str] ): # Recursive traversal if root: inorder(root.left ,snake_case_ ) res.append(root.val ) inorder(root.right ,snake_case_ ) def a__ ( _UpperCamelCase : Tuple ): # Build BST if len(snake_case_ ) == 0: return arr __lowerCamelCase = Node(arr[0] ) for i in range(1 ,len(snake_case_ ) ): root.insert(arr[i] ) # Traverse BST in order. __lowerCamelCase = [] inorder(snake_case_ ,snake_case_ ) return res if __name__ == "__main__": print(tree_sort([10, 1, 3, 2, 9, 14, 13]))
710
import argparse import datetime import json import time import warnings from logging import getLogger from pathlib import Path from typing import Dict, List import torch from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import calculate_bleu, calculate_rouge, chunks, parse_numeric_n_bool_cl_kwargs, use_task_specific_params a_ = getLogger(__name__) a_ = """cuda""" if torch.cuda.is_available() else """cpu""" def a__ ( _UpperCamelCase : List[str] ,_UpperCamelCase : str ,_UpperCamelCase : str ,_UpperCamelCase : int = 8 ,_UpperCamelCase : str = DEFAULT_DEVICE ,_UpperCamelCase : Dict=False ,_UpperCamelCase : Dict="summarization" ,_UpperCamelCase : Optional[int]=None ,**_UpperCamelCase : Dict ,): __lowerCamelCase = Path(_UpperCamelCase ).open('''w''' ,encoding='''utf-8''' ) __lowerCamelCase = str(_UpperCamelCase ) __lowerCamelCase = AutoModelForSeqaSeqLM.from_pretrained(_UpperCamelCase ).to(_UpperCamelCase ) if fpaa: __lowerCamelCase = model.half() __lowerCamelCase = AutoTokenizer.from_pretrained(_UpperCamelCase ) logger.info(F"""Inferred tokenizer type: {tokenizer.__class__}""" ) # if this is wrong, check config.model_type. __lowerCamelCase = time.time() # update config with task specific params use_task_specific_params(_UpperCamelCase ,_UpperCamelCase ) if prefix is None: __lowerCamelCase = prefix or getattr(model.config ,'''prefix''' ,'''''' ) or '''''' for examples_chunk in tqdm(list(chunks(_UpperCamelCase ,_UpperCamelCase ) ) ): __lowerCamelCase = [prefix + text for text in examples_chunk] __lowerCamelCase = tokenizer(_UpperCamelCase ,return_tensors='''pt''' ,truncation=_UpperCamelCase ,padding='''longest''' ).to(_UpperCamelCase ) __lowerCamelCase = model.generate( input_ids=batch.input_ids ,attention_mask=batch.attention_mask ,**_UpperCamelCase ,) __lowerCamelCase = tokenizer.batch_decode(_UpperCamelCase ,skip_special_tokens=_UpperCamelCase ,clean_up_tokenization_spaces=_UpperCamelCase ) for hypothesis in dec: fout.write(hypothesis + '''\n''' ) fout.flush() fout.close() __lowerCamelCase = int(time.time() - start_time ) # seconds __lowerCamelCase = len(_UpperCamelCase ) return {"n_obs": n_obs, "runtime": runtime, "seconds_per_sample": round(runtime / n_obs ,4 )} def a__ ( ): return datetime.datetime.now().strftime('''%Y-%m-%d %H:%M:%S''' ) def a__ ( _UpperCamelCase : Union[str, Any]=True ): __lowerCamelCase = argparse.ArgumentParser() parser.add_argument('''model_name''' ,type=_UpperCamelCase ,help='''like facebook/bart-large-cnn,t5-base, etc.''' ) parser.add_argument('''input_path''' ,type=_UpperCamelCase ,help='''like cnn_dm/test.source''' ) parser.add_argument('''save_path''' ,type=_UpperCamelCase ,help='''where to save summaries''' ) parser.add_argument('''--reference_path''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,help='''like cnn_dm/test.target''' ) parser.add_argument('''--score_path''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,default='''metrics.json''' ,help='''where to save metrics''' ) parser.add_argument('''--device''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,default=_UpperCamelCase ,help='''cuda, cuda:1, cpu etc.''' ) parser.add_argument( '''--prefix''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,default=_UpperCamelCase ,help='''will be added to the begininng of src examples''' ) parser.add_argument('''--task''' ,type=_UpperCamelCase ,default='''summarization''' ,help='''used for task_specific_params + metrics''' ) parser.add_argument('''--bs''' ,type=_UpperCamelCase ,default=8 ,required=_UpperCamelCase ,help='''batch size''' ) parser.add_argument( '''--n_obs''' ,type=_UpperCamelCase ,default=-1 ,required=_UpperCamelCase ,help='''How many observations. Defaults to all.''' ) parser.add_argument('''--fp16''' ,action='''store_true''' ) parser.add_argument('''--dump-args''' ,action='''store_true''' ,help='''print the custom hparams with the results''' ) parser.add_argument( '''--info''' ,nargs='''?''' ,type=_UpperCamelCase ,const=datetime_now() ,help=( '''use in conjunction w/ --dump-args to print with the results whatever other info you\'d like, e.g.''' ''' lang=en-ru. If no value is passed, the current datetime string will be used.''' ) ,) # Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate __lowerCamelCase ,__lowerCamelCase = parser.parse_known_args() __lowerCamelCase = parse_numeric_n_bool_cl_kwargs(_UpperCamelCase ) if parsed_args and verbose: print(F"""parsed the following generate kwargs: {parsed_args}""" ) __lowerCamelCase = [''' ''' + x.rstrip() if '''t5''' in args.model_name else x.rstrip() for x in open(args.input_path ).readlines()] if args.n_obs > 0: __lowerCamelCase = examples[: args.n_obs] Path(args.save_path ).parent.mkdir(exist_ok=_UpperCamelCase ) if args.reference_path is None and Path(args.score_path ).exists(): warnings.warn(F"""score_path {args.score_path} will be overwritten unless you type ctrl-c.""" ) if args.device == "cpu" and args.fpaa: # this mix leads to RuntimeError: "threshold_cpu" not implemented for 'Half' raise ValueError('''Can\'t mix --fp16 and --device cpu''' ) __lowerCamelCase = generate_summaries_or_translations( _UpperCamelCase ,args.save_path ,args.model_name ,batch_size=args.bs ,device=args.device ,fpaa=args.fpaa ,task=args.task ,prefix=args.prefix ,**_UpperCamelCase ,) if args.reference_path is None: return {} # Compute scores __lowerCamelCase = calculate_bleu if '''translation''' in args.task else calculate_rouge __lowerCamelCase = [x.rstrip() for x in open(args.save_path ).readlines()] __lowerCamelCase = [x.rstrip() for x in open(args.reference_path ).readlines()][: len(_UpperCamelCase )] __lowerCamelCase = score_fn(_UpperCamelCase ,_UpperCamelCase ) scores.update(_UpperCamelCase ) if args.dump_args: scores.update(_UpperCamelCase ) if args.info: __lowerCamelCase = args.info if verbose: print(_UpperCamelCase ) if args.score_path is not None: json.dump(_UpperCamelCase ,open(args.score_path ,'''w''' ) ) return scores if __name__ == "__main__": # Usage for MT: # python run_eval.py MODEL_NAME $DATA_DIR/test.source $save_dir/test_translations.txt --reference_path $DATA_DIR/test.target --score_path $save_dir/test_bleu.json --task translation $@ run_generate(verbose=True)
622
0
'''simple docstring''' import os import sys import unittest a_ = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path a_ = os.path.join(git_repo_path, """src""", """transformers""") a_ = '\n{0} = None\n' a_ = '\nclass {0}(metaclass=DummyObject):\n _backends = {1}\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, {1})\n' a_ = '\ndef {0}(*args, **kwargs):\n requires_backends({0}, {1})\n' class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = find_backend(''' _import_structure["models.albert"].append("AlbertTokenizerFast")''' ) self.assertIsNone(UpperCamelCase__ ) __lowerCamelCase = find_backend(''' if not is_tokenizers_available():''' ) self.assertEqual(UpperCamelCase__ , '''tokenizers''' ) __lowerCamelCase = find_backend(''' if not is_tensorflow_text_available():''' ) self.assertEqual(UpperCamelCase__ , '''tensorflow_text''' ) __lowerCamelCase = find_backend(''' if not (is_sentencepiece_available() and is_tokenizers_available()):''' ) self.assertEqual(UpperCamelCase__ , '''sentencepiece_and_tokenizers''' ) __lowerCamelCase = find_backend( ''' if not (is_sentencepiece_available() and is_tensorflow_text_available()):''' ) self.assertEqual(UpperCamelCase__ , '''sentencepiece_and_tensorflow_text''' ) __lowerCamelCase = find_backend( ''' if not (is_sentencepiece_available() and is_tokenizers_available() and is_vision_available()):''' ) self.assertEqual(UpperCamelCase__ , '''sentencepiece_and_tokenizers_and_vision''' ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn('''torch''' , UpperCamelCase__ ) self.assertIn('''tensorflow_text''' , UpperCamelCase__ ) self.assertIn('''sentencepiece_and_tokenizers''' , UpperCamelCase__ ) # Likewise, we can't assert on the exact content of a key self.assertIn('''BertModel''' , objects['''torch'''] ) self.assertIn('''TFBertModel''' , objects['''tf'''] ) self.assertIn('''FlaxBertModel''' , objects['''flax'''] ) self.assertIn('''BertModel''' , objects['''torch'''] ) self.assertIn('''TFBertTokenizer''' , objects['''tensorflow_text'''] ) self.assertIn('''convert_slow_tokenizer''' , objects['''sentencepiece_and_tokenizers'''] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = create_dummy_object('''CONSTANT''' , '''\'torch\'''' ) self.assertEqual(UpperCamelCase__ , '''\nCONSTANT = None\n''' ) __lowerCamelCase = create_dummy_object('''function''' , '''\'torch\'''' ) self.assertEqual( UpperCamelCase__ , '''\ndef function(*args, **kwargs):\n requires_backends(function, \'torch\')\n''' ) __lowerCamelCase = ''' class FakeClass(metaclass=DummyObject): _backends = \'torch\' def __init__(self, *args, **kwargs): requires_backends(self, \'torch\') ''' __lowerCamelCase = create_dummy_object('''FakeClass''' , '''\'torch\'''' ) self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''# This file is autogenerated by the command `make fix-copies`, do not edit. from ..utils import DummyObject, requires_backends CONSTANT = None def function(*args, **kwargs): requires_backends(function, ["torch"]) class FakeClass(metaclass=DummyObject): _backends = ["torch"] def __init__(self, *args, **kwargs): requires_backends(self, ["torch"]) ''' __lowerCamelCase = create_dummy_files({'''torch''': ['''CONSTANT''', '''function''', '''FakeClass''']} ) self.assertEqual(dummy_files['''torch'''] , UpperCamelCase__ )
711
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def a__ ( _UpperCamelCase : Optional[int] ,_UpperCamelCase : List[str] ,_UpperCamelCase : List[Any]=None ,_UpperCamelCase : Any=None ): if attention_mask is None: __lowerCamelCase = tf.cast(tf.math.not_equal(_UpperCamelCase ,config.pad_token_id ) ,tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class __lowerCAmelCase : lowerCAmelCase__ = OPTConfig lowerCAmelCase__ = {} lowerCAmelCase__ = """gelu""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=99 , __UpperCAmelCase=16 , __UpperCAmelCase=2 , __UpperCAmelCase=4 , __UpperCAmelCase=4 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=20 , __UpperCAmelCase=2 , __UpperCAmelCase=1 , __UpperCAmelCase=0 , __UpperCAmelCase=16 , __UpperCAmelCase=16 , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = eos_token_id __lowerCamelCase = pad_token_id __lowerCamelCase = bos_token_id __lowerCamelCase = embed_dim __lowerCamelCase = word_embed_proj_dim __lowerCamelCase = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowerCamelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowerCamelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowerCamelCase = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=__UpperCAmelCase , **self.config_updates , ) __lowerCamelCase = prepare_opt_inputs_dict(__UpperCAmelCase , __UpperCAmelCase ) return config, inputs_dict def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = TFOPTModel(config=__UpperCAmelCase ) __lowerCamelCase = inputs_dict['''input_ids'''] __lowerCamelCase = input_ids[:1, :] __lowerCamelCase = inputs_dict['''attention_mask'''][:1, :] __lowerCamelCase = 1 # first forward pass __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , use_cache=__UpperCAmelCase ) __lowerCamelCase ,__lowerCamelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowerCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowerCamelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowerCamelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowerCamelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase )[0] __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , past_key_values=__UpperCAmelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowerCamelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowerCamelCase = output_from_no_past[:, -3:, random_slice_idx] __lowerCamelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__UpperCAmelCase , __UpperCAmelCase , rtol=1E-3 ) @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () lowerCAmelCase__ = (TFOPTForCausalLM,) if is_tf_available() else () lowerCAmelCase__ = ( {"""feature-extraction""": TFOPTModel, """text-generation""": TFOPTForCausalLM} if is_tf_available() else {} ) lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = 1_0 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(__UpperCAmelCase , __UpperCAmelCase ): if hasattr(__UpperCAmelCase , '''weight''' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(__UpperCAmelCase , '''weight''' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 10, config.vocab_size + 10]: # build the embeddings __lowerCamelCase = model_class(config=__UpperCAmelCase ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_input_embeddings() ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(__UpperCAmelCase ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_input_embeddings() ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. __lowerCamelCase = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , __UpperCAmelCase ) # check that weights remain the same after resizing __lowerCamelCase = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCamelCase = False self.assertTrue(__UpperCAmelCase ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , __UpperCAmelCase ) __lowerCamelCase = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCamelCase = False self.assertTrue(__UpperCAmelCase ) def a__ ( _UpperCamelCase : Optional[Any] ): return tf.constant(_UpperCamelCase ,dtype=tf.intaa ) @require_tf class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = 9_9 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = tf.ones((4, 1) , dtype=tf.intaa ) * 2 __lowerCamelCase = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) __lowerCamelCase = input_ids.shape[0] __lowerCamelCase = OPTConfig( vocab_size=self.vocab_size , hidden_size=24 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class __lowerCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTModel.from_pretrained('''facebook/opt-350m''' ) __lowerCamelCase = _long_tensor([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] ) __lowerCamelCase = tf.not_equal(__UpperCAmelCase , model.config.pad_token_id ) with tf.GradientTape(): __lowerCamelCase = model(input_ids=__UpperCAmelCase , attention_mask=__UpperCAmelCase ).last_hidden_state __lowerCamelCase = (1, 11, 512) self.assertEqual(output.shape , __UpperCAmelCase ) __lowerCamelCase = tf.constant( [[-0.2_873, -1.9_218, -0.3_033], [-1.2_710, -0.1_338, -0.1_902], [0.4_095, 0.1_214, -1.3_121]] ) self.assertTrue(np.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=4E-3 ) ) __lowerCamelCase = tf.function(__UpperCAmelCase , jit_compile=__UpperCAmelCase ) __lowerCamelCase = xla_generate(__UpperCAmelCase , __UpperCAmelCase )[0] self.assertTrue(np.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=4E-2 ) ) @require_tf @slow class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' super().setUp() __lowerCamelCase = '''facebook/opt-350m''' def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTForCausalLM.from_pretrained(self.path_model ) __lowerCamelCase = GPTaTokenizer.from_pretrained(self.path_model ) __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of''', '''Paris is the capital of France and''', '''Computers and mobile phones have taken''', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' , padding=__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) __lowerCamelCase = tf.constant( [ [1.3_851, -13.8_923, -10.5_229, -10.7_533, -0.2_309, -10.2_384, -0.5_365, -9.0_947, -5.1_670], [-4.7_073, -10.6_276, -3.9_415, -21.5_242, -0.2_822, -0.2_822, -0.2_822, -0.2_822, -0.2_822], [0.6_247, -3.4_229, -8.9_179, -1.4_297, -14.1_650, 1.4_146, -9.0_218, -0.2_703, -0.2_703], [6.4_783, -1.9_913, -10.7_926, -2.3_336, 1.5_092, -0.9_974, -6.8_213, 1.3_477, 1.3_477], ] ) self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-4 ) ) __lowerCamelCase = tf.function(__UpperCAmelCase , jit_compile=__UpperCAmelCase ) __lowerCamelCase = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-4 ) ) @require_tf @slow class __lowerCAmelCase ( unittest.TestCase ): @property def lowerCamelCase ( self ): '''simple docstring''' return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-125m''' __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of New York, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] __lowerCamelCase = [] __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) for prompt in self.prompts: __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(__UpperCAmelCase , max_length=10 ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) predicted_outputs += generated_string self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-350m''' __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = '''left''' # use different length sentences to test batching __lowerCamelCase = [ '''Hello, my dog is a little''', '''Today, I''', ] __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' , padding=__UpperCAmelCase ) __lowerCamelCase = inputs['''input_ids'''] __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase , attention_mask=inputs['''attention_mask'''] ) __lowerCamelCase = tokenizer(sentences[0] , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase ) __lowerCamelCase = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['''attention_mask'''][-1] , tf.intaa ) ) __lowerCamelCase = tokenizer(sentences[1] , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase , max_length=model.config.max_length - num_paddings ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = [ '''Hello, my dog is a little bit of a dork.\nI\'m a little bit''', '''Today, I was in the middle of a conversation with a friend about the''', ] self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , [non_padded_sentence, padded_sentence] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-350m''' __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of San Francisco, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] __lowerCamelCase = [] __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) for prompt in self.prompts: __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(__UpperCAmelCase , max_length=10 ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) predicted_outputs += generated_string self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase )
622
0
def a__ ( _UpperCamelCase : int ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : List[str] ): if height >= 1: move_tower(height - 1 ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) move_disk(lowerCamelCase_ ,lowerCamelCase_ ) move_tower(height - 1 ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) def a__ ( _UpperCamelCase : List[Any] ,_UpperCamelCase : List[Any] ): print('''moving disk from''' ,lowerCamelCase_ ,'''to''' ,lowerCamelCase_ ) def a__ ( ): __lowerCamelCase = int(input('''Height of hanoi: ''' ).strip() ) move_tower(lowerCamelCase_ ,'''A''' ,'''B''' ,'''C''' ) if __name__ == "__main__": main()
712
import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) a_ = logging.getLogger(__name__) def a__ ( _UpperCamelCase : str ,_UpperCamelCase : List[Any] ): __lowerCamelCase = np.argmax(_UpperCamelCase ,axis=1 ) return np.sum(outputs == labels ) def a__ ( _UpperCamelCase : Optional[int] ): with open(_UpperCamelCase ,encoding='''utf_8''' ) as f: __lowerCamelCase = csv.reader(_UpperCamelCase ) __lowerCamelCase = [] next(_UpperCamelCase ) # skip the first line for line in tqdm(_UpperCamelCase ): output.append((''' '''.join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def a__ ( _UpperCamelCase : Any ,_UpperCamelCase : Dict ,_UpperCamelCase : str ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Any ,_UpperCamelCase : Dict ): __lowerCamelCase = [] for dataset in encoded_datasets: __lowerCamelCase = len(_UpperCamelCase ) __lowerCamelCase = np.zeros((n_batch, 2, input_len) ,dtype=np.intaa ) __lowerCamelCase = np.zeros((n_batch, 2) ,dtype=np.intaa ) __lowerCamelCase = np.full((n_batch, 2, input_len) ,fill_value=-1_00 ,dtype=np.intaa ) __lowerCamelCase = np.zeros((n_batch,) ,dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(_UpperCamelCase ): __lowerCamelCase = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __lowerCamelCase = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __lowerCamelCase = with_conta __lowerCamelCase = with_conta __lowerCamelCase = len(_UpperCamelCase ) - 1 __lowerCamelCase = len(_UpperCamelCase ) - 1 __lowerCamelCase = with_conta __lowerCamelCase = with_conta __lowerCamelCase = mc_label __lowerCamelCase = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(_UpperCamelCase ) for t in all_inputs ) ) return tensor_datasets def a__ ( ): __lowerCamelCase = argparse.ArgumentParser() parser.add_argument('''--model_name''' ,type=_UpperCamelCase ,default='''openai-gpt''' ,help='''pretrained model name''' ) parser.add_argument('''--do_train''' ,action='''store_true''' ,help='''Whether to run training.''' ) parser.add_argument('''--do_eval''' ,action='''store_true''' ,help='''Whether to run eval on the dev set.''' ) parser.add_argument( '''--output_dir''' ,default=_UpperCamelCase ,type=_UpperCamelCase ,required=_UpperCamelCase ,help='''The output directory where the model predictions and checkpoints will be written.''' ,) parser.add_argument('''--train_dataset''' ,type=_UpperCamelCase ,default='''''' ) parser.add_argument('''--eval_dataset''' ,type=_UpperCamelCase ,default='''''' ) parser.add_argument('''--seed''' ,type=_UpperCamelCase ,default=42 ) parser.add_argument('''--num_train_epochs''' ,type=_UpperCamelCase ,default=3 ) parser.add_argument('''--train_batch_size''' ,type=_UpperCamelCase ,default=8 ) parser.add_argument('''--eval_batch_size''' ,type=_UpperCamelCase ,default=16 ) parser.add_argument('''--adam_epsilon''' ,default=1e-8 ,type=_UpperCamelCase ,help='''Epsilon for Adam optimizer.''' ) parser.add_argument('''--max_grad_norm''' ,type=_UpperCamelCase ,default=1 ) parser.add_argument( '''--max_steps''' ,default=-1 ,type=_UpperCamelCase ,help=( '''If > 0: set total number of training steps to perform. Override num_train_epochs.''' ) ,) parser.add_argument( '''--gradient_accumulation_steps''' ,type=_UpperCamelCase ,default=1 ,help='''Number of updates steps to accumulate before performing a backward/update pass.''' ,) parser.add_argument('''--learning_rate''' ,type=_UpperCamelCase ,default=6.25e-5 ) parser.add_argument('''--warmup_steps''' ,default=0 ,type=_UpperCamelCase ,help='''Linear warmup over warmup_steps.''' ) parser.add_argument('''--lr_schedule''' ,type=_UpperCamelCase ,default='''warmup_linear''' ) parser.add_argument('''--weight_decay''' ,type=_UpperCamelCase ,default=0.01 ) parser.add_argument('''--lm_coef''' ,type=_UpperCamelCase ,default=0.9 ) parser.add_argument('''--n_valid''' ,type=_UpperCamelCase ,default=3_74 ) parser.add_argument('''--server_ip''' ,type=_UpperCamelCase ,default='''''' ,help='''Can be used for distant debugging.''' ) parser.add_argument('''--server_port''' ,type=_UpperCamelCase ,default='''''' ,help='''Can be used for distant debugging.''' ) __lowerCamelCase = parser.parse_args() print(_UpperCamelCase ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('''Waiting for debugger attach''' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) ,redirect_output=_UpperCamelCase ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) __lowerCamelCase = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) __lowerCamelCase = torch.cuda.device_count() logger.info('''device: {}, n_gpu {}'''.format(_UpperCamelCase ,_UpperCamelCase ) ) if not args.do_train and not args.do_eval: raise ValueError('''At least one of `do_train` or `do_eval` must be True.''' ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset __lowerCamelCase = ['''_start_''', '''_delimiter_''', '''_classify_'''] __lowerCamelCase = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(_UpperCamelCase ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(_UpperCamelCase ) __lowerCamelCase = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(_UpperCamelCase ) ) model.to(_UpperCamelCase ) # Load and encode the datasets def tokenize_and_encode(_UpperCamelCase : Dict ): if isinstance(_UpperCamelCase ,_UpperCamelCase ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(_UpperCamelCase ) ) elif isinstance(_UpperCamelCase ,_UpperCamelCase ): return obj return [tokenize_and_encode(_UpperCamelCase ) for o in obj] logger.info('''Encoding dataset...''' ) __lowerCamelCase = load_rocstories_dataset(args.train_dataset ) __lowerCamelCase = load_rocstories_dataset(args.eval_dataset ) __lowerCamelCase = (train_dataset, eval_dataset) __lowerCamelCase = tokenize_and_encode(_UpperCamelCase ) # Compute the max input length for the Transformer __lowerCamelCase = model.config.n_positions // 2 - 2 __lowerCamelCase = max( len(story[:max_length] ) + max(len(conta[:max_length] ) ,len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) __lowerCamelCase = min(_UpperCamelCase ,model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders __lowerCamelCase = pre_process_datasets(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ,*_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = tensor_datasets[0], tensor_datasets[1] __lowerCamelCase = TensorDataset(*_UpperCamelCase ) __lowerCamelCase = RandomSampler(_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,sampler=_UpperCamelCase ,batch_size=args.train_batch_size ) __lowerCamelCase = TensorDataset(*_UpperCamelCase ) __lowerCamelCase = SequentialSampler(_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,sampler=_UpperCamelCase ,batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: __lowerCamelCase = args.max_steps __lowerCamelCase = args.max_steps // (len(_UpperCamelCase ) // args.gradient_accumulation_steps) + 1 else: __lowerCamelCase = len(_UpperCamelCase ) // args.gradient_accumulation_steps * args.num_train_epochs __lowerCamelCase = list(model.named_parameters() ) __lowerCamelCase = ['''bias''', '''LayerNorm.bias''', '''LayerNorm.weight'''] __lowerCamelCase = [ { '''params''': [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], '''weight_decay''': args.weight_decay, }, {'''params''': [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0}, ] __lowerCamelCase = AdamW(_UpperCamelCase ,lr=args.learning_rate ,eps=args.adam_epsilon ) __lowerCamelCase = get_linear_schedule_with_warmup( _UpperCamelCase ,num_warmup_steps=args.warmup_steps ,num_training_steps=_UpperCamelCase ) if args.do_train: __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) ,desc='''Epoch''' ): __lowerCamelCase = 0 __lowerCamelCase = 0 __lowerCamelCase = tqdm(_UpperCamelCase ,desc='''Training''' ) for step, batch in enumerate(_UpperCamelCase ): __lowerCamelCase = tuple(t.to(_UpperCamelCase ) for t in batch ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = batch __lowerCamelCase = model(_UpperCamelCase ,mc_token_ids=_UpperCamelCase ,lm_labels=_UpperCamelCase ,mc_labels=_UpperCamelCase ) __lowerCamelCase = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() __lowerCamelCase = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 __lowerCamelCase = '''Training loss: {:.2e} lr: {:.2e}'''.format(_UpperCamelCase ,scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer __lowerCamelCase = model.module if hasattr(_UpperCamelCase ,'''module''' ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` __lowerCamelCase = os.path.join(args.output_dir ,_UpperCamelCase ) __lowerCamelCase = os.path.join(args.output_dir ,_UpperCamelCase ) torch.save(model_to_save.state_dict() ,_UpperCamelCase ) model_to_save.config.to_json_file(_UpperCamelCase ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned __lowerCamelCase = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) __lowerCamelCase = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(_UpperCamelCase ) if args.do_eval: model.eval() __lowerCamelCase ,__lowerCamelCase = 0, 0 __lowerCamelCase ,__lowerCamelCase = 0, 0 for batch in tqdm(_UpperCamelCase ,desc='''Evaluating''' ): __lowerCamelCase = tuple(t.to(_UpperCamelCase ) for t in batch ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = batch with torch.no_grad(): __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = model( _UpperCamelCase ,mc_token_ids=_UpperCamelCase ,lm_labels=_UpperCamelCase ,mc_labels=_UpperCamelCase ) __lowerCamelCase = mc_logits.detach().cpu().numpy() __lowerCamelCase = mc_labels.to('''cpu''' ).numpy() __lowerCamelCase = accuracy(_UpperCamelCase ,_UpperCamelCase ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 __lowerCamelCase = eval_loss / nb_eval_steps __lowerCamelCase = eval_accuracy / nb_eval_examples __lowerCamelCase = tr_loss / nb_tr_steps if args.do_train else None __lowerCamelCase = {'''eval_loss''': eval_loss, '''eval_accuracy''': eval_accuracy, '''train_loss''': train_loss} __lowerCamelCase = os.path.join(args.output_dir ,'''eval_results.txt''' ) with open(_UpperCamelCase ,'''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key in sorted(result.keys() ): logger.info(''' %s = %s''' ,_UpperCamelCase ,str(result[key] ) ) writer.write('''%s = %s\n''' % (key, str(result[key] )) ) if __name__ == "__main__": main()
622
0
def a__ ( _UpperCamelCase : int = 10_00 ): __lowerCamelCase = 2**power __lowerCamelCase = str(__snake_case ) __lowerCamelCase = list(__snake_case ) __lowerCamelCase = 0 for i in list_num: sum_of_num += int(__snake_case ) return sum_of_num if __name__ == "__main__": a_ = int(input("""Enter the power of 2: """).strip()) print("""2 ^ """, power, """ = """, 2**power) a_ = solution(power) print("""Sum of the digits is: """, result)
713
import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class __lowerCAmelCase ( lowerCAmelCase__ ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=1024 , __UpperCAmelCase=1024 , __UpperCAmelCase=3.6 ): '''simple docstring''' __lowerCamelCase = tokenizer __lowerCamelCase = tokenizer.bos_token_id __lowerCamelCase = dataset __lowerCamelCase = seq_length __lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self ): '''simple docstring''' __lowerCamelCase = iter(self.dataset ) __lowerCamelCase = True while more_examples: __lowerCamelCase ,__lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(__UpperCAmelCase )['''content'''] ) buffer_len += len(buffer[-1] ) except StopIteration: __lowerCamelCase = False break __lowerCamelCase = tokenizer(__UpperCAmelCase , truncation=__UpperCAmelCase )['''input_ids'''] __lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(__UpperCAmelCase ) , self.seq_length ): __lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(__UpperCAmelCase ) == self.seq_length: yield torch.tensor(__UpperCAmelCase ) def a__ ( _UpperCamelCase : List[Any] ): __lowerCamelCase = {'''streaming''': True} __lowerCamelCase = load_dataset(args.dataset_name ,split='''train''' ,**_UpperCamelCase ) __lowerCamelCase = ConstantLengthDataset(_UpperCamelCase ,_UpperCamelCase ,seq_length=args.seq_length ) __lowerCamelCase = DataLoader(_UpperCamelCase ,batch_size=args.batch_size ) return eval_dataloader def a__ ( _UpperCamelCase : str ): model.eval() __lowerCamelCase = [] for step, batch in enumerate(_UpperCamelCase ): with torch.no_grad(): __lowerCamelCase = model(_UpperCamelCase ,labels=_UpperCamelCase ) __lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(_UpperCamelCase ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break __lowerCamelCase = torch.mean(torch.cat(_UpperCamelCase ) ) try: __lowerCamelCase = torch.exp(_UpperCamelCase ) except OverflowError: __lowerCamelCase = float('''inf''' ) return loss.item(), perplexity.item() # Setup Accelerator a_ = Accelerator() # Parse configuration a_ = HfArgumentParser(EvaluationArguments) a_ = parser.parse_args() set_seed(args.seed) # Logging a_ = logging.getLogger(__name__) logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) # Load model and tokenizer a_ = AutoModelForCausalLM.from_pretrained(args.model_ckpt) a_ = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader a_ = create_dataloader(args) # Prepare everything with our `accelerator`. a_ , a_ = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info("""Evaluating and saving model after training""") a_ , a_ = evaluate(args) logger.info(f"loss/eval: {eval_loss}, perplexity: {perplexity}")
622
0
import numpy # List of input, output pairs a_ = ( ((5, 2, 3), 15), ((6, 5, 9), 25), ((11, 12, 13), 41), ((1, 1, 1), 8), ((11, 12, 13), 41), ) a_ = (((515, 22, 13), 555), ((61, 35, 49), 150)) a_ = [2, 4, 1, 5] a_ = len(train_data) a_ = 0.0_09 def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Optional[int]="train" ): return calculate_hypothesis_value(a__ ,a__ ) - output( a__ ,a__ ) def a__ ( _UpperCamelCase : Any ): __lowerCamelCase = 0 for i in range(len(a__ ) - 1 ): hyp_val += data_input_tuple[i] * parameter_vector[i + 1] hyp_val += parameter_vector[0] return hyp_val def a__ ( _UpperCamelCase : Optional[Any] ,_UpperCamelCase : int ): if data_set == "train": return train_data[example_no][1] elif data_set == "test": return test_data[example_no][1] return None def a__ ( _UpperCamelCase : str ,_UpperCamelCase : List[str] ): if data_set == "train": return _hypothesis_value(train_data[example_no][0] ) elif data_set == "test": return _hypothesis_value(test_data[example_no][0] ) return None def a__ ( _UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Optional[Any]=m ): __lowerCamelCase = 0 for i in range(a__ ): if index == -1: summation_value += _error(a__ ) else: summation_value += _error(a__ ) * train_data[i][0][index] return summation_value def a__ ( _UpperCamelCase : Tuple ): __lowerCamelCase = summation_of_cost_derivative(a__ ,a__ ) / m return cost_derivative_value def a__ ( ): global parameter_vector # Tune these values to set a tolerance value for predicted output __lowerCamelCase = 0.000_002 __lowerCamelCase = 0 __lowerCamelCase = 0 while True: j += 1 __lowerCamelCase = [0, 0, 0, 0] for i in range(0 ,len(a__ ) ): __lowerCamelCase = get_cost_derivative(i - 1 ) __lowerCamelCase = ( parameter_vector[i] - LEARNING_RATE * cost_derivative ) if numpy.allclose( a__ ,a__ ,atol=a__ ,rtol=a__ ,): break __lowerCamelCase = temp_parameter_vector print(('''Number of iterations:''', j) ) def a__ ( ): for i in range(len(a__ ) ): print(('''Actual output value:''', output(a__ ,'''test''' )) ) print(('''Hypothesis output:''', calculate_hypothesis_value(a__ ,'''test''' )) ) if __name__ == "__main__": run_gradient_descent() print("""\nTesting gradient descent for a linear hypothesis function.\n""") test_gradient_descent()
714
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """lxmert""" lowerCAmelCase__ = {} def __init__( self , __UpperCAmelCase=30522 , __UpperCAmelCase=768 , __UpperCAmelCase=12 , __UpperCAmelCase=9500 , __UpperCAmelCase=1600 , __UpperCAmelCase=400 , __UpperCAmelCase=3072 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=9 , __UpperCAmelCase=5 , __UpperCAmelCase=5 , __UpperCAmelCase=2048 , __UpperCAmelCase=4 , __UpperCAmelCase=6.67 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = num_qa_labels __lowerCamelCase = num_object_labels __lowerCamelCase = num_attr_labels __lowerCamelCase = l_layers __lowerCamelCase = x_layers __lowerCamelCase = r_layers __lowerCamelCase = visual_feat_dim __lowerCamelCase = visual_pos_dim __lowerCamelCase = visual_loss_normalizer __lowerCamelCase = task_matched __lowerCamelCase = task_mask_lm __lowerCamelCase = task_obj_predict __lowerCamelCase = task_qa __lowerCamelCase = visual_obj_loss __lowerCamelCase = visual_attr_loss __lowerCamelCase = visual_feat_loss __lowerCamelCase = {'''vision''': r_layers, '''cross_encoder''': x_layers, '''language''': l_layers} super().__init__(**__UpperCAmelCase )
622
0
'''simple docstring''' import itertools import string from collections.abc import Generator, Iterable def a__ ( _UpperCamelCase : Iterable[str] ,_UpperCamelCase : int ): __lowerCamelCase = iter(UpperCamelCase__ ) while True: __lowerCamelCase = tuple(itertools.islice(UpperCamelCase__ ,UpperCamelCase__ ) ) if not chunk: return yield chunk def a__ ( _UpperCamelCase : str ): __lowerCamelCase = ''''''.join([c.upper() for c in dirty if c in string.ascii_letters] ) __lowerCamelCase = '''''' if len(UpperCamelCase__ ) < 2: return dirty for i in range(len(UpperCamelCase__ ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(UpperCamelCase__ ) & 1: clean += "X" return clean def a__ ( _UpperCamelCase : str ): # I and J are used interchangeably to allow # us to use a 5x5 table (25 letters) __lowerCamelCase = '''ABCDEFGHIKLMNOPQRSTUVWXYZ''' # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler __lowerCamelCase = [] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(UpperCamelCase__ ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(UpperCamelCase__ ) return table def a__ ( _UpperCamelCase : str ,_UpperCamelCase : str ): __lowerCamelCase = generate_table(UpperCamelCase__ ) __lowerCamelCase = prepare_input(UpperCamelCase__ ) __lowerCamelCase = '''''' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(UpperCamelCase__ ,2 ): __lowerCamelCase ,__lowerCamelCase = divmod(table.index(UpperCamelCase__ ) ,5 ) __lowerCamelCase ,__lowerCamelCase = divmod(table.index(UpperCamelCase__ ) ,5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def a__ ( _UpperCamelCase : str ,_UpperCamelCase : str ): __lowerCamelCase = generate_table(UpperCamelCase__ ) __lowerCamelCase = '''''' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(UpperCamelCase__ ,2 ): __lowerCamelCase ,__lowerCamelCase = divmod(table.index(UpperCamelCase__ ) ,5 ) __lowerCamelCase ,__lowerCamelCase = divmod(table.index(UpperCamelCase__ ) ,5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
715
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Any ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Any ): if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = np.full((len(_UpperCamelCase ), sequence_length, 2) ,_UpperCamelCase ) else: __lowerCamelCase = np.full((len(_UpperCamelCase ), sequence_length) ,_UpperCamelCase ) for i, tensor in enumerate(_UpperCamelCase ): if padding_side == "right": if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = tensor[:sequence_length] else: __lowerCamelCase = tensor[:sequence_length] else: if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = tensor[:sequence_length] else: __lowerCamelCase = tensor[:sequence_length] return out_tensor.tolist() def a__ ( _UpperCamelCase : Dict ): __lowerCamelCase = ord(_UpperCamelCase ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 1_23 and cp <= 1_26): return True __lowerCamelCase = unicodedata.category(_UpperCamelCase ) if cat.startswith('''P''' ): return True return False @dataclass class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = 42 lowerCAmelCase__ = True lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = -1_0_0 lowerCAmelCase__ = "pt" def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' import torch __lowerCamelCase = '''label''' if '''label''' in features[0].keys() else '''labels''' __lowerCamelCase = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __lowerCamelCase = self.tokenizer.pad( __UpperCAmelCase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch __lowerCamelCase = torch.tensor(batch['''entity_ids'''] ).shape[1] __lowerCamelCase = self.tokenizer.padding_side if padding_side == "right": __lowerCamelCase = [ list(__UpperCAmelCase ) + [self.label_pad_token_id] * (sequence_length - len(__UpperCAmelCase )) for label in labels ] else: __lowerCamelCase = [ [self.label_pad_token_id] * (sequence_length - len(__UpperCAmelCase )) + list(__UpperCAmelCase ) for label in labels ] __lowerCamelCase = [feature['''ner_tags'''] for feature in features] __lowerCamelCase = padding_tensor(__UpperCAmelCase , -1 , __UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = [feature['''original_entity_spans'''] for feature in features] __lowerCamelCase = padding_tensor(__UpperCAmelCase , (-1, -1) , __UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = {k: torch.tensor(__UpperCAmelCase , dtype=torch.intaa ) for k, v in batch.items()} return batch
622
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available a_ = { '''configuration_roc_bert''': ['''ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''RoCBertConfig'''], '''tokenization_roc_bert''': ['''RoCBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: pass try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ '''ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''RoCBertForCausalLM''', '''RoCBertForMaskedLM''', '''RoCBertForMultipleChoice''', '''RoCBertForPreTraining''', '''RoCBertForQuestionAnswering''', '''RoCBertForSequenceClassification''', '''RoCBertForTokenClassification''', '''RoCBertLayer''', '''RoCBertModel''', '''RoCBertPreTrainedModel''', '''load_tf_weights_in_roc_bert''', ] if TYPE_CHECKING: from .configuration_roc_bert import ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RoCBertConfig from .tokenization_roc_bert import RoCBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: raise OptionalDependencyNotAvailable() try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roc_bert import ( ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, RoCBertForCausalLM, RoCBertForMaskedLM, RoCBertForMultipleChoice, RoCBertForPreTraining, RoCBertForQuestionAnswering, RoCBertForSequenceClassification, RoCBertForTokenClassification, RoCBertLayer, RoCBertModel, RoCBertPreTrainedModel, load_tf_weights_in_roc_bert, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
716
from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=99 , __UpperCAmelCase=[1, 1, 2] , __UpperCAmelCase=1 , __UpperCAmelCase=32 , __UpperCAmelCase=4 , __UpperCAmelCase=8 , __UpperCAmelCase=37 , __UpperCAmelCase="gelu_new" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.0 , __UpperCAmelCase=512 , __UpperCAmelCase=3 , __UpperCAmelCase=0.02 , __UpperCAmelCase=3 , __UpperCAmelCase=4 , __UpperCAmelCase=None , __UpperCAmelCase=False , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_input_mask __lowerCamelCase = use_token_type_ids __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = block_sizes __lowerCamelCase = num_decoder_layers __lowerCamelCase = d_model __lowerCamelCase = n_head __lowerCamelCase = d_head __lowerCamelCase = d_inner __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout __lowerCamelCase = attention_dropout __lowerCamelCase = activation_dropout __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = 2 __lowerCamelCase = num_labels __lowerCamelCase = num_choices __lowerCamelCase = scope __lowerCamelCase = initializer_std # Used in the tests to check the size of the first attention layer __lowerCamelCase = n_head # Used in the tests to check the size of the first hidden state __lowerCamelCase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowerCamelCase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowerCamelCase = self.num_hidden_layers + 2 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCamelCase = None if self.use_token_type_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCamelCase = FunnelConfig( vocab_size=self.vocab_size , block_sizes=self.block_sizes , num_decoder_layers=self.num_decoder_layers , d_model=self.d_model , n_head=self.n_head , d_head=self.d_head , d_inner=self.d_inner , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , activation_dropout=self.activation_dropout , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_std=self.initializer_std , ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = [input_ids, input_mask] __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = [input_ids, input_mask] __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 3, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForPreTraining(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForMaskedLM(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = TFFunnelForSequenceClassification(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_choices __lowerCamelCase = TFFunnelForMultipleChoice(config=__UpperCAmelCase ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = TFFunnelForTokenClassification(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForQuestionAnswering(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.prepare_config_and_inputs() ( ( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) , ) = config_and_inputs __lowerCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) lowerCAmelCase__ = ( { """feature-extraction""": (TFFunnelBaseModel, TFFunnelModel), """fill-mask""": TFFunnelForMaskedLM, """question-answering""": TFFunnelForQuestionAnswering, """text-classification""": TFFunnelForSequenceClassification, """token-classification""": TFFunnelForTokenClassification, """zero-shot""": TFFunnelForSequenceClassification, } if is_tf_available() else {} ) lowerCAmelCase__ = False lowerCAmelCase__ = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFFunnelModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCAmelCase ) @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) lowerCAmelCase__ = False lowerCAmelCase__ = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFFunnelModelTester(self , base=__UpperCAmelCase ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCAmelCase )
622
0
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase ): lowerCAmelCase__ = StableDiffusionXLImgaImgPipeline lowerCAmelCase__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} lowerCAmelCase__ = PipelineTesterMixin.required_optional_params - {"""latents"""} lowerCAmelCase__ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowerCAmelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS lowerCAmelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowerCamelCase ( self ): '''simple docstring''' torch.manual_seed(0 ) __lowerCamelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=__a , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) __lowerCamelCase = EulerDiscreteScheduler( beta_start=0.00_085 , beta_end=0.012 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) __lowerCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='''gelu''' , projection_dim=32 , ) __lowerCamelCase = CLIPTextModel(__a ) __lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__a ) __lowerCamelCase = CLIPTextModelWithProjection(__a ) __lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__a ) __lowerCamelCase = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'text_encoder_2': text_encoder_a, 'tokenizer_2': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase=0 ): '''simple docstring''' __lowerCamelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__a ) ).to(__a ) __lowerCamelCase = image / 2 + 0.5 if str(__a ).startswith('''mps''' ): __lowerCamelCase = torch.manual_seed(__a ) else: __lowerCamelCase = torch.Generator(device=__a ).manual_seed(__a ) __lowerCamelCase = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 5.0, 'output_type': 'numpy', 'strength': 0.75, } return inputs def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = StableDiffusionXLImgaImgPipeline(**__a ) __lowerCamelCase = sd_pipe.to(__a ) sd_pipe.set_progress_bar_config(disable=__a ) __lowerCamelCase = self.get_dummy_inputs(__a ) __lowerCamelCase = sd_pipe(**__a ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCamelCase = np.array([0.4_656, 0.4_840, 0.4_439, 0.6_698, 0.5_574, 0.4_524, 0.5_799, 0.5_943, 0.5_165] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase ( self ): '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def lowerCamelCase ( self ): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = StableDiffusionXLImgaImgPipeline(**__a ) __lowerCamelCase = sd_pipe.to(__a ) __lowerCamelCase = sd_pipe.to(__a ) sd_pipe.set_progress_bar_config(disable=__a ) # forward without prompt embeds __lowerCamelCase = self.get_dummy_inputs(__a ) __lowerCamelCase = 3 * ['this is a negative prompt'] __lowerCamelCase = negative_prompt __lowerCamelCase = 3 * [inputs['prompt']] __lowerCamelCase = sd_pipe(**__a ) __lowerCamelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds __lowerCamelCase = self.get_dummy_inputs(__a ) __lowerCamelCase = 3 * ['this is a negative prompt'] __lowerCamelCase = 3 * [inputs.pop('''prompt''' )] ( __lowerCamelCase ) = sd_pipe.encode_prompt(__a , negative_prompt=__a ) __lowerCamelCase = sd_pipe( **__a , prompt_embeds=__a , negative_prompt_embeds=__a , pooled_prompt_embeds=__a , negative_pooled_prompt_embeds=__a , ) __lowerCamelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase="cpu" , __UpperCAmelCase=torch.floataa , __UpperCAmelCase=0 ): '''simple docstring''' __lowerCamelCase = torch.Generator(device=__a ).manual_seed(__a ) __lowerCamelCase = np.random.RandomState(__a ).standard_normal((1, 4, 64, 64) ) __lowerCamelCase = torch.from_numpy(__a ).to(device=__a , dtype=__a ) __lowerCamelCase = { 'prompt': 'a photograph of an astronaut riding a horse', 'latents': latents, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(__a ) pipe.set_progress_bar_config(disable=__a ) __lowerCamelCase = self.get_inputs(__a ) __lowerCamelCase = pipe(**__a ).images __lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.49_493, 0.47_896, 0.40_798, 0.54_214, 0.53_212, 0.48_202, 0.47_656, 0.46_329, 0.48_506] ) assert np.abs(image_slice - expected_slice ).max() < 7E-3
717
from collections import namedtuple import requests from lxml import html # type: ignore a_ = namedtuple("""covid_data""", """cases deaths recovered""") def a__ ( _UpperCamelCase : str = "https://www.worldometers.info/coronavirus/" ): __lowerCamelCase = '''//div[@class = "maincounter-number"]/span/text()''' return covid_data(*html.fromstring(requests.get(_UpperCamelCase ).content ).xpath(_UpperCamelCase ) ) a_ = """Total COVID-19 cases in the world: {} Total deaths due to COVID-19 in the world: {} Total COVID-19 patients recovered in the world: {}""" print(fmt.format(*covid_stats()))
622
0
import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer a_ = """bart""" a_ = True @st.cache(allow_output_mutation=_lowercase ) def a__ ( ): if LOAD_DENSE_INDEX: __lowerCamelCase = AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) __lowerCamelCase = AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) __lowerCamelCase = qar_model.eval() else: __lowerCamelCase ,__lowerCamelCase = (None, None) if MODEL_TYPE == "bart": __lowerCamelCase = AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) __lowerCamelCase = AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) __lowerCamelCase = torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) __lowerCamelCase = sas_model.eval() else: __lowerCamelCase ,__lowerCamelCase = make_qa_sas_model( model_name='''t5-small''' ,from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' ,device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=_lowercase ) def a__ ( ): if LOAD_DENSE_INDEX: __lowerCamelCase = faiss.StandardGpuResources() __lowerCamelCase = datasets.load_dataset(path='''wiki_snippets''' ,name='''wiki40b_en_100_0''' )['''train'''] __lowerCamelCase = np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' ,dtype='''float32''' ,mode='''r''' ,shape=(wikiaab_passages.num_rows, 1_28) ,) __lowerCamelCase = faiss.IndexFlatIP(1_28 ) __lowerCamelCase = faiss.index_cpu_to_gpu(_lowercase ,1 ,_lowercase ) wikiaab_gpu_index_flat.add(_lowercase ) # TODO fix for larger GPU else: __lowerCamelCase ,__lowerCamelCase = (None, None) __lowerCamelCase = Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=_lowercase ) def a__ ( ): __lowerCamelCase = datasets.load_dataset('''eli5''' ,name='''LFQA_reddit''' ) __lowerCamelCase = elia['''train_eli5'''] __lowerCamelCase = np.memmap( '''eli5_questions_reps.dat''' ,dtype='''float32''' ,mode='''r''' ,shape=(elia_train.num_rows, 1_28) ) __lowerCamelCase = faiss.IndexFlatIP(1_28 ) eli5_train_q_index.add(_lowercase ) return (elia_train, eli5_train_q_index) a_ , a_ , a_ = load_indexes() a_ , a_ , a_ , a_ = load_models() a_ , a_ = load_train_data() def a__ ( _UpperCamelCase : Any ,_UpperCamelCase : Optional[int]=10 ): __lowerCamelCase = embed_questions_for_retrieval([question] ,_lowercase ,_lowercase ) __lowerCamelCase ,__lowerCamelCase = eli5_train_q_index.search(_lowercase ,_lowercase ) __lowerCamelCase = [elia_train[int(_lowercase )] for i in I[0]] return nn_examples def a__ ( _UpperCamelCase : List[str] ,_UpperCamelCase : Optional[Any]="wiki40b" ,_UpperCamelCase : Optional[Any]="dense" ,_UpperCamelCase : int=10 ): if source == "none": __lowerCamelCase ,__lowerCamelCase = (''' <P> '''.join(['''''' for _ in range(11 )] ).strip(), []) else: if method == "dense": __lowerCamelCase ,__lowerCamelCase = query_qa_dense_index( _lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ) else: __lowerCamelCase ,__lowerCamelCase = query_es_index( _lowercase ,_lowercase ,index_name='''english_wiki40b_snippets_100w''' ,n_results=_lowercase ,) __lowerCamelCase = [ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] __lowerCamelCase = '''question: {} context: {}'''.format(_lowercase ,_lowercase ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda _UpperCamelCase : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda _UpperCamelCase : None), } ) def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : int=64 ,_UpperCamelCase : List[str]=2_56 ,_UpperCamelCase : Tuple=False ,_UpperCamelCase : Tuple=2 ,_UpperCamelCase : Any=0.95 ,_UpperCamelCase : Optional[int]=0.8 ): with torch.no_grad(): __lowerCamelCase = qa_sas_generate( _lowercase ,_lowercase ,_lowercase ,num_answers=1 ,num_beams=_lowercase ,min_len=_lowercase ,max_len=_lowercase ,do_sample=_lowercase ,temp=_lowercase ,top_p=_lowercase ,top_k=_lowercase ,max_input_length=10_24 ,device='''cuda:0''' ,)[0] return (answer, support_list) st.title("""Long Form Question Answering with ELI5""") # Start sidebar a_ = """<img src=\'https://huggingface.co/front/assets/huggingface_logo.svg\'>""" a_ = """\n<html>\n <head>\n <style>\n .img-container {\n padding-left: 90px;\n padding-right: 90px;\n padding-top: 50px;\n padding-bottom: 50px;\n background-color: #f0f3f9;\n }\n </style>\n </head>\n <body>\n <span class=\"img-container\"> <!-- Inline parent element -->\n %s\n </span>\n </body>\n</html>\n""" % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia a_ = """\nThis demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html).\nFirst, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset,\na pre-processed fixed snapshot of Wikipedia.\n""" st.sidebar.markdown(description, unsafe_allow_html=True) a_ = [ """Answer the question""", """View the retrieved document only""", """View the most similar ELI5 question and answer""", """Show me everything, please!""", ] a_ = st.sidebar.checkbox("""Demo options""") if demo_options: a_ = st.sidebar.selectbox( """""", action_list, index=3, ) a_ = action_list.index(action_st) a_ = st.sidebar.selectbox( """""", ["""Show full text of passages""", """Show passage section titles"""], index=0, ) a_ = show_type == """Show full text of passages""" else: a_ = 3 a_ = True a_ = st.sidebar.checkbox("""Retrieval options""") if retrieval_options: a_ = """\n ### Information retriever options\n\n The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding\n trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs.\n The answer is then generated by sequence to sequence model which takes the question and retrieved document as input.\n """ st.sidebar.markdown(retriever_info) a_ = st.sidebar.selectbox("""Which Wikipedia format should the model use?""", ["""wiki40b""", """none"""]) a_ = st.sidebar.selectbox("""Which Wikipedia indexer should the model use?""", ["""dense""", """sparse""", """mixed"""]) else: a_ = """wiki40b""" a_ = """dense""" a_ = """beam""" a_ = 2 a_ = 64 a_ = 256 a_ = None a_ = None a_ = st.sidebar.checkbox("""Generation options""") if generate_options: a_ = """\n ### Answer generation options\n\n The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large)\n weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with\n **beam** search, or **sample** from the decoder\'s output probabilities.\n """ st.sidebar.markdown(generate_info) a_ = st.sidebar.selectbox("""Would you like to use beam search or sample an answer?""", ["""beam""", """sampled"""]) a_ = st.sidebar.slider( """Minimum generation length""", min_value=8, max_value=256, value=64, step=8, format=None, key=None ) a_ = st.sidebar.slider( """Maximum generation length""", min_value=64, max_value=512, value=256, step=16, format=None, key=None ) if sampled == "beam": a_ = st.sidebar.slider("""Beam size""", min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: a_ = st.sidebar.slider( """Nucleus sampling p""", min_value=0.1, max_value=1.0, value=0.95, step=0.01, format=None, key=None ) a_ = st.sidebar.slider( """Temperature""", min_value=0.1, max_value=1.0, value=0.7, step=0.01, format=None, key=None ) a_ = None # start main text a_ = [ """<MY QUESTION>""", """How do people make chocolate?""", """Why do we get a fever when we are sick?""", """How can different animals perceive different colors?""", """What is natural language processing?""", """What\'s the best way to treat a sunburn?""", """What exactly are vitamins ?""", """How does nuclear energy provide electricity?""", """What\'s the difference between viruses and bacteria?""", """Why are flutes classified as woodwinds when most of them are made out of metal ?""", """Why do people like drinking coffee even though it tastes so bad?""", """What happens when wine ages? How does it make the wine taste better?""", """If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?""", """How can we set a date to the beginning or end of an artistic period? Doesn\'t the change happen gradually?""", """How does New Zealand have so many large bird predators?""", ] a_ = st.selectbox( """What would you like to ask? ---- select <MY QUESTION> to enter a new query""", questions_list, index=1, ) if question_s == "<MY QUESTION>": a_ = st.text_input("""Enter your question here:""", """""") else: a_ = question_s if st.button("""Show me!"""): if action in [0, 1, 3]: if index_type == "mixed": a_ , a_ = make_support(question, source=wiki_source, method="""dense""", n_results=10) a_ , a_ = make_support(question, source=wiki_source, method="""sparse""", n_results=10) a_ = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] a_ = support_list[:10] a_ = """<P> """ + """ <P> """.join([res[-1] for res in support_list]) else: a_ , a_ = make_support(question, source=wiki_source, method=index_type, n_results=10) if action in [0, 3]: a_ , a_ = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == """sampled"""), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown("""### The model generated answer is:""") st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown("""--- \n ### The model is drawing information from the following Wikipedia passages:""") for i, res in enumerate(support_list): a_ = """https://en.wikipedia.org/wiki/{}""".format(res[0].replace(""" """, """_""")) a_ = res[1].strip() if sec_titles == "": a_ = """[{}]({})""".format(res[0], wiki_url) else: a_ = sec_titles.split(""" & """) a_ = """ & """.join( ["""[{}]({}#{})""".format(sec.strip(), wiki_url, sec.strip().replace(""" """, """_""")) for sec in sec_list] ) st.markdown( """{0:02d} - **Article**: {1:<18} <br> _Section_: {2}""".format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( """> <span style=\"font-family:arial; font-size:10pt;\">""" + res[-1] + """</span>""", unsafe_allow_html=True ) if action in [2, 3]: a_ = find_nearest_training(question) a_ = nn_train_list[0] st.markdown( """--- \n ### The most similar question in the ELI5 training set was: \n\n {}""".format(train_exple["""title"""]) ) a_ = [ """{}. {}""".format(i + 1, """ \n""".join([line.strip() for line in ans.split("""\n""") if line.strip() != """"""])) for i, (ans, sc) in enumerate(zip(train_exple["""answers"""]["""text"""], train_exple["""answers"""]["""score"""])) if i == 0 or sc > 2 ] st.markdown("""##### Its answers were: \n\n {}""".format("""\n""".join(answers_st))) a_ = """\n---\n\n**Disclaimer**\n\n*The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system.\nEvaluating biases of such a model and ensuring factual generations are still very much open research problems.\nTherefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.*\n""" st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
718
def a__ ( _UpperCamelCase : str ,_UpperCamelCase : str = " " ): __lowerCamelCase = [] __lowerCamelCase = 0 for index, char in enumerate(_UpperCamelCase ): if char == separator: split_words.append(string[last_index:index] ) __lowerCamelCase = index + 1 elif index + 1 == len(_UpperCamelCase ): split_words.append(string[last_index : index + 1] ) return split_words if __name__ == "__main__": from doctest import testmod testmod()
622
0
a_ = """\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n""" a_ = [{"""type""": """code""", """content""": INSTALL_CONTENT}] a_ = { """{processor_class}""": """FakeProcessorClass""", """{model_class}""": """FakeModelClass""", """{object_class}""": """FakeObjectClass""", }
719
import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class __lowerCAmelCase ( lowerCAmelCase__ ): def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = tempfile.mkdtemp() __lowerCamelCase = 8 # DPR tok __lowerCamelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] __lowerCamelCase = os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) __lowerCamelCase = os.path.join(__UpperCAmelCase , DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok __lowerCamelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] __lowerCamelCase = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) __lowerCamelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] __lowerCamelCase = {'''unk_token''': '''<unk>'''} __lowerCamelCase = os.path.join(self.tmpdirname , '''bart_tokenizer''' ) os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) __lowerCamelCase = os.path.join(__UpperCAmelCase , BART_VOCAB_FILES_NAMES['''vocab_file'''] ) __lowerCamelCase = os.path.join(__UpperCAmelCase , BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(__UpperCAmelCase ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(__UpperCAmelCase ) ) def lowerCamelCase ( self ): '''simple docstring''' return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''bart_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_dataset() __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: __lowerCamelCase = dataset __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.get_dummy_dataset() __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''custom''' , ) if from_disk: __lowerCamelCase = os.path.join(self.tmpdirname , '''dataset''' ) __lowerCamelCase = os.path.join(self.tmpdirname , '''index.faiss''' ) dataset.get_index('''embeddings''' ).save(os.path.join(self.tmpdirname , '''index.faiss''' ) ) dataset.drop_index('''embeddings''' ) dataset.save_to_disk(os.path.join(self.tmpdirname , '''dataset''' ) ) del dataset __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , __UpperCAmelCase ) , ) return retriever def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) __lowerCamelCase = os.path.join(self.tmpdirname , '''hf_bert_base.hnswSQ8_correct_phi_128.c_index''' ) dataset.save_faiss_index('''embeddings''' , index_file_name + '''.index.dpr''' ) pickle.dump(dataset['''id'''] , open(index_file_name + '''.index_meta.dpr''' , '''wb''' ) ) __lowerCamelCase = os.path.join(self.tmpdirname , '''psgs_w100.tsv.pkl''' ) __lowerCamelCase = {sample['''id''']: [sample['''text'''], sample['''title''']] for sample in dataset} pickle.dump(__UpperCAmelCase , open(__UpperCAmelCase , '''wb''' ) ) __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''legacy''' , index_path=self.tmpdirname , ) __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: __lowerCamelCase = self.get_dummy_dataset() retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_legacy_index_retriever() __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''text'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''text'''][0] , '''bar''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''text'''][0] , '''foo''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def lowerCamelCase ( self ): '''simple docstring''' import torch __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() __lowerCamelCase = [[5, 7], [10, 11]] __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever(__UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = ( out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , np.ndarray ) __lowerCamelCase = retriever( __UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase , return_tensors='''pt''' , ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = ( # noqa: F841 out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], out['''doc_ids'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dpr_ctx_encoder_tokenizer() __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) retriever.set_ctx_encoder_tokenizer(__UpperCAmelCase ) __lowerCamelCase = [[5, 7], [10, 11]] __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever(__UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase ) self.assertEqual( len(__UpperCAmelCase ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('''tokenized_doc_ids''', '''tokenized_doc_attention_mask''') ) , __UpperCAmelCase ) # check for doc token related keys in dictionary.
622
0
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore a_ = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" a_ = [file for file in filepaths if file != file.lower()] if upper_files: print(f"{len(upper_files)} files contain uppercase characters:") print("""\n""".join(upper_files) + """\n""") a_ = [file for file in filepaths if ' ' in file] if space_files: print(f"{len(space_files)} files contain space characters:") print("""\n""".join(space_files) + """\n""") a_ = [file for file in filepaths if '-' in file] if hyphen_files: print(f"{len(hyphen_files)} files contain hyphen characters:") print("""\n""".join(hyphen_files) + """\n""") a_ = [file for file in filepaths if os.sep not in file] if nodir_files: print(f"{len(nodir_files)} files are not in a directory:") print("""\n""".join(nodir_files) + """\n""") a_ = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
720
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """sail/poolformer_s12""": """https://huggingface.co/sail/poolformer_s12/resolve/main/config.json""", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """poolformer""" def __init__( self , __UpperCAmelCase=3 , __UpperCAmelCase=16 , __UpperCAmelCase=16 , __UpperCAmelCase=3 , __UpperCAmelCase=4.0 , __UpperCAmelCase=[2, 2, 6, 2] , __UpperCAmelCase=[64, 128, 320, 512] , __UpperCAmelCase=[7, 3, 3, 3] , __UpperCAmelCase=[4, 2, 2, 2] , __UpperCAmelCase=[2, 1, 1, 1] , __UpperCAmelCase=4 , __UpperCAmelCase=0.0 , __UpperCAmelCase="gelu" , __UpperCAmelCase=True , __UpperCAmelCase=1E-5 , __UpperCAmelCase=0.02 , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = num_channels __lowerCamelCase = patch_size __lowerCamelCase = stride __lowerCamelCase = padding __lowerCamelCase = pool_size __lowerCamelCase = hidden_sizes __lowerCamelCase = mlp_ratio __lowerCamelCase = depths __lowerCamelCase = patch_sizes __lowerCamelCase = strides __lowerCamelCase = num_encoder_blocks __lowerCamelCase = drop_path_rate __lowerCamelCase = hidden_act __lowerCamelCase = use_layer_scale __lowerCamelCase = layer_scale_init_value __lowerCamelCase = initializer_range super().__init__(**__UpperCAmelCase ) class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = version.parse("""1.11""" ) @property def lowerCamelCase ( self ): '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def lowerCamelCase ( self ): '''simple docstring''' return 2E-3
622
0
import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class __lowerCAmelCase : @staticmethod def lowerCamelCase ( *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' pass def a__ ( _UpperCamelCase : Image ): __lowerCamelCase = hashlib.mda(image.tobytes() ) return m.hexdigest()[:10] def a__ ( _UpperCamelCase : Image ): __lowerCamelCase = np.array(lowerCAmelCase__ ) __lowerCamelCase = npimg.shape return {"hash": hashimage(lowerCAmelCase__ ), "shape": shape} @is_pipeline_test @require_vision @require_torch class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) lowerCAmelCase__ = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = MaskGenerationPipeline(model=_a , image_processor=_a ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' pass @require_tf @unittest.skip('''Image segmentation not implemented in TF''' ) def lowerCamelCase ( self ): '''simple docstring''' pass @slow @require_torch def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline('''mask-generation''' , model='''facebook/sam-vit-huge''' ) __lowerCamelCase = image_segmenter('''http://images.cocodataset.org/val2017/000000039769.jpg''' , points_per_batch=256 ) # Shortening by hashing __lowerCamelCase = [] for i, o in enumerate(outputs['''masks'''] ): new_outupt += [{"mask": mask_to_test_readable(_a ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(_a , decimals=4 ) , [ {'''mask''': {'''hash''': '''115ad19f5f''', '''shape''': (480, 640)}, '''scores''': 1.0_444}, {'''mask''': {'''hash''': '''6affa964c6''', '''shape''': (480, 640)}, '''scores''': 1.021}, {'''mask''': {'''hash''': '''dfe28a0388''', '''shape''': (480, 640)}, '''scores''': 1.0_167}, {'''mask''': {'''hash''': '''c0a5f4a318''', '''shape''': (480, 640)}, '''scores''': 1.0_132}, {'''mask''': {'''hash''': '''fe8065c197''', '''shape''': (480, 640)}, '''scores''': 1.0_053}, {'''mask''': {'''hash''': '''e2d0b7a0b7''', '''shape''': (480, 640)}, '''scores''': 0.9_967}, {'''mask''': {'''hash''': '''453c7844bd''', '''shape''': (480, 640)}, '''scores''': 0.993}, {'''mask''': {'''hash''': '''3d44f2926d''', '''shape''': (480, 640)}, '''scores''': 0.9_909}, {'''mask''': {'''hash''': '''64033ddc3f''', '''shape''': (480, 640)}, '''scores''': 0.9_879}, {'''mask''': {'''hash''': '''801064ff79''', '''shape''': (480, 640)}, '''scores''': 0.9_834}, {'''mask''': {'''hash''': '''6172f276ef''', '''shape''': (480, 640)}, '''scores''': 0.9_716}, {'''mask''': {'''hash''': '''b49e60e084''', '''shape''': (480, 640)}, '''scores''': 0.9_612}, {'''mask''': {'''hash''': '''a811e775fd''', '''shape''': (480, 640)}, '''scores''': 0.9_599}, {'''mask''': {'''hash''': '''a6a8ebcf4b''', '''shape''': (480, 640)}, '''scores''': 0.9_552}, {'''mask''': {'''hash''': '''9d8257e080''', '''shape''': (480, 640)}, '''scores''': 0.9_532}, {'''mask''': {'''hash''': '''32de6454a8''', '''shape''': (480, 640)}, '''scores''': 0.9_516}, {'''mask''': {'''hash''': '''af3d4af2c8''', '''shape''': (480, 640)}, '''scores''': 0.9_499}, {'''mask''': {'''hash''': '''3c6db475fb''', '''shape''': (480, 640)}, '''scores''': 0.9_483}, {'''mask''': {'''hash''': '''c290813fb9''', '''shape''': (480, 640)}, '''scores''': 0.9_464}, {'''mask''': {'''hash''': '''b6f0b8f606''', '''shape''': (480, 640)}, '''scores''': 0.943}, {'''mask''': {'''hash''': '''92ce16bfdf''', '''shape''': (480, 640)}, '''scores''': 0.943}, {'''mask''': {'''hash''': '''c749b25868''', '''shape''': (480, 640)}, '''scores''': 0.9_408}, {'''mask''': {'''hash''': '''efb6cab859''', '''shape''': (480, 640)}, '''scores''': 0.9_335}, {'''mask''': {'''hash''': '''1ff2eafb30''', '''shape''': (480, 640)}, '''scores''': 0.9_326}, {'''mask''': {'''hash''': '''788b798e24''', '''shape''': (480, 640)}, '''scores''': 0.9_262}, {'''mask''': {'''hash''': '''abea804f0e''', '''shape''': (480, 640)}, '''scores''': 0.8_999}, {'''mask''': {'''hash''': '''7b9e8ddb73''', '''shape''': (480, 640)}, '''scores''': 0.8_986}, {'''mask''': {'''hash''': '''cd24047c8a''', '''shape''': (480, 640)}, '''scores''': 0.8_984}, {'''mask''': {'''hash''': '''6943e6bcbd''', '''shape''': (480, 640)}, '''scores''': 0.8_873}, {'''mask''': {'''hash''': '''b5f47c9191''', '''shape''': (480, 640)}, '''scores''': 0.8_871} ] , ) # fmt: on @require_torch @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/sam-vit-huge''' __lowerCamelCase = pipeline('''mask-generation''' , model=_a ) __lowerCamelCase = image_segmenter( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , pred_iou_thresh=1 , points_per_batch=256 ) # Shortening by hashing __lowerCamelCase = [] for i, o in enumerate(outputs['''masks'''] ): new_outupt += [{"mask": mask_to_test_readable(_a ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(_a , decimals=4 ) , [ {'''mask''': {'''hash''': '''115ad19f5f''', '''shape''': (480, 640)}, '''scores''': 1.0_444}, {'''mask''': {'''hash''': '''6affa964c6''', '''shape''': (480, 640)}, '''scores''': 1.0_210}, {'''mask''': {'''hash''': '''dfe28a0388''', '''shape''': (480, 640)}, '''scores''': 1.0_167}, {'''mask''': {'''hash''': '''c0a5f4a318''', '''shape''': (480, 640)}, '''scores''': 1.0_132}, {'''mask''': {'''hash''': '''fe8065c197''', '''shape''': (480, 640)}, '''scores''': 1.0_053}, ] , )
721
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """uclanlp/visualbert-vqa""": """https://huggingface.co/uclanlp/visualbert-vqa/resolve/main/config.json""", """uclanlp/visualbert-vqa-pre""": """https://huggingface.co/uclanlp/visualbert-vqa-pre/resolve/main/config.json""", """uclanlp/visualbert-vqa-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-vqa-coco-pre/resolve/main/config.json""" ), """uclanlp/visualbert-vcr""": """https://huggingface.co/uclanlp/visualbert-vcr/resolve/main/config.json""", """uclanlp/visualbert-vcr-pre""": """https://huggingface.co/uclanlp/visualbert-vcr-pre/resolve/main/config.json""", """uclanlp/visualbert-vcr-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-vcr-coco-pre/resolve/main/config.json""" ), """uclanlp/visualbert-nlvr2""": """https://huggingface.co/uclanlp/visualbert-nlvr2/resolve/main/config.json""", """uclanlp/visualbert-nlvr2-pre""": """https://huggingface.co/uclanlp/visualbert-nlvr2-pre/resolve/main/config.json""", """uclanlp/visualbert-nlvr2-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-nlvr2-coco-pre/resolve/main/config.json""" ) # See all VisualBERT models at https://huggingface.co/models?filter=visual_bert } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """visual_bert""" def __init__( self , __UpperCAmelCase=30522 , __UpperCAmelCase=768 , __UpperCAmelCase=512 , __UpperCAmelCase=12 , __UpperCAmelCase=12 , __UpperCAmelCase=3072 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase=1 , __UpperCAmelCase=0 , __UpperCAmelCase=2 , **__UpperCAmelCase , ): '''simple docstring''' super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) __lowerCamelCase = vocab_size __lowerCamelCase = max_position_embeddings __lowerCamelCase = hidden_size __lowerCamelCase = visual_embedding_dim __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = initializer_range __lowerCamelCase = type_vocab_size __lowerCamelCase = layer_norm_eps __lowerCamelCase = bypass_transformer __lowerCamelCase = special_visual_initialize
622
0
import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() a_ = logging.get_logger(__name__) a_ = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", } a_ = [ """lm_head""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def a__ ( _UpperCamelCase : int ,_UpperCamelCase : Any ,_UpperCamelCase : Optional[Any] ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : str ): for attribute in key.split('''.''' ): __lowerCamelCase = getattr(_lowerCamelCase ,_lowerCamelCase ) if weight_type is not None: __lowerCamelCase = getattr(_lowerCamelCase ,_lowerCamelCase ).shape else: __lowerCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( F"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": __lowerCamelCase = value elif weight_type == "weight_g": __lowerCamelCase = value elif weight_type == "weight_v": __lowerCamelCase = value elif weight_type == "bias": __lowerCamelCase = value else: __lowerCamelCase = value logger.info(F"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : Optional[int] ): __lowerCamelCase = [] __lowerCamelCase = fairseq_model.state_dict() __lowerCamelCase = hf_model.feature_extractor __lowerCamelCase = hf_model.adapter for name, value in fairseq_dict.items(): __lowerCamelCase = False if "conv_layers" in name: load_conv_layer( _lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,hf_model.config.feat_extract_norm == '''group''' ,) __lowerCamelCase = True elif any(x in name for x in ['''adaptor''', '''w2v_encoder.proj.''', '''w2v_proj_ln.'''] ): load_adapter(_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ) __lowerCamelCase = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: __lowerCamelCase = True if "*" in mapped_key: __lowerCamelCase = name.split(_lowerCamelCase )[0].split('''.''' )[-2] __lowerCamelCase = mapped_key.replace('''*''' ,_lowerCamelCase ) if "weight_g" in name: __lowerCamelCase = '''weight_g''' elif "weight_v" in name: __lowerCamelCase = '''weight_v''' elif "bias" in name: __lowerCamelCase = '''bias''' elif "weight" in name: __lowerCamelCase = '''weight''' else: __lowerCamelCase = None set_recursively(_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ) continue if not is_used: unused_weights.append(_lowerCamelCase ) logger.warning(F"""Unused weights: {unused_weights}""" ) def a__ ( _UpperCamelCase : List[str] ,_UpperCamelCase : Optional[Any] ,_UpperCamelCase : int ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : str ): __lowerCamelCase = full_name.split('''conv_layers.''' )[-1] __lowerCamelCase = name.split('''.''' ) __lowerCamelCase = int(items[0] ) __lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __lowerCamelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __lowerCamelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) __lowerCamelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) __lowerCamelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(_lowerCamelCase ) def a__ ( _UpperCamelCase : Any ,_UpperCamelCase : str ,_UpperCamelCase : Dict ,_UpperCamelCase : Any ): __lowerCamelCase = full_name.split('''adaptor.''' )[-1] __lowerCamelCase = name.split('''.''' ) if items[1].isdigit(): __lowerCamelCase = int(items[1] ) else: __lowerCamelCase = None if "adaptor" not in full_name: if "proj_ln" in full_name: # has to be layer norm if "bias" in name: assert ( value.shape == adapter.proj_layer_norm.bias.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj_layer_norm.bias.data.shape} was found.""" __lowerCamelCase = value logger.info(F"""Adapter proj layer norm bias was initialized from {full_name}.""" ) if "weight" in name: assert ( value.shape == adapter.proj_layer_norm.weight.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj_layer_norm.weight.data.shape} was found.""" __lowerCamelCase = value else: # has to be projection layer if "bias" in name: assert ( value.shape == adapter.proj.bias.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj.bias.data.shape} was found.""" __lowerCamelCase = value logger.info(F"""Adapter proj layer bias was initialized from {full_name}.""" ) if "weight" in name: assert ( value.shape == adapter.proj.weight.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj.weight.data.shape} was found.""" __lowerCamelCase = value logger.info(F"""Adapter proj layer weight was initialized from {full_name}.""" ) elif isinstance(_lowerCamelCase ,_lowerCamelCase ): if "bias" in name: assert ( value.shape == adapter.layers[layer_id].conv.bias.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.bias.data.shape} was found.""" __lowerCamelCase = value logger.info(F"""Adapter layer {layer_id} bias was initialized from {full_name}.""" ) elif "weight" in name: assert ( value.shape == adapter.layers[layer_id].conv.weight.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.weight.data.shape} was found.""" __lowerCamelCase = value logger.info(F"""Adapter layer {layer_id} bias was initialized from {full_name}.""" ) else: unused_weights.append(_lowerCamelCase ) def a__ ( _UpperCamelCase : List[str] ): __lowerCamelCase ,__lowerCamelCase = emb.weight.shape __lowerCamelCase = nn.Linear(_lowerCamelCase ,_lowerCamelCase ,bias=_lowerCamelCase ) __lowerCamelCase = emb.weight.data return lin_layer @torch.no_grad() def a__ ( _UpperCamelCase : Optional[int] ,_UpperCamelCase : str ,_UpperCamelCase : Optional[Any] ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : int ,_UpperCamelCase : str ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : str ,_UpperCamelCase : Tuple ,_UpperCamelCase : Optional[Any] ,_UpperCamelCase : Union[str, Any] ,): __lowerCamelCase = WavaVecaConfig.from_pretrained( _lowerCamelCase ,add_adapter=_lowerCamelCase ,adapter_stride=_lowerCamelCase ,adapter_kernel_size=_lowerCamelCase ,use_auth_token=_lowerCamelCase ,output_hidden_size=_lowerCamelCase ,) __lowerCamelCase = MBartConfig.from_pretrained(_lowerCamelCase ) # load model __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] ,arg_overrides={ '''config_yaml''': config_yaml_path, '''data''': '''/'''.join(dict_path.split('''/''' )[:-1] ), '''w2v_path''': checkpoint_path, '''load_pretrained_decoder_from''': None, } ,) __lowerCamelCase = model[0].eval() # load feature extractor __lowerCamelCase = WavaVecaFeatureExtractor.from_pretrained(_lowerCamelCase ,use_auth_token=_lowerCamelCase ) # set weights for wav2vec2 encoder __lowerCamelCase = WavaVecaModel(_lowerCamelCase ) recursively_load_weights_wavaveca(model.encoder ,_lowerCamelCase ) # load decoder weights __lowerCamelCase = MBartForCausalLM(_lowerCamelCase ) __lowerCamelCase ,__lowerCamelCase = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() ,strict=_lowerCamelCase ) logger.warning(F"""The following keys are missing when loading the decoder weights: {missing_keys}""" ) logger.warning(F"""The following keys are unexpected when loading the decoder weights: {unexpected_keys}""" ) __lowerCamelCase = SpeechEncoderDecoderModel(encoder=_lowerCamelCase ,decoder=_lowerCamelCase ) __lowerCamelCase = False __lowerCamelCase = MBartaaTokenizer(_lowerCamelCase ) tokenizer.save_pretrained(_lowerCamelCase ) __lowerCamelCase = hf_wavavec.config.to_dict() __lowerCamelCase = tokenizer.pad_token_id __lowerCamelCase = tokenizer.bos_token_id __lowerCamelCase = tokenizer.eos_token_id __lowerCamelCase = '''mbart50''' __lowerCamelCase = '''wav2vec2''' __lowerCamelCase = tokenizer.eos_token_id __lowerCamelCase = 25_00_04 __lowerCamelCase = tokenizer.eos_token_id __lowerCamelCase = SpeechEncoderDecoderConfig.from_dict(_lowerCamelCase ) hf_wavavec.save_pretrained(_lowerCamelCase ) feature_extractor.save_pretrained(_lowerCamelCase ) if __name__ == "__main__": a_ = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_yaml_path""", default=None, type=str, help="""Path to yaml file of fine-tuned model""") parser.add_argument( """--encoder_config_path""", default="""facebook/wav2vec2-xls-r-1b""", type=str, help="""Path to hf encoder wav2vec2 checkpoint config""", ) parser.add_argument( """--decoder_config_path""", default="""facebook/mbart-large-50-one-to-many-mmt""", type=str, help="""Path to hf decoder checkpoint config""", ) parser.add_argument("""--add_adapter""", default=True, type=bool, help="""whethere to add model adapter layers""") parser.add_argument("""--adapter_stride""", default=2, type=int, help="""stride of adapter layers""") parser.add_argument("""--adapter_kernel_size""", default=3, type=int, help="""kernel size of adapter layers""") parser.add_argument("""--encoder_output_dim""", default=1_024, type=int, help="""encoder output dim""") parser.add_argument("""--start_token_id""", default=250_004, type=int, help="""`decoder_start_token_id` of model config""") a_ = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, args.config_yaml_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, add_adapter=args.add_adapter, adapter_kernel_size=args.adapter_kernel_size, adapter_stride=args.adapter_stride, decoder_start_token_id=args.start_token_id, encoder_output_dim=args.encoder_output_dim, )
700
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging a_ = logging.get_logger(__name__) a_ = {"""vocab_file""": """spiece.model"""} a_ = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", } } a_ = { """albert-base-v1""": 512, """albert-large-v1""": 512, """albert-xlarge-v1""": 512, """albert-xxlarge-v1""": 512, """albert-base-v2""": 512, """albert-large-v2""": 512, """albert-xlarge-v2""": 512, """albert-xxlarge-v2""": 512, } a_ = """▁""" class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , __UpperCAmelCase , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="<unk>" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="<pad>" , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[MASK]" , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. __lowerCamelCase = ( AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase , normalized=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token ) __lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__UpperCAmelCase , remove_space=__UpperCAmelCase , keep_accents=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) __lowerCamelCase = do_lower_case __lowerCamelCase = remove_space __lowerCamelCase = keep_accents __lowerCamelCase = vocab_file __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCAmelCase ) @property def lowerCamelCase ( self ): '''simple docstring''' return len(self.sp_model ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): '''simple docstring''' __lowerCamelCase = self.__dict__.copy() __lowerCamelCase = None return state def __setstate__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCamelCase = {} __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' if self.remove_space: __lowerCamelCase = ''' '''.join(inputs.strip().split() ) else: __lowerCamelCase = inputs __lowerCamelCase = outputs.replace('''``''' , '''"''' ).replace('''\'\'''' , '''"''' ) if not self.keep_accents: __lowerCamelCase = unicodedata.normalize('''NFKD''' , __UpperCAmelCase ) __lowerCamelCase = ''''''.join([c for c in outputs if not unicodedata.combining(__UpperCAmelCase )] ) if self.do_lower_case: __lowerCamelCase = outputs.lower() return outputs def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.preprocess_text(__UpperCAmelCase ) __lowerCamelCase = self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) __lowerCamelCase = [] for piece in pieces: if len(__UpperCAmelCase ) > 1 and piece[-1] == str(''',''' ) and piece[-2].isdigit(): __lowerCamelCase = self.sp_model.EncodeAsPieces(piece[:-1].replace(__UpperCAmelCase , '''''' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __lowerCamelCase = cur_pieces[1:] else: __lowerCamelCase = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__UpperCAmelCase ) else: new_pieces.append(__UpperCAmelCase ) return new_pieces def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.PieceToId(__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.IdToPiece(__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = '''''' __lowerCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__UpperCAmelCase ) + token __lowerCamelCase = True __lowerCamelCase = [] else: current_sub_tokens.append(__UpperCAmelCase ) __lowerCamelCase = False out_string += self.sp_model.decode(__UpperCAmelCase ) return out_string.strip() def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase ) if token_ids_a is not None: return [1] + ([0] * len(__UpperCAmelCase )) + [1] + ([0] * len(__UpperCAmelCase )) + [1] return [1] + ([0] * len(__UpperCAmelCase )) + [1] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' if not os.path.isdir(__UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCamelCase = os.path.join( __UpperCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , '''wb''' ) as fi: __lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,)
622
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a_ = { """configuration_table_transformer""": [ """TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TableTransformerConfig""", """TableTransformerOnnxConfig""", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ """TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TableTransformerForObjectDetection""", """TableTransformerModel""", """TableTransformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
701
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed a_ = """true""" def a__ ( _UpperCamelCase : Union[str, Any] ,_UpperCamelCase : List[str]=82 ,_UpperCamelCase : Optional[Any]=16 ): set_seed(42 ) __lowerCamelCase = RegressionModel() __lowerCamelCase = deepcopy(_UpperCamelCase ) __lowerCamelCase = RegressionDataset(length=_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,batch_size=_UpperCamelCase ) model.to(accelerator.device ) __lowerCamelCase ,__lowerCamelCase = accelerator.prepare(_UpperCamelCase ,_UpperCamelCase ) return model, ddp_model, dataloader def a__ ( _UpperCamelCase : Accelerator ,_UpperCamelCase : str=False ): __lowerCamelCase = AutoTokenizer.from_pretrained('''hf-internal-testing/mrpc-bert-base-cased''' ) __lowerCamelCase = load_dataset('''glue''' ,'''mrpc''' ,split='''validation''' ) def tokenize_function(_UpperCamelCase : int ): __lowerCamelCase = tokenizer(examples['''sentence1'''] ,examples['''sentence2'''] ,truncation=_UpperCamelCase ,max_length=_UpperCamelCase ) return outputs with accelerator.main_process_first(): __lowerCamelCase = dataset.map( _UpperCamelCase ,batched=_UpperCamelCase ,remove_columns=['''idx''', '''sentence1''', '''sentence2'''] ,) __lowerCamelCase = tokenized_datasets.rename_column('''label''' ,'''labels''' ) def collate_fn(_UpperCamelCase : Any ): if use_longest: return tokenizer.pad(_UpperCamelCase ,padding='''longest''' ,return_tensors='''pt''' ) return tokenizer.pad(_UpperCamelCase ,padding='''max_length''' ,max_length=1_28 ,return_tensors='''pt''' ) return DataLoader(_UpperCamelCase ,shuffle=_UpperCamelCase ,collate_fn=_UpperCamelCase ,batch_size=16 ) def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : List[str] ): __lowerCamelCase = Accelerator(dispatch_batches=_UpperCamelCase ,split_batches=_UpperCamelCase ) __lowerCamelCase = get_dataloader(_UpperCamelCase ,not dispatch_batches ) __lowerCamelCase = AutoModelForSequenceClassification.from_pretrained( '''hf-internal-testing/mrpc-bert-base-cased''' ,return_dict=_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = accelerator.prepare(_UpperCamelCase ,_UpperCamelCase ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : Optional[Any] ,_UpperCamelCase : Union[str, Any] ): __lowerCamelCase = [] for batch in dataloader: __lowerCamelCase ,__lowerCamelCase = batch.values() with torch.no_grad(): __lowerCamelCase = model(_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) __lowerCamelCase ,__lowerCamelCase = [], [] for logit, targ in logits_and_targets: logits.append(_UpperCamelCase ) targs.append(_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = torch.cat(_UpperCamelCase ), torch.cat(_UpperCamelCase ) return logits, targs def a__ ( _UpperCamelCase : Accelerator ,_UpperCamelCase : List[Any]=82 ,_UpperCamelCase : str=False ,_UpperCamelCase : List[str]=False ,_UpperCamelCase : Optional[int]=16 ): __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = get_basic_setup(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = generate_predictions(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ) assert ( len(_UpperCamelCase ) == num_samples ), F"""Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(_UpperCamelCase )}""" def a__ ( _UpperCamelCase : bool = False ,_UpperCamelCase : bool = False ): __lowerCamelCase = evaluate.load('''glue''' ,'''mrpc''' ) __lowerCamelCase ,__lowerCamelCase = get_mrpc_setup(_UpperCamelCase ,_UpperCamelCase ) # First do baseline __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = setup['''no'''] model.to(_UpperCamelCase ) model.eval() for batch in dataloader: batch.to(_UpperCamelCase ) with torch.inference_mode(): __lowerCamelCase = model(**_UpperCamelCase ) __lowerCamelCase = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=_UpperCamelCase ,references=batch['''labels'''] ) __lowerCamelCase = metric.compute() # Then do distributed __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = setup['''ddp'''] model.eval() for batch in dataloader: with torch.inference_mode(): __lowerCamelCase = model(**_UpperCamelCase ) __lowerCamelCase = outputs.logits.argmax(dim=-1 ) __lowerCamelCase = batch['''labels'''] __lowerCamelCase ,__lowerCamelCase = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=_UpperCamelCase ,references=_UpperCamelCase ) __lowerCamelCase = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] ,distributed[key] ), F"""Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n""" def a__ ( ): __lowerCamelCase = Accelerator(split_batches=_UpperCamelCase ,dispatch_batches=_UpperCamelCase ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print('''**Testing gather_for_metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(F"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`""" ) test_mrpc(_UpperCamelCase ,_UpperCamelCase ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test torch metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: __lowerCamelCase = Accelerator(split_batches=_UpperCamelCase ,dispatch_batches=_UpperCamelCase ) if accelerator.is_local_main_process: print(F"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99""" ) test_torch_metrics(_UpperCamelCase ,99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test last batch is not dropped when perfectly divisible**''' ) __lowerCamelCase = Accelerator() test_torch_metrics(_UpperCamelCase ,5_12 ) accelerator.state._reset_state() def a__ ( _UpperCamelCase : Optional[int] ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
622
0
from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers a_ = [ """python""", """tqdm""", """regex""", """requests""", """packaging""", """filelock""", """numpy""", """tokenizers""", """huggingface-hub""", """safetensors""", """accelerate""", """pyyaml""", ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f"can\'t find {pkg} in {deps.keys()}, check dependency_versions_table.py") def a__ ( _UpperCamelCase : Optional[int] ,_UpperCamelCase : Dict=None ): require_version(deps[pkg] ,SCREAMING_SNAKE_CASE_ )
702
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = StableDiffusionXLImgaImgPipeline lowerCAmelCase__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} lowerCAmelCase__ = PipelineTesterMixin.required_optional_params - {"""latents"""} lowerCAmelCase__ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowerCAmelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS lowerCAmelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowerCamelCase ( self ): '''simple docstring''' torch.manual_seed(0 ) __lowerCamelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=__UpperCAmelCase , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) __lowerCamelCase = EulerDiscreteScheduler( beta_start=0.00_085 , beta_end=0.012 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) __lowerCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='''gelu''' , projection_dim=32 , ) __lowerCamelCase = CLIPTextModel(__UpperCAmelCase ) __lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__UpperCAmelCase ) __lowerCamelCase = CLIPTextModelWithProjection(__UpperCAmelCase ) __lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__UpperCAmelCase ) __lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase=0 ): '''simple docstring''' __lowerCamelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__UpperCAmelCase ) ).to(__UpperCAmelCase ) __lowerCamelCase = image / 2 + 0.5 if str(__UpperCAmelCase ).startswith('''mps''' ): __lowerCamelCase = torch.manual_seed(__UpperCAmelCase ) else: __lowerCamelCase = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) __lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.75, } return inputs def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = StableDiffusionXLImgaImgPipeline(**__UpperCAmelCase ) __lowerCamelCase = sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase = sd_pipe(**__UpperCAmelCase ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCamelCase = np.array([0.4_656, 0.4_840, 0.4_439, 0.6_698, 0.5_574, 0.4_524, 0.5_799, 0.5_943, 0.5_165] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase ( self ): '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def lowerCamelCase ( self ): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = StableDiffusionXLImgaImgPipeline(**__UpperCAmelCase ) __lowerCamelCase = sd_pipe.to(__UpperCAmelCase ) __lowerCamelCase = sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) # forward without prompt embeds __lowerCamelCase = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase = 3 * ['''this is a negative prompt'''] __lowerCamelCase = negative_prompt __lowerCamelCase = 3 * [inputs['''prompt''']] __lowerCamelCase = sd_pipe(**__UpperCAmelCase ) __lowerCamelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds __lowerCamelCase = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase = 3 * ['''this is a negative prompt'''] __lowerCamelCase = 3 * [inputs.pop('''prompt''' )] ( ( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) , ) = sd_pipe.encode_prompt(__UpperCAmelCase , negative_prompt=__UpperCAmelCase ) __lowerCamelCase = sd_pipe( **__UpperCAmelCase , prompt_embeds=__UpperCAmelCase , negative_prompt_embeds=__UpperCAmelCase , pooled_prompt_embeds=__UpperCAmelCase , negative_pooled_prompt_embeds=__UpperCAmelCase , ) __lowerCamelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase="cpu" , __UpperCAmelCase=torch.floataa , __UpperCAmelCase=0 ): '''simple docstring''' __lowerCamelCase = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) __lowerCamelCase = np.random.RandomState(__UpperCAmelCase ).standard_normal((1, 4, 64, 64) ) __lowerCamelCase = torch.from_numpy(__UpperCAmelCase ).to(device=__UpperCAmelCase , dtype=__UpperCAmelCase ) __lowerCamelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = self.get_inputs(__UpperCAmelCase ) __lowerCamelCase = pipe(**__UpperCAmelCase ).images __lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.49_493, 0.47_896, 0.40_798, 0.54_214, 0.53_212, 0.48_202, 0.47_656, 0.46_329, 0.48_506] ) assert np.abs(image_slice - expected_slice ).max() < 7E-3
622
0
import numpy class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. __lowerCamelCase = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. __lowerCamelCase = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. __lowerCamelCase = numpy.random.rand(3 , 1 ) # Real output values provided. __lowerCamelCase = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. __lowerCamelCase = numpy.zeros(output_array.shape ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. __lowerCamelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. __lowerCamelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) __lowerCamelCase = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) __lowerCamelCase = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' for iteration in range(1 , iterations + 1 ): __lowerCamelCase = self.feedforward() self.back_propagation() if give_loss: __lowerCamelCase = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F"""Iteration {iteration} Loss: {loss}""" ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = input_arr __lowerCamelCase = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) __lowerCamelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) __lowerCamelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def a__ ( _UpperCamelCase : Union[str, Any] ): return 1 / (1 + numpy.exp(-value )) def a__ ( _UpperCamelCase : str ): return (value) * (1 - (value)) def a__ ( ): __lowerCamelCase = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) ,dtype=numpy.floataa ,) # True output values for the given input values. __lowerCamelCase = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) ,dtype=numpy.floataa ) # Calling neural network class. __lowerCamelCase = TwoHiddenLayerNeuralNetwork( input_array=SCREAMING_SNAKE_CASE_ ,output_array=SCREAMING_SNAKE_CASE_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=SCREAMING_SNAKE_CASE_ ,iterations=10 ,give_loss=SCREAMING_SNAKE_CASE_ ) return neural_network.predict(numpy.array(([1, 1, 1]) ,dtype=numpy.floataa ) ) if __name__ == "__main__": example()
703
import torch from diffusers import StableDiffusionPipeline a_ = """path-to-your-trained-model""" a_ = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to("""cuda""") a_ = """A photo of sks dog in a bucket""" a_ = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save("""dog-bucket.png""")
622
0
import requests a_ = """""" # <-- Put your OpenWeatherMap appid here! a_ = """https://api.openweathermap.org/data/2.5/""" def a__ ( _UpperCamelCase : str = "Chicago" ,_UpperCamelCase : str = APPID ): return requests.get(URL_BASE + '''weather''' ,params=locals() ).json() def a__ ( _UpperCamelCase : str = "Kolkata, India" ,_UpperCamelCase : str = APPID ): return requests.get(URL_BASE + '''forecast''' ,params=locals() ).json() def a__ ( _UpperCamelCase : float = 55.68 ,_UpperCamelCase : float = 12.57 ,_UpperCamelCase : str = APPID ): return requests.get(URL_BASE + '''onecall''' ,params=locals() ).json() if __name__ == "__main__": from pprint import pprint while True: a_ = input("""Enter a location:""").strip() if location: pprint(current_weather(location)) else: break
704
import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class __lowerCAmelCase : @staticmethod def lowerCamelCase ( *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' pass def a__ ( _UpperCamelCase : List[str] ): return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. a_ = ( """https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png""" ) @is_pipeline_test @require_torch @require_vision class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model=__UpperCAmelCase , tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '''''' ) ) ) __lowerCamelCase = '''What is the placebo?''' __lowerCamelCase = [ { '''image''': load_image(__UpperCAmelCase ), '''question''': question, }, { '''image''': image, '''question''': question, }, { '''image''': image, '''question''': question, '''word_boxes''': word_boxes, }, ] return dqa_pipeline, examples def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = dqa_pipeline(__UpperCAmelCase , top_k=2 ) self.assertEqual( __UpperCAmelCase , [ [ {'''score''': ANY(__UpperCAmelCase ), '''answer''': ANY(__UpperCAmelCase ), '''start''': ANY(__UpperCAmelCase ), '''end''': ANY(__UpperCAmelCase )}, {'''score''': ANY(__UpperCAmelCase ), '''answer''': ANY(__UpperCAmelCase ), '''start''': ANY(__UpperCAmelCase ), '''end''': ANY(__UpperCAmelCase )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline('''document-question-answering''' , model='''hf-internal-testing/tiny-random-layoutlmv2''' ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''How many cats are there?''' __lowerCamelCase = [ {'''score''': 0.0_001, '''answer''': '''oy 2312/2019''', '''start''': 38, '''end''': 39}, {'''score''': 0.0_001, '''answer''': '''oy 2312/2019 DUE''', '''start''': 38, '''end''': 40}, ] __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , __UpperCAmelCase ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , __UpperCAmelCase ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably __lowerCamelCase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(__UpperCAmelCase , [] ) # We can optionnally pass directly the words and bounding boxes __lowerCamelCase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' __lowerCamelCase = [] __lowerCamelCase = [] __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , words=__UpperCAmelCase , boxes=__UpperCAmelCase , top_k=2 ) self.assertEqual(__UpperCAmelCase , [] ) @slow @require_torch @require_detectrona @require_pytesseract def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_944, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_009, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_944, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_009, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.9_944, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_009, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , max_seq_len=50 , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_974, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_948, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_974, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_948, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.9_974, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_948, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__UpperCAmelCase ) __lowerCamelCase = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__UpperCAmelCase , revision='''3dc6de3''' , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] ] * 2 , ) __lowerCamelCase = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '''''' ) ) ) # This model should also work if `image` is set to None __lowerCamelCase = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__UpperCAmelCase ) __lowerCamelCase = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__UpperCAmelCase , revision='''3dc6de3''' , max_seq_len=50 , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_999, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_998, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.9_999, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_998, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] ] * 2 , ) __lowerCamelCase = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '''''' ) ) ) # This model should also work if `image` is set to None __lowerCamelCase = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_999, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_998, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) @slow @require_torch def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model='''naver-clova-ix/donut-base-finetuned-docvqa''' , tokenizer=AutoTokenizer.from_pretrained('''naver-clova-ix/donut-base-finetuned-docvqa''' ) , feature_extractor='''naver-clova-ix/donut-base-finetuned-docvqa''' , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , [{'''answer''': '''us-001'''}] ) @require_tf @unittest.skip('''Document question answering not implemented in TF''' ) def lowerCamelCase ( self ): '''simple docstring''' pass
622
0
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable a_ = { """configuration_gpt_neox_japanese""": ["""GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTNeoXJapaneseConfig"""], """tokenization_gpt_neox_japanese""": ["""GPTNeoXJapaneseTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ """GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST""", """GPTNeoXJapaneseForCausalLM""", """GPTNeoXJapaneseLayer""", """GPTNeoXJapaneseModel""", """GPTNeoXJapanesePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_gpt_neox_japanese import GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXJapaneseConfig from .tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox_japanese import ( GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseLayer, GPTNeoXJapaneseModel, GPTNeoXJapanesePreTrainedModel, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
705
import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin a_ = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = XLMProphetNetTokenizer lowerCAmelCase__ = False lowerCAmelCase__ = True def lowerCamelCase ( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __lowerCamelCase = XLMProphetNetTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''[PAD]''' __lowerCamelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCAmelCase ) , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''[PAD]''' ) self.assertEqual(vocab_keys[1] , '''[CLS]''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(__UpperCAmelCase ) , 1012 ) def lowerCamelCase ( self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1012 ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = XLMProphetNetTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) __lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__UpperCAmelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __lowerCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] , ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''[UNK]''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''[UNK]''', '''.''', ] , ) @cached_property def lowerCamelCase ( self ): '''simple docstring''' return XLMProphetNetTokenizer.from_pretrained('''microsoft/xprophetnet-large-wiki100-cased''' ) @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''Hello World!''' __lowerCamelCase = [35389, 6672, 49, 2] self.assertListEqual(__UpperCAmelCase , self.big_tokenizer.encode(__UpperCAmelCase ) ) @slow def lowerCamelCase ( self ): '''simple docstring''' # fmt: off __lowerCamelCase = {'''input_ids''': [[11073, 82783, 18, 26, 82783, 549, 51540, 248, 17209, 1301, 217, 20, 215186, 1325, 147, 17209, 1301, 217, 20, 56370, 53, 122020, 20, 16477, 27, 87355, 4548, 20, 4728, 78392, 17, 159969, 18, 26, 24491, 629, 15, 538, 22704, 5439, 15, 2788, 24491, 9885, 15, 43534, 605, 15, 814, 18403, 33200, 29, 15, 43534, 24458, 12410, 111, 24966, 83669, 9637, 144068, 26, 850, 22346, 27, 147, 24966, 83669, 83490, 26, 39113, 735, 27, 689, 656, 2800, 1339, 4600, 53, 122020, 115785, 34, 816, 1339, 46887, 18, 147, 53905, 1951, 42238, 41170, 17732, 834, 436, 15, 27523, 98733, 217, 147, 5542, 4981, 930, 17347, 16, 2], [20091, 629, 94, 82786, 58, 490, 20, 1528, 84, 53905, 344, 80592, 110128, 18822, 5267, 1306, 62, 152537, 308, 7997, 401, 124427, 549, 35442, 225, 109, 15055, 25748, 147, 7119, 43712, 34, 767, 135366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 63784, 119466, 17, 147808, 88214, 18, 656, 81, 32, 3296, 10280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__UpperCAmelCase , model_name='''microsoft/xprophetnet-large-wiki100-cased''' , revision='''1acad1643ddd54a44df6a1b797ada8373685d90e''' , )
622
0
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging a_ = logging.get_logger(__name__) a_ = {'''vocab_file''': '''spiece.model'''} a_ = { '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', } } a_ = { '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } # Segments (not really needed) a_ = 0 a_ = 1 a_ = 2 a_ = 3 a_ = 4 class __lowerCAmelCase ( __a ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = '''left''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase="<s>" , __UpperCAmelCase="</s>" , __UpperCAmelCase="<unk>" , __UpperCAmelCase="<sep>" , __UpperCAmelCase="<pad>" , __UpperCAmelCase="<cls>" , __UpperCAmelCase="<mask>" , __UpperCAmelCase=["<eop>", "<eod>"] , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it __lowerCamelCase = AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else mask_token __lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=snake_case__ , remove_space=snake_case__ , keep_accents=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , sep_token=snake_case__ , pad_token=snake_case__ , cls_token=snake_case__ , mask_token=snake_case__ , additional_special_tokens=snake_case__ , sp_model_kwargs=self.sp_model_kwargs , **snake_case__ , ) __lowerCamelCase = 3 __lowerCamelCase = do_lower_case __lowerCamelCase = remove_space __lowerCamelCase = keep_accents __lowerCamelCase = vocab_file __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case__ ) @property def lowerCamelCase ( self ): '''simple docstring''' return len(self.sp_model ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = {self.convert_ids_to_tokens(snake_case__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): '''simple docstring''' __lowerCamelCase = self.__dict__.copy() __lowerCamelCase = None return state def __setstate__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCamelCase = {} __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' if self.remove_space: __lowerCamelCase = ''' '''.join(inputs.strip().split() ) else: __lowerCamelCase = inputs __lowerCamelCase = outputs.replace('''``''' , '''\"''' ).replace('''\'\'''' , '''\"''' ) if not self.keep_accents: __lowerCamelCase = unicodedata.normalize('''NFKD''' , snake_case__ ) __lowerCamelCase = ''''''.join([c for c in outputs if not unicodedata.combining(snake_case__ )] ) if self.do_lower_case: __lowerCamelCase = outputs.lower() return outputs def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.preprocess_text(snake_case__ ) __lowerCamelCase = self.sp_model.encode(snake_case__ , out_type=snake_case__ ) __lowerCamelCase = [] for piece in pieces: if len(snake_case__ ) > 1 and piece[-1] == str(''',''' ) and piece[-2].isdigit(): __lowerCamelCase = self.sp_model.EncodeAsPieces(piece[:-1].replace(snake_case__ , '''''' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __lowerCamelCase = cur_pieces[1:] else: __lowerCamelCase = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(snake_case__ ) else: new_pieces.append(snake_case__ ) return new_pieces def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.PieceToId(snake_case__ ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case__ ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = ''''''.join(snake_case__ ).replace(snake_case__ , ''' ''' ).strip() return out_string def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = False , __UpperCAmelCase = None , __UpperCAmelCase = True , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = kwargs.pop('''use_source_tokenizer''' , snake_case__ ) __lowerCamelCase = self.convert_ids_to_tokens(snake_case__ , skip_special_tokens=snake_case__ ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 __lowerCamelCase = [] __lowerCamelCase = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case__ ) ) __lowerCamelCase = [] sub_texts.append(snake_case__ ) else: current_sub_text.append(snake_case__ ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case__ ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens __lowerCamelCase = ''''''.join(snake_case__ ) __lowerCamelCase = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: __lowerCamelCase = self.clean_up_tokenization(snake_case__ ) return clean_text else: return text def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case__ , token_ids_a=snake_case__ , already_has_special_tokens=snake_case__ ) if token_ids_a is not None: return ([0] * len(snake_case__ )) + [1] + ([0] * len(snake_case__ )) + [1, 1] return ([0] * len(snake_case__ )) + [1, 1] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' if not os.path.isdir(snake_case__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCamelCase = os.path.join( snake_case__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case__ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case__ , '''wb''' ) as fi: __lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(snake_case__ ) return (out_vocab_file,)
706
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py a_ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. a_ = direct_transformers_import(PATH_TO_TRANSFORMERS) a_ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` a_ = re.compile(R"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") a_ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def a__ ( _UpperCamelCase : Union[str, Any] ): __lowerCamelCase = None # source code of `config_class` __lowerCamelCase = inspect.getsource(_UpperCamelCase ) __lowerCamelCase = _re_checkpoint.findall(_UpperCamelCase ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('''/''' ): __lowerCamelCase = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link __lowerCamelCase = F"""https://huggingface.co/{ckpt_name}""" if ckpt_link == ckpt_link_from_name: __lowerCamelCase = ckpt_name break return checkpoint def a__ ( ): __lowerCamelCase = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue __lowerCamelCase = get_checkpoint_from_config_class(_UpperCamelCase ) __lowerCamelCase = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(_UpperCamelCase ) if len(_UpperCamelCase ) > 0: __lowerCamelCase = '''\n'''.join(sorted(_UpperCamelCase ) ) raise ValueError(F"""The following configurations don't contain any valid checkpoint:\n{message}""" ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
622
0
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class __lowerCAmelCase ( __lowerCamelCase ): lowerCAmelCase__ = '''ctrl''' lowerCAmelCase__ = ['''past_key_values'''] lowerCAmelCase__ = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self , __UpperCAmelCase=246534 , __UpperCAmelCase=256 , __UpperCAmelCase=1280 , __UpperCAmelCase=8192 , __UpperCAmelCase=48 , __UpperCAmelCase=16 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=1E-6 , __UpperCAmelCase=0.02 , __UpperCAmelCase=True , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = vocab_size __lowerCamelCase = n_positions __lowerCamelCase = n_embd __lowerCamelCase = n_layer __lowerCamelCase = n_head __lowerCamelCase = dff __lowerCamelCase = resid_pdrop __lowerCamelCase = embd_pdrop __lowerCamelCase = layer_norm_epsilon __lowerCamelCase = initializer_range __lowerCamelCase = use_cache super().__init__(**a_ )
707
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a_ = { """configuration_clipseg""": [ """CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP""", """CLIPSegConfig""", """CLIPSegTextConfig""", """CLIPSegVisionConfig""", ], """processing_clipseg""": ["""CLIPSegProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ """CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST""", """CLIPSegModel""", """CLIPSegPreTrainedModel""", """CLIPSegTextModel""", """CLIPSegVisionModel""", """CLIPSegForImageSegmentation""", ] if TYPE_CHECKING: from .configuration_clipseg import ( CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPSegConfig, CLIPSegTextConfig, CLIPSegVisionConfig, ) from .processing_clipseg import CLIPSegProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clipseg import ( CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPSegForImageSegmentation, CLIPSegModel, CLIPSegPreTrainedModel, CLIPSegTextModel, CLIPSegVisionModel, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
622
0
import copy import fnmatch import json import os import pickle as pkl import shutil import sys import tarfile import tempfile from collections import OrderedDict from contextlib import contextmanager from functools import partial from hashlib import shaaaa from io import BytesIO from pathlib import Path from urllib.parse import urlparse from zipfile import ZipFile, is_zipfile import cva import numpy as np import requests import wget from filelock import FileLock from PIL import Image from tqdm.auto import tqdm from yaml import Loader, dump, load try: import torch a_ = True except ImportError: a_ = False try: from torch.hub import _get_torch_home a_ = _get_torch_home() except ImportError: a_ = os.path.expanduser( os.getenv("""TORCH_HOME""", os.path.join(os.getenv("""XDG_CACHE_HOME""", """~/.cache"""), """torch""")) ) a_ = os.path.join(torch_cache_home, """transformers""") a_ = """https://cdn.huggingface.co""" a_ = """https://s3.amazonaws.com/models.huggingface.co/bert""" a_ = """/""".join(str(Path(__file__).resolve()).split("""/""")[:-1]) a_ = os.path.join(PATH, """config.yaml""") a_ = os.path.join(PATH, """attributes.txt""") a_ = os.path.join(PATH, """objects.txt""") a_ = os.getenv("""PYTORCH_PRETRAINED_BERT_CACHE""", default_cache_path) a_ = os.getenv("""PYTORCH_TRANSFORMERS_CACHE""", PYTORCH_PRETRAINED_BERT_CACHE) a_ = os.getenv("""TRANSFORMERS_CACHE""", PYTORCH_TRANSFORMERS_CACHE) a_ = """pytorch_model.bin""" a_ = """config.yaml""" def a__ ( _UpperCamelCase : Dict=OBJECTS ,_UpperCamelCase : Optional[int]=ATTRIBUTES ): __lowerCamelCase = [] with open(_snake_case ) as f: for object in f.readlines(): vg_classes.append(object.split(''',''' )[0].lower().strip() ) __lowerCamelCase = [] with open(_snake_case ) as f: for object in f.readlines(): vg_attrs.append(object.split(''',''' )[0].lower().strip() ) return vg_classes, vg_attrs def a__ ( _UpperCamelCase : Optional[int] ): __lowerCamelCase = OrderedDict() with open(_snake_case ,'''rb''' ) as f: __lowerCamelCase = pkl.load(_snake_case )['''model'''] for k in copy.deepcopy(list(ckp.keys() ) ): __lowerCamelCase = ckp.pop(_snake_case ) if isinstance(_snake_case ,np.ndarray ): __lowerCamelCase = torch.tensor(_snake_case ) else: assert isinstance(_snake_case ,torch.tensor ), type(_snake_case ) __lowerCamelCase = v return r class __lowerCAmelCase : lowerCAmelCase__ = {} def __init__( self , __UpperCAmelCase , __UpperCAmelCase = "root" , __UpperCAmelCase=0 ): '''simple docstring''' __lowerCamelCase = name __lowerCamelCase = level __lowerCamelCase = {} for k, v in dictionary.items(): if v is None: raise ValueError() __lowerCamelCase = copy.deepcopy(__UpperCAmelCase ) __lowerCamelCase = copy.deepcopy(__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ): __lowerCamelCase = Config(__UpperCAmelCase , name=__UpperCAmelCase , level=level + 1 ) __lowerCamelCase = v setattr(self , __UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = d def __repr__( self ): '''simple docstring''' return str(list((self._pointer.keys()) ) ) def __setattr__( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = val __lowerCamelCase = val __lowerCamelCase = key.split('''.''' ) __lowerCamelCase = len(__UpperCAmelCase ) - 1 __lowerCamelCase = self._pointer if len(__UpperCAmelCase ) > 1: for i, l in enumerate(__UpperCAmelCase ): if hasattr(self , __UpperCAmelCase ) and isinstance(getattr(self , __UpperCAmelCase ) , __UpperCAmelCase ): setattr(getattr(self , __UpperCAmelCase ) , '''.'''.join(levels[i:] ) , __UpperCAmelCase ) if l == last_level: __lowerCamelCase = val else: __lowerCamelCase = pointer[l] def lowerCamelCase ( self ): '''simple docstring''' return self._pointer def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' with open(F"""{file_name}""" , '''w''' ) as stream: dump(__UpperCAmelCase , __UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' with open(F"""{file_name}""" , '''w''' ) as stream: json.dump(__UpperCAmelCase , __UpperCAmelCase ) @staticmethod def lowerCamelCase ( __UpperCAmelCase ): '''simple docstring''' with open(__UpperCAmelCase ) as stream: __lowerCamelCase = load(__UpperCAmelCase , Loader=__UpperCAmelCase ) return data def __str__( self ): '''simple docstring''' __lowerCamelCase = ''' ''' if self._name != "root": __lowerCamelCase = F"""{t * (self._level-1)}{self._name}:\n""" else: __lowerCamelCase = '''''' __lowerCamelCase = self._level for i, (k, v) in enumerate(self._pointer.items() ): if isinstance(__UpperCAmelCase , __UpperCAmelCase ): r += F"""{t * (self._level)}{v}\n""" self._level += 1 else: r += F"""{t * (self._level)}{k}: {v} ({type(__UpperCAmelCase ).__name__})\n""" __lowerCamelCase = level return r[:-1] @classmethod def lowerCamelCase ( cls , __UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' __lowerCamelCase ,__lowerCamelCase = cls.get_config_dict(__UpperCAmelCase , **__UpperCAmelCase ) return cls(__UpperCAmelCase ) @classmethod def lowerCamelCase ( cls , __UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = kwargs.pop('''cache_dir''' , __UpperCAmelCase ) __lowerCamelCase = kwargs.pop('''force_download''' , __UpperCAmelCase ) __lowerCamelCase = kwargs.pop('''resume_download''' , __UpperCAmelCase ) __lowerCamelCase = kwargs.pop('''proxies''' , __UpperCAmelCase ) __lowerCamelCase = kwargs.pop('''local_files_only''' , __UpperCAmelCase ) if os.path.isdir(__UpperCAmelCase ): __lowerCamelCase = os.path.join(__UpperCAmelCase , __UpperCAmelCase ) elif os.path.isfile(__UpperCAmelCase ) or is_remote_url(__UpperCAmelCase ): __lowerCamelCase = pretrained_model_name_or_path else: __lowerCamelCase = hf_bucket_url(__UpperCAmelCase , filename=__UpperCAmelCase , use_cdn=__UpperCAmelCase ) try: # Load from URL or cache if already cached __lowerCamelCase = cached_path( __UpperCAmelCase , cache_dir=__UpperCAmelCase , force_download=__UpperCAmelCase , proxies=__UpperCAmelCase , resume_download=__UpperCAmelCase , local_files_only=__UpperCAmelCase , ) # Load config dict if resolved_config_file is None: raise EnvironmentError __lowerCamelCase = Config.load_yaml(__UpperCAmelCase ) except EnvironmentError: __lowerCamelCase = '''Can\'t load config for''' raise EnvironmentError(__UpperCAmelCase ) if resolved_config_file == config_file: print('''loading configuration file from path''' ) else: print('''loading configuration file cache''' ) return Config.load_yaml(__UpperCAmelCase ), kwargs def a__ ( _UpperCamelCase : Dict ): __lowerCamelCase = torch.load('''dump.pt''' ,map_location=in_tensor.device ) __lowerCamelCase = in_tensor.numpy() __lowerCamelCase = out_tensor.numpy()[0] print(na.shape ,na[0, 0, :5] ) print(na.shape ,na[0, 0, :5] ) assert np.allclose(_snake_case ,_snake_case ,rtol=0.01 ,atol=0.1 ), ( F"""{sum([1 for x in np.isclose(_snake_case ,_snake_case ,rtol=0.01 ,atol=0.1 ).flatten() if x is False] )/len(na.flatten() )*1_00:.4f} %""" " element-wise mismatch" ) raise Exception('''tensors are all good''' ) # Hugging face functions below def a__ ( _UpperCamelCase : Optional[Any] ): __lowerCamelCase = urlparse(_snake_case ) return parsed.scheme in ("http", "https") def a__ ( _UpperCamelCase : int ,_UpperCamelCase : List[str] ,_UpperCamelCase : List[Any]=True ): __lowerCamelCase = CLOUDFRONT_DISTRIB_PREFIX if use_cdn else S3_BUCKET_PREFIX __lowerCamelCase = '''/''' not in model_id if legacy_format: return F"""{endpoint}/{model_id}-{filename}""" else: return F"""{endpoint}/{model_id}/{filename}""" def a__ ( _UpperCamelCase : str ,_UpperCamelCase : Dict ,_UpperCamelCase : Optional[Any]=None ,_UpperCamelCase : Optional[Any]=0 ,_UpperCamelCase : str=None ,): __lowerCamelCase = '''python/{}'''.format(sys.version.split()[0] ) if _torch_available: ua += "; torch/{}".format(torch.__version__ ) if isinstance(_snake_case ,_snake_case ): ua += "; " + "; ".join('''{}/{}'''.format(_snake_case ,_snake_case ) for k, v in user_agent.items() ) elif isinstance(_snake_case ,_snake_case ): ua += "; " + user_agent __lowerCamelCase = {'''user-agent''': ua} if resume_size > 0: __lowerCamelCase = '''bytes=%d-''' % (resume_size,) __lowerCamelCase = requests.get(_snake_case ,stream=_snake_case ,proxies=_snake_case ,headers=_snake_case ) if response.status_code == 4_16: # Range not satisfiable return __lowerCamelCase = response.headers.get('''Content-Length''' ) __lowerCamelCase = resume_size + int(_snake_case ) if content_length is not None else None __lowerCamelCase = tqdm( unit='''B''' ,unit_scale=_snake_case ,total=_snake_case ,initial=_snake_case ,desc='''Downloading''' ,) for chunk in response.iter_content(chunk_size=10_24 ): if chunk: # filter out keep-alive new chunks progress.update(len(_snake_case ) ) temp_file.write(_snake_case ) progress.close() def a__ ( _UpperCamelCase : int ,_UpperCamelCase : Tuple=None ,_UpperCamelCase : Union[str, Any]=False ,_UpperCamelCase : List[Any]=None ,_UpperCamelCase : str=10 ,_UpperCamelCase : List[Any]=False ,_UpperCamelCase : str=None ,_UpperCamelCase : int=False ,): if cache_dir is None: __lowerCamelCase = TRANSFORMERS_CACHE if isinstance(_snake_case ,_snake_case ): __lowerCamelCase = str(_snake_case ) os.makedirs(_snake_case ,exist_ok=_snake_case ) __lowerCamelCase = None if not local_files_only: try: __lowerCamelCase = requests.head(_snake_case ,allow_redirects=_snake_case ,proxies=_snake_case ,timeout=_snake_case ) if response.status_code == 2_00: __lowerCamelCase = response.headers.get('''ETag''' ) except (EnvironmentError, requests.exceptions.Timeout): # etag is already None pass __lowerCamelCase = url_to_filename(_snake_case ,_snake_case ) # get cache path to put the file __lowerCamelCase = os.path.join(_snake_case ,_snake_case ) # etag is None = we don't have a connection, or url doesn't exist, or is otherwise inaccessible. # try to get the last downloaded one if etag is None: if os.path.exists(_snake_case ): return cache_path else: __lowerCamelCase = [ file for file in fnmatch.filter(os.listdir(_snake_case ) ,filename + '''.*''' ) if not file.endswith('''.json''' ) and not file.endswith('''.lock''' ) ] if len(_snake_case ) > 0: return os.path.join(_snake_case ,matching_files[-1] ) else: # If files cannot be found and local_files_only=True, # the models might've been found if local_files_only=False # Notify the user about that if local_files_only: raise ValueError( '''Cannot find the requested files in the cached path and outgoing traffic has been''' ''' disabled. To enable model look-ups and downloads online, set \'local_files_only\'''' ''' to False.''' ) return None # From now on, etag is not None. if os.path.exists(_snake_case ) and not force_download: return cache_path # Prevent parallel downloads of the same file with a lock. __lowerCamelCase = cache_path + '''.lock''' with FileLock(_snake_case ): # If the download just completed while the lock was activated. if os.path.exists(_snake_case ) and not force_download: # Even if returning early like here, the lock will be released. return cache_path if resume_download: __lowerCamelCase = cache_path + '''.incomplete''' @contextmanager def _resumable_file_manager(): with open(_snake_case ,'''a+b''' ) as f: yield f __lowerCamelCase = _resumable_file_manager if os.path.exists(_snake_case ): __lowerCamelCase = os.stat(_snake_case ).st_size else: __lowerCamelCase = 0 else: __lowerCamelCase = partial(tempfile.NamedTemporaryFile ,dir=_snake_case ,delete=_snake_case ) __lowerCamelCase = 0 # Download to temporary file, then copy to cache dir once finished. # Otherwise you get corrupt cache entries if the download gets interrupted. with temp_file_manager() as temp_file: print( '''%s not found in cache or force_download set to True, downloading to %s''' ,_snake_case ,temp_file.name ,) http_get( _snake_case ,_snake_case ,proxies=_snake_case ,resume_size=_snake_case ,user_agent=_snake_case ,) os.replace(temp_file.name ,_snake_case ) __lowerCamelCase = {'''url''': url, '''etag''': etag} __lowerCamelCase = cache_path + '''.json''' with open(_snake_case ,'''w''' ) as meta_file: json.dump(_snake_case ,_snake_case ) return cache_path def a__ ( _UpperCamelCase : List[Any] ,_UpperCamelCase : List[str]=None ): __lowerCamelCase = url.encode('''utf-8''' ) __lowerCamelCase = shaaaa(_snake_case ) __lowerCamelCase = url_hash.hexdigest() if etag: __lowerCamelCase = etag.encode('''utf-8''' ) __lowerCamelCase = shaaaa(_snake_case ) filename += "." + etag_hash.hexdigest() if url.endswith('''.h5''' ): filename += ".h5" return filename def a__ ( _UpperCamelCase : Optional[Any] ,_UpperCamelCase : int=None ,_UpperCamelCase : Optional[Any]=False ,_UpperCamelCase : int=None ,_UpperCamelCase : Any=False ,_UpperCamelCase : Any=None ,_UpperCamelCase : List[str]=False ,_UpperCamelCase : List[Any]=False ,_UpperCamelCase : Any=False ,): if cache_dir is None: __lowerCamelCase = TRANSFORMERS_CACHE if isinstance(_snake_case ,_snake_case ): __lowerCamelCase = str(_snake_case ) if isinstance(_snake_case ,_snake_case ): __lowerCamelCase = str(_snake_case ) if is_remote_url(_snake_case ): # URL, so get it from the cache (downloading if necessary) __lowerCamelCase = get_from_cache( _snake_case ,cache_dir=_snake_case ,force_download=_snake_case ,proxies=_snake_case ,resume_download=_snake_case ,user_agent=_snake_case ,local_files_only=_snake_case ,) elif os.path.exists(_snake_case ): # File, and it exists. __lowerCamelCase = url_or_filename elif urlparse(_snake_case ).scheme == "": # File, but it doesn't exist. raise EnvironmentError('''file {} not found'''.format(_snake_case ) ) else: # Something unknown raise ValueError('''unable to parse {} as a URL or as a local path'''.format(_snake_case ) ) if extract_compressed_file: if not is_zipfile(_snake_case ) and not tarfile.is_tarfile(_snake_case ): return output_path # Path where we extract compressed archives # We avoid '.' in dir name and add "-extracted" at the end: "./model.zip" => "./model-zip-extracted/" __lowerCamelCase ,__lowerCamelCase = os.path.split(_snake_case ) __lowerCamelCase = output_file.replace('''.''' ,'''-''' ) + '''-extracted''' __lowerCamelCase = os.path.join(_snake_case ,_snake_case ) if os.path.isdir(_snake_case ) and os.listdir(_snake_case ) and not force_extract: return output_path_extracted # Prevent parallel extractions __lowerCamelCase = output_path + '''.lock''' with FileLock(_snake_case ): shutil.rmtree(_snake_case ,ignore_errors=_snake_case ) os.makedirs(_snake_case ) if is_zipfile(_snake_case ): with ZipFile(_snake_case ,'''r''' ) as zip_file: zip_file.extractall(_snake_case ) zip_file.close() elif tarfile.is_tarfile(_snake_case ): __lowerCamelCase = tarfile.open(_snake_case ) tar_file.extractall(_snake_case ) tar_file.close() else: raise EnvironmentError('''Archive format of {} could not be identified'''.format(_snake_case ) ) return output_path_extracted return output_path def a__ ( _UpperCamelCase : Any ,_UpperCamelCase : int="," ): assert isinstance(_snake_case ,_snake_case ) if os.path.isfile(_snake_case ): with open(_snake_case ) as f: __lowerCamelCase = eval(f.read() ) else: __lowerCamelCase = requests.get(_snake_case ) try: __lowerCamelCase = requests.json() except Exception: __lowerCamelCase = req.content.decode() assert data is not None, "could not connect" try: __lowerCamelCase = eval(_snake_case ) except Exception: __lowerCamelCase = data.split('''\n''' ) req.close() return data def a__ ( _UpperCamelCase : List[Any] ): __lowerCamelCase = requests.get(_snake_case ) __lowerCamelCase = np.array(Image.open(BytesIO(response.content ) ) ) return img def a__ ( _UpperCamelCase : Any ): __lowerCamelCase = url.split('''/''' )[-1] if fn not in os.listdir(os.getcwd() ): wget.download(_snake_case ) with open(_snake_case ,'''rb''' ) as stream: __lowerCamelCase = pkl.load(_snake_case ) __lowerCamelCase = weights.pop('''model''' ) __lowerCamelCase = {} for k, v in model.items(): __lowerCamelCase = torch.from_numpy(_snake_case ) if "running_var" in k: __lowerCamelCase = torch.tensor([0] ) __lowerCamelCase = k.replace('''running_var''' ,'''num_batches_tracked''' ) __lowerCamelCase = zero return new def a__ ( ): print(F"""{os.path.abspath(os.path.join(_snake_case ,os.pardir ) )}/demo.ipynb""" ) def a__ ( _UpperCamelCase : List[Any] ,_UpperCamelCase : str="RGB" ): assert isinstance(_snake_case ,_snake_case ) if os.path.isfile(_snake_case ): __lowerCamelCase = cva.imread(_snake_case ) else: __lowerCamelCase = get_image_from_url(_snake_case ) assert img is not None, F"""could not connect to: {im}""" __lowerCamelCase = cva.cvtColor(_snake_case ,cva.COLOR_BGR2RGB ) if input_format == "RGB": __lowerCamelCase = img[:, :, ::-1] return img def a__ ( _UpperCamelCase : List[Any] ,_UpperCamelCase : Tuple=1 ): return (images[i : i + batch] for i in range(0 ,len(_snake_case ) ,_snake_case ))
708
import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = RoFormerTokenizer lowerCAmelCase__ = RoFormerTokenizerFast lowerCAmelCase__ = True lowerCAmelCase__ = True def lowerCamelCase ( self ): '''simple docstring''' super().setUp() def lowerCamelCase ( self , **__UpperCAmelCase ): '''simple docstring''' return self.tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **__UpperCAmelCase ) def lowerCamelCase ( self , **__UpperCAmelCase ): '''simple docstring''' return self.rust_tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''永和服装饰品有限公司,今天天气非常好''' __lowerCamelCase = '''永和 服装 饰品 有限公司 , 今 天 天 气 非常 好''' return input_text, output_text def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_tokenizer() __lowerCamelCase ,__lowerCamelCase = self.get_chinese_input_output_texts() __lowerCamelCase = tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , output_text.split() ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [22943, 21332, 34431, 45904, 117, 306, 1231, 1231, 2653, 33994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_rust_tokenizer() __lowerCamelCase ,__lowerCamelCase = self.get_chinese_input_output_texts() __lowerCamelCase = tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , output_text.split() ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [22943, 21332, 34431, 45904, 117, 306, 1231, 1231, 2653, 33994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' pass
622
0
from __future__ import annotations import numpy as np def a__ ( _UpperCamelCase : list[float] ): return np.maximum(0 ,_A ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
709
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device a_ = False class __lowerCAmelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained('''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt='''first prompt''' , image=__UpperCAmelCase , text_to_image_strength=0.75 , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained(__UpperCAmelCase , torch_dtype=torch.floataa ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = generator.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt='''first prompt''' , image=__UpperCAmelCase , text_to_image_strength=0.75 , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained('''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = '''cyberpunk 2077''' __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt=__UpperCAmelCase , image=__UpperCAmelCase , text_to_image_strength=0.75 , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' , ).images __lowerCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __lowerCamelCase = '''A painting of a squirrel eating a burger ''' __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.text_to_image( prompt=__UpperCAmelCase , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' ).images __lowerCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __lowerCamelCase = pipe.image_variation(__UpperCAmelCase , generator=__UpperCAmelCase , output_type='''numpy''' ).images __lowerCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
622
0
import random from .binary_exp_mod import bin_exp_mod def a__ ( _UpperCamelCase : Union[str, Any] ,_UpperCamelCase : List[str]=10_00 ): if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd __lowerCamelCase = n - 1 __lowerCamelCase = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) __lowerCamelCase = 0 while count < prec: __lowerCamelCase = random.randint(2 ,n - 1 ) __lowerCamelCase = bin_exp_mod(A_ ,A_ ,A_ ) if b != 1: __lowerCamelCase = True for _ in range(A_ ): if b == n - 1: __lowerCamelCase = False break __lowerCamelCase = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": a_ = abs(int(input("""Enter bound : """).strip())) print("""Here\'s the list of primes:""") print(""", """.join(str(i) for i in range(n + 1) if is_prime_big(i)))
710
import argparse import datetime import json import time import warnings from logging import getLogger from pathlib import Path from typing import Dict, List import torch from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import calculate_bleu, calculate_rouge, chunks, parse_numeric_n_bool_cl_kwargs, use_task_specific_params a_ = getLogger(__name__) a_ = """cuda""" if torch.cuda.is_available() else """cpu""" def a__ ( _UpperCamelCase : List[str] ,_UpperCamelCase : str ,_UpperCamelCase : str ,_UpperCamelCase : int = 8 ,_UpperCamelCase : str = DEFAULT_DEVICE ,_UpperCamelCase : Dict=False ,_UpperCamelCase : Dict="summarization" ,_UpperCamelCase : Optional[int]=None ,**_UpperCamelCase : Dict ,): __lowerCamelCase = Path(_UpperCamelCase ).open('''w''' ,encoding='''utf-8''' ) __lowerCamelCase = str(_UpperCamelCase ) __lowerCamelCase = AutoModelForSeqaSeqLM.from_pretrained(_UpperCamelCase ).to(_UpperCamelCase ) if fpaa: __lowerCamelCase = model.half() __lowerCamelCase = AutoTokenizer.from_pretrained(_UpperCamelCase ) logger.info(F"""Inferred tokenizer type: {tokenizer.__class__}""" ) # if this is wrong, check config.model_type. __lowerCamelCase = time.time() # update config with task specific params use_task_specific_params(_UpperCamelCase ,_UpperCamelCase ) if prefix is None: __lowerCamelCase = prefix or getattr(model.config ,'''prefix''' ,'''''' ) or '''''' for examples_chunk in tqdm(list(chunks(_UpperCamelCase ,_UpperCamelCase ) ) ): __lowerCamelCase = [prefix + text for text in examples_chunk] __lowerCamelCase = tokenizer(_UpperCamelCase ,return_tensors='''pt''' ,truncation=_UpperCamelCase ,padding='''longest''' ).to(_UpperCamelCase ) __lowerCamelCase = model.generate( input_ids=batch.input_ids ,attention_mask=batch.attention_mask ,**_UpperCamelCase ,) __lowerCamelCase = tokenizer.batch_decode(_UpperCamelCase ,skip_special_tokens=_UpperCamelCase ,clean_up_tokenization_spaces=_UpperCamelCase ) for hypothesis in dec: fout.write(hypothesis + '''\n''' ) fout.flush() fout.close() __lowerCamelCase = int(time.time() - start_time ) # seconds __lowerCamelCase = len(_UpperCamelCase ) return {"n_obs": n_obs, "runtime": runtime, "seconds_per_sample": round(runtime / n_obs ,4 )} def a__ ( ): return datetime.datetime.now().strftime('''%Y-%m-%d %H:%M:%S''' ) def a__ ( _UpperCamelCase : Union[str, Any]=True ): __lowerCamelCase = argparse.ArgumentParser() parser.add_argument('''model_name''' ,type=_UpperCamelCase ,help='''like facebook/bart-large-cnn,t5-base, etc.''' ) parser.add_argument('''input_path''' ,type=_UpperCamelCase ,help='''like cnn_dm/test.source''' ) parser.add_argument('''save_path''' ,type=_UpperCamelCase ,help='''where to save summaries''' ) parser.add_argument('''--reference_path''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,help='''like cnn_dm/test.target''' ) parser.add_argument('''--score_path''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,default='''metrics.json''' ,help='''where to save metrics''' ) parser.add_argument('''--device''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,default=_UpperCamelCase ,help='''cuda, cuda:1, cpu etc.''' ) parser.add_argument( '''--prefix''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,default=_UpperCamelCase ,help='''will be added to the begininng of src examples''' ) parser.add_argument('''--task''' ,type=_UpperCamelCase ,default='''summarization''' ,help='''used for task_specific_params + metrics''' ) parser.add_argument('''--bs''' ,type=_UpperCamelCase ,default=8 ,required=_UpperCamelCase ,help='''batch size''' ) parser.add_argument( '''--n_obs''' ,type=_UpperCamelCase ,default=-1 ,required=_UpperCamelCase ,help='''How many observations. Defaults to all.''' ) parser.add_argument('''--fp16''' ,action='''store_true''' ) parser.add_argument('''--dump-args''' ,action='''store_true''' ,help='''print the custom hparams with the results''' ) parser.add_argument( '''--info''' ,nargs='''?''' ,type=_UpperCamelCase ,const=datetime_now() ,help=( '''use in conjunction w/ --dump-args to print with the results whatever other info you\'d like, e.g.''' ''' lang=en-ru. If no value is passed, the current datetime string will be used.''' ) ,) # Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate __lowerCamelCase ,__lowerCamelCase = parser.parse_known_args() __lowerCamelCase = parse_numeric_n_bool_cl_kwargs(_UpperCamelCase ) if parsed_args and verbose: print(F"""parsed the following generate kwargs: {parsed_args}""" ) __lowerCamelCase = [''' ''' + x.rstrip() if '''t5''' in args.model_name else x.rstrip() for x in open(args.input_path ).readlines()] if args.n_obs > 0: __lowerCamelCase = examples[: args.n_obs] Path(args.save_path ).parent.mkdir(exist_ok=_UpperCamelCase ) if args.reference_path is None and Path(args.score_path ).exists(): warnings.warn(F"""score_path {args.score_path} will be overwritten unless you type ctrl-c.""" ) if args.device == "cpu" and args.fpaa: # this mix leads to RuntimeError: "threshold_cpu" not implemented for 'Half' raise ValueError('''Can\'t mix --fp16 and --device cpu''' ) __lowerCamelCase = generate_summaries_or_translations( _UpperCamelCase ,args.save_path ,args.model_name ,batch_size=args.bs ,device=args.device ,fpaa=args.fpaa ,task=args.task ,prefix=args.prefix ,**_UpperCamelCase ,) if args.reference_path is None: return {} # Compute scores __lowerCamelCase = calculate_bleu if '''translation''' in args.task else calculate_rouge __lowerCamelCase = [x.rstrip() for x in open(args.save_path ).readlines()] __lowerCamelCase = [x.rstrip() for x in open(args.reference_path ).readlines()][: len(_UpperCamelCase )] __lowerCamelCase = score_fn(_UpperCamelCase ,_UpperCamelCase ) scores.update(_UpperCamelCase ) if args.dump_args: scores.update(_UpperCamelCase ) if args.info: __lowerCamelCase = args.info if verbose: print(_UpperCamelCase ) if args.score_path is not None: json.dump(_UpperCamelCase ,open(args.score_path ,'''w''' ) ) return scores if __name__ == "__main__": # Usage for MT: # python run_eval.py MODEL_NAME $DATA_DIR/test.source $save_dir/test_translations.txt --reference_path $DATA_DIR/test.target --score_path $save_dir/test_bleu.json --task translation $@ run_generate(verbose=True)
622
0
'''simple docstring''' from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a_ = {'''configuration_mmbt''': ['''MMBTConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ['''MMBTForClassification''', '''MMBTModel''', '''ModalEmbeddings'''] if TYPE_CHECKING: from .configuration_mmbt import MMBTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mmbt import MMBTForClassification, MMBTModel, ModalEmbeddings else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
711
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def a__ ( _UpperCamelCase : Optional[int] ,_UpperCamelCase : List[str] ,_UpperCamelCase : List[Any]=None ,_UpperCamelCase : Any=None ): if attention_mask is None: __lowerCamelCase = tf.cast(tf.math.not_equal(_UpperCamelCase ,config.pad_token_id ) ,tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class __lowerCAmelCase : lowerCAmelCase__ = OPTConfig lowerCAmelCase__ = {} lowerCAmelCase__ = """gelu""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=99 , __UpperCAmelCase=16 , __UpperCAmelCase=2 , __UpperCAmelCase=4 , __UpperCAmelCase=4 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=20 , __UpperCAmelCase=2 , __UpperCAmelCase=1 , __UpperCAmelCase=0 , __UpperCAmelCase=16 , __UpperCAmelCase=16 , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = eos_token_id __lowerCamelCase = pad_token_id __lowerCamelCase = bos_token_id __lowerCamelCase = embed_dim __lowerCamelCase = word_embed_proj_dim __lowerCamelCase = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowerCamelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowerCamelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowerCamelCase = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=__UpperCAmelCase , **self.config_updates , ) __lowerCamelCase = prepare_opt_inputs_dict(__UpperCAmelCase , __UpperCAmelCase ) return config, inputs_dict def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = TFOPTModel(config=__UpperCAmelCase ) __lowerCamelCase = inputs_dict['''input_ids'''] __lowerCamelCase = input_ids[:1, :] __lowerCamelCase = inputs_dict['''attention_mask'''][:1, :] __lowerCamelCase = 1 # first forward pass __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , use_cache=__UpperCAmelCase ) __lowerCamelCase ,__lowerCamelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowerCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowerCamelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowerCamelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowerCamelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase )[0] __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , past_key_values=__UpperCAmelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowerCamelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowerCamelCase = output_from_no_past[:, -3:, random_slice_idx] __lowerCamelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__UpperCAmelCase , __UpperCAmelCase , rtol=1E-3 ) @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () lowerCAmelCase__ = (TFOPTForCausalLM,) if is_tf_available() else () lowerCAmelCase__ = ( {"""feature-extraction""": TFOPTModel, """text-generation""": TFOPTForCausalLM} if is_tf_available() else {} ) lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = 1_0 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(__UpperCAmelCase , __UpperCAmelCase ): if hasattr(__UpperCAmelCase , '''weight''' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(__UpperCAmelCase , '''weight''' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 10, config.vocab_size + 10]: # build the embeddings __lowerCamelCase = model_class(config=__UpperCAmelCase ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_input_embeddings() ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(__UpperCAmelCase ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_input_embeddings() ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. __lowerCamelCase = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , __UpperCAmelCase ) # check that weights remain the same after resizing __lowerCamelCase = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCamelCase = False self.assertTrue(__UpperCAmelCase ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , __UpperCAmelCase ) __lowerCamelCase = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCamelCase = False self.assertTrue(__UpperCAmelCase ) def a__ ( _UpperCamelCase : Optional[Any] ): return tf.constant(_UpperCamelCase ,dtype=tf.intaa ) @require_tf class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = 9_9 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = tf.ones((4, 1) , dtype=tf.intaa ) * 2 __lowerCamelCase = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) __lowerCamelCase = input_ids.shape[0] __lowerCamelCase = OPTConfig( vocab_size=self.vocab_size , hidden_size=24 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class __lowerCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTModel.from_pretrained('''facebook/opt-350m''' ) __lowerCamelCase = _long_tensor([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] ) __lowerCamelCase = tf.not_equal(__UpperCAmelCase , model.config.pad_token_id ) with tf.GradientTape(): __lowerCamelCase = model(input_ids=__UpperCAmelCase , attention_mask=__UpperCAmelCase ).last_hidden_state __lowerCamelCase = (1, 11, 512) self.assertEqual(output.shape , __UpperCAmelCase ) __lowerCamelCase = tf.constant( [[-0.2_873, -1.9_218, -0.3_033], [-1.2_710, -0.1_338, -0.1_902], [0.4_095, 0.1_214, -1.3_121]] ) self.assertTrue(np.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=4E-3 ) ) __lowerCamelCase = tf.function(__UpperCAmelCase , jit_compile=__UpperCAmelCase ) __lowerCamelCase = xla_generate(__UpperCAmelCase , __UpperCAmelCase )[0] self.assertTrue(np.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=4E-2 ) ) @require_tf @slow class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' super().setUp() __lowerCamelCase = '''facebook/opt-350m''' def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTForCausalLM.from_pretrained(self.path_model ) __lowerCamelCase = GPTaTokenizer.from_pretrained(self.path_model ) __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of''', '''Paris is the capital of France and''', '''Computers and mobile phones have taken''', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' , padding=__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) __lowerCamelCase = tf.constant( [ [1.3_851, -13.8_923, -10.5_229, -10.7_533, -0.2_309, -10.2_384, -0.5_365, -9.0_947, -5.1_670], [-4.7_073, -10.6_276, -3.9_415, -21.5_242, -0.2_822, -0.2_822, -0.2_822, -0.2_822, -0.2_822], [0.6_247, -3.4_229, -8.9_179, -1.4_297, -14.1_650, 1.4_146, -9.0_218, -0.2_703, -0.2_703], [6.4_783, -1.9_913, -10.7_926, -2.3_336, 1.5_092, -0.9_974, -6.8_213, 1.3_477, 1.3_477], ] ) self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-4 ) ) __lowerCamelCase = tf.function(__UpperCAmelCase , jit_compile=__UpperCAmelCase ) __lowerCamelCase = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-4 ) ) @require_tf @slow class __lowerCAmelCase ( unittest.TestCase ): @property def lowerCamelCase ( self ): '''simple docstring''' return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-125m''' __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of New York, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] __lowerCamelCase = [] __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) for prompt in self.prompts: __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(__UpperCAmelCase , max_length=10 ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) predicted_outputs += generated_string self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-350m''' __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = '''left''' # use different length sentences to test batching __lowerCamelCase = [ '''Hello, my dog is a little''', '''Today, I''', ] __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' , padding=__UpperCAmelCase ) __lowerCamelCase = inputs['''input_ids'''] __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase , attention_mask=inputs['''attention_mask'''] ) __lowerCamelCase = tokenizer(sentences[0] , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase ) __lowerCamelCase = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['''attention_mask'''][-1] , tf.intaa ) ) __lowerCamelCase = tokenizer(sentences[1] , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase , max_length=model.config.max_length - num_paddings ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = [ '''Hello, my dog is a little bit of a dork.\nI\'m a little bit''', '''Today, I was in the middle of a conversation with a friend about the''', ] self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , [non_padded_sentence, padded_sentence] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-350m''' __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of San Francisco, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] __lowerCamelCase = [] __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) for prompt in self.prompts: __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(__UpperCAmelCase , max_length=10 ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) predicted_outputs += generated_string self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase )
622
0
from collections import OrderedDict from typing import Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...feature_extraction_utils import FeatureExtractionMixin from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType, logging a_ = logging.get_logger(__name__) a_ = { """deepmind/language-perceiver""": """https://huggingface.co/deepmind/language-perceiver/resolve/main/config.json""", # See all Perceiver models at https://huggingface.co/models?filter=perceiver } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """perceiver""" def __init__( self , __UpperCAmelCase=256 , __UpperCAmelCase=1280 , __UpperCAmelCase=768 , __UpperCAmelCase=1 , __UpperCAmelCase=26 , __UpperCAmelCase=8 , __UpperCAmelCase=8 , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase="kv" , __UpperCAmelCase=1 , __UpperCAmelCase=1 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=True , __UpperCAmelCase=262 , __UpperCAmelCase=2048 , __UpperCAmelCase=56 , __UpperCAmelCase=[368, 496] , __UpperCAmelCase=16 , __UpperCAmelCase=1920 , __UpperCAmelCase=16 , __UpperCAmelCase=[1, 16, 224, 224] , **__UpperCAmelCase , ): '''simple docstring''' super().__init__(**__a ) __lowerCamelCase = num_latents __lowerCamelCase = d_latents __lowerCamelCase = d_model __lowerCamelCase = num_blocks __lowerCamelCase = num_self_attends_per_block __lowerCamelCase = num_self_attention_heads __lowerCamelCase = num_cross_attention_heads __lowerCamelCase = qk_channels __lowerCamelCase = v_channels __lowerCamelCase = cross_attention_shape_for_attention __lowerCamelCase = self_attention_widening_factor __lowerCamelCase = cross_attention_widening_factor __lowerCamelCase = hidden_act __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = use_query_residual # masked language modeling attributes __lowerCamelCase = vocab_size __lowerCamelCase = max_position_embeddings # image classification attributes __lowerCamelCase = image_size # flow attributes __lowerCamelCase = train_size # multimodal autoencoding attributes __lowerCamelCase = num_frames __lowerCamelCase = audio_samples_per_frame __lowerCamelCase = samples_per_patch __lowerCamelCase = output_shape class __lowerCAmelCase ( lowerCAmelCase__ ): @property def lowerCamelCase ( self ): '''simple docstring''' if self.task == "multiple-choice": __lowerCamelCase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: __lowerCamelCase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''inputs''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] ) @property def lowerCamelCase ( self ): '''simple docstring''' return 1E-4 def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = -1 , __UpperCAmelCase = -1 , __UpperCAmelCase = -1 , __UpperCAmelCase = False , __UpperCAmelCase = None , __UpperCAmelCase = 3 , __UpperCAmelCase = 40 , __UpperCAmelCase = 40 , ): '''simple docstring''' # copied from `transformers.onnx.config.OnnxConfig` and slightly altered/simplified if isinstance(__a , __a ): # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX __lowerCamelCase = compute_effective_axis_dimension( __a , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __lowerCamelCase = preprocessor.num_special_tokens_to_add(__a ) __lowerCamelCase = compute_effective_axis_dimension( __a , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__a ) # Generate dummy inputs according to compute batch and sequence __lowerCamelCase = [''' '''.join(['''a'''] ) * seq_length] * batch_size __lowerCamelCase = dict(preprocessor(__a , return_tensors=__a ) ) __lowerCamelCase = inputs.pop('''input_ids''' ) return inputs elif isinstance(__a , __a ) and preprocessor.model_input_names[0] == "pixel_values": # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX __lowerCamelCase = compute_effective_axis_dimension(__a , fixed_dimension=OnnxConfig.default_fixed_batch ) __lowerCamelCase = self._generate_dummy_images(__a , __a , __a , __a ) __lowerCamelCase = dict(preprocessor(images=__a , return_tensors=__a ) ) __lowerCamelCase = inputs.pop('''pixel_values''' ) return inputs else: raise ValueError( '''Unable to generate dummy inputs for the model. Please provide a tokenizer or a preprocessor.''' )
712
import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) a_ = logging.getLogger(__name__) def a__ ( _UpperCamelCase : str ,_UpperCamelCase : List[Any] ): __lowerCamelCase = np.argmax(_UpperCamelCase ,axis=1 ) return np.sum(outputs == labels ) def a__ ( _UpperCamelCase : Optional[int] ): with open(_UpperCamelCase ,encoding='''utf_8''' ) as f: __lowerCamelCase = csv.reader(_UpperCamelCase ) __lowerCamelCase = [] next(_UpperCamelCase ) # skip the first line for line in tqdm(_UpperCamelCase ): output.append((''' '''.join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def a__ ( _UpperCamelCase : Any ,_UpperCamelCase : Dict ,_UpperCamelCase : str ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Any ,_UpperCamelCase : Dict ): __lowerCamelCase = [] for dataset in encoded_datasets: __lowerCamelCase = len(_UpperCamelCase ) __lowerCamelCase = np.zeros((n_batch, 2, input_len) ,dtype=np.intaa ) __lowerCamelCase = np.zeros((n_batch, 2) ,dtype=np.intaa ) __lowerCamelCase = np.full((n_batch, 2, input_len) ,fill_value=-1_00 ,dtype=np.intaa ) __lowerCamelCase = np.zeros((n_batch,) ,dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(_UpperCamelCase ): __lowerCamelCase = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __lowerCamelCase = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __lowerCamelCase = with_conta __lowerCamelCase = with_conta __lowerCamelCase = len(_UpperCamelCase ) - 1 __lowerCamelCase = len(_UpperCamelCase ) - 1 __lowerCamelCase = with_conta __lowerCamelCase = with_conta __lowerCamelCase = mc_label __lowerCamelCase = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(_UpperCamelCase ) for t in all_inputs ) ) return tensor_datasets def a__ ( ): __lowerCamelCase = argparse.ArgumentParser() parser.add_argument('''--model_name''' ,type=_UpperCamelCase ,default='''openai-gpt''' ,help='''pretrained model name''' ) parser.add_argument('''--do_train''' ,action='''store_true''' ,help='''Whether to run training.''' ) parser.add_argument('''--do_eval''' ,action='''store_true''' ,help='''Whether to run eval on the dev set.''' ) parser.add_argument( '''--output_dir''' ,default=_UpperCamelCase ,type=_UpperCamelCase ,required=_UpperCamelCase ,help='''The output directory where the model predictions and checkpoints will be written.''' ,) parser.add_argument('''--train_dataset''' ,type=_UpperCamelCase ,default='''''' ) parser.add_argument('''--eval_dataset''' ,type=_UpperCamelCase ,default='''''' ) parser.add_argument('''--seed''' ,type=_UpperCamelCase ,default=42 ) parser.add_argument('''--num_train_epochs''' ,type=_UpperCamelCase ,default=3 ) parser.add_argument('''--train_batch_size''' ,type=_UpperCamelCase ,default=8 ) parser.add_argument('''--eval_batch_size''' ,type=_UpperCamelCase ,default=16 ) parser.add_argument('''--adam_epsilon''' ,default=1e-8 ,type=_UpperCamelCase ,help='''Epsilon for Adam optimizer.''' ) parser.add_argument('''--max_grad_norm''' ,type=_UpperCamelCase ,default=1 ) parser.add_argument( '''--max_steps''' ,default=-1 ,type=_UpperCamelCase ,help=( '''If > 0: set total number of training steps to perform. Override num_train_epochs.''' ) ,) parser.add_argument( '''--gradient_accumulation_steps''' ,type=_UpperCamelCase ,default=1 ,help='''Number of updates steps to accumulate before performing a backward/update pass.''' ,) parser.add_argument('''--learning_rate''' ,type=_UpperCamelCase ,default=6.25e-5 ) parser.add_argument('''--warmup_steps''' ,default=0 ,type=_UpperCamelCase ,help='''Linear warmup over warmup_steps.''' ) parser.add_argument('''--lr_schedule''' ,type=_UpperCamelCase ,default='''warmup_linear''' ) parser.add_argument('''--weight_decay''' ,type=_UpperCamelCase ,default=0.01 ) parser.add_argument('''--lm_coef''' ,type=_UpperCamelCase ,default=0.9 ) parser.add_argument('''--n_valid''' ,type=_UpperCamelCase ,default=3_74 ) parser.add_argument('''--server_ip''' ,type=_UpperCamelCase ,default='''''' ,help='''Can be used for distant debugging.''' ) parser.add_argument('''--server_port''' ,type=_UpperCamelCase ,default='''''' ,help='''Can be used for distant debugging.''' ) __lowerCamelCase = parser.parse_args() print(_UpperCamelCase ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('''Waiting for debugger attach''' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) ,redirect_output=_UpperCamelCase ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) __lowerCamelCase = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) __lowerCamelCase = torch.cuda.device_count() logger.info('''device: {}, n_gpu {}'''.format(_UpperCamelCase ,_UpperCamelCase ) ) if not args.do_train and not args.do_eval: raise ValueError('''At least one of `do_train` or `do_eval` must be True.''' ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset __lowerCamelCase = ['''_start_''', '''_delimiter_''', '''_classify_'''] __lowerCamelCase = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(_UpperCamelCase ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(_UpperCamelCase ) __lowerCamelCase = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(_UpperCamelCase ) ) model.to(_UpperCamelCase ) # Load and encode the datasets def tokenize_and_encode(_UpperCamelCase : Dict ): if isinstance(_UpperCamelCase ,_UpperCamelCase ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(_UpperCamelCase ) ) elif isinstance(_UpperCamelCase ,_UpperCamelCase ): return obj return [tokenize_and_encode(_UpperCamelCase ) for o in obj] logger.info('''Encoding dataset...''' ) __lowerCamelCase = load_rocstories_dataset(args.train_dataset ) __lowerCamelCase = load_rocstories_dataset(args.eval_dataset ) __lowerCamelCase = (train_dataset, eval_dataset) __lowerCamelCase = tokenize_and_encode(_UpperCamelCase ) # Compute the max input length for the Transformer __lowerCamelCase = model.config.n_positions // 2 - 2 __lowerCamelCase = max( len(story[:max_length] ) + max(len(conta[:max_length] ) ,len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) __lowerCamelCase = min(_UpperCamelCase ,model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders __lowerCamelCase = pre_process_datasets(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ,*_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = tensor_datasets[0], tensor_datasets[1] __lowerCamelCase = TensorDataset(*_UpperCamelCase ) __lowerCamelCase = RandomSampler(_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,sampler=_UpperCamelCase ,batch_size=args.train_batch_size ) __lowerCamelCase = TensorDataset(*_UpperCamelCase ) __lowerCamelCase = SequentialSampler(_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,sampler=_UpperCamelCase ,batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: __lowerCamelCase = args.max_steps __lowerCamelCase = args.max_steps // (len(_UpperCamelCase ) // args.gradient_accumulation_steps) + 1 else: __lowerCamelCase = len(_UpperCamelCase ) // args.gradient_accumulation_steps * args.num_train_epochs __lowerCamelCase = list(model.named_parameters() ) __lowerCamelCase = ['''bias''', '''LayerNorm.bias''', '''LayerNorm.weight'''] __lowerCamelCase = [ { '''params''': [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], '''weight_decay''': args.weight_decay, }, {'''params''': [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0}, ] __lowerCamelCase = AdamW(_UpperCamelCase ,lr=args.learning_rate ,eps=args.adam_epsilon ) __lowerCamelCase = get_linear_schedule_with_warmup( _UpperCamelCase ,num_warmup_steps=args.warmup_steps ,num_training_steps=_UpperCamelCase ) if args.do_train: __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) ,desc='''Epoch''' ): __lowerCamelCase = 0 __lowerCamelCase = 0 __lowerCamelCase = tqdm(_UpperCamelCase ,desc='''Training''' ) for step, batch in enumerate(_UpperCamelCase ): __lowerCamelCase = tuple(t.to(_UpperCamelCase ) for t in batch ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = batch __lowerCamelCase = model(_UpperCamelCase ,mc_token_ids=_UpperCamelCase ,lm_labels=_UpperCamelCase ,mc_labels=_UpperCamelCase ) __lowerCamelCase = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() __lowerCamelCase = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 __lowerCamelCase = '''Training loss: {:.2e} lr: {:.2e}'''.format(_UpperCamelCase ,scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer __lowerCamelCase = model.module if hasattr(_UpperCamelCase ,'''module''' ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` __lowerCamelCase = os.path.join(args.output_dir ,_UpperCamelCase ) __lowerCamelCase = os.path.join(args.output_dir ,_UpperCamelCase ) torch.save(model_to_save.state_dict() ,_UpperCamelCase ) model_to_save.config.to_json_file(_UpperCamelCase ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned __lowerCamelCase = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) __lowerCamelCase = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(_UpperCamelCase ) if args.do_eval: model.eval() __lowerCamelCase ,__lowerCamelCase = 0, 0 __lowerCamelCase ,__lowerCamelCase = 0, 0 for batch in tqdm(_UpperCamelCase ,desc='''Evaluating''' ): __lowerCamelCase = tuple(t.to(_UpperCamelCase ) for t in batch ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = batch with torch.no_grad(): __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = model( _UpperCamelCase ,mc_token_ids=_UpperCamelCase ,lm_labels=_UpperCamelCase ,mc_labels=_UpperCamelCase ) __lowerCamelCase = mc_logits.detach().cpu().numpy() __lowerCamelCase = mc_labels.to('''cpu''' ).numpy() __lowerCamelCase = accuracy(_UpperCamelCase ,_UpperCamelCase ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 __lowerCamelCase = eval_loss / nb_eval_steps __lowerCamelCase = eval_accuracy / nb_eval_examples __lowerCamelCase = tr_loss / nb_tr_steps if args.do_train else None __lowerCamelCase = {'''eval_loss''': eval_loss, '''eval_accuracy''': eval_accuracy, '''train_loss''': train_loss} __lowerCamelCase = os.path.join(args.output_dir ,'''eval_results.txt''' ) with open(_UpperCamelCase ,'''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key in sorted(result.keys() ): logger.info(''' %s = %s''' ,_UpperCamelCase ,str(result[key] ) ) writer.write('''%s = %s\n''' % (key, str(result[key] )) ) if __name__ == "__main__": main()
622
0
from __future__ import annotations a_ = 10 def a__ ( _UpperCamelCase : str ): __lowerCamelCase = 1 __lowerCamelCase = max(lowerCamelCase_ ) while placement <= max_digit: # declare and initialize empty buckets __lowerCamelCase = [[] for _ in range(lowerCamelCase_ )] # split list_of_ints between the buckets for i in list_of_ints: __lowerCamelCase = int((i / placement) % RADIX ) buckets[tmp].append(lowerCamelCase_ ) # put each buckets' contents into list_of_ints __lowerCamelCase = 0 for b in range(lowerCamelCase_ ): for i in buckets[b]: __lowerCamelCase = i a += 1 # move to next placement *= RADIX return list_of_ints if __name__ == "__main__": import doctest doctest.testmod()
713
import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class __lowerCAmelCase ( lowerCAmelCase__ ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=1024 , __UpperCAmelCase=1024 , __UpperCAmelCase=3.6 ): '''simple docstring''' __lowerCamelCase = tokenizer __lowerCamelCase = tokenizer.bos_token_id __lowerCamelCase = dataset __lowerCamelCase = seq_length __lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self ): '''simple docstring''' __lowerCamelCase = iter(self.dataset ) __lowerCamelCase = True while more_examples: __lowerCamelCase ,__lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(__UpperCAmelCase )['''content'''] ) buffer_len += len(buffer[-1] ) except StopIteration: __lowerCamelCase = False break __lowerCamelCase = tokenizer(__UpperCAmelCase , truncation=__UpperCAmelCase )['''input_ids'''] __lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(__UpperCAmelCase ) , self.seq_length ): __lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(__UpperCAmelCase ) == self.seq_length: yield torch.tensor(__UpperCAmelCase ) def a__ ( _UpperCamelCase : List[Any] ): __lowerCamelCase = {'''streaming''': True} __lowerCamelCase = load_dataset(args.dataset_name ,split='''train''' ,**_UpperCamelCase ) __lowerCamelCase = ConstantLengthDataset(_UpperCamelCase ,_UpperCamelCase ,seq_length=args.seq_length ) __lowerCamelCase = DataLoader(_UpperCamelCase ,batch_size=args.batch_size ) return eval_dataloader def a__ ( _UpperCamelCase : str ): model.eval() __lowerCamelCase = [] for step, batch in enumerate(_UpperCamelCase ): with torch.no_grad(): __lowerCamelCase = model(_UpperCamelCase ,labels=_UpperCamelCase ) __lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(_UpperCamelCase ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break __lowerCamelCase = torch.mean(torch.cat(_UpperCamelCase ) ) try: __lowerCamelCase = torch.exp(_UpperCamelCase ) except OverflowError: __lowerCamelCase = float('''inf''' ) return loss.item(), perplexity.item() # Setup Accelerator a_ = Accelerator() # Parse configuration a_ = HfArgumentParser(EvaluationArguments) a_ = parser.parse_args() set_seed(args.seed) # Logging a_ = logging.getLogger(__name__) logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) # Load model and tokenizer a_ = AutoModelForCausalLM.from_pretrained(args.model_ckpt) a_ = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader a_ = create_dataloader(args) # Prepare everything with our `accelerator`. a_ , a_ = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info("""Evaluating and saving model after training""") a_ , a_ = evaluate(args) logger.info(f"loss/eval: {eval_loss}, perplexity: {perplexity}")
622
0
# # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def a__ ( *_UpperCamelCase : Any ): with open(A__ ,'''r''' ) as fh: fcntl.flock(A__ ,fcntl.LOCK_EX ) try: print(*A__ ) finally: fcntl.flock(A__ ,fcntl.LOCK_UN ) a_ = int(os.environ["""LOCAL_RANK"""]) torch.cuda.set_device(local_rank) a_ = torch.device("""cuda""", local_rank) a_ = socket.gethostname() a_ = f"[{hostname}-{local_rank}]" try: # test distributed dist.init_process_group("""nccl""") dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank a_ = dist.get_rank() a_ = dist.get_world_size() printflock(f"{gpu} is OK (global rank: {rank}/{world_size})") dist.barrier() if rank == 0: printflock(f"pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}") except Exception: printflock(f"{gpu} is broken") raise
714
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """lxmert""" lowerCAmelCase__ = {} def __init__( self , __UpperCAmelCase=30522 , __UpperCAmelCase=768 , __UpperCAmelCase=12 , __UpperCAmelCase=9500 , __UpperCAmelCase=1600 , __UpperCAmelCase=400 , __UpperCAmelCase=3072 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=9 , __UpperCAmelCase=5 , __UpperCAmelCase=5 , __UpperCAmelCase=2048 , __UpperCAmelCase=4 , __UpperCAmelCase=6.67 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = num_qa_labels __lowerCamelCase = num_object_labels __lowerCamelCase = num_attr_labels __lowerCamelCase = l_layers __lowerCamelCase = x_layers __lowerCamelCase = r_layers __lowerCamelCase = visual_feat_dim __lowerCamelCase = visual_pos_dim __lowerCamelCase = visual_loss_normalizer __lowerCamelCase = task_matched __lowerCamelCase = task_mask_lm __lowerCamelCase = task_obj_predict __lowerCamelCase = task_qa __lowerCamelCase = visual_obj_loss __lowerCamelCase = visual_attr_loss __lowerCamelCase = visual_feat_loss __lowerCamelCase = {'''vision''': r_layers, '''cross_encoder''': x_layers, '''language''': l_layers} super().__init__(**__UpperCAmelCase )
622
0
'''simple docstring''' import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class __lowerCAmelCase : @staticmethod def lowerCamelCase ( *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' pass def a__ ( _UpperCamelCase : List[str] ): __lowerCamelCase = hashlib.mda(image.tobytes() ) return m.hexdigest()[:10] def a__ ( _UpperCamelCase : Any ): __lowerCamelCase = np.array(lowerCamelCase__ ) __lowerCamelCase = npimg.shape return {"hash": hashimage(lowerCamelCase__ ), "shape": shape} @is_pipeline_test @require_vision @require_torch class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) lowerCAmelCase__ = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = MaskGenerationPipeline(model=__UpperCAmelCase , image_processor=__UpperCAmelCase ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' pass @require_tf @unittest.skip('''Image segmentation not implemented in TF''' ) def lowerCamelCase ( self ): '''simple docstring''' pass @slow @require_torch def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline('''mask-generation''' , model='''facebook/sam-vit-huge''' ) __lowerCamelCase = image_segmenter('''http://images.cocodataset.org/val2017/000000039769.jpg''' , points_per_batch=256 ) # Shortening by hashing __lowerCamelCase = [] for i, o in enumerate(outputs['''masks'''] ): new_outupt += [{"mask": mask_to_test_readable(__UpperCAmelCase ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''mask''': {'''hash''': '''115ad19f5f''', '''shape''': (480, 640)}, '''scores''': 1.0_444}, {'''mask''': {'''hash''': '''6affa964c6''', '''shape''': (480, 640)}, '''scores''': 1.021}, {'''mask''': {'''hash''': '''dfe28a0388''', '''shape''': (480, 640)}, '''scores''': 1.0_167}, {'''mask''': {'''hash''': '''c0a5f4a318''', '''shape''': (480, 640)}, '''scores''': 1.0_132}, {'''mask''': {'''hash''': '''fe8065c197''', '''shape''': (480, 640)}, '''scores''': 1.0_053}, {'''mask''': {'''hash''': '''e2d0b7a0b7''', '''shape''': (480, 640)}, '''scores''': 0.9_967}, {'''mask''': {'''hash''': '''453c7844bd''', '''shape''': (480, 640)}, '''scores''': 0.993}, {'''mask''': {'''hash''': '''3d44f2926d''', '''shape''': (480, 640)}, '''scores''': 0.9_909}, {'''mask''': {'''hash''': '''64033ddc3f''', '''shape''': (480, 640)}, '''scores''': 0.9_879}, {'''mask''': {'''hash''': '''801064ff79''', '''shape''': (480, 640)}, '''scores''': 0.9_834}, {'''mask''': {'''hash''': '''6172f276ef''', '''shape''': (480, 640)}, '''scores''': 0.9_716}, {'''mask''': {'''hash''': '''b49e60e084''', '''shape''': (480, 640)}, '''scores''': 0.9_612}, {'''mask''': {'''hash''': '''a811e775fd''', '''shape''': (480, 640)}, '''scores''': 0.9_599}, {'''mask''': {'''hash''': '''a6a8ebcf4b''', '''shape''': (480, 640)}, '''scores''': 0.9_552}, {'''mask''': {'''hash''': '''9d8257e080''', '''shape''': (480, 640)}, '''scores''': 0.9_532}, {'''mask''': {'''hash''': '''32de6454a8''', '''shape''': (480, 640)}, '''scores''': 0.9_516}, {'''mask''': {'''hash''': '''af3d4af2c8''', '''shape''': (480, 640)}, '''scores''': 0.9_499}, {'''mask''': {'''hash''': '''3c6db475fb''', '''shape''': (480, 640)}, '''scores''': 0.9_483}, {'''mask''': {'''hash''': '''c290813fb9''', '''shape''': (480, 640)}, '''scores''': 0.9_464}, {'''mask''': {'''hash''': '''b6f0b8f606''', '''shape''': (480, 640)}, '''scores''': 0.943}, {'''mask''': {'''hash''': '''92ce16bfdf''', '''shape''': (480, 640)}, '''scores''': 0.943}, {'''mask''': {'''hash''': '''c749b25868''', '''shape''': (480, 640)}, '''scores''': 0.9_408}, {'''mask''': {'''hash''': '''efb6cab859''', '''shape''': (480, 640)}, '''scores''': 0.9_335}, {'''mask''': {'''hash''': '''1ff2eafb30''', '''shape''': (480, 640)}, '''scores''': 0.9_326}, {'''mask''': {'''hash''': '''788b798e24''', '''shape''': (480, 640)}, '''scores''': 0.9_262}, {'''mask''': {'''hash''': '''abea804f0e''', '''shape''': (480, 640)}, '''scores''': 0.8_999}, {'''mask''': {'''hash''': '''7b9e8ddb73''', '''shape''': (480, 640)}, '''scores''': 0.8_986}, {'''mask''': {'''hash''': '''cd24047c8a''', '''shape''': (480, 640)}, '''scores''': 0.8_984}, {'''mask''': {'''hash''': '''6943e6bcbd''', '''shape''': (480, 640)}, '''scores''': 0.8_873}, {'''mask''': {'''hash''': '''b5f47c9191''', '''shape''': (480, 640)}, '''scores''': 0.8_871} ] , ) # fmt: on @require_torch @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = "facebook/sam-vit-huge" __lowerCamelCase = pipeline('''mask-generation''' , model=__UpperCAmelCase ) __lowerCamelCase = image_segmenter( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , pred_iou_thresh=1 , points_per_batch=256 ) # Shortening by hashing __lowerCamelCase = [] for i, o in enumerate(outputs['''masks'''] ): new_outupt += [{"mask": mask_to_test_readable(__UpperCAmelCase ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''mask''': {'''hash''': '''115ad19f5f''', '''shape''': (480, 640)}, '''scores''': 1.0_444}, {'''mask''': {'''hash''': '''6affa964c6''', '''shape''': (480, 640)}, '''scores''': 1.0_210}, {'''mask''': {'''hash''': '''dfe28a0388''', '''shape''': (480, 640)}, '''scores''': 1.0_167}, {'''mask''': {'''hash''': '''c0a5f4a318''', '''shape''': (480, 640)}, '''scores''': 1.0_132}, {'''mask''': {'''hash''': '''fe8065c197''', '''shape''': (480, 640)}, '''scores''': 1.0_053}, ] , )
715
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Any ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Any ): if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = np.full((len(_UpperCamelCase ), sequence_length, 2) ,_UpperCamelCase ) else: __lowerCamelCase = np.full((len(_UpperCamelCase ), sequence_length) ,_UpperCamelCase ) for i, tensor in enumerate(_UpperCamelCase ): if padding_side == "right": if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = tensor[:sequence_length] else: __lowerCamelCase = tensor[:sequence_length] else: if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = tensor[:sequence_length] else: __lowerCamelCase = tensor[:sequence_length] return out_tensor.tolist() def a__ ( _UpperCamelCase : Dict ): __lowerCamelCase = ord(_UpperCamelCase ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 1_23 and cp <= 1_26): return True __lowerCamelCase = unicodedata.category(_UpperCamelCase ) if cat.startswith('''P''' ): return True return False @dataclass class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = 42 lowerCAmelCase__ = True lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = -1_0_0 lowerCAmelCase__ = "pt" def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' import torch __lowerCamelCase = '''label''' if '''label''' in features[0].keys() else '''labels''' __lowerCamelCase = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __lowerCamelCase = self.tokenizer.pad( __UpperCAmelCase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch __lowerCamelCase = torch.tensor(batch['''entity_ids'''] ).shape[1] __lowerCamelCase = self.tokenizer.padding_side if padding_side == "right": __lowerCamelCase = [ list(__UpperCAmelCase ) + [self.label_pad_token_id] * (sequence_length - len(__UpperCAmelCase )) for label in labels ] else: __lowerCamelCase = [ [self.label_pad_token_id] * (sequence_length - len(__UpperCAmelCase )) + list(__UpperCAmelCase ) for label in labels ] __lowerCamelCase = [feature['''ner_tags'''] for feature in features] __lowerCamelCase = padding_tensor(__UpperCAmelCase , -1 , __UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = [feature['''original_entity_spans'''] for feature in features] __lowerCamelCase = padding_tensor(__UpperCAmelCase , (-1, -1) , __UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = {k: torch.tensor(__UpperCAmelCase , dtype=torch.intaa ) for k, v in batch.items()} return batch
622
0
import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_MAPPING, LevitForImageClassification, LevitForImageClassificationWithTeacher, LevitModel, ) from transformers.models.levit.modeling_levit import LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class __lowerCAmelCase ( lowerCAmelCase__ ): def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase_ , '''hidden_sizes''' ) ) self.parent.assertTrue(hasattr(lowercase_ , '''num_attention_heads''' ) ) class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=64 , __UpperCAmelCase=3 , __UpperCAmelCase=3 , __UpperCAmelCase=2 , __UpperCAmelCase=1 , __UpperCAmelCase=16 , __UpperCAmelCase=[128, 256, 384] , __UpperCAmelCase=[4, 6, 8] , __UpperCAmelCase=[2, 3, 4] , __UpperCAmelCase=[16, 16, 16] , __UpperCAmelCase=0 , __UpperCAmelCase=[2, 2, 2] , __UpperCAmelCase=[2, 2, 2] , __UpperCAmelCase=0.02 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=2 , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = image_size __lowerCamelCase = num_channels __lowerCamelCase = kernel_size __lowerCamelCase = stride __lowerCamelCase = padding __lowerCamelCase = hidden_sizes __lowerCamelCase = num_attention_heads __lowerCamelCase = depths __lowerCamelCase = key_dim __lowerCamelCase = drop_path_rate __lowerCamelCase = patch_size __lowerCamelCase = attention_ratio __lowerCamelCase = mlp_ratio __lowerCamelCase = initializer_range __lowerCamelCase = [ ['''Subsample''', key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ['''Subsample''', key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] __lowerCamelCase = is_training __lowerCamelCase = use_labels __lowerCamelCase = num_labels __lowerCamelCase = initializer_range def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.num_labels ) __lowerCamelCase = self.get_config() return config, pixel_values, labels def lowerCamelCase ( self ): '''simple docstring''' return LevitConfig( image_size=self.image_size , num_channels=self.num_channels , kernel_size=self.kernel_size , stride=self.stride , padding=self.padding , patch_size=self.patch_size , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , depths=self.depths , key_dim=self.key_dim , drop_path_rate=self.drop_path_rate , mlp_ratio=self.mlp_ratio , attention_ratio=self.attention_ratio , initializer_range=self.initializer_range , down_ops=self.down_ops , ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = LevitModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() __lowerCamelCase = model(lowercase_ ) __lowerCamelCase = (self.image_size, self.image_size) __lowerCamelCase ,__lowerCamelCase = image_size[0], image_size[1] for _ in range(4 ): __lowerCamelCase = floor(((height + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) __lowerCamelCase = floor(((width + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, ceil(height / 4 ) * ceil(width / 4 ), self.hidden_sizes[-1]) , ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = LevitForImageClassification(lowercase_ ) model.to(lowercase_ ) model.eval() __lowerCamelCase = model(lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.prepare_config_and_inputs() __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = config_and_inputs __lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = ( (LevitModel, LevitForImageClassification, LevitForImageClassificationWithTeacher) if is_torch_available() else () ) lowerCAmelCase__ = ( { """feature-extraction""": LevitModel, """image-classification""": (LevitForImageClassification, LevitForImageClassificationWithTeacher), } if is_torch_available() else {} ) lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = LevitModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=lowercase_ , has_text_modality=lowercase_ , hidden_size=37 ) def lowerCamelCase ( self ): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCamelCase ( self ): '''simple docstring''' return @unittest.skip(reason='''Levit does not use inputs_embeds''' ) def lowerCamelCase ( self ): '''simple docstring''' pass @unittest.skip(reason='''Levit does not support input and output embeddings''' ) def lowerCamelCase ( self ): '''simple docstring''' pass @unittest.skip(reason='''Levit does not output attentions''' ) def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = model_class(lowercase_ ) __lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCamelCase = [*signature.parameters.keys()] __lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowercase_ ) def lowerCamelCase ( self ): '''simple docstring''' def check_hidden_states_output(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): __lowerCamelCase = model_class(lowercase_ ) model.to(lowercase_ ) model.eval() with torch.no_grad(): __lowerCamelCase = model(**self._prepare_for_class(lowercase_ , lowercase_ ) ) __lowerCamelCase = outputs.hidden_states __lowerCamelCase = len(self.model_tester.depths ) + 1 self.assertEqual(len(lowercase_ ) , lowercase_ ) __lowerCamelCase = (self.model_tester.image_size, self.model_tester.image_size) __lowerCamelCase ,__lowerCamelCase = image_size[0], image_size[1] for _ in range(4 ): __lowerCamelCase = floor( ( (height + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) __lowerCamelCase = floor( ( (width + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [ height * width, self.model_tester.hidden_sizes[0], ] , ) __lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase = True check_hidden_states_output(lowercase_ , lowercase_ , lowercase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowerCamelCase = True check_hidden_states_output(lowercase_ , lowercase_ , lowercase_ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=False ): '''simple docstring''' __lowerCamelCase = super()._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_ ) if return_labels: if model_class.__name__ == "LevitForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase_ ) def lowerCamelCase ( self ): '''simple docstring''' if not self.model_tester.is_training: return __lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = True for model_class in self.all_model_classes: # LevitForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowercase_ ) or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue __lowerCamelCase = model_class(lowercase_ ) model.to(lowercase_ ) model.train() __lowerCamelCase = self._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_ ) __lowerCamelCase = model(**lowercase_ ).loss loss.backward() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return __lowerCamelCase = False __lowerCamelCase = True for model_class in self.all_model_classes: if model_class in get_values(lowercase_ ) or not model_class.supports_gradient_checkpointing: continue # LevitForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "LevitForImageClassificationWithTeacher": continue __lowerCamelCase = model_class(lowercase_ ) model.gradient_checkpointing_enable() model.to(lowercase_ ) model.train() __lowerCamelCase = self._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_ ) __lowerCamelCase = model(**lowercase_ ).loss loss.backward() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = [ {'''title''': '''multi_label_classification''', '''num_labels''': 2, '''dtype''': torch.float}, {'''title''': '''single_label_classification''', '''num_labels''': 1, '''dtype''': torch.long}, {'''title''': '''regression''', '''num_labels''': 1, '''dtype''': torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowercase_ ), ] or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F"""Testing {model_class} with {problem_type["title"]}""" ): __lowerCamelCase = problem_type['''title'''] __lowerCamelCase = problem_type['''num_labels'''] __lowerCamelCase = model_class(lowercase_ ) model.to(lowercase_ ) model.train() __lowerCamelCase = self._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_ ) if problem_type["num_labels"] > 1: __lowerCamelCase = inputs['''labels'''].unsqueeze(1 ).repeat(1 , problem_type['''num_labels'''] ) __lowerCamelCase = inputs['''labels'''].to(problem_type['''dtype'''] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowercase_ ) as warning_list: __lowerCamelCase = model(**lowercase_ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F"""Something is going wrong in the regression problem: intercepted {w.message}""" ) loss.backward() @slow def lowerCamelCase ( self ): '''simple docstring''' for model_name in LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = LevitModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) def a__ ( ): __lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __lowerCAmelCase ( unittest.TestCase ): @cached_property def lowerCamelCase ( self ): '''simple docstring''' return LevitImageProcessor.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = LevitForImageClassificationWithTeacher.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( lowercase_ ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = prepare_img() __lowerCamelCase = image_processor(images=lowercase_ , return_tensors='''pt''' ).to(lowercase_ ) # forward pass with torch.no_grad(): __lowerCamelCase = model(**lowercase_ ) # verify the logits __lowerCamelCase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowercase_ ) __lowerCamelCase = torch.tensor([1.0_448, -0.3_745, -1.8_317] ).to(lowercase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase_ , atol=1E-4 ) )
716
from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=99 , __UpperCAmelCase=[1, 1, 2] , __UpperCAmelCase=1 , __UpperCAmelCase=32 , __UpperCAmelCase=4 , __UpperCAmelCase=8 , __UpperCAmelCase=37 , __UpperCAmelCase="gelu_new" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.0 , __UpperCAmelCase=512 , __UpperCAmelCase=3 , __UpperCAmelCase=0.02 , __UpperCAmelCase=3 , __UpperCAmelCase=4 , __UpperCAmelCase=None , __UpperCAmelCase=False , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_input_mask __lowerCamelCase = use_token_type_ids __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = block_sizes __lowerCamelCase = num_decoder_layers __lowerCamelCase = d_model __lowerCamelCase = n_head __lowerCamelCase = d_head __lowerCamelCase = d_inner __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout __lowerCamelCase = attention_dropout __lowerCamelCase = activation_dropout __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = 2 __lowerCamelCase = num_labels __lowerCamelCase = num_choices __lowerCamelCase = scope __lowerCamelCase = initializer_std # Used in the tests to check the size of the first attention layer __lowerCamelCase = n_head # Used in the tests to check the size of the first hidden state __lowerCamelCase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowerCamelCase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowerCamelCase = self.num_hidden_layers + 2 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCamelCase = None if self.use_token_type_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCamelCase = FunnelConfig( vocab_size=self.vocab_size , block_sizes=self.block_sizes , num_decoder_layers=self.num_decoder_layers , d_model=self.d_model , n_head=self.n_head , d_head=self.d_head , d_inner=self.d_inner , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , activation_dropout=self.activation_dropout , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_std=self.initializer_std , ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = [input_ids, input_mask] __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = [input_ids, input_mask] __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 3, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForPreTraining(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForMaskedLM(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = TFFunnelForSequenceClassification(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_choices __lowerCamelCase = TFFunnelForMultipleChoice(config=__UpperCAmelCase ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = TFFunnelForTokenClassification(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForQuestionAnswering(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.prepare_config_and_inputs() ( ( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) , ) = config_and_inputs __lowerCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) lowerCAmelCase__ = ( { """feature-extraction""": (TFFunnelBaseModel, TFFunnelModel), """fill-mask""": TFFunnelForMaskedLM, """question-answering""": TFFunnelForQuestionAnswering, """text-classification""": TFFunnelForSequenceClassification, """token-classification""": TFFunnelForTokenClassification, """zero-shot""": TFFunnelForSequenceClassification, } if is_tf_available() else {} ) lowerCAmelCase__ = False lowerCAmelCase__ = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFFunnelModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCAmelCase ) @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) lowerCAmelCase__ = False lowerCAmelCase__ = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFFunnelModelTester(self , base=__UpperCAmelCase ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCAmelCase )
622
0
import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() a_ = logging.get_logger(__name__) def a__ ( _UpperCamelCase : str ): __lowerCamelCase = torch.load(snake_case__ ,map_location='''cpu''' ) if "model" in sd.keys(): __lowerCamelCase = torch.load(snake_case__ ,map_location='''cpu''' )['''model'''] # pop unnecessary weights __lowerCamelCase = [ '''decoder.version''', '''decoder.output_projection.weight''', ] for key in keys_to_delete: if key in sd: sd.pop(snake_case__ ) __lowerCamelCase = { '''decoder.project_in_dim.weight''': '''decoder.project_in.weight''', '''decoder.project_out_dim.weight''': '''decoder.project_out.weight''', '''decoder.layer_norm.weight''': '''decoder.final_layer_norm.weight''', '''decoder.layer_norm.bias''': '''decoder.final_layer_norm.bias''', } for old_key, new_key in keys_to_rename.items(): if old_key in sd: __lowerCamelCase = sd.pop(snake_case__ ) __lowerCamelCase = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: __lowerCamelCase = sd[key] # We split QKV in separate Q,K,V __lowerCamelCase = key.replace('''.qkv_proj.''' ,'''.q_proj.''' ) __lowerCamelCase = key.replace('''.qkv_proj.''' ,'''.k_proj.''' ) __lowerCamelCase = key.replace('''.qkv_proj.''' ,'''.v_proj.''' ) __lowerCamelCase = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = torch.split(snake_case__ ,depth // 3 ,dim=0 ) __lowerCamelCase = q __lowerCamelCase = k __lowerCamelCase = v del sd[key] return sd @torch.no_grad() def a__ ( _UpperCamelCase : List[Any] ,_UpperCamelCase : Tuple ,_UpperCamelCase : List[str]=None ): __lowerCamelCase = load_checkpoint(snake_case__ ) if config is not None: __lowerCamelCase = OPTConfig.from_pretrained(snake_case__ ) else: __lowerCamelCase = OPTConfig() __lowerCamelCase = OPTModel(snake_case__ ).half().eval() model.load_state_dict(snake_case__ ) # Check results Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) model.save_pretrained(snake_case__ ) if __name__ == "__main__": a_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--fairseq_path""", type=str, help=( """path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:""" """ https://huggingface.co/models?other=opt_metasq""" ), ) parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--hf_config""", default=None, type=str, help="""Define HF config.""") a_ = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
717
from collections import namedtuple import requests from lxml import html # type: ignore a_ = namedtuple("""covid_data""", """cases deaths recovered""") def a__ ( _UpperCamelCase : str = "https://www.worldometers.info/coronavirus/" ): __lowerCamelCase = '''//div[@class = "maincounter-number"]/span/text()''' return covid_data(*html.fromstring(requests.get(_UpperCamelCase ).content ).xpath(_UpperCamelCase ) ) a_ = """Total COVID-19 cases in the world: {} Total deaths due to COVID-19 in the world: {} Total COVID-19 patients recovered in the world: {}""" print(fmt.format(*covid_stats()))
622
0
import string import numpy def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Dict ): return b if a == 0 else greatest_common_divisor(b % a ,_UpperCamelCase ) class __lowerCAmelCase : lowerCAmelCase__ = string.ascii_uppercase + string.digits # This cipher takes alphanumerics into account # i.e. a total of 36 characters # take x and return x % len(key_string) lowerCAmelCase__ = numpy.vectorize(lambda lowerCAmelCase__ : x % 3_6 ) lowerCAmelCase__ = numpy.vectorize(lowerCAmelCase__ ) def __init__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.modulus(UpperCAmelCase__ ) # mod36 calc's on the encrypt key self.check_determinant() # validate the determinant of the encryption key __lowerCamelCase = encrypt_key.shape[0] def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.key_string.index(UpperCAmelCase__ ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.key_string[round(UpperCAmelCase__ )] def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: __lowerCamelCase = det % len(self.key_string ) __lowerCamelCase = len(self.key_string ) if greatest_common_divisor(UpperCAmelCase__ , len(self.key_string ) ) != 1: __lowerCamelCase = ( F"""determinant modular {req_l} of encryption key({det}) """ F"""is not co prime w.r.t {req_l}.\nTry another key.""" ) raise ValueError(UpperCAmelCase__ ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = [char for char in text.upper() if char in self.key_string] __lowerCamelCase = chars[-1] while len(UpperCAmelCase__ ) % self.break_key != 0: chars.append(UpperCAmelCase__ ) return "".join(UpperCAmelCase__ ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.process_text(text.upper() ) __lowerCamelCase = '''''' for i in range(0 , len(UpperCAmelCase__ ) - self.break_key + 1 , self.break_key ): __lowerCamelCase = text[i : i + self.break_key] __lowerCamelCase = [self.replace_letters(UpperCAmelCase__ ) for char in batch] __lowerCamelCase = numpy.array([vec] ).T __lowerCamelCase = self.modulus(self.encrypt_key.dot(UpperCAmelCase__ ) ).T.tolist()[ 0 ] __lowerCamelCase = ''''''.join( self.replace_digits(UpperCAmelCase__ ) for num in batch_encrypted ) encrypted += encrypted_batch return encrypted def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: __lowerCamelCase = det % len(self.key_string ) __lowerCamelCase = None for i in range(len(self.key_string ) ): if (det * i) % len(self.key_string ) == 1: __lowerCamelCase = i break __lowerCamelCase = ( det_inv * numpy.linalg.det(self.encrypt_key ) * numpy.linalg.inv(self.encrypt_key ) ) return self.to_int(self.modulus(UpperCAmelCase__ ) ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.make_decrypt_key() __lowerCamelCase = self.process_text(text.upper() ) __lowerCamelCase = '''''' for i in range(0 , len(UpperCAmelCase__ ) - self.break_key + 1 , self.break_key ): __lowerCamelCase = text[i : i + self.break_key] __lowerCamelCase = [self.replace_letters(UpperCAmelCase__ ) for char in batch] __lowerCamelCase = numpy.array([vec] ).T __lowerCamelCase = self.modulus(decrypt_key.dot(UpperCAmelCase__ ) ).T.tolist()[0] __lowerCamelCase = ''''''.join( self.replace_digits(UpperCAmelCase__ ) for num in batch_decrypted ) decrypted += decrypted_batch return decrypted def a__ ( ): __lowerCamelCase = int(input('''Enter the order of the encryption key: ''' ) ) __lowerCamelCase = [] print('''Enter each row of the encryption key with space separated integers''' ) for _ in range(_UpperCamelCase ): __lowerCamelCase = [int(_UpperCamelCase ) for x in input().split()] hill_matrix.append(_UpperCamelCase ) __lowerCamelCase = HillCipher(numpy.array(_UpperCamelCase ) ) print('''Would you like to encrypt or decrypt some text? (1 or 2)''' ) __lowerCamelCase = input('''\n1. Encrypt\n2. Decrypt\n''' ) if option == "1": __lowerCamelCase = input('''What text would you like to encrypt?: ''' ) print('''Your encrypted text is:''' ) print(hc.encrypt(_UpperCamelCase ) ) elif option == "2": __lowerCamelCase = input('''What text would you like to decrypt?: ''' ) print('''Your decrypted text is:''' ) print(hc.decrypt(_UpperCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
718
def a__ ( _UpperCamelCase : str ,_UpperCamelCase : str = " " ): __lowerCamelCase = [] __lowerCamelCase = 0 for index, char in enumerate(_UpperCamelCase ): if char == separator: split_words.append(string[last_index:index] ) __lowerCamelCase = index + 1 elif index + 1 == len(_UpperCamelCase ): split_words.append(string[last_index : index + 1] ) return split_words if __name__ == "__main__": from doctest import testmod testmod()
622
0
import warnings from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): lowerCAmelCase__ = ["""image_processor""", """tokenizer"""] lowerCAmelCase__ = """FlavaImageProcessor""" lowerCAmelCase__ = ("""BertTokenizer""", """BertTokenizerFast""") def __init__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , **__UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , snake_case__ , ) __lowerCamelCase = kwargs.pop('''feature_extractor''' ) __lowerCamelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(snake_case__ , snake_case__ ) __lowerCamelCase = self.image_processor def __call__( self , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = True , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = None , __UpperCAmelCase = 0 , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = True , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: __lowerCamelCase = self.tokenizer( text=snake_case__ , add_special_tokens=snake_case__ , padding=snake_case__ , truncation=snake_case__ , max_length=snake_case__ , stride=snake_case__ , pad_to_multiple_of=snake_case__ , return_token_type_ids=snake_case__ , return_attention_mask=snake_case__ , return_overflowing_tokens=snake_case__ , return_special_tokens_mask=snake_case__ , return_offsets_mapping=snake_case__ , return_length=snake_case__ , verbose=snake_case__ , return_tensors=snake_case__ , **snake_case__ , ) if images is not None: __lowerCamelCase = self.image_processor( snake_case__ , return_image_mask=snake_case__ , return_codebook_pixels=snake_case__ , return_tensors=snake_case__ , **snake_case__ , ) if text is not None and images is not None: encoding.update(snake_case__ ) return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**snake_case__ ) , tensor_type=snake_case__ ) def lowerCamelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' return self.tokenizer.batch_decode(*snake_case__ , **snake_case__ ) def lowerCamelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' return self.tokenizer.decode(*snake_case__ , **snake_case__ ) @property def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.tokenizer.model_input_names __lowerCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCamelCase ( self ): '''simple docstring''' warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , snake_case__ , ) return self.image_processor_class @property def lowerCamelCase ( self ): '''simple docstring''' warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , snake_case__ , ) return self.image_processor
719
import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class __lowerCAmelCase ( lowerCAmelCase__ ): def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = tempfile.mkdtemp() __lowerCamelCase = 8 # DPR tok __lowerCamelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] __lowerCamelCase = os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) __lowerCamelCase = os.path.join(__UpperCAmelCase , DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok __lowerCamelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] __lowerCamelCase = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) __lowerCamelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] __lowerCamelCase = {'''unk_token''': '''<unk>'''} __lowerCamelCase = os.path.join(self.tmpdirname , '''bart_tokenizer''' ) os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) __lowerCamelCase = os.path.join(__UpperCAmelCase , BART_VOCAB_FILES_NAMES['''vocab_file'''] ) __lowerCamelCase = os.path.join(__UpperCAmelCase , BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(__UpperCAmelCase ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(__UpperCAmelCase ) ) def lowerCamelCase ( self ): '''simple docstring''' return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''bart_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_dataset() __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: __lowerCamelCase = dataset __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.get_dummy_dataset() __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''custom''' , ) if from_disk: __lowerCamelCase = os.path.join(self.tmpdirname , '''dataset''' ) __lowerCamelCase = os.path.join(self.tmpdirname , '''index.faiss''' ) dataset.get_index('''embeddings''' ).save(os.path.join(self.tmpdirname , '''index.faiss''' ) ) dataset.drop_index('''embeddings''' ) dataset.save_to_disk(os.path.join(self.tmpdirname , '''dataset''' ) ) del dataset __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , __UpperCAmelCase ) , ) return retriever def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) __lowerCamelCase = os.path.join(self.tmpdirname , '''hf_bert_base.hnswSQ8_correct_phi_128.c_index''' ) dataset.save_faiss_index('''embeddings''' , index_file_name + '''.index.dpr''' ) pickle.dump(dataset['''id'''] , open(index_file_name + '''.index_meta.dpr''' , '''wb''' ) ) __lowerCamelCase = os.path.join(self.tmpdirname , '''psgs_w100.tsv.pkl''' ) __lowerCamelCase = {sample['''id''']: [sample['''text'''], sample['''title''']] for sample in dataset} pickle.dump(__UpperCAmelCase , open(__UpperCAmelCase , '''wb''' ) ) __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''legacy''' , index_path=self.tmpdirname , ) __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: __lowerCamelCase = self.get_dummy_dataset() retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_legacy_index_retriever() __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''text'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''text'''][0] , '''bar''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''text'''][0] , '''foo''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def lowerCamelCase ( self ): '''simple docstring''' import torch __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() __lowerCamelCase = [[5, 7], [10, 11]] __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever(__UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = ( out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , np.ndarray ) __lowerCamelCase = retriever( __UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase , return_tensors='''pt''' , ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = ( # noqa: F841 out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], out['''doc_ids'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dpr_ctx_encoder_tokenizer() __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) retriever.set_ctx_encoder_tokenizer(__UpperCAmelCase ) __lowerCamelCase = [[5, 7], [10, 11]] __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever(__UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase ) self.assertEqual( len(__UpperCAmelCase ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('''tokenized_doc_ids''', '''tokenized_doc_attention_mask''') ) , __UpperCAmelCase ) # check for doc token related keys in dictionary.
622
0
import math def a__ ( _UpperCamelCase : int ): __lowerCamelCase = 0 __lowerCamelCase = 0 while num > 0: __lowerCamelCase = num % 8 __lowerCamelCase = octal + (remainder * math.floor(math.pow(10 ,__a ) )) counter += 1 __lowerCamelCase = math.floor(num / 8 ) # basically /= 8 without remainder if any # This formatting removes trailing '.0' from `octal`. return F"""0o{int(__a )}""" def a__ ( ): print('''\n2 in octal is:''' ) print(decimal_to_octal(2 ) ) # = 2 print('''\n8 in octal is:''' ) print(decimal_to_octal(8 ) ) # = 10 print('''\n65 in octal is:''' ) print(decimal_to_octal(65 ) ) # = 101 print('''\n216 in octal is:''' ) print(decimal_to_octal(2_16 ) ) # = 330 print('''\n512 in octal is:''' ) print(decimal_to_octal(5_12 ) ) # = 1000 print('''\n''' ) if __name__ == "__main__": main()
720
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """sail/poolformer_s12""": """https://huggingface.co/sail/poolformer_s12/resolve/main/config.json""", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """poolformer""" def __init__( self , __UpperCAmelCase=3 , __UpperCAmelCase=16 , __UpperCAmelCase=16 , __UpperCAmelCase=3 , __UpperCAmelCase=4.0 , __UpperCAmelCase=[2, 2, 6, 2] , __UpperCAmelCase=[64, 128, 320, 512] , __UpperCAmelCase=[7, 3, 3, 3] , __UpperCAmelCase=[4, 2, 2, 2] , __UpperCAmelCase=[2, 1, 1, 1] , __UpperCAmelCase=4 , __UpperCAmelCase=0.0 , __UpperCAmelCase="gelu" , __UpperCAmelCase=True , __UpperCAmelCase=1E-5 , __UpperCAmelCase=0.02 , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = num_channels __lowerCamelCase = patch_size __lowerCamelCase = stride __lowerCamelCase = padding __lowerCamelCase = pool_size __lowerCamelCase = hidden_sizes __lowerCamelCase = mlp_ratio __lowerCamelCase = depths __lowerCamelCase = patch_sizes __lowerCamelCase = strides __lowerCamelCase = num_encoder_blocks __lowerCamelCase = drop_path_rate __lowerCamelCase = hidden_act __lowerCamelCase = use_layer_scale __lowerCamelCase = layer_scale_init_value __lowerCamelCase = initializer_range super().__init__(**__UpperCAmelCase ) class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = version.parse("""1.11""" ) @property def lowerCamelCase ( self ): '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def lowerCamelCase ( self ): '''simple docstring''' return 2E-3
622
0
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available a_ = { """configuration_cpmant""": ["""CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """CpmAntConfig"""], """tokenization_cpmant""": ["""CpmAntTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ """CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST""", """CpmAntForCausalLM""", """CpmAntModel""", """CpmAntPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_cpmant import CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP, CpmAntConfig from .tokenization_cpmant import CpmAntTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_cpmant import ( CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST, CpmAntForCausalLM, CpmAntModel, CpmAntPreTrainedModel, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
721
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """uclanlp/visualbert-vqa""": """https://huggingface.co/uclanlp/visualbert-vqa/resolve/main/config.json""", """uclanlp/visualbert-vqa-pre""": """https://huggingface.co/uclanlp/visualbert-vqa-pre/resolve/main/config.json""", """uclanlp/visualbert-vqa-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-vqa-coco-pre/resolve/main/config.json""" ), """uclanlp/visualbert-vcr""": """https://huggingface.co/uclanlp/visualbert-vcr/resolve/main/config.json""", """uclanlp/visualbert-vcr-pre""": """https://huggingface.co/uclanlp/visualbert-vcr-pre/resolve/main/config.json""", """uclanlp/visualbert-vcr-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-vcr-coco-pre/resolve/main/config.json""" ), """uclanlp/visualbert-nlvr2""": """https://huggingface.co/uclanlp/visualbert-nlvr2/resolve/main/config.json""", """uclanlp/visualbert-nlvr2-pre""": """https://huggingface.co/uclanlp/visualbert-nlvr2-pre/resolve/main/config.json""", """uclanlp/visualbert-nlvr2-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-nlvr2-coco-pre/resolve/main/config.json""" ) # See all VisualBERT models at https://huggingface.co/models?filter=visual_bert } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """visual_bert""" def __init__( self , __UpperCAmelCase=30522 , __UpperCAmelCase=768 , __UpperCAmelCase=512 , __UpperCAmelCase=12 , __UpperCAmelCase=12 , __UpperCAmelCase=3072 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase=1 , __UpperCAmelCase=0 , __UpperCAmelCase=2 , **__UpperCAmelCase , ): '''simple docstring''' super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) __lowerCamelCase = vocab_size __lowerCamelCase = max_position_embeddings __lowerCamelCase = hidden_size __lowerCamelCase = visual_embedding_dim __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = initializer_range __lowerCamelCase = type_vocab_size __lowerCamelCase = layer_norm_eps __lowerCamelCase = bypass_transformer __lowerCamelCase = special_visual_initialize
622
0
import argparse import os import re a_ = """src/transformers""" # Pattern that looks at the indentation in a line. a_ = re.compile(R"""^(\s*)\S""") # Pattern that matches `"key":" and puts `key` in group 0. a_ = re.compile(R"""^\s*\"([^\"]+)\":""") # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. a_ = re.compile(R"""^\s*_import_structure\[\"([^\"]+)\"\]""") # Pattern that matches `"key",` and puts `key` in group 0. a_ = re.compile(R"""^\s*\"([^\"]+)\",\s*$""") # Pattern that matches any `[stuff]` and puts `stuff` in group 0. a_ = re.compile(R"""\[([^\]]+)\]""") def a__ ( _UpperCamelCase : int ): __lowerCamelCase = _re_indent.search(_UpperCamelCase ) return "" if search is None else search.groups()[0] def a__ ( _UpperCamelCase : List[str] ,_UpperCamelCase : Dict="" ,_UpperCamelCase : Dict=None ,_UpperCamelCase : Dict=None ): __lowerCamelCase = 0 __lowerCamelCase = code.split('''\n''' ) if start_prompt is not None: while not lines[index].startswith(_UpperCamelCase ): index += 1 __lowerCamelCase = ['''\n'''.join(lines[:index] )] else: __lowerCamelCase = [] # We split into blocks until we get to the `end_prompt` (or the end of the block). __lowerCamelCase = [lines[index]] index += 1 while index < len(_UpperCamelCase ) and (end_prompt is None or not lines[index].startswith(_UpperCamelCase )): if len(lines[index] ) > 0 and get_indent(lines[index] ) == indent_level: if len(_UpperCamelCase ) > 0 and get_indent(current_block[-1] ).startswith(indent_level + ''' ''' ): current_block.append(lines[index] ) blocks.append('''\n'''.join(_UpperCamelCase ) ) if index < len(_UpperCamelCase ) - 1: __lowerCamelCase = [lines[index + 1]] index += 1 else: __lowerCamelCase = [] else: blocks.append('''\n'''.join(_UpperCamelCase ) ) __lowerCamelCase = [lines[index]] else: current_block.append(lines[index] ) index += 1 # Adds current block if it's nonempty. if len(_UpperCamelCase ) > 0: blocks.append('''\n'''.join(_UpperCamelCase ) ) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(_UpperCamelCase ): blocks.append('''\n'''.join(lines[index:] ) ) return blocks def a__ ( _UpperCamelCase : Optional[Any] ): def _inner(_UpperCamelCase : List[Any] ): return key(_UpperCamelCase ).lower().replace('''_''' ,'''''' ) return _inner def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Optional[int]=None ): # If no key is provided, we use a noop. def noop(_UpperCamelCase : Optional[Any] ): return x if key is None: __lowerCamelCase = noop # Constants are all uppercase, they go first. __lowerCamelCase = [obj for obj in objects if key(_UpperCamelCase ).isupper()] # Classes are not all uppercase but start with a capital, they go second. __lowerCamelCase = [obj for obj in objects if key(_UpperCamelCase )[0].isupper() and not key(_UpperCamelCase ).isupper()] # Functions begin with a lowercase, they go last. __lowerCamelCase = [obj for obj in objects if not key(_UpperCamelCase )[0].isupper()] __lowerCamelCase = ignore_underscore(_UpperCamelCase ) return sorted(_UpperCamelCase ,key=_UpperCamelCase ) + sorted(_UpperCamelCase ,key=_UpperCamelCase ) + sorted(_UpperCamelCase ,key=_UpperCamelCase ) def a__ ( _UpperCamelCase : Tuple ): # This inner function sort imports between [ ]. def _replace(_UpperCamelCase : int ): __lowerCamelCase = match.groups()[0] if "," not in imports: return F"""[{imports}]""" __lowerCamelCase = [part.strip().replace('''"''' ,'''''' ) for part in imports.split(''',''' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: __lowerCamelCase = keys[:-1] return "[" + ", ".join([F"""\"{k}\"""" for k in sort_objects(_UpperCamelCase )] ) + "]" __lowerCamelCase = import_statement.split('''\n''' ) if len(_UpperCamelCase ) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. __lowerCamelCase = 2 if lines[1].strip() == '''[''' else 1 __lowerCamelCase = [(i, _re_strip_line.search(_UpperCamelCase ).groups()[0]) for i, line in enumerate(lines[idx:-idx] )] __lowerCamelCase = sort_objects(_UpperCamelCase ,key=lambda _UpperCamelCase : x[1] ) __lowerCamelCase = [lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:] ) elif len(_UpperCamelCase ) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1] ) is not None: __lowerCamelCase = _re_bracket_content.sub(_replace ,lines[1] ) else: __lowerCamelCase = [part.strip().replace('''"''' ,'''''' ) for part in lines[1].split(''',''' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: __lowerCamelCase = keys[:-1] __lowerCamelCase = get_indent(lines[1] ) + ''', '''.join([F"""\"{k}\"""" for k in sort_objects(_UpperCamelCase )] ) return "\n".join(_UpperCamelCase ) else: # Finally we have to deal with imports fitting on one line __lowerCamelCase = _re_bracket_content.sub(_replace ,_UpperCamelCase ) return import_statement def a__ ( _UpperCamelCase : List[Any] ,_UpperCamelCase : Tuple=True ): with open(_UpperCamelCase ,encoding='''utf-8''' ) as f: __lowerCamelCase = f.read() if "_import_structure" not in code: return # Blocks of indent level 0 __lowerCamelCase = split_code_in_indented_blocks( _UpperCamelCase ,start_prompt='''_import_structure = {''' ,end_prompt='''if TYPE_CHECKING:''' ) # We ignore block 0 (everything untils start_prompt) and the last block (everything after end_prompt). for block_idx in range(1 ,len(_UpperCamelCase ) - 1 ): # Check if the block contains some `_import_structure`s thingy to sort. __lowerCamelCase = main_blocks[block_idx] __lowerCamelCase = block.split('''\n''' ) # Get to the start of the imports. __lowerCamelCase = 0 while line_idx < len(_UpperCamelCase ) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: __lowerCamelCase = len(_UpperCamelCase ) else: line_idx += 1 if line_idx >= len(_UpperCamelCase ): continue # Ignore beginning and last line: they don't contain anything. __lowerCamelCase = '''\n'''.join(block_lines[line_idx:-1] ) __lowerCamelCase = get_indent(block_lines[1] ) # Slit the internal block into blocks of indent level 1. __lowerCamelCase = split_code_in_indented_blocks(_UpperCamelCase ,indent_level=_UpperCamelCase ) # We have two categories of import key: list or _import_structure[key].append/extend __lowerCamelCase = _re_direct_key if '''_import_structure = {''' in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. __lowerCamelCase = [(pattern.search(_UpperCamelCase ).groups()[0] if pattern.search(_UpperCamelCase ) is not None else None) for b in internal_blocks] # We only sort the lines with a key. __lowerCamelCase = [(i, key) for i, key in enumerate(_UpperCamelCase ) if key is not None] __lowerCamelCase = [x[0] for x in sorted(_UpperCamelCase ,key=lambda _UpperCamelCase : x[1] )] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. __lowerCamelCase = 0 __lowerCamelCase = [] for i in range(len(_UpperCamelCase ) ): if keys[i] is None: reorderded_blocks.append(internal_blocks[i] ) else: __lowerCamelCase = sort_objects_in_import(internal_blocks[sorted_indices[count]] ) reorderded_blocks.append(_UpperCamelCase ) count += 1 # And we put our main block back together with its first and last line. __lowerCamelCase = '''\n'''.join(block_lines[:line_idx] + reorderded_blocks + [block_lines[-1]] ) if code != "\n".join(_UpperCamelCase ): if check_only: return True else: print(F"""Overwriting {file}.""" ) with open(_UpperCamelCase ,'''w''' ,encoding='''utf-8''' ) as f: f.write('''\n'''.join(_UpperCamelCase ) ) def a__ ( _UpperCamelCase : Union[str, Any]=True ): __lowerCamelCase = [] for root, _, files in os.walk(_UpperCamelCase ): if "__init__.py" in files: __lowerCamelCase = sort_imports(os.path.join(_UpperCamelCase ,'''__init__.py''' ) ,check_only=_UpperCamelCase ) if result: __lowerCamelCase = [os.path.join(_UpperCamelCase ,'''__init__.py''' )] if len(_UpperCamelCase ) > 0: raise ValueError(F"""Would overwrite {len(_UpperCamelCase )} files, run `make style`.""" ) if __name__ == "__main__": a_ = argparse.ArgumentParser() parser.add_argument("""--check_only""", action="""store_true""", help="""Whether to only check or fix style.""") a_ = parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
700
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging a_ = logging.get_logger(__name__) a_ = {"""vocab_file""": """spiece.model"""} a_ = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", } } a_ = { """albert-base-v1""": 512, """albert-large-v1""": 512, """albert-xlarge-v1""": 512, """albert-xxlarge-v1""": 512, """albert-base-v2""": 512, """albert-large-v2""": 512, """albert-xlarge-v2""": 512, """albert-xxlarge-v2""": 512, } a_ = """▁""" class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , __UpperCAmelCase , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="<unk>" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="<pad>" , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[MASK]" , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. __lowerCamelCase = ( AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase , normalized=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token ) __lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__UpperCAmelCase , remove_space=__UpperCAmelCase , keep_accents=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) __lowerCamelCase = do_lower_case __lowerCamelCase = remove_space __lowerCamelCase = keep_accents __lowerCamelCase = vocab_file __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCAmelCase ) @property def lowerCamelCase ( self ): '''simple docstring''' return len(self.sp_model ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): '''simple docstring''' __lowerCamelCase = self.__dict__.copy() __lowerCamelCase = None return state def __setstate__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCamelCase = {} __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' if self.remove_space: __lowerCamelCase = ''' '''.join(inputs.strip().split() ) else: __lowerCamelCase = inputs __lowerCamelCase = outputs.replace('''``''' , '''"''' ).replace('''\'\'''' , '''"''' ) if not self.keep_accents: __lowerCamelCase = unicodedata.normalize('''NFKD''' , __UpperCAmelCase ) __lowerCamelCase = ''''''.join([c for c in outputs if not unicodedata.combining(__UpperCAmelCase )] ) if self.do_lower_case: __lowerCamelCase = outputs.lower() return outputs def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.preprocess_text(__UpperCAmelCase ) __lowerCamelCase = self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) __lowerCamelCase = [] for piece in pieces: if len(__UpperCAmelCase ) > 1 and piece[-1] == str(''',''' ) and piece[-2].isdigit(): __lowerCamelCase = self.sp_model.EncodeAsPieces(piece[:-1].replace(__UpperCAmelCase , '''''' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __lowerCamelCase = cur_pieces[1:] else: __lowerCamelCase = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__UpperCAmelCase ) else: new_pieces.append(__UpperCAmelCase ) return new_pieces def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.PieceToId(__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.IdToPiece(__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = '''''' __lowerCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__UpperCAmelCase ) + token __lowerCamelCase = True __lowerCamelCase = [] else: current_sub_tokens.append(__UpperCAmelCase ) __lowerCamelCase = False out_string += self.sp_model.decode(__UpperCAmelCase ) return out_string.strip() def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase ) if token_ids_a is not None: return [1] + ([0] * len(__UpperCAmelCase )) + [1] + ([0] * len(__UpperCAmelCase )) + [1] return [1] + ([0] * len(__UpperCAmelCase )) + [1] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' if not os.path.isdir(__UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCamelCase = os.path.join( __UpperCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , '''wb''' ) as fi: __lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,)
622
0
from __future__ import annotations def snake_case_ ( _UpperCamelCase : List[Any] ): if len(_lowerCAmelCase ) == 0: return [] __lowerCamelCase ,__lowerCamelCase = min(_lowerCAmelCase ), max(_lowerCAmelCase ) __lowerCamelCase = int(max_value - min_value ) + 1 __lowerCamelCase = [[] for _ in range(_lowerCAmelCase )] for i in my_list: buckets[int(i - min_value )].append(_lowerCAmelCase ) return [v for bucket in buckets for v in sorted(_lowerCAmelCase )] if __name__ == "__main__": from doctest import testmod testmod() assert bucket_sort([4, 5, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bucket_sort([0, 1, -10, 15, 2, -2]) == [-10, -2, 0, 1, 2, 15]
701
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed a_ = """true""" def a__ ( _UpperCamelCase : Union[str, Any] ,_UpperCamelCase : List[str]=82 ,_UpperCamelCase : Optional[Any]=16 ): set_seed(42 ) __lowerCamelCase = RegressionModel() __lowerCamelCase = deepcopy(_UpperCamelCase ) __lowerCamelCase = RegressionDataset(length=_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,batch_size=_UpperCamelCase ) model.to(accelerator.device ) __lowerCamelCase ,__lowerCamelCase = accelerator.prepare(_UpperCamelCase ,_UpperCamelCase ) return model, ddp_model, dataloader def a__ ( _UpperCamelCase : Accelerator ,_UpperCamelCase : str=False ): __lowerCamelCase = AutoTokenizer.from_pretrained('''hf-internal-testing/mrpc-bert-base-cased''' ) __lowerCamelCase = load_dataset('''glue''' ,'''mrpc''' ,split='''validation''' ) def tokenize_function(_UpperCamelCase : int ): __lowerCamelCase = tokenizer(examples['''sentence1'''] ,examples['''sentence2'''] ,truncation=_UpperCamelCase ,max_length=_UpperCamelCase ) return outputs with accelerator.main_process_first(): __lowerCamelCase = dataset.map( _UpperCamelCase ,batched=_UpperCamelCase ,remove_columns=['''idx''', '''sentence1''', '''sentence2'''] ,) __lowerCamelCase = tokenized_datasets.rename_column('''label''' ,'''labels''' ) def collate_fn(_UpperCamelCase : Any ): if use_longest: return tokenizer.pad(_UpperCamelCase ,padding='''longest''' ,return_tensors='''pt''' ) return tokenizer.pad(_UpperCamelCase ,padding='''max_length''' ,max_length=1_28 ,return_tensors='''pt''' ) return DataLoader(_UpperCamelCase ,shuffle=_UpperCamelCase ,collate_fn=_UpperCamelCase ,batch_size=16 ) def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : List[str] ): __lowerCamelCase = Accelerator(dispatch_batches=_UpperCamelCase ,split_batches=_UpperCamelCase ) __lowerCamelCase = get_dataloader(_UpperCamelCase ,not dispatch_batches ) __lowerCamelCase = AutoModelForSequenceClassification.from_pretrained( '''hf-internal-testing/mrpc-bert-base-cased''' ,return_dict=_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = accelerator.prepare(_UpperCamelCase ,_UpperCamelCase ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : Optional[Any] ,_UpperCamelCase : Union[str, Any] ): __lowerCamelCase = [] for batch in dataloader: __lowerCamelCase ,__lowerCamelCase = batch.values() with torch.no_grad(): __lowerCamelCase = model(_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) __lowerCamelCase ,__lowerCamelCase = [], [] for logit, targ in logits_and_targets: logits.append(_UpperCamelCase ) targs.append(_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = torch.cat(_UpperCamelCase ), torch.cat(_UpperCamelCase ) return logits, targs def a__ ( _UpperCamelCase : Accelerator ,_UpperCamelCase : List[Any]=82 ,_UpperCamelCase : str=False ,_UpperCamelCase : List[str]=False ,_UpperCamelCase : Optional[int]=16 ): __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = get_basic_setup(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = generate_predictions(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ) assert ( len(_UpperCamelCase ) == num_samples ), F"""Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(_UpperCamelCase )}""" def a__ ( _UpperCamelCase : bool = False ,_UpperCamelCase : bool = False ): __lowerCamelCase = evaluate.load('''glue''' ,'''mrpc''' ) __lowerCamelCase ,__lowerCamelCase = get_mrpc_setup(_UpperCamelCase ,_UpperCamelCase ) # First do baseline __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = setup['''no'''] model.to(_UpperCamelCase ) model.eval() for batch in dataloader: batch.to(_UpperCamelCase ) with torch.inference_mode(): __lowerCamelCase = model(**_UpperCamelCase ) __lowerCamelCase = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=_UpperCamelCase ,references=batch['''labels'''] ) __lowerCamelCase = metric.compute() # Then do distributed __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = setup['''ddp'''] model.eval() for batch in dataloader: with torch.inference_mode(): __lowerCamelCase = model(**_UpperCamelCase ) __lowerCamelCase = outputs.logits.argmax(dim=-1 ) __lowerCamelCase = batch['''labels'''] __lowerCamelCase ,__lowerCamelCase = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=_UpperCamelCase ,references=_UpperCamelCase ) __lowerCamelCase = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] ,distributed[key] ), F"""Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n""" def a__ ( ): __lowerCamelCase = Accelerator(split_batches=_UpperCamelCase ,dispatch_batches=_UpperCamelCase ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print('''**Testing gather_for_metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(F"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`""" ) test_mrpc(_UpperCamelCase ,_UpperCamelCase ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test torch metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: __lowerCamelCase = Accelerator(split_batches=_UpperCamelCase ,dispatch_batches=_UpperCamelCase ) if accelerator.is_local_main_process: print(F"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99""" ) test_torch_metrics(_UpperCamelCase ,99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test last batch is not dropped when perfectly divisible**''' ) __lowerCamelCase = Accelerator() test_torch_metrics(_UpperCamelCase ,5_12 ) accelerator.state._reset_state() def a__ ( _UpperCamelCase : Optional[int] ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
622
0
import doctest from collections import deque import numpy as np class __lowerCAmelCase : def __init__( self ): '''simple docstring''' __lowerCamelCase = [2, 1, 2, -1] __lowerCamelCase = [1, 2, 3, 4] def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = len(self.first_signal ) __lowerCamelCase = len(self.second_signal ) __lowerCamelCase = max(__lowerCamelCase , __lowerCamelCase ) # create a zero matrix of max_length x max_length __lowerCamelCase = [[0] * max_length for i in range(__lowerCamelCase )] # fills the smaller signal with zeros to make both signals of same length if length_first_signal < length_second_signal: self.first_signal += [0] * (max_length - length_first_signal) elif length_first_signal > length_second_signal: self.second_signal += [0] * (max_length - length_second_signal) for i in range(__lowerCamelCase ): __lowerCamelCase = deque(self.second_signal ) rotated_signal.rotate(__lowerCamelCase ) for j, item in enumerate(__lowerCamelCase ): matrix[i][j] += item # multiply the matrix with the first signal __lowerCamelCase = np.matmul(np.transpose(__lowerCamelCase ) , np.transpose(self.first_signal ) ) # rounding-off to two decimal places return [round(__lowerCamelCase , 2 ) for i in final_signal] if __name__ == "__main__": doctest.testmod()
702
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = StableDiffusionXLImgaImgPipeline lowerCAmelCase__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} lowerCAmelCase__ = PipelineTesterMixin.required_optional_params - {"""latents"""} lowerCAmelCase__ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowerCAmelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS lowerCAmelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowerCamelCase ( self ): '''simple docstring''' torch.manual_seed(0 ) __lowerCamelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=__UpperCAmelCase , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=80 , cross_attention_dim=64 , ) __lowerCamelCase = EulerDiscreteScheduler( beta_start=0.00_085 , beta_end=0.012 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) __lowerCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='''gelu''' , projection_dim=32 , ) __lowerCamelCase = CLIPTextModel(__UpperCAmelCase ) __lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__UpperCAmelCase ) __lowerCamelCase = CLIPTextModelWithProjection(__UpperCAmelCase ) __lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=__UpperCAmelCase ) __lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase=0 ): '''simple docstring''' __lowerCamelCase = floats_tensor((1, 3, 32, 32) , rng=random.Random(__UpperCAmelCase ) ).to(__UpperCAmelCase ) __lowerCamelCase = image / 2 + 0.5 if str(__UpperCAmelCase ).startswith('''mps''' ): __lowerCamelCase = torch.manual_seed(__UpperCAmelCase ) else: __lowerCamelCase = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) __lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.75, } return inputs def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = StableDiffusionXLImgaImgPipeline(**__UpperCAmelCase ) __lowerCamelCase = sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase = sd_pipe(**__UpperCAmelCase ).images __lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCamelCase = np.array([0.4_656, 0.4_840, 0.4_439, 0.6_698, 0.5_574, 0.4_524, 0.5_799, 0.5_943, 0.5_165] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase ( self ): '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def lowerCamelCase ( self ): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = StableDiffusionXLImgaImgPipeline(**__UpperCAmelCase ) __lowerCamelCase = sd_pipe.to(__UpperCAmelCase ) __lowerCamelCase = sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) # forward without prompt embeds __lowerCamelCase = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase = 3 * ['''this is a negative prompt'''] __lowerCamelCase = negative_prompt __lowerCamelCase = 3 * [inputs['''prompt''']] __lowerCamelCase = sd_pipe(**__UpperCAmelCase ) __lowerCamelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds __lowerCamelCase = self.get_dummy_inputs(__UpperCAmelCase ) __lowerCamelCase = 3 * ['''this is a negative prompt'''] __lowerCamelCase = 3 * [inputs.pop('''prompt''' )] ( ( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) , ) = sd_pipe.encode_prompt(__UpperCAmelCase , negative_prompt=__UpperCAmelCase ) __lowerCamelCase = sd_pipe( **__UpperCAmelCase , prompt_embeds=__UpperCAmelCase , negative_prompt_embeds=__UpperCAmelCase , pooled_prompt_embeds=__UpperCAmelCase , negative_pooled_prompt_embeds=__UpperCAmelCase , ) __lowerCamelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase="cpu" , __UpperCAmelCase=torch.floataa , __UpperCAmelCase=0 ): '''simple docstring''' __lowerCamelCase = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) __lowerCamelCase = np.random.RandomState(__UpperCAmelCase ).standard_normal((1, 4, 64, 64) ) __lowerCamelCase = torch.from_numpy(__UpperCAmelCase ).to(device=__UpperCAmelCase , dtype=__UpperCAmelCase ) __lowerCamelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = self.get_inputs(__UpperCAmelCase ) __lowerCamelCase = pipe(**__UpperCAmelCase ).images __lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.49_493, 0.47_896, 0.40_798, 0.54_214, 0.53_212, 0.48_202, 0.47_656, 0.46_329, 0.48_506] ) assert np.abs(image_slice - expected_slice ).max() < 7E-3
622
0
import os import zipfile import pytest from datasets.utils.extract import ( BzipaExtractor, Extractor, GzipExtractor, LzaExtractor, SevenZipExtractor, TarExtractor, XzExtractor, ZipExtractor, ZstdExtractor, ) from .utils import require_lza, require_pyazr, require_zstandard @pytest.mark.parametrize( '''compression_format, is_archive''' ,[ ('''7z''', True), ('''bz2''', False), ('''gzip''', False), ('''lz4''', False), ('''tar''', True), ('''xz''', False), ('''zip''', True), ('''zstd''', False), ] ,) def a__ ( _UpperCamelCase : str ,_UpperCamelCase : Any ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : Any ,_UpperCamelCase : Dict ,_UpperCamelCase : List[Any] ,_UpperCamelCase : str ,_UpperCamelCase : Dict ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : Optional[Any] ,_UpperCamelCase : Optional[Any] ,_UpperCamelCase : Tuple ,): __lowerCamelCase = { '''7z''': (seven_zip_file, SevenZipExtractor), '''bz2''': (bza_file, BzipaExtractor), '''gzip''': (gz_file, GzipExtractor), '''lz4''': (lza_file, LzaExtractor), '''tar''': (tar_file, TarExtractor), '''xz''': (xz_file, XzExtractor), '''zip''': (zip_file, ZipExtractor), '''zstd''': (zstd_file, ZstdExtractor), } __lowerCamelCase ,__lowerCamelCase = input_paths_and_base_extractors[compression_format] if input_path is None: __lowerCamelCase = F"""for \'{compression_format}\' compression_format, """ if compression_format == "7z": reason += require_pyazr.kwargs["reason"] elif compression_format == "lz4": reason += require_lza.kwargs["reason"] elif compression_format == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(UpperCamelCase__ ) assert base_extractor.is_extractable(UpperCamelCase__ ) __lowerCamelCase = tmp_path / ('''extracted''' if is_archive else '''extracted.txt''') base_extractor.extract(UpperCamelCase__ ,UpperCamelCase__ ) if is_archive: assert output_path.is_dir() for file_path in output_path.iterdir(): assert file_path.name == text_file.name __lowerCamelCase = file_path.read_text(encoding='''utf-8''' ) else: __lowerCamelCase = output_path.read_text(encoding='''utf-8''' ) __lowerCamelCase = text_file.read_text(encoding='''utf-8''' ) assert extracted_file_content == expected_file_content @pytest.mark.parametrize( '''compression_format, is_archive''' ,[ ('''7z''', True), ('''bz2''', False), ('''gzip''', False), ('''lz4''', False), ('''tar''', True), ('''xz''', False), ('''zip''', True), ('''zstd''', False), ] ,) def a__ ( _UpperCamelCase : int ,_UpperCamelCase : List[str] ,_UpperCamelCase : List[Any] ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : str ,_UpperCamelCase : Optional[Any] ,_UpperCamelCase : List[str] ,_UpperCamelCase : Dict ,_UpperCamelCase : Tuple ,_UpperCamelCase : Tuple ,_UpperCamelCase : int ,_UpperCamelCase : List[str] ,): __lowerCamelCase = { '''7z''': seven_zip_file, '''bz2''': bza_file, '''gzip''': gz_file, '''lz4''': lza_file, '''tar''': tar_file, '''xz''': xz_file, '''zip''': zip_file, '''zstd''': zstd_file, } __lowerCamelCase = input_paths[compression_format] if input_path is None: __lowerCamelCase = F"""for \'{compression_format}\' compression_format, """ if compression_format == "7z": reason += require_pyazr.kwargs["reason"] elif compression_format == "lz4": reason += require_lza.kwargs["reason"] elif compression_format == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(UpperCamelCase__ ) __lowerCamelCase = Extractor.infer_extractor_format(UpperCamelCase__ ) assert extractor_format is not None __lowerCamelCase = tmp_path / ('''extracted''' if is_archive else '''extracted.txt''') Extractor.extract(UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ ) if is_archive: assert output_path.is_dir() for file_path in output_path.iterdir(): assert file_path.name == text_file.name __lowerCamelCase = file_path.read_text(encoding='''utf-8''' ) else: __lowerCamelCase = output_path.read_text(encoding='''utf-8''' ) __lowerCamelCase = text_file.read_text(encoding='''utf-8''' ) assert extracted_file_content == expected_file_content @pytest.fixture def a__ ( _UpperCamelCase : List[Any] ,_UpperCamelCase : List[Any] ): import tarfile __lowerCamelCase = tmp_path / '''data_dot_dot''' directory.mkdir() __lowerCamelCase = directory / '''tar_file_with_dot_dot.tar''' with tarfile.TarFile(UpperCamelCase__ ,'''w''' ) as f: f.add(UpperCamelCase__ ,arcname=os.path.join('''..''' ,text_file.name ) ) return path @pytest.fixture def a__ ( _UpperCamelCase : Tuple ): import tarfile __lowerCamelCase = tmp_path / '''data_sym_link''' directory.mkdir() __lowerCamelCase = directory / '''tar_file_with_sym_link.tar''' os.symlink('''..''' ,directory / '''subdir''' ,target_is_directory=UpperCamelCase__ ) with tarfile.TarFile(UpperCamelCase__ ,'''w''' ) as f: f.add(str(directory / '''subdir''' ) ,arcname='''subdir''' ) # str required by os.readlink on Windows and Python < 3.8 return path @pytest.mark.parametrize( '''insecure_tar_file, error_log''' ,[('''tar_file_with_dot_dot''', '''illegal path'''), ('''tar_file_with_sym_link''', '''Symlink''')] ,) def a__ ( _UpperCamelCase : List[str] ,_UpperCamelCase : Tuple ,_UpperCamelCase : List[Any] ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : Any ,_UpperCamelCase : str ): __lowerCamelCase = { '''tar_file_with_dot_dot''': tar_file_with_dot_dot, '''tar_file_with_sym_link''': tar_file_with_sym_link, } __lowerCamelCase = insecure_tar_files[insecure_tar_file] __lowerCamelCase = tmp_path / '''extracted''' TarExtractor.extract(UpperCamelCase__ ,UpperCamelCase__ ) assert caplog.text for record in caplog.records: assert record.levelname == "ERROR" assert error_log in record.msg def a__ ( _UpperCamelCase : Optional[Any] ): __lowerCamelCase = tmpdir / '''not_a_zip_file''' # From: https://github.com/python/cpython/pull/5053 __lowerCamelCase = ( b'''\x89PNG\r\n\x1a\n\x00\x00\x00\rIHDR\x00\x00\x00\x01\x00\x00''' b'''\x00\x02\x08\x06\x00\x00\x00\x99\x81\xb6\'\x00\x00\x00\x15I''' b'''DATx\x01\x01\n\x00\xf5\xff\x00PK\x05\x06\x00PK\x06\x06\x07''' b'''\xac\x01N\xc6|a\r\x00\x00\x00\x00IEND\xaeB`\x82''' ) with not_a_zip_file.open('''wb''' ) as f: f.write(UpperCamelCase__ ) assert zipfile.is_zipfile(str(UpperCamelCase__ ) ) # is a false positive for `zipfile` assert not ZipExtractor.is_extractable(UpperCamelCase__ ) # but we're right
703
import torch from diffusers import StableDiffusionPipeline a_ = """path-to-your-trained-model""" a_ = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to("""cuda""") a_ = """A photo of sks dog in a bucket""" a_ = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save("""dog-bucket.png""")
622
0
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """google/fnet-base""": """https://huggingface.co/google/fnet-base/resolve/main/config.json""", """google/fnet-large""": """https://huggingface.co/google/fnet-large/resolve/main/config.json""" # See all FNet models at https://huggingface.co/models?filter=fnet } class __lowerCAmelCase ( snake_case__ ): lowerCAmelCase__ = """fnet""" def __init__( self , __UpperCAmelCase=32000 , __UpperCAmelCase=768 , __UpperCAmelCase=12 , __UpperCAmelCase=3072 , __UpperCAmelCase="gelu_new" , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=4 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=False , __UpperCAmelCase=512 , __UpperCAmelCase=3 , __UpperCAmelCase=1 , __UpperCAmelCase=2 , **__UpperCAmelCase , ): '''simple docstring''' super().__init__(pad_token_id=lowercase_ , bos_token_id=lowercase_ , eos_token_id=lowercase_ , **lowercase_ ) __lowerCamelCase = vocab_size __lowerCamelCase = max_position_embeddings __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = initializer_range __lowerCamelCase = type_vocab_size __lowerCamelCase = layer_norm_eps __lowerCamelCase = use_tpu_fourier_optimizations __lowerCamelCase = tpu_short_seq_length
704
import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class __lowerCAmelCase : @staticmethod def lowerCamelCase ( *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' pass def a__ ( _UpperCamelCase : List[str] ): return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. a_ = ( """https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png""" ) @is_pipeline_test @require_torch @require_vision class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model=__UpperCAmelCase , tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '''''' ) ) ) __lowerCamelCase = '''What is the placebo?''' __lowerCamelCase = [ { '''image''': load_image(__UpperCAmelCase ), '''question''': question, }, { '''image''': image, '''question''': question, }, { '''image''': image, '''question''': question, '''word_boxes''': word_boxes, }, ] return dqa_pipeline, examples def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = dqa_pipeline(__UpperCAmelCase , top_k=2 ) self.assertEqual( __UpperCAmelCase , [ [ {'''score''': ANY(__UpperCAmelCase ), '''answer''': ANY(__UpperCAmelCase ), '''start''': ANY(__UpperCAmelCase ), '''end''': ANY(__UpperCAmelCase )}, {'''score''': ANY(__UpperCAmelCase ), '''answer''': ANY(__UpperCAmelCase ), '''start''': ANY(__UpperCAmelCase ), '''end''': ANY(__UpperCAmelCase )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline('''document-question-answering''' , model='''hf-internal-testing/tiny-random-layoutlmv2''' ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''How many cats are there?''' __lowerCamelCase = [ {'''score''': 0.0_001, '''answer''': '''oy 2312/2019''', '''start''': 38, '''end''': 39}, {'''score''': 0.0_001, '''answer''': '''oy 2312/2019 DUE''', '''start''': 38, '''end''': 40}, ] __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , __UpperCAmelCase ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , __UpperCAmelCase ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably __lowerCamelCase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(__UpperCAmelCase , [] ) # We can optionnally pass directly the words and bounding boxes __lowerCamelCase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' __lowerCamelCase = [] __lowerCamelCase = [] __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , words=__UpperCAmelCase , boxes=__UpperCAmelCase , top_k=2 ) self.assertEqual(__UpperCAmelCase , [] ) @slow @require_torch @require_detectrona @require_pytesseract def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_944, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_009, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_944, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_009, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.9_944, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_009, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , max_seq_len=50 , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_974, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_948, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_974, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_948, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.9_974, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_948, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__UpperCAmelCase ) __lowerCamelCase = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__UpperCAmelCase , revision='''3dc6de3''' , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) __lowerCamelCase = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] ] * 2 , ) __lowerCamelCase = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '''''' ) ) ) # This model should also work if `image` is set to None __lowerCamelCase = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.4_251, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_819, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__UpperCAmelCase ) __lowerCamelCase = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__UpperCAmelCase , revision='''3dc6de3''' , max_seq_len=50 , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_999, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_998, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) __lowerCamelCase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ [ {'''score''': 0.9_999, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_998, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] ] * 2 , ) __lowerCamelCase = list(zip(*apply_tesseract(load_image(__UpperCAmelCase ) , __UpperCAmelCase , '''''' ) ) ) # This model should also work if `image` is set to None __lowerCamelCase = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__UpperCAmelCase , decimals=4 ) , [ {'''score''': 0.9_999, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_998, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) @slow @require_torch def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = pipeline( '''document-question-answering''' , model='''naver-clova-ix/donut-base-finetuned-docvqa''' , tokenizer=AutoTokenizer.from_pretrained('''naver-clova-ix/donut-base-finetuned-docvqa''' ) , feature_extractor='''naver-clova-ix/donut-base-finetuned-docvqa''' , ) __lowerCamelCase = INVOICE_URL __lowerCamelCase = '''What is the invoice number?''' __lowerCamelCase = dqa_pipeline(image=__UpperCAmelCase , question=__UpperCAmelCase , top_k=2 ) self.assertEqual(nested_simplify(__UpperCAmelCase , decimals=4 ) , [{'''answer''': '''us-001'''}] ) @require_tf @unittest.skip('''Document question answering not implemented in TF''' ) def lowerCamelCase ( self ): '''simple docstring''' pass
622
0
import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() a_ = logging.get_logger(__name__) a_ = "https://openaipublic.azureedge.net/jukebox/models/" a_ = { "jukebox-1b-lyrics": [ "5b/vqvae.pth.tar", "5b/prior_level_0.pth.tar", "5b/prior_level_1.pth.tar", "1b_lyrics/prior_level_2.pth.tar", ], "jukebox-5b-lyrics": [ "5b/vqvae.pth.tar", "5b/prior_level_0.pth.tar", "5b/prior_level_1.pth.tar", "5b_lyrics/prior_level_2.pth.tar", ], } def a__ ( _UpperCamelCase : Tuple ): if key.endswith('''.model.1.bias''' ) and len(key.split('''.''' ) ) > 10: __lowerCamelCase = key.replace('''.model.1.bias''' ,'''.conv1d_1.bias''' ) elif key.endswith('''.model.1.weight''' ) and len(key.split('''.''' ) ) > 10: __lowerCamelCase = key.replace('''.model.1.weight''' ,'''.conv1d_1.weight''' ) elif key.endswith('''.model.3.bias''' ) and len(key.split('''.''' ) ) > 10: __lowerCamelCase = key.replace('''.model.3.bias''' ,'''.conv1d_2.bias''' ) elif key.endswith('''.model.3.weight''' ) and len(key.split('''.''' ) ) > 10: __lowerCamelCase = key.replace('''.model.3.weight''' ,'''.conv1d_2.weight''' ) if "conditioner_blocks.0." in key: __lowerCamelCase = key.replace('''conditioner_blocks.0''' ,'''conditioner_blocks''' ) if "prime_prior" in key: __lowerCamelCase = key.replace('''prime_prior''' ,'''encoder''' ) if ".emb." in key and "total" not in key and "absolute" not in key and "relative" not in key: __lowerCamelCase = key.replace('''.emb.''' ,'''.''' ) if key.endswith('''k''' ): # replace vqvae.X.k with vqvae.X.codebook return key.replace('''.k''' ,'''.codebook''' ) if "y_emb." in key: return key.replace('''y_emb.''' ,'''metadata_embedding.''' ) if "x_emb.emb." in key: __lowerCamelCase = key.replace('''0.x_emb.emb''' ,'''embed_tokens''' ) if "prime_state_ln" in key: return key.replace('''prime_state_ln''' ,'''encoder.final_layer_norm''' ) if ".ln" in key: return key.replace('''.ln''' ,'''.layer_norm''' ) if "_ln" in key: return key.replace('''_ln''' ,'''_layer_norm''' ) if "prime_state_proj" in key: return key.replace('''prime_state_proj''' ,'''encoder.proj_in''' ) if "prime_x_out" in key: return key.replace('''prime_x_out''' ,'''encoder.lm_head''' ) if "prior.x_out" in key: return key.replace('''x_out''' ,'''fc_proj_out''' ) if "x_emb" in key: return key.replace('''x_emb''' ,'''embed_tokens''' ) return key def a__ ( _UpperCamelCase : str ,_UpperCamelCase : Tuple ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : List[Any] ): __lowerCamelCase = {} import re __lowerCamelCase = re.compile(R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)''' ) __lowerCamelCase = re.compile( R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) __lowerCamelCase = re.compile(R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)''' ) __lowerCamelCase = re.compile(R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)''' ) __lowerCamelCase = re.compile( R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) __lowerCamelCase = re.compile(R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)''' ) __lowerCamelCase = re.compile(R'''conditioner_blocks.(\d*).cond.model.(\d*).(\d).(bias|weight)''' ) __lowerCamelCase = re.compile( R'''conditioner_blocks.(\d*).cond.model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) __lowerCamelCase = re.compile(R'''conditioner_blocks.(\d*).cond.model.(\d*).(bias|weight)''' ) for original_key, value in state_dict.items(): # rename vqvae.encoder keys if re_encoder_block_conv_in.fullmatch(_UpperCamelCase ): __lowerCamelCase = re_encoder_block_conv_in.match(_UpperCamelCase ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = int(groups[2] ) * 2 + int(groups[3] ) __lowerCamelCase = F"""encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.{groups[-1]}""" __lowerCamelCase = re_encoder_block_conv_in.sub(_UpperCamelCase ,_UpperCamelCase ) elif re_encoder_block_resnet.fullmatch(_UpperCamelCase ): __lowerCamelCase = re_encoder_block_resnet.match(_UpperCamelCase ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = int(groups[2] ) * 2 + int(groups[3] ) __lowerCamelCase = {'''1''': 1, '''3''': 2}[groups[-2]] __lowerCamelCase = F"""encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.""" __lowerCamelCase = F"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" __lowerCamelCase = prefix + resnet_block __lowerCamelCase = re_encoder_block_resnet.sub(_UpperCamelCase ,_UpperCamelCase ) elif re_encoder_block_proj_out.fullmatch(_UpperCamelCase ): __lowerCamelCase = re_encoder_block_proj_out.match(_UpperCamelCase ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = F"""encoders.{groups[0]}.level_blocks.{groups[1]}.proj_out.{groups[-1]}""" __lowerCamelCase = re_encoder_block_proj_out.sub(_UpperCamelCase ,_UpperCamelCase ) # rename vqvae.decoder keys elif re_decoder_block_conv_out.fullmatch(_UpperCamelCase ): __lowerCamelCase = re_decoder_block_conv_out.match(_UpperCamelCase ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = int(groups[2] ) * 2 + int(groups[3] ) - 2 __lowerCamelCase = F"""decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.{groups[-1]}""" __lowerCamelCase = re_decoder_block_conv_out.sub(_UpperCamelCase ,_UpperCamelCase ) elif re_decoder_block_resnet.fullmatch(_UpperCamelCase ): __lowerCamelCase = re_decoder_block_resnet.match(_UpperCamelCase ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = int(groups[2] ) * 2 + int(groups[3] ) - 2 __lowerCamelCase = {'''1''': 1, '''3''': 2}[groups[-2]] __lowerCamelCase = F"""decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.""" __lowerCamelCase = F"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" __lowerCamelCase = prefix + resnet_block __lowerCamelCase = re_decoder_block_resnet.sub(_UpperCamelCase ,_UpperCamelCase ) elif re_decoder_block_proj_in.fullmatch(_UpperCamelCase ): __lowerCamelCase = re_decoder_block_proj_in.match(_UpperCamelCase ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = F"""decoders.{groups[0]}.level_blocks.{groups[1]}.proj_in.{groups[-1]}""" __lowerCamelCase = re_decoder_block_proj_in.sub(_UpperCamelCase ,_UpperCamelCase ) # rename prior cond.model to upsampler.upsample_block and resnet elif re_prior_cond_conv_out.fullmatch(_UpperCamelCase ): __lowerCamelCase = re_prior_cond_conv_out.match(_UpperCamelCase ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = int(groups[1] ) * 2 + int(groups[2] ) - 2 __lowerCamelCase = F"""conditioner_blocks.upsampler.upsample_block.{block_index}.{groups[-1]}""" __lowerCamelCase = re_prior_cond_conv_out.sub(_UpperCamelCase ,_UpperCamelCase ) elif re_prior_cond_resnet.fullmatch(_UpperCamelCase ): __lowerCamelCase = re_prior_cond_resnet.match(_UpperCamelCase ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = int(groups[1] ) * 2 + int(groups[2] ) - 2 __lowerCamelCase = {'''1''': 1, '''3''': 2}[groups[-2]] __lowerCamelCase = F"""conditioner_blocks.upsampler.upsample_block.{block_index}.""" __lowerCamelCase = F"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" __lowerCamelCase = prefix + resnet_block __lowerCamelCase = re_prior_cond_resnet.sub(_UpperCamelCase ,_UpperCamelCase ) elif re_prior_cond_proj_in.fullmatch(_UpperCamelCase ): __lowerCamelCase = re_prior_cond_proj_in.match(_UpperCamelCase ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = F"""conditioner_blocks.upsampler.proj_in.{groups[-1]}""" __lowerCamelCase = re_prior_cond_proj_in.sub(_UpperCamelCase ,_UpperCamelCase ) # keep original key else: __lowerCamelCase = original_key __lowerCamelCase = replace_key(_UpperCamelCase ) if F"""{key_prefix}.{key}""" not in model_state_dict or key is None: print(F"""failed converting {original_key} to {key}, does not match""" ) # handle missmatched shape elif value.shape != model_state_dict[F"""{key_prefix}.{key}"""].shape: __lowerCamelCase = model_state_dict[F"""{key_prefix}.{key}"""] print(F"""{original_key}-> {key} : \nshape {val.shape} and { value.shape}, do not match""" ) __lowerCamelCase = original_key __lowerCamelCase = original_key __lowerCamelCase = value return new_dict @torch.no_grad() def a__ ( _UpperCamelCase : List[Any]=None ,_UpperCamelCase : int=None ): for file in MODEL_MAPPING[model_name]: if not os.path.isfile(F"""{pytorch_dump_folder_path}/{file.split("/" )[-1]}""" ): __lowerCamelCase = requests.get(F"""{PREFIX}{file}""" ,allow_redirects=_UpperCamelCase ) os.makedirs(F"""{pytorch_dump_folder_path}/""" ,exist_ok=_UpperCamelCase ) open(F"""{pytorch_dump_folder_path}/{file.split("/" )[-1]}""" ,'''wb''' ).write(r.content ) __lowerCamelCase = MODEL_MAPPING[model_name.split('''/''' )[-1]] __lowerCamelCase = JukeboxConfig.from_pretrained(_UpperCamelCase ) __lowerCamelCase = JukeboxModel(_UpperCamelCase ) __lowerCamelCase = [] __lowerCamelCase = {} for i, dict_name in enumerate(_UpperCamelCase ): __lowerCamelCase = torch.load(F"""{pytorch_dump_folder_path}/{dict_name.split("/" )[-1]}""" )['''model'''] __lowerCamelCase = {} for k in old_dic.keys(): if k.endswith('''.b''' ): __lowerCamelCase = old_dic[k] elif k.endswith('''.w''' ): __lowerCamelCase = old_dic[k] elif "level_2" not in dict_name and "cond.model." in k: __lowerCamelCase = old_dic[k] else: __lowerCamelCase = old_dic[k] __lowerCamelCase = '''vqvae''' if i == 0 else F"""priors.{3 - i}""" __lowerCamelCase = fix_jukebox_keys(_UpperCamelCase ,model.state_dict() ,_UpperCamelCase ,_UpperCamelCase ) weight_dict.append(_UpperCamelCase ) __lowerCamelCase = weight_dict.pop(0 ) model.vqvae.load_state_dict(_UpperCamelCase ) for i in range(len(_UpperCamelCase ) ): model.priors[i].load_state_dict(weight_dict[2 - i] ) Path(_UpperCamelCase ).mkdir(exist_ok=_UpperCamelCase ) with open(F"""{pytorch_dump_folder_path}/mapping.json""" ,'''w''' ) as txtfile: json.dump(_UpperCamelCase ,_UpperCamelCase ) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCamelCase ) return weight_dict if __name__ == "__main__": a_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""jukebox-5b-lyrics""", type=str, help="""Name of the model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default="""jukebox-5b-lyrics-converted""", type=str, help="""Path to the output PyTorch model directory.""", ) a_ = parser.parse_args() convert_openai_checkpoint(args.model_name, args.pytorch_dump_folder_path)
705
import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin a_ = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = XLMProphetNetTokenizer lowerCAmelCase__ = False lowerCAmelCase__ = True def lowerCamelCase ( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __lowerCamelCase = XLMProphetNetTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''[PAD]''' __lowerCamelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCAmelCase ) , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''[PAD]''' ) self.assertEqual(vocab_keys[1] , '''[CLS]''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(__UpperCAmelCase ) , 1012 ) def lowerCamelCase ( self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1012 ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = XLMProphetNetTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) __lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__UpperCAmelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __lowerCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] , ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''[UNK]''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''[UNK]''', '''.''', ] , ) @cached_property def lowerCamelCase ( self ): '''simple docstring''' return XLMProphetNetTokenizer.from_pretrained('''microsoft/xprophetnet-large-wiki100-cased''' ) @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''Hello World!''' __lowerCamelCase = [35389, 6672, 49, 2] self.assertListEqual(__UpperCAmelCase , self.big_tokenizer.encode(__UpperCAmelCase ) ) @slow def lowerCamelCase ( self ): '''simple docstring''' # fmt: off __lowerCamelCase = {'''input_ids''': [[11073, 82783, 18, 26, 82783, 549, 51540, 248, 17209, 1301, 217, 20, 215186, 1325, 147, 17209, 1301, 217, 20, 56370, 53, 122020, 20, 16477, 27, 87355, 4548, 20, 4728, 78392, 17, 159969, 18, 26, 24491, 629, 15, 538, 22704, 5439, 15, 2788, 24491, 9885, 15, 43534, 605, 15, 814, 18403, 33200, 29, 15, 43534, 24458, 12410, 111, 24966, 83669, 9637, 144068, 26, 850, 22346, 27, 147, 24966, 83669, 83490, 26, 39113, 735, 27, 689, 656, 2800, 1339, 4600, 53, 122020, 115785, 34, 816, 1339, 46887, 18, 147, 53905, 1951, 42238, 41170, 17732, 834, 436, 15, 27523, 98733, 217, 147, 5542, 4981, 930, 17347, 16, 2], [20091, 629, 94, 82786, 58, 490, 20, 1528, 84, 53905, 344, 80592, 110128, 18822, 5267, 1306, 62, 152537, 308, 7997, 401, 124427, 549, 35442, 225, 109, 15055, 25748, 147, 7119, 43712, 34, 767, 135366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 63784, 119466, 17, 147808, 88214, 18, 656, 81, 32, 3296, 10280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__UpperCAmelCase , model_name='''microsoft/xprophetnet-large-wiki100-cased''' , revision='''1acad1643ddd54a44df6a1b797ada8373685d90e''' , )
622
0
def a__ ( _UpperCamelCase : int ,_UpperCamelCase : int ): if not isinstance(_lowerCamelCase ,_lowerCamelCase ): raise ValueError('''iterations must be defined as integers''' ) if not isinstance(_lowerCamelCase ,_lowerCamelCase ) or not number >= 1: raise ValueError( '''starting number must be\n and integer and be more than 0''' ) if not iterations >= 1: raise ValueError('''Iterations must be done more than 0 times to play FizzBuzz''' ) __lowerCamelCase = "" while number <= iterations: if number % 3 == 0: out += "Fizz" if number % 5 == 0: out += "Buzz" if 0 not in (number % 3, number % 5): out += str(_lowerCamelCase ) # print(out) number += 1 out += " " return out if __name__ == "__main__": import doctest doctest.testmod()
706
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py a_ = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. a_ = direct_transformers_import(PATH_TO_TRANSFORMERS) a_ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` a_ = re.compile(R"""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") a_ = { """DecisionTransformerConfig""", """EncoderDecoderConfig""", """MusicgenConfig""", """RagConfig""", """SpeechEncoderDecoderConfig""", """TimmBackboneConfig""", """VisionEncoderDecoderConfig""", """VisionTextDualEncoderConfig""", """LlamaConfig""", } def a__ ( _UpperCamelCase : Union[str, Any] ): __lowerCamelCase = None # source code of `config_class` __lowerCamelCase = inspect.getsource(_UpperCamelCase ) __lowerCamelCase = _re_checkpoint.findall(_UpperCamelCase ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('''/''' ): __lowerCamelCase = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link __lowerCamelCase = F"""https://huggingface.co/{ckpt_name}""" if ckpt_link == ckpt_link_from_name: __lowerCamelCase = ckpt_name break return checkpoint def a__ ( ): __lowerCamelCase = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue __lowerCamelCase = get_checkpoint_from_config_class(_UpperCamelCase ) __lowerCamelCase = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(_UpperCamelCase ) if len(_UpperCamelCase ) > 0: __lowerCamelCase = '''\n'''.join(sorted(_UpperCamelCase ) ) raise ValueError(F"""The following configurations don't contain any valid checkpoint:\n{message}""" ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
622
0
import argparse import struct import unittest class __lowerCAmelCase : def __init__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = data # Initialize hash values __lowerCamelCase = [ 0X6a_09e_667, 0Xbb_67a_e85, 0X3c_6ef_372, 0Xa5_4ff_53a, 0X51_0e5_27f, 0X9b_056_88c, 0X1f_83d_9ab, 0X5b_e0c_d19, ] # Initialize round constants __lowerCamelCase = [ 0X42_8a2_f98, 0X71_374_491, 0Xb5_c0f_bcf, 0Xe9_b5d_ba5, 0X39_56c_25b, 0X59_f11_1f1, 0X92_3f8_2a4, 0Xab_1c5_ed5, 0Xd8_07a_a98, 0X12_835_b01, 0X24_318_5be, 0X55_0c7_dc3, 0X72_be5_d74, 0X80_deb_1fe, 0X9b_dc0_6a7, 0Xc1_9bf_174, 0Xe4_9b6_9c1, 0Xef_be4_786, 0X0f_c19_dc6, 0X24_0ca_1cc, 0X2d_e92_c6f, 0X4a_748_4aa, 0X5c_b0a_9dc, 0X76_f98_8da, 0X98_3e5_152, 0Xa8_31c_66d, 0Xb0_032_7c8, 0Xbf_597_fc7, 0Xc6_e00_bf3, 0Xd5_a79_147, 0X06_ca6_351, 0X14_292_967, 0X27_b70_a85, 0X2e_1b2_138, 0X4d_2c6_dfc, 0X53_380_d13, 0X65_0a7_354, 0X76_6a0_abb, 0X81_c2c_92e, 0X92_722_c85, 0Xa2_bfe_8a1, 0Xa8_1a6_64b, 0Xc2_4b8_b70, 0Xc7_6c5_1a3, 0Xd1_92e_819, 0Xd6_990_624, 0Xf4_0e3_585, 0X10_6aa_070, 0X19_a4c_116, 0X1e_376_c08, 0X27_487_74c, 0X34_b0b_cb5, 0X39_1c0_cb3, 0X4e_d8a_a4a, 0X5b_9cc_a4f, 0X68_2e6_ff3, 0X74_8f8_2ee, 0X78_a56_36f, 0X84_c87_814, 0X8c_c70_208, 0X90_bef_ffa, 0Xa4_506_ceb, 0Xbe_f9a_3f7, 0Xc6_717_8f2, ] __lowerCamelCase = self.preprocessing(self.data ) self.final_hash() @staticmethod def lowerCamelCase ( __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = b'''\x80''' + (b'''\x00''' * (63 - (len(__UpperCamelCase ) + 8) % 64)) __lowerCamelCase = struct.pack('''>Q''' , (len(__UpperCamelCase ) * 8) ) return data + padding + big_endian_integer def lowerCamelCase ( self ): '''simple docstring''' # Convert into blocks of 64 bytes __lowerCamelCase = [ self.preprocessed_data[x : x + 64] for x in range(0 , len(self.preprocessed_data ) , 64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers __lowerCamelCase = list(struct.unpack('''>16L''' , __UpperCamelCase ) ) # add 48 0-ed integers words += [0] * 48 __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = self.hashes for index in range(0 , 64 ): if index > 15: # modify the zero-ed indexes at the end of the array __lowerCamelCase = ( self.ror(words[index - 15] , 7 ) ^ self.ror(words[index - 15] , 18 ) ^ (words[index - 15] >> 3) ) __lowerCamelCase = ( self.ror(words[index - 2] , 17 ) ^ self.ror(words[index - 2] , 19 ) ^ (words[index - 2] >> 10) ) __lowerCamelCase = ( words[index - 16] + sa + words[index - 7] + sa ) % 0X100_000_000 # Compression __lowerCamelCase = self.ror(__UpperCamelCase , 6 ) ^ self.ror(__UpperCamelCase , 11 ) ^ self.ror(__UpperCamelCase , 25 ) __lowerCamelCase = (e & f) ^ ((~e & 0Xff_fff_fff) & g) __lowerCamelCase = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0X100_000_000 __lowerCamelCase = self.ror(__UpperCamelCase , 2 ) ^ self.ror(__UpperCamelCase , 13 ) ^ self.ror(__UpperCamelCase , 22 ) __lowerCamelCase = (a & b) ^ (a & c) ^ (b & c) __lowerCamelCase = (sa + maj) % 0X100_000_000 __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = ( g, f, e, ((d + tempa) % 0X100_000_000), c, b, a, ((tempa + tempa) % 0X100_000_000), ) __lowerCamelCase = [a, b, c, d, e, f, g, h] # Modify final values __lowerCamelCase = [ ((element + mutated_hash_values[index]) % 0X100_000_000) for index, element in enumerate(self.hashes ) ] __lowerCamelCase = ''''''.join([hex(__UpperCamelCase )[2:].zfill(8 ) for value in self.hashes] ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' return 0Xff_fff_fff & (value << (32 - rotations)) | (value >> rotations) class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' import hashlib __lowerCamelCase = bytes('''Test String''' , '''utf-8''' ) self.assertEqual(SHAaaa(__UpperCamelCase ).hash , hashlib.shaaaa(__UpperCamelCase ).hexdigest() ) def a__ ( ): import doctest doctest.testmod() __lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '''-s''' ,'''--string''' ,dest='''input_string''' ,default='''Hello World!! Welcome to Cryptography''' ,help='''Hash the string''' ,) parser.add_argument( '''-f''' ,'''--file''' ,dest='''input_file''' ,help='''Hash contents of a file''' ) __lowerCamelCase = parser.parse_args() __lowerCamelCase = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file ,'''rb''' ) as f: __lowerCamelCase = f.read() else: __lowerCamelCase = bytes(_SCREAMING_SNAKE_CASE ,'''utf-8''' ) print(SHAaaa(_SCREAMING_SNAKE_CASE ).hash ) if __name__ == "__main__": main()
707
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a_ = { """configuration_clipseg""": [ """CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP""", """CLIPSegConfig""", """CLIPSegTextConfig""", """CLIPSegVisionConfig""", ], """processing_clipseg""": ["""CLIPSegProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ """CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST""", """CLIPSegModel""", """CLIPSegPreTrainedModel""", """CLIPSegTextModel""", """CLIPSegVisionModel""", """CLIPSegForImageSegmentation""", ] if TYPE_CHECKING: from .configuration_clipseg import ( CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPSegConfig, CLIPSegTextConfig, CLIPSegVisionConfig, ) from .processing_clipseg import CLIPSegProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clipseg import ( CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPSegForImageSegmentation, CLIPSegModel, CLIPSegPreTrainedModel, CLIPSegTextModel, CLIPSegVisionModel, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
622
0
import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class __lowerCAmelCase ( A_ ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = dataset __lowerCamelCase = process __lowerCamelCase = params def __len__( self ): '''simple docstring''' return len(self.dataset ) def __getitem__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.dataset[i] __lowerCamelCase = self.process(__UpperCAmelCase , **self.params ) return processed class __lowerCAmelCase ( A_ ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None ): '''simple docstring''' __lowerCamelCase = loader __lowerCamelCase = infer __lowerCamelCase = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether __lowerCamelCase = None __lowerCamelCase = loader_batch_size # Internal bookkeeping __lowerCamelCase = None __lowerCamelCase = None def __len__( self ): '''simple docstring''' return len(self.loader ) def __iter__( self ): '''simple docstring''' __lowerCamelCase = iter(self.loader ) return self def lowerCamelCase ( self ): '''simple docstring''' if isinstance(self._loader_batch_data , torch.Tensor ): # Batch data is simple tensor, just fetch the slice __lowerCamelCase = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) __lowerCamelCase = {} for k, element in self._loader_batch_data.items(): if isinstance(__UpperCAmelCase , __UpperCAmelCase ): # Convert ModelOutput to tuple first __lowerCamelCase = element.to_tuple() if isinstance(element[0] , torch.Tensor ): __lowerCamelCase = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): __lowerCamelCase = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(__UpperCAmelCase , __UpperCAmelCase ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] , torch.Tensor ): __lowerCamelCase = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): __lowerCamelCase = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around __lowerCamelCase = None elif isinstance(element[self._loader_batch_index] , torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers __lowerCamelCase = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] , np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers __lowerCamelCase = np.expand_dims(element[self._loader_batch_index] , 0 ) else: # This is typically a list, so no need to `unsqueeze`. __lowerCamelCase = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 __lowerCamelCase = self._loader_batch_data.__class__(__UpperCAmelCase ) self._loader_batch_index += 1 return result def lowerCamelCase ( self ): '''simple docstring''' if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch __lowerCamelCase = next(self.iterator ) __lowerCamelCase = self.infer(__UpperCAmelCase , **self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(__UpperCAmelCase , torch.Tensor ): __lowerCamelCase = processed else: __lowerCamelCase = list(processed.keys() )[0] __lowerCamelCase = processed[key] if isinstance(__UpperCAmelCase , __UpperCAmelCase ): __lowerCamelCase = len(__UpperCAmelCase ) else: __lowerCamelCase = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. __lowerCamelCase = observed_batch_size # Setting internal index to unwrap the batch __lowerCamelCase = processed __lowerCamelCase = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class __lowerCAmelCase ( A_ ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None ): '''simple docstring''' super().__init__(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) def __iter__( self ): '''simple docstring''' __lowerCamelCase = iter(self.loader ) __lowerCamelCase = None return self def lowerCamelCase ( self ): '''simple docstring''' if self.subiterator is None: __lowerCamelCase = self.infer(next(self.iterator ) , **self.params ) try: # Try to return next item __lowerCamelCase = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators __lowerCamelCase = self.infer(next(self.iterator ) , **self.params ) __lowerCamelCase = next(self.subiterator ) return processed class __lowerCAmelCase ( A_ ): def __iter__( self ): '''simple docstring''' __lowerCamelCase = iter(self.loader ) return self def lowerCamelCase ( self ): '''simple docstring''' # Extremely similar to PipelineIterator in its unpacking mechanism # BUT, we have an extra required item which is the presence of `is_last` # That is because everything is flattened by `PipelineChunkIterator` we # need to keep track of how to regroup here in the original `process` # boundaries so that `process` and `postprocess` see the same data. # This iterator accumulates items (possibly while unbatching) until it # its a `is_last` and then just passes it on to the caller. __lowerCamelCase = False __lowerCamelCase = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: __lowerCamelCase = self.loader_batch_item() __lowerCamelCase = item.pop('''is_last''' ) accumulator.append(__UpperCAmelCase ) if is_last: return accumulator while not is_last: __lowerCamelCase = self.infer(next(self.iterator ) , **self.params ) if self.loader_batch_size is not None: if isinstance(__UpperCAmelCase , torch.Tensor ): __lowerCamelCase = processed else: __lowerCamelCase = list(processed.keys() )[0] __lowerCamelCase = processed[key] if isinstance(__UpperCAmelCase , __UpperCAmelCase ): __lowerCamelCase = len(__UpperCAmelCase ) else: __lowerCamelCase = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. __lowerCamelCase = observed_batch_size __lowerCamelCase = processed __lowerCamelCase = 0 while self._loader_batch_index < self.loader_batch_size: __lowerCamelCase = self.loader_batch_item() __lowerCamelCase = item.pop('''is_last''' ) accumulator.append(__UpperCAmelCase ) if is_last: return accumulator else: __lowerCamelCase = processed __lowerCamelCase = item.pop('''is_last''' ) accumulator.append(__UpperCAmelCase ) return accumulator class __lowerCAmelCase ( A_ ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = dataset __lowerCamelCase = key def __len__( self ): '''simple docstring''' return len(self.dataset ) def __getitem__( self , __UpperCAmelCase ): '''simple docstring''' return self.dataset[i][self.key] class __lowerCAmelCase ( A_ ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = dataset __lowerCamelCase = keya __lowerCamelCase = keya def __len__( self ): '''simple docstring''' return len(self.dataset ) def __getitem__( self , __UpperCAmelCase ): '''simple docstring''' return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
708
import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = RoFormerTokenizer lowerCAmelCase__ = RoFormerTokenizerFast lowerCAmelCase__ = True lowerCAmelCase__ = True def lowerCamelCase ( self ): '''simple docstring''' super().setUp() def lowerCamelCase ( self , **__UpperCAmelCase ): '''simple docstring''' return self.tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **__UpperCAmelCase ) def lowerCamelCase ( self , **__UpperCAmelCase ): '''simple docstring''' return self.rust_tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''永和服装饰品有限公司,今天天气非常好''' __lowerCamelCase = '''永和 服装 饰品 有限公司 , 今 天 天 气 非常 好''' return input_text, output_text def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_tokenizer() __lowerCamelCase ,__lowerCamelCase = self.get_chinese_input_output_texts() __lowerCamelCase = tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , output_text.split() ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [22943, 21332, 34431, 45904, 117, 306, 1231, 1231, 2653, 33994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_rust_tokenizer() __lowerCamelCase ,__lowerCamelCase = self.get_chinese_input_output_texts() __lowerCamelCase = tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , output_text.split() ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [22943, 21332, 34431, 45904, 117, 306, 1231, 1231, 2653, 33994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' pass def lowerCamelCase ( self ): '''simple docstring''' pass
622
0
from __future__ import annotations a_ = """#""" class __lowerCAmelCase : def __init__( self ): '''simple docstring''' __lowerCamelCase = {} def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self._trie for char in text: if char not in trie: __lowerCamelCase = {} __lowerCamelCase = trie[char] __lowerCamelCase = True def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self._trie for char in prefix: if char in trie: __lowerCamelCase = trie[char] else: return [] return self._elements(UpperCamelCase__ ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = [] for c, v in d.items(): __lowerCamelCase = [' '] if c == END else [(c + s) for s in self._elements(UpperCamelCase__ )] result.extend(UpperCamelCase__ ) return tuple(UpperCamelCase__ ) a_ = Trie() a_ = ("""depart""", """detergent""", """daring""", """dog""", """deer""", """deal""") for word in words: trie.insert_word(word) def a__ ( _UpperCamelCase : str ): __lowerCamelCase = trie.find_word(lowerCAmelCase__ ) return tuple(string + word for word in suffixes ) def a__ ( ): print(autocomplete_using_trie('''de''' ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
709
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device a_ = False class __lowerCAmelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained('''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt='''first prompt''' , image=__UpperCAmelCase , text_to_image_strength=0.75 , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained(__UpperCAmelCase , torch_dtype=torch.floataa ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = generator.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt='''first prompt''' , image=__UpperCAmelCase , text_to_image_strength=0.75 , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = VersatileDiffusionPipeline.from_pretrained('''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __lowerCamelCase = '''cyberpunk 2077''' __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.dual_guided( prompt=__UpperCAmelCase , image=__UpperCAmelCase , text_to_image_strength=0.75 , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' , ).images __lowerCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.1_448, 0.1_619, 0.1_741, 0.1_086, 0.1_147, 0.1_128, 0.1_199, 0.1_165, 0.1_001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __lowerCamelCase = '''A painting of a squirrel eating a burger ''' __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe.text_to_image( prompt=__UpperCAmelCase , generator=__UpperCAmelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' ).images __lowerCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.3_367, 0.3_169, 0.2_656, 0.3_870, 0.4_790, 0.3_796, 0.4_009, 0.4_878, 0.4_778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __lowerCamelCase = pipe.image_variation(__UpperCAmelCase , generator=__UpperCAmelCase , output_type='''numpy''' ).images __lowerCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __lowerCamelCase = np.array([0.3_076, 0.3_123, 0.3_284, 0.3_782, 0.3_770, 0.3_894, 0.4_297, 0.4_331, 0.4_456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
622
0
from abc import ABC, abstractmethod from argparse import ArgumentParser class __lowerCAmelCase ( a__ ): @staticmethod @abstractmethod def lowerCamelCase ( __UpperCAmelCase ): '''simple docstring''' raise NotImplementedError() @abstractmethod def lowerCamelCase ( self ): '''simple docstring''' raise NotImplementedError()
710
import argparse import datetime import json import time import warnings from logging import getLogger from pathlib import Path from typing import Dict, List import torch from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import calculate_bleu, calculate_rouge, chunks, parse_numeric_n_bool_cl_kwargs, use_task_specific_params a_ = getLogger(__name__) a_ = """cuda""" if torch.cuda.is_available() else """cpu""" def a__ ( _UpperCamelCase : List[str] ,_UpperCamelCase : str ,_UpperCamelCase : str ,_UpperCamelCase : int = 8 ,_UpperCamelCase : str = DEFAULT_DEVICE ,_UpperCamelCase : Dict=False ,_UpperCamelCase : Dict="summarization" ,_UpperCamelCase : Optional[int]=None ,**_UpperCamelCase : Dict ,): __lowerCamelCase = Path(_UpperCamelCase ).open('''w''' ,encoding='''utf-8''' ) __lowerCamelCase = str(_UpperCamelCase ) __lowerCamelCase = AutoModelForSeqaSeqLM.from_pretrained(_UpperCamelCase ).to(_UpperCamelCase ) if fpaa: __lowerCamelCase = model.half() __lowerCamelCase = AutoTokenizer.from_pretrained(_UpperCamelCase ) logger.info(F"""Inferred tokenizer type: {tokenizer.__class__}""" ) # if this is wrong, check config.model_type. __lowerCamelCase = time.time() # update config with task specific params use_task_specific_params(_UpperCamelCase ,_UpperCamelCase ) if prefix is None: __lowerCamelCase = prefix or getattr(model.config ,'''prefix''' ,'''''' ) or '''''' for examples_chunk in tqdm(list(chunks(_UpperCamelCase ,_UpperCamelCase ) ) ): __lowerCamelCase = [prefix + text for text in examples_chunk] __lowerCamelCase = tokenizer(_UpperCamelCase ,return_tensors='''pt''' ,truncation=_UpperCamelCase ,padding='''longest''' ).to(_UpperCamelCase ) __lowerCamelCase = model.generate( input_ids=batch.input_ids ,attention_mask=batch.attention_mask ,**_UpperCamelCase ,) __lowerCamelCase = tokenizer.batch_decode(_UpperCamelCase ,skip_special_tokens=_UpperCamelCase ,clean_up_tokenization_spaces=_UpperCamelCase ) for hypothesis in dec: fout.write(hypothesis + '''\n''' ) fout.flush() fout.close() __lowerCamelCase = int(time.time() - start_time ) # seconds __lowerCamelCase = len(_UpperCamelCase ) return {"n_obs": n_obs, "runtime": runtime, "seconds_per_sample": round(runtime / n_obs ,4 )} def a__ ( ): return datetime.datetime.now().strftime('''%Y-%m-%d %H:%M:%S''' ) def a__ ( _UpperCamelCase : Union[str, Any]=True ): __lowerCamelCase = argparse.ArgumentParser() parser.add_argument('''model_name''' ,type=_UpperCamelCase ,help='''like facebook/bart-large-cnn,t5-base, etc.''' ) parser.add_argument('''input_path''' ,type=_UpperCamelCase ,help='''like cnn_dm/test.source''' ) parser.add_argument('''save_path''' ,type=_UpperCamelCase ,help='''where to save summaries''' ) parser.add_argument('''--reference_path''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,help='''like cnn_dm/test.target''' ) parser.add_argument('''--score_path''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,default='''metrics.json''' ,help='''where to save metrics''' ) parser.add_argument('''--device''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,default=_UpperCamelCase ,help='''cuda, cuda:1, cpu etc.''' ) parser.add_argument( '''--prefix''' ,type=_UpperCamelCase ,required=_UpperCamelCase ,default=_UpperCamelCase ,help='''will be added to the begininng of src examples''' ) parser.add_argument('''--task''' ,type=_UpperCamelCase ,default='''summarization''' ,help='''used for task_specific_params + metrics''' ) parser.add_argument('''--bs''' ,type=_UpperCamelCase ,default=8 ,required=_UpperCamelCase ,help='''batch size''' ) parser.add_argument( '''--n_obs''' ,type=_UpperCamelCase ,default=-1 ,required=_UpperCamelCase ,help='''How many observations. Defaults to all.''' ) parser.add_argument('''--fp16''' ,action='''store_true''' ) parser.add_argument('''--dump-args''' ,action='''store_true''' ,help='''print the custom hparams with the results''' ) parser.add_argument( '''--info''' ,nargs='''?''' ,type=_UpperCamelCase ,const=datetime_now() ,help=( '''use in conjunction w/ --dump-args to print with the results whatever other info you\'d like, e.g.''' ''' lang=en-ru. If no value is passed, the current datetime string will be used.''' ) ,) # Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate __lowerCamelCase ,__lowerCamelCase = parser.parse_known_args() __lowerCamelCase = parse_numeric_n_bool_cl_kwargs(_UpperCamelCase ) if parsed_args and verbose: print(F"""parsed the following generate kwargs: {parsed_args}""" ) __lowerCamelCase = [''' ''' + x.rstrip() if '''t5''' in args.model_name else x.rstrip() for x in open(args.input_path ).readlines()] if args.n_obs > 0: __lowerCamelCase = examples[: args.n_obs] Path(args.save_path ).parent.mkdir(exist_ok=_UpperCamelCase ) if args.reference_path is None and Path(args.score_path ).exists(): warnings.warn(F"""score_path {args.score_path} will be overwritten unless you type ctrl-c.""" ) if args.device == "cpu" and args.fpaa: # this mix leads to RuntimeError: "threshold_cpu" not implemented for 'Half' raise ValueError('''Can\'t mix --fp16 and --device cpu''' ) __lowerCamelCase = generate_summaries_or_translations( _UpperCamelCase ,args.save_path ,args.model_name ,batch_size=args.bs ,device=args.device ,fpaa=args.fpaa ,task=args.task ,prefix=args.prefix ,**_UpperCamelCase ,) if args.reference_path is None: return {} # Compute scores __lowerCamelCase = calculate_bleu if '''translation''' in args.task else calculate_rouge __lowerCamelCase = [x.rstrip() for x in open(args.save_path ).readlines()] __lowerCamelCase = [x.rstrip() for x in open(args.reference_path ).readlines()][: len(_UpperCamelCase )] __lowerCamelCase = score_fn(_UpperCamelCase ,_UpperCamelCase ) scores.update(_UpperCamelCase ) if args.dump_args: scores.update(_UpperCamelCase ) if args.info: __lowerCamelCase = args.info if verbose: print(_UpperCamelCase ) if args.score_path is not None: json.dump(_UpperCamelCase ,open(args.score_path ,'''w''' ) ) return scores if __name__ == "__main__": # Usage for MT: # python run_eval.py MODEL_NAME $DATA_DIR/test.source $save_dir/test_translations.txt --reference_path $DATA_DIR/test.target --score_path $save_dir/test_bleu.json --task translation $@ run_generate(verbose=True)
622
0
'''simple docstring''' def a__ ( _UpperCamelCase : str ,_UpperCamelCase : str ): __lowerCamelCase = len(_UpperCamelCase ) __lowerCamelCase = [] for i in range(len(_UpperCamelCase ) - pat_len + 1 ): __lowerCamelCase = True for j in range(_UpperCamelCase ): if s[i + j] != pattern[j]: __lowerCamelCase = False break if match_found: position.append(_UpperCamelCase ) return position if __name__ == "__main__": assert naive_pattern_search("""ABCDEFG""", """DE""") == [3] print(naive_pattern_search("""ABAAABCDBBABCDDEBCABC""", """ABC"""))
711
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def a__ ( _UpperCamelCase : Optional[int] ,_UpperCamelCase : List[str] ,_UpperCamelCase : List[Any]=None ,_UpperCamelCase : Any=None ): if attention_mask is None: __lowerCamelCase = tf.cast(tf.math.not_equal(_UpperCamelCase ,config.pad_token_id ) ,tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class __lowerCAmelCase : lowerCAmelCase__ = OPTConfig lowerCAmelCase__ = {} lowerCAmelCase__ = """gelu""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase=99 , __UpperCAmelCase=16 , __UpperCAmelCase=2 , __UpperCAmelCase=4 , __UpperCAmelCase=4 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=20 , __UpperCAmelCase=2 , __UpperCAmelCase=1 , __UpperCAmelCase=0 , __UpperCAmelCase=16 , __UpperCAmelCase=16 , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = eos_token_id __lowerCamelCase = pad_token_id __lowerCamelCase = bos_token_id __lowerCamelCase = embed_dim __lowerCamelCase = word_embed_proj_dim __lowerCamelCase = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowerCamelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowerCamelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowerCamelCase = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=__UpperCAmelCase , **self.config_updates , ) __lowerCamelCase = prepare_opt_inputs_dict(__UpperCAmelCase , __UpperCAmelCase ) return config, inputs_dict def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = TFOPTModel(config=__UpperCAmelCase ) __lowerCamelCase = inputs_dict['''input_ids'''] __lowerCamelCase = input_ids[:1, :] __lowerCamelCase = inputs_dict['''attention_mask'''][:1, :] __lowerCamelCase = 1 # first forward pass __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , use_cache=__UpperCAmelCase ) __lowerCamelCase ,__lowerCamelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowerCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowerCamelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowerCamelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowerCamelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase )[0] __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , past_key_values=__UpperCAmelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowerCamelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowerCamelCase = output_from_no_past[:, -3:, random_slice_idx] __lowerCamelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__UpperCAmelCase , __UpperCAmelCase , rtol=1E-3 ) @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () lowerCAmelCase__ = (TFOPTForCausalLM,) if is_tf_available() else () lowerCAmelCase__ = ( {"""feature-extraction""": TFOPTModel, """text-generation""": TFOPTForCausalLM} if is_tf_available() else {} ) lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = 1_0 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase ,__lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(__UpperCAmelCase , __UpperCAmelCase ): if hasattr(__UpperCAmelCase , '''weight''' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(__UpperCAmelCase , '''weight''' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 10, config.vocab_size + 10]: # build the embeddings __lowerCamelCase = model_class(config=__UpperCAmelCase ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_input_embeddings() ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(__UpperCAmelCase ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_input_embeddings() ) __lowerCamelCase = _get_word_embedding_weight(__UpperCAmelCase , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. __lowerCamelCase = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , __UpperCAmelCase ) # check that weights remain the same after resizing __lowerCamelCase = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCamelCase = False self.assertTrue(__UpperCAmelCase ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , __UpperCAmelCase ) __lowerCamelCase = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCamelCase = False self.assertTrue(__UpperCAmelCase ) def a__ ( _UpperCamelCase : Optional[Any] ): return tf.constant(_UpperCamelCase ,dtype=tf.intaa ) @require_tf class __lowerCAmelCase ( unittest.TestCase ): lowerCAmelCase__ = 9_9 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = tf.ones((4, 1) , dtype=tf.intaa ) * 2 __lowerCamelCase = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) __lowerCamelCase = input_ids.shape[0] __lowerCamelCase = OPTConfig( vocab_size=self.vocab_size , hidden_size=24 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class __lowerCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTModel.from_pretrained('''facebook/opt-350m''' ) __lowerCamelCase = _long_tensor([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] ) __lowerCamelCase = tf.not_equal(__UpperCAmelCase , model.config.pad_token_id ) with tf.GradientTape(): __lowerCamelCase = model(input_ids=__UpperCAmelCase , attention_mask=__UpperCAmelCase ).last_hidden_state __lowerCamelCase = (1, 11, 512) self.assertEqual(output.shape , __UpperCAmelCase ) __lowerCamelCase = tf.constant( [[-0.2_873, -1.9_218, -0.3_033], [-1.2_710, -0.1_338, -0.1_902], [0.4_095, 0.1_214, -1.3_121]] ) self.assertTrue(np.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=4E-3 ) ) __lowerCamelCase = tf.function(__UpperCAmelCase , jit_compile=__UpperCAmelCase ) __lowerCamelCase = xla_generate(__UpperCAmelCase , __UpperCAmelCase )[0] self.assertTrue(np.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=4E-2 ) ) @require_tf @slow class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): '''simple docstring''' super().setUp() __lowerCamelCase = '''facebook/opt-350m''' def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFOPTForCausalLM.from_pretrained(self.path_model ) __lowerCamelCase = GPTaTokenizer.from_pretrained(self.path_model ) __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of''', '''Paris is the capital of France and''', '''Computers and mobile phones have taken''', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' , padding=__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) __lowerCamelCase = tf.constant( [ [1.3_851, -13.8_923, -10.5_229, -10.7_533, -0.2_309, -10.2_384, -0.5_365, -9.0_947, -5.1_670], [-4.7_073, -10.6_276, -3.9_415, -21.5_242, -0.2_822, -0.2_822, -0.2_822, -0.2_822, -0.2_822], [0.6_247, -3.4_229, -8.9_179, -1.4_297, -14.1_650, 1.4_146, -9.0_218, -0.2_703, -0.2_703], [6.4_783, -1.9_913, -10.7_926, -2.3_336, 1.5_092, -0.9_974, -6.8_213, 1.3_477, 1.3_477], ] ) self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-4 ) ) __lowerCamelCase = tf.function(__UpperCAmelCase , jit_compile=__UpperCAmelCase ) __lowerCamelCase = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(__UpperCAmelCase , __UpperCAmelCase , atol=1E-4 ) ) @require_tf @slow class __lowerCAmelCase ( unittest.TestCase ): @property def lowerCamelCase ( self ): '''simple docstring''' return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-125m''' __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of New York, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] __lowerCamelCase = [] __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) for prompt in self.prompts: __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(__UpperCAmelCase , max_length=10 ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) predicted_outputs += generated_string self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-350m''' __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = '''left''' # use different length sentences to test batching __lowerCamelCase = [ '''Hello, my dog is a little''', '''Today, I''', ] __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' , padding=__UpperCAmelCase ) __lowerCamelCase = inputs['''input_ids'''] __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase , attention_mask=inputs['''attention_mask'''] ) __lowerCamelCase = tokenizer(sentences[0] , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase ) __lowerCamelCase = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['''attention_mask'''][-1] , tf.intaa ) ) __lowerCamelCase = tokenizer(sentences[1] , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(input_ids=__UpperCAmelCase , max_length=model.config.max_length - num_paddings ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=__UpperCAmelCase ) __lowerCamelCase = [ '''Hello, my dog is a little bit of a dork.\nI\'m a little bit''', '''Today, I was in the middle of a conversation with a friend about the''', ] self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , [non_padded_sentence, padded_sentence] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = '''facebook/opt-350m''' __lowerCamelCase = [ '''Today is a beautiful day and I want to''', '''In the city of San Francisco, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] __lowerCamelCase = [] __lowerCamelCase = GPTaTokenizer.from_pretrained(__UpperCAmelCase ) __lowerCamelCase = TFOPTForCausalLM.from_pretrained(__UpperCAmelCase ) for prompt in self.prompts: __lowerCamelCase = tokenizer(__UpperCAmelCase , return_tensors='''tf''' ).input_ids __lowerCamelCase = model.generate(__UpperCAmelCase , max_length=10 ) __lowerCamelCase = tokenizer.batch_decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase ) predicted_outputs += generated_string self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase )
622
0
import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelForTokenClassification, AutoModelWithLMHead, AutoTokenizer, PretrainedConfig, PreTrainedTokenizer, ) from transformers.optimization import ( Adafactor, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.utils.versions import require_version a_ = logging.getLogger(__name__) require_version("""pytorch_lightning>=1.0.4""") a_ = { "base": AutoModel, "sequence-classification": AutoModelForSequenceClassification, "question-answering": AutoModelForQuestionAnswering, "pretraining": AutoModelForPreTraining, "token-classification": AutoModelForTokenClassification, "language-modeling": AutoModelWithLMHead, "summarization": AutoModelForSeqaSeqLM, "translation": AutoModelForSeqaSeqLM, } # update this and the import above to support new schedulers from transformers.optimization a_ = { "linear": get_linear_schedule_with_warmup, "cosine": get_cosine_schedule_with_warmup, "cosine_w_restarts": get_cosine_with_hard_restarts_schedule_with_warmup, "polynomial": get_polynomial_decay_schedule_with_warmup, # '': get_constant_schedule, # not supported for now # '': get_constant_schedule_with_warmup, # not supported for now } a_ = sorted(arg_to_scheduler.keys()) a_ = "{" + ", ".join(arg_to_scheduler_choices) + "}" class __lowerCAmelCase ( pl.LightningModule ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase="base" , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=None , **__UpperCAmelCase , ): '''simple docstring''' super().__init__() # TODO: move to self.save_hyperparameters() # self.save_hyperparameters() # can also expand arguments into trainer signature for easier reading self.save_hyperparameters(lowerCAmelCase_ ) __lowerCamelCase = 0 __lowerCamelCase = Path(self.hparams.output_dir ) __lowerCamelCase = self.hparams.cache_dir if self.hparams.cache_dir else None if config is None: __lowerCamelCase = AutoConfig.from_pretrained( self.hparams.config_name if self.hparams.config_name else self.hparams.model_name_or_path , **({'''num_labels''': num_labels} if num_labels is not None else {}) , cache_dir=lowerCAmelCase_ , **lowerCAmelCase_ , ) else: __lowerCamelCase = config __lowerCamelCase = ('''encoder_layerdrop''', '''decoder_layerdrop''', '''dropout''', '''attention_dropout''') for p in extra_model_params: if getattr(self.hparams , lowerCAmelCase_ , lowerCAmelCase_ ): assert hasattr(self.config , lowerCAmelCase_ ), F"""model config doesn\'t have a `{p}` attribute""" setattr(self.config , lowerCAmelCase_ , getattr(self.hparams , lowerCAmelCase_ ) ) if tokenizer is None: __lowerCamelCase = AutoTokenizer.from_pretrained( self.hparams.tokenizer_name if self.hparams.tokenizer_name else self.hparams.model_name_or_path , cache_dir=lowerCAmelCase_ , ) else: __lowerCamelCase = tokenizer __lowerCamelCase = MODEL_MODES[mode] if model is None: __lowerCamelCase = self.model_type.from_pretrained( self.hparams.model_name_or_path , from_tf=bool('''.ckpt''' in self.hparams.model_name_or_path ) , config=self.config , cache_dir=lowerCAmelCase_ , ) else: __lowerCamelCase = model def lowerCamelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.model_type.from_pretrained(*lowerCAmelCase_ , **lowerCAmelCase_ ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = arg_to_scheduler[self.hparams.lr_scheduler] __lowerCamelCase = get_schedule_func( self.opt , num_warmup_steps=self.hparams.warmup_steps , num_training_steps=self.total_steps() ) __lowerCamelCase = {'''scheduler''': scheduler, '''interval''': '''step''', '''frequency''': 1} return scheduler def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model __lowerCamelCase = ['''bias''', '''LayerNorm.weight'''] __lowerCamelCase = [ { '''params''': [ p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay ) ], # check this named paramters '''weight_decay''': self.hparams.weight_decay, }, { '''params''': [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0, }, ] if self.hparams.adafactor: __lowerCamelCase = Adafactor( lowerCAmelCase_ , lr=self.hparams.learning_rate , scale_parameter=lowerCAmelCase_ , relative_step=lowerCAmelCase_ ) else: __lowerCamelCase = AdamW( lowerCAmelCase_ , lr=self.hparams.learning_rate , eps=self.hparams.adam_epsilon ) __lowerCamelCase = optimizer __lowerCamelCase = self.get_lr_scheduler() return [optimizer], [scheduler] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' return self.validation_step(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.validation_end(lowerCAmelCase_ ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = max(1 , self.hparams.gpus ) # TODO: consider num_tpu_cores __lowerCamelCase = self.hparams.train_batch_size * self.hparams.accumulate_grad_batches * num_devices return (self.dataset_size / effective_batch_size) * self.hparams.max_epochs def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' if stage == "test": __lowerCamelCase = len(self.test_dataloader().dataset ) else: __lowerCamelCase = self.get_dataloader('''train''' , self.hparams.train_batch_size , shuffle=lowerCAmelCase_ ) __lowerCamelCase = len(self.train_dataloader().dataset ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = False ): '''simple docstring''' raise NotImplementedError('''You must implement this for your task''' ) def lowerCamelCase ( self ): '''simple docstring''' return self.train_loader def lowerCamelCase ( self ): '''simple docstring''' return self.get_dataloader('''dev''' , self.hparams.eval_batch_size , shuffle=lowerCAmelCase_ ) def lowerCamelCase ( self ): '''simple docstring''' return self.get_dataloader('''test''' , self.hparams.eval_batch_size , shuffle=lowerCAmelCase_ ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return os.path.join( self.hparams.data_dir , '''cached_{}_{}_{}'''.format( lowerCAmelCase_ , list(filter(lowerCAmelCase_ , self.hparams.model_name_or_path.split('''/''' ) ) ).pop() , str(self.hparams.max_seq_length ) , ) , ) @pl.utilities.rank_zero_only def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.output_dir.joinpath('''best_tfmr''' ) __lowerCamelCase = self.step_count self.model.save_pretrained(lowerCAmelCase_ ) self.tokenizer.save_pretrained(lowerCAmelCase_ ) @staticmethod def lowerCamelCase ( __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' parser.add_argument( '''--model_name_or_path''' , default=lowerCAmelCase_ , type=lowerCAmelCase_ , required=lowerCAmelCase_ , help='''Path to pretrained model or model identifier from huggingface.co/models''' , ) parser.add_argument( '''--config_name''' , default='''''' , type=lowerCAmelCase_ , help='''Pretrained config name or path if not the same as model_name''' ) parser.add_argument( '''--tokenizer_name''' , default=lowerCAmelCase_ , type=lowerCAmelCase_ , help='''Pretrained tokenizer name or path if not the same as model_name''' , ) parser.add_argument( '''--cache_dir''' , default=str(Path(lowerCAmelCase_ ).parent / '''test_run''' / '''cache''' ) , type=lowerCAmelCase_ , help='''Where do you want to store the pre-trained models downloaded from huggingface.co''' , ) parser.add_argument( '''--encoder_layerdrop''' , type=lowerCAmelCase_ , help='''Encoder layer dropout probability (Optional). Goes into model.config''' , ) parser.add_argument( '''--decoder_layerdrop''' , type=lowerCAmelCase_ , help='''Decoder layer dropout probability (Optional). Goes into model.config''' , ) parser.add_argument( '''--dropout''' , type=lowerCAmelCase_ , help='''Dropout probability (Optional). Goes into model.config''' , ) parser.add_argument( '''--attention_dropout''' , type=lowerCAmelCase_ , help='''Attention dropout probability (Optional). Goes into model.config''' , ) parser.add_argument('''--learning_rate''' , default=5E-5 , type=lowerCAmelCase_ , help='''The initial learning rate for Adam.''' ) parser.add_argument( '''--lr_scheduler''' , default='''linear''' , choices=lowerCAmelCase_ , metavar=lowerCAmelCase_ , type=lowerCAmelCase_ , help='''Learning rate scheduler''' , ) parser.add_argument('''--weight_decay''' , default=0.0 , type=lowerCAmelCase_ , help='''Weight decay if we apply some.''' ) parser.add_argument('''--adam_epsilon''' , default=1E-8 , type=lowerCAmelCase_ , help='''Epsilon for Adam optimizer.''' ) parser.add_argument('''--warmup_steps''' , default=0 , type=lowerCAmelCase_ , help='''Linear warmup over warmup_steps.''' ) parser.add_argument('''--num_workers''' , default=4 , type=lowerCAmelCase_ , help='''kwarg passed to DataLoader''' ) parser.add_argument('''--num_train_epochs''' , dest='''max_epochs''' , default=3 , type=lowerCAmelCase_ ) parser.add_argument('''--train_batch_size''' , default=32 , type=lowerCAmelCase_ ) parser.add_argument('''--eval_batch_size''' , default=32 , type=lowerCAmelCase_ ) parser.add_argument('''--adafactor''' , action='''store_true''' ) class __lowerCAmelCase ( pl.Callback ): def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' if ( trainer.is_global_zero and trainer.global_rank == 0 ): # we initialize the retriever only on master worker with RAY. In new pytorch-lightning accelorators are removed. pl_module.model.rag.retriever.init_retrieval() # better to use hook functions. class __lowerCAmelCase ( pl.Callback ): def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' for name, param in pl_module.model.rag.named_parameters(): if param.grad is None: print(lowerCAmelCase_ ) class __lowerCAmelCase ( pl.Callback ): def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = trainer.lr_schedulers[0]['''scheduler'''] __lowerCamelCase = {F"""lr_group_{i}""": lr for i, lr in enumerate(lr_scheduler.get_lr() )} pl_module.logger.log_metrics(lowerCAmelCase_ ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' rank_zero_info('''***** Validation results *****''' ) __lowerCamelCase = trainer.callback_metrics # Log results for key in sorted(lowerCAmelCase_ ): if key not in ["log", "progress_bar"]: rank_zero_info('''{} = {}\n'''.format(lowerCAmelCase_ , str(metrics[key] ) ) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' rank_zero_info('''***** Test results *****''' ) __lowerCamelCase = trainer.callback_metrics # Log and save results to file __lowerCamelCase = os.path.join(pl_module.hparams.output_dir , '''test_results.txt''' ) with open(lowerCAmelCase_ , '''w''' ) as writer: for key in sorted(lowerCAmelCase_ ): if key not in ["log", "progress_bar"]: rank_zero_info('''{} = {}\n'''.format(lowerCAmelCase_ , str(metrics[key] ) ) ) writer.write('''{} = {}\n'''.format(lowerCAmelCase_ , str(metrics[key] ) ) ) def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : str ): parser.add_argument( '''--output_dir''' ,default=str(Path(__lowerCAmelCase ).parent / '''test_run''' / '''model_checkpoints''' ) ,type=__lowerCAmelCase ,help='''The output directory where the model predictions and checkpoints will be written.''' ,) parser.add_argument( '''--fp16''' ,action='''store_true''' ,help='''Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit''' ,) parser.add_argument( '''--fp16_opt_level''' ,type=__lowerCAmelCase ,default='''O2''' ,help=( '''For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\'].''' '''See details at https://nvidia.github.io/apex/amp.html''' ) ,) parser.add_argument('''--n_tpu_cores''' ,dest='''tpu_cores''' ,type=__lowerCAmelCase ) parser.add_argument('''--max_grad_norm''' ,dest='''gradient_clip_val''' ,default=1.0 ,type=__lowerCAmelCase ,help='''Max gradient norm''' ) parser.add_argument('''--do_train''' ,action='''store_true''' ,help='''Whether to run training.''' ) parser.add_argument('''--do_predict''' ,action='''store_true''' ,help='''Whether to run predictions on the test set.''' ) parser.add_argument( '''--gradient_accumulation_steps''' ,dest='''accumulate_grad_batches''' ,type=__lowerCAmelCase ,default=1 ,help='''Number of updates steps to accumulate before performing a backward/update pass.''' ,) parser.add_argument('''--seed''' ,type=__lowerCAmelCase ,default=42 ,help='''random seed for initialization''' ) parser.add_argument( '''--data_dir''' ,default=str(Path(__lowerCAmelCase ).parent / '''test_run''' / '''dummy-train-data''' ) ,type=__lowerCAmelCase ,help='''The input data dir. Should contain the training files for the CoNLL-2003 NER task.''' ,) def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : str=None ,_UpperCamelCase : List[Any]=True ,_UpperCamelCase : List[Any]=[] ,_UpperCamelCase : Union[str, Any]=None ,_UpperCamelCase : int=None ,**_UpperCamelCase : int ,): pl.seed_everything(args.seed ) # init model __lowerCamelCase = Path(model.hparams.output_dir ) odir.mkdir(exist_ok=__lowerCAmelCase ) # add custom checkpoints if checkpoint_callback is None: __lowerCamelCase = pl.callbacks.ModelCheckpoint( filepath=args.output_dir ,prefix='''checkpoint''' ,monitor='''val_loss''' ,mode='''min''' ,save_top_k=1 ) if early_stopping_callback: extra_callbacks.append(__lowerCAmelCase ) if logging_callback is None: __lowerCamelCase = LoggingCallback() __lowerCamelCase = {} if args.fpaa: __lowerCamelCase = 16 if args.gpus > 1: __lowerCamelCase = '''auto''' __lowerCamelCase = '''ddp''' __lowerCamelCase = args.accumulate_grad_batches __lowerCamelCase = None __lowerCamelCase = '''auto''' __lowerCamelCase = pl.Trainer.from_argparse_args( __lowerCAmelCase ,weights_summary=__lowerCAmelCase ,callbacks=[logging_callback] + extra_callbacks + [InitCallback()] + [checkpoint_callback] ,logger=__lowerCAmelCase ,val_check_interval=1 ,num_sanity_val_steps=2 ,**__lowerCAmelCase ,) if args.do_train: trainer.fit(__lowerCAmelCase ) else: print('''RAG modeling tests with new set functions successfuly executed!''' ) return trainer
712
import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) a_ = logging.getLogger(__name__) def a__ ( _UpperCamelCase : str ,_UpperCamelCase : List[Any] ): __lowerCamelCase = np.argmax(_UpperCamelCase ,axis=1 ) return np.sum(outputs == labels ) def a__ ( _UpperCamelCase : Optional[int] ): with open(_UpperCamelCase ,encoding='''utf_8''' ) as f: __lowerCamelCase = csv.reader(_UpperCamelCase ) __lowerCamelCase = [] next(_UpperCamelCase ) # skip the first line for line in tqdm(_UpperCamelCase ): output.append((''' '''.join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def a__ ( _UpperCamelCase : Any ,_UpperCamelCase : Dict ,_UpperCamelCase : str ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Any ,_UpperCamelCase : Dict ): __lowerCamelCase = [] for dataset in encoded_datasets: __lowerCamelCase = len(_UpperCamelCase ) __lowerCamelCase = np.zeros((n_batch, 2, input_len) ,dtype=np.intaa ) __lowerCamelCase = np.zeros((n_batch, 2) ,dtype=np.intaa ) __lowerCamelCase = np.full((n_batch, 2, input_len) ,fill_value=-1_00 ,dtype=np.intaa ) __lowerCamelCase = np.zeros((n_batch,) ,dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(_UpperCamelCase ): __lowerCamelCase = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __lowerCamelCase = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __lowerCamelCase = with_conta __lowerCamelCase = with_conta __lowerCamelCase = len(_UpperCamelCase ) - 1 __lowerCamelCase = len(_UpperCamelCase ) - 1 __lowerCamelCase = with_conta __lowerCamelCase = with_conta __lowerCamelCase = mc_label __lowerCamelCase = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(_UpperCamelCase ) for t in all_inputs ) ) return tensor_datasets def a__ ( ): __lowerCamelCase = argparse.ArgumentParser() parser.add_argument('''--model_name''' ,type=_UpperCamelCase ,default='''openai-gpt''' ,help='''pretrained model name''' ) parser.add_argument('''--do_train''' ,action='''store_true''' ,help='''Whether to run training.''' ) parser.add_argument('''--do_eval''' ,action='''store_true''' ,help='''Whether to run eval on the dev set.''' ) parser.add_argument( '''--output_dir''' ,default=_UpperCamelCase ,type=_UpperCamelCase ,required=_UpperCamelCase ,help='''The output directory where the model predictions and checkpoints will be written.''' ,) parser.add_argument('''--train_dataset''' ,type=_UpperCamelCase ,default='''''' ) parser.add_argument('''--eval_dataset''' ,type=_UpperCamelCase ,default='''''' ) parser.add_argument('''--seed''' ,type=_UpperCamelCase ,default=42 ) parser.add_argument('''--num_train_epochs''' ,type=_UpperCamelCase ,default=3 ) parser.add_argument('''--train_batch_size''' ,type=_UpperCamelCase ,default=8 ) parser.add_argument('''--eval_batch_size''' ,type=_UpperCamelCase ,default=16 ) parser.add_argument('''--adam_epsilon''' ,default=1e-8 ,type=_UpperCamelCase ,help='''Epsilon for Adam optimizer.''' ) parser.add_argument('''--max_grad_norm''' ,type=_UpperCamelCase ,default=1 ) parser.add_argument( '''--max_steps''' ,default=-1 ,type=_UpperCamelCase ,help=( '''If > 0: set total number of training steps to perform. Override num_train_epochs.''' ) ,) parser.add_argument( '''--gradient_accumulation_steps''' ,type=_UpperCamelCase ,default=1 ,help='''Number of updates steps to accumulate before performing a backward/update pass.''' ,) parser.add_argument('''--learning_rate''' ,type=_UpperCamelCase ,default=6.25e-5 ) parser.add_argument('''--warmup_steps''' ,default=0 ,type=_UpperCamelCase ,help='''Linear warmup over warmup_steps.''' ) parser.add_argument('''--lr_schedule''' ,type=_UpperCamelCase ,default='''warmup_linear''' ) parser.add_argument('''--weight_decay''' ,type=_UpperCamelCase ,default=0.01 ) parser.add_argument('''--lm_coef''' ,type=_UpperCamelCase ,default=0.9 ) parser.add_argument('''--n_valid''' ,type=_UpperCamelCase ,default=3_74 ) parser.add_argument('''--server_ip''' ,type=_UpperCamelCase ,default='''''' ,help='''Can be used for distant debugging.''' ) parser.add_argument('''--server_port''' ,type=_UpperCamelCase ,default='''''' ,help='''Can be used for distant debugging.''' ) __lowerCamelCase = parser.parse_args() print(_UpperCamelCase ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('''Waiting for debugger attach''' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) ,redirect_output=_UpperCamelCase ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) __lowerCamelCase = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) __lowerCamelCase = torch.cuda.device_count() logger.info('''device: {}, n_gpu {}'''.format(_UpperCamelCase ,_UpperCamelCase ) ) if not args.do_train and not args.do_eval: raise ValueError('''At least one of `do_train` or `do_eval` must be True.''' ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset __lowerCamelCase = ['''_start_''', '''_delimiter_''', '''_classify_'''] __lowerCamelCase = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(_UpperCamelCase ) __lowerCamelCase = tokenizer.convert_tokens_to_ids(_UpperCamelCase ) __lowerCamelCase = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(_UpperCamelCase ) ) model.to(_UpperCamelCase ) # Load and encode the datasets def tokenize_and_encode(_UpperCamelCase : Dict ): if isinstance(_UpperCamelCase ,_UpperCamelCase ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(_UpperCamelCase ) ) elif isinstance(_UpperCamelCase ,_UpperCamelCase ): return obj return [tokenize_and_encode(_UpperCamelCase ) for o in obj] logger.info('''Encoding dataset...''' ) __lowerCamelCase = load_rocstories_dataset(args.train_dataset ) __lowerCamelCase = load_rocstories_dataset(args.eval_dataset ) __lowerCamelCase = (train_dataset, eval_dataset) __lowerCamelCase = tokenize_and_encode(_UpperCamelCase ) # Compute the max input length for the Transformer __lowerCamelCase = model.config.n_positions // 2 - 2 __lowerCamelCase = max( len(story[:max_length] ) + max(len(conta[:max_length] ) ,len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) __lowerCamelCase = min(_UpperCamelCase ,model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders __lowerCamelCase = pre_process_datasets(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ,*_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase = tensor_datasets[0], tensor_datasets[1] __lowerCamelCase = TensorDataset(*_UpperCamelCase ) __lowerCamelCase = RandomSampler(_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,sampler=_UpperCamelCase ,batch_size=args.train_batch_size ) __lowerCamelCase = TensorDataset(*_UpperCamelCase ) __lowerCamelCase = SequentialSampler(_UpperCamelCase ) __lowerCamelCase = DataLoader(_UpperCamelCase ,sampler=_UpperCamelCase ,batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: __lowerCamelCase = args.max_steps __lowerCamelCase = args.max_steps // (len(_UpperCamelCase ) // args.gradient_accumulation_steps) + 1 else: __lowerCamelCase = len(_UpperCamelCase ) // args.gradient_accumulation_steps * args.num_train_epochs __lowerCamelCase = list(model.named_parameters() ) __lowerCamelCase = ['''bias''', '''LayerNorm.bias''', '''LayerNorm.weight'''] __lowerCamelCase = [ { '''params''': [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], '''weight_decay''': args.weight_decay, }, {'''params''': [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0}, ] __lowerCamelCase = AdamW(_UpperCamelCase ,lr=args.learning_rate ,eps=args.adam_epsilon ) __lowerCamelCase = get_linear_schedule_with_warmup( _UpperCamelCase ,num_warmup_steps=args.warmup_steps ,num_training_steps=_UpperCamelCase ) if args.do_train: __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) ,desc='''Epoch''' ): __lowerCamelCase = 0 __lowerCamelCase = 0 __lowerCamelCase = tqdm(_UpperCamelCase ,desc='''Training''' ) for step, batch in enumerate(_UpperCamelCase ): __lowerCamelCase = tuple(t.to(_UpperCamelCase ) for t in batch ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = batch __lowerCamelCase = model(_UpperCamelCase ,mc_token_ids=_UpperCamelCase ,lm_labels=_UpperCamelCase ,mc_labels=_UpperCamelCase ) __lowerCamelCase = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() __lowerCamelCase = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 __lowerCamelCase = '''Training loss: {:.2e} lr: {:.2e}'''.format(_UpperCamelCase ,scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer __lowerCamelCase = model.module if hasattr(_UpperCamelCase ,'''module''' ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` __lowerCamelCase = os.path.join(args.output_dir ,_UpperCamelCase ) __lowerCamelCase = os.path.join(args.output_dir ,_UpperCamelCase ) torch.save(model_to_save.state_dict() ,_UpperCamelCase ) model_to_save.config.to_json_file(_UpperCamelCase ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned __lowerCamelCase = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) __lowerCamelCase = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(_UpperCamelCase ) if args.do_eval: model.eval() __lowerCamelCase ,__lowerCamelCase = 0, 0 __lowerCamelCase ,__lowerCamelCase = 0, 0 for batch in tqdm(_UpperCamelCase ,desc='''Evaluating''' ): __lowerCamelCase = tuple(t.to(_UpperCamelCase ) for t in batch ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = batch with torch.no_grad(): __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = model( _UpperCamelCase ,mc_token_ids=_UpperCamelCase ,lm_labels=_UpperCamelCase ,mc_labels=_UpperCamelCase ) __lowerCamelCase = mc_logits.detach().cpu().numpy() __lowerCamelCase = mc_labels.to('''cpu''' ).numpy() __lowerCamelCase = accuracy(_UpperCamelCase ,_UpperCamelCase ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 __lowerCamelCase = eval_loss / nb_eval_steps __lowerCamelCase = eval_accuracy / nb_eval_examples __lowerCamelCase = tr_loss / nb_tr_steps if args.do_train else None __lowerCamelCase = {'''eval_loss''': eval_loss, '''eval_accuracy''': eval_accuracy, '''train_loss''': train_loss} __lowerCamelCase = os.path.join(args.output_dir ,'''eval_results.txt''' ) with open(_UpperCamelCase ,'''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key in sorted(result.keys() ): logger.info(''' %s = %s''' ,_UpperCamelCase ,str(result[key] ) ) writer.write('''%s = %s\n''' % (key, str(result[key] )) ) if __name__ == "__main__": main()
622
0
import random def a__ ( _UpperCamelCase : Union[str, Any] ,_UpperCamelCase : List[Any] ,_UpperCamelCase : List[Any] ): __lowerCamelCase = a[left_index] __lowerCamelCase = left_index + 1 for j in range(left_index + 1 ,a_ ): if a[j] < pivot: __lowerCamelCase = a[i], a[j] i += 1 __lowerCamelCase = a[i - 1], a[left_index] return i - 1 def a__ ( _UpperCamelCase : List[Any] ,_UpperCamelCase : List[Any] ,_UpperCamelCase : List[Any] ): if left < right: __lowerCamelCase = random.randint(a_ ,right - 1 ) __lowerCamelCase = ( a[left], a[pivot], ) # switches the pivot with the left most bound __lowerCamelCase = partition(a_ ,a_ ,a_ ) quick_sort_random( a_ ,a_ ,a_ ) # recursive quicksort to the left of the pivot point quick_sort_random( a_ ,pivot_index + 1 ,a_ ) # recursive quicksort to the right of the pivot point def a__ ( ): __lowerCamelCase = input('''Enter numbers separated by a comma:\n''' ).strip() __lowerCamelCase = [int(a_ ) for item in user_input.split(''',''' )] quick_sort_random(a_ ,0 ,len(a_ ) ) print(a_ ) if __name__ == "__main__": main()
713
import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class __lowerCAmelCase ( lowerCAmelCase__ ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=1024 , __UpperCAmelCase=1024 , __UpperCAmelCase=3.6 ): '''simple docstring''' __lowerCamelCase = tokenizer __lowerCamelCase = tokenizer.bos_token_id __lowerCamelCase = dataset __lowerCamelCase = seq_length __lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self ): '''simple docstring''' __lowerCamelCase = iter(self.dataset ) __lowerCamelCase = True while more_examples: __lowerCamelCase ,__lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(__UpperCAmelCase )['''content'''] ) buffer_len += len(buffer[-1] ) except StopIteration: __lowerCamelCase = False break __lowerCamelCase = tokenizer(__UpperCAmelCase , truncation=__UpperCAmelCase )['''input_ids'''] __lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(__UpperCAmelCase ) , self.seq_length ): __lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(__UpperCAmelCase ) == self.seq_length: yield torch.tensor(__UpperCAmelCase ) def a__ ( _UpperCamelCase : List[Any] ): __lowerCamelCase = {'''streaming''': True} __lowerCamelCase = load_dataset(args.dataset_name ,split='''train''' ,**_UpperCamelCase ) __lowerCamelCase = ConstantLengthDataset(_UpperCamelCase ,_UpperCamelCase ,seq_length=args.seq_length ) __lowerCamelCase = DataLoader(_UpperCamelCase ,batch_size=args.batch_size ) return eval_dataloader def a__ ( _UpperCamelCase : str ): model.eval() __lowerCamelCase = [] for step, batch in enumerate(_UpperCamelCase ): with torch.no_grad(): __lowerCamelCase = model(_UpperCamelCase ,labels=_UpperCamelCase ) __lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(_UpperCamelCase ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break __lowerCamelCase = torch.mean(torch.cat(_UpperCamelCase ) ) try: __lowerCamelCase = torch.exp(_UpperCamelCase ) except OverflowError: __lowerCamelCase = float('''inf''' ) return loss.item(), perplexity.item() # Setup Accelerator a_ = Accelerator() # Parse configuration a_ = HfArgumentParser(EvaluationArguments) a_ = parser.parse_args() set_seed(args.seed) # Logging a_ = logging.getLogger(__name__) logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO ) # Load model and tokenizer a_ = AutoModelForCausalLM.from_pretrained(args.model_ckpt) a_ = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader a_ = create_dataloader(args) # Prepare everything with our `accelerator`. a_ , a_ = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info("""Evaluating and saving model after training""") a_ , a_ = evaluate(args) logger.info(f"loss/eval: {eval_loss}, perplexity: {perplexity}")
622
0
import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging a_ = logging.get_logger(__name__) # pylint: disable=invalid-name class __lowerCAmelCase ( _a ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase=768 ): '''simple docstring''' super().__init__(_A ) __lowerCamelCase = proj_size __lowerCamelCase = CLIPVisionModel(_A ) __lowerCamelCase = PaintByExampleMapper(_A ) __lowerCamelCase = nn.LayerNorm(config.hidden_size ) __lowerCamelCase = nn.Linear(config.hidden_size , self.proj_size ) # uncondition for scaling __lowerCamelCase = nn.Parameter(torch.randn((1, 1, self.proj_size) ) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase=False ): '''simple docstring''' __lowerCamelCase = self.model(pixel_values=_A ) __lowerCamelCase = clip_output.pooler_output __lowerCamelCase = self.mapper(latent_states[:, None] ) __lowerCamelCase = self.final_layer_norm(_A ) __lowerCamelCase = self.proj_out(_A ) if return_uncond_vector: return latent_states, self.uncond_vector return latent_states class __lowerCAmelCase ( nn.Module ): def __init__( self , __UpperCAmelCase ): '''simple docstring''' super().__init__() __lowerCamelCase = (config.num_hidden_layers + 1) // 5 __lowerCamelCase = config.hidden_size __lowerCamelCase = 1 __lowerCamelCase = nn.ModuleList( [ BasicTransformerBlock(_A , _A , _A , activation_fn='''gelu''' , attention_bias=_A ) for _ in range(_A ) ] ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' for block in self.blocks: __lowerCamelCase = block(_A ) return hidden_states
714
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """lxmert""" lowerCAmelCase__ = {} def __init__( self , __UpperCAmelCase=30522 , __UpperCAmelCase=768 , __UpperCAmelCase=12 , __UpperCAmelCase=9500 , __UpperCAmelCase=1600 , __UpperCAmelCase=400 , __UpperCAmelCase=3072 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=9 , __UpperCAmelCase=5 , __UpperCAmelCase=5 , __UpperCAmelCase=2048 , __UpperCAmelCase=4 , __UpperCAmelCase=6.67 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = num_qa_labels __lowerCamelCase = num_object_labels __lowerCamelCase = num_attr_labels __lowerCamelCase = l_layers __lowerCamelCase = x_layers __lowerCamelCase = r_layers __lowerCamelCase = visual_feat_dim __lowerCamelCase = visual_pos_dim __lowerCamelCase = visual_loss_normalizer __lowerCamelCase = task_matched __lowerCamelCase = task_mask_lm __lowerCamelCase = task_obj_predict __lowerCamelCase = task_qa __lowerCamelCase = visual_obj_loss __lowerCamelCase = visual_attr_loss __lowerCamelCase = visual_feat_loss __lowerCamelCase = {'''vision''': r_layers, '''cross_encoder''': x_layers, '''language''': l_layers} super().__init__(**__UpperCAmelCase )
622
0
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() a_ = logging.get_logger(__name__) a_ = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """adapter_layer""": """encoder.layers.*.adapter_layer""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", """pooling_layer.linear""": """projector""", """pooling_layer.projection""": """classifier""", } a_ = [ """lm_head""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", """projector""", """classifier""", ] def a__ ( _UpperCamelCase : Optional[int] ): __lowerCamelCase = {} with open(UpperCAmelCase__ ,'''r''' ) as file: for line_number, line in enumerate(UpperCAmelCase__ ): __lowerCamelCase = line.strip() if line: __lowerCamelCase = line.split() __lowerCamelCase = line_number __lowerCamelCase = words[0] __lowerCamelCase = value return result def a__ ( _UpperCamelCase : Optional[int] ,_UpperCamelCase : Tuple ,_UpperCamelCase : str ,_UpperCamelCase : List[str] ,_UpperCamelCase : int ): for attribute in key.split('''.''' ): __lowerCamelCase = getattr(UpperCAmelCase__ ,UpperCAmelCase__ ) __lowerCamelCase = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(UpperCAmelCase__ ): __lowerCamelCase = PARAM_MAPPING[full_name.split('''.''' )[-1]] __lowerCamelCase = '''param''' if weight_type is not None and weight_type != "param": __lowerCamelCase = getattr(UpperCAmelCase__ ,UpperCAmelCase__ ).shape elif weight_type is not None and weight_type == "param": __lowerCamelCase = hf_pointer for attribute in hf_param_name.split('''.''' ): __lowerCamelCase = getattr(UpperCAmelCase__ ,UpperCAmelCase__ ) __lowerCamelCase = shape_pointer.shape # let's reduce dimension __lowerCamelCase = value[0] else: __lowerCamelCase = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": __lowerCamelCase = value elif weight_type == "weight_g": __lowerCamelCase = value elif weight_type == "weight_v": __lowerCamelCase = value elif weight_type == "bias": __lowerCamelCase = value elif weight_type == "param": for attribute in hf_param_name.split('''.''' ): __lowerCamelCase = getattr(UpperCAmelCase__ ,UpperCAmelCase__ ) __lowerCamelCase = value else: __lowerCamelCase = value logger.info(F"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : List[Any] ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Dict ): __lowerCamelCase = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(UpperCAmelCase__ ): __lowerCamelCase = PARAM_MAPPING[full_name.split('''.''' )[-1]] __lowerCamelCase = '''param''' if weight_type is not None and weight_type != "param": __lowerCamelCase = '''.'''.join([key, weight_type] ) elif weight_type is not None and weight_type == "param": __lowerCamelCase = '''.'''.join([key, hf_param_name] ) else: __lowerCamelCase = key __lowerCamelCase = value if '''lm_head''' in full_key else value[0] a_ = { """W_a""": """linear_1.weight""", """W_b""": """linear_2.weight""", """b_a""": """linear_1.bias""", """b_b""": """linear_2.bias""", """ln_W""": """norm.weight""", """ln_b""": """norm.bias""", } def a__ ( _UpperCamelCase : Optional[Any] ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : Optional[Any]=None ,_UpperCamelCase : Any=None ): __lowerCamelCase = False for key, mapped_key in MAPPING.items(): __lowerCamelCase = '''wav2vec2.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: __lowerCamelCase = True if "*" in mapped_key: __lowerCamelCase = name.split(UpperCAmelCase__ )[0].split('''.''' )[-2] __lowerCamelCase = mapped_key.replace('''*''' ,UpperCAmelCase__ ) if "weight_g" in name: __lowerCamelCase = '''weight_g''' elif "weight_v" in name: __lowerCamelCase = '''weight_v''' elif "bias" in name: __lowerCamelCase = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj __lowerCamelCase = '''weight''' else: __lowerCamelCase = None if hf_dict is not None: rename_dict(UpperCAmelCase__ ,UpperCAmelCase__ ,UpperCAmelCase__ ,UpperCAmelCase__ ,UpperCAmelCase__ ) else: set_recursively(UpperCAmelCase__ ,UpperCAmelCase__ ,UpperCAmelCase__ ,UpperCAmelCase__ ,UpperCAmelCase__ ) return is_used return is_used def a__ ( _UpperCamelCase : Optional[Any] ,_UpperCamelCase : List[str] ,_UpperCamelCase : Tuple ): __lowerCamelCase = [] __lowerCamelCase = fairseq_model.state_dict() __lowerCamelCase = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): __lowerCamelCase = False if "conv_layers" in name: load_conv_layer( UpperCAmelCase__ ,UpperCAmelCase__ ,UpperCAmelCase__ ,UpperCAmelCase__ ,hf_model.config.feat_extract_norm == '''group''' ,) __lowerCamelCase = True else: __lowerCamelCase = load_wavaveca_layer(UpperCAmelCase__ ,UpperCAmelCase__ ,UpperCAmelCase__ ) if not is_used: unused_weights.append(UpperCAmelCase__ ) logger.warning(F"""Unused weights: {unused_weights}""" ) def a__ ( _UpperCamelCase : List[Any] ,_UpperCamelCase : Dict ,_UpperCamelCase : Any ,_UpperCamelCase : int ,_UpperCamelCase : int ): __lowerCamelCase = full_name.split('''conv_layers.''' )[-1] __lowerCamelCase = name.split('''.''' ) __lowerCamelCase = int(items[0] ) __lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __lowerCamelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __lowerCamelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.""" ) __lowerCamelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.""" ) __lowerCamelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(UpperCAmelCase__ ) @torch.no_grad() def a__ ( _UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Any ,_UpperCamelCase : List[str]=None ,_UpperCamelCase : Optional[Any]=None ,_UpperCamelCase : Union[str, Any]=True ,_UpperCamelCase : List[str]=False ): if config_path is not None: __lowerCamelCase = WavaVecaConfig.from_pretrained(UpperCAmelCase__ ) else: __lowerCamelCase = WavaVecaConfig() if is_seq_class: __lowerCamelCase = read_txt_into_dict(UpperCAmelCase__ ) __lowerCamelCase = idalabel __lowerCamelCase = WavaVecaForSequenceClassification(UpperCAmelCase__ ) __lowerCamelCase = WavaVecaFeatureExtractor( feature_size=1 ,sampling_rate=1_60_00 ,padding_value=0 ,do_normalize=UpperCAmelCase__ ,return_attention_mask=UpperCAmelCase__ ,) feature_extractor.save_pretrained(UpperCAmelCase__ ) elif is_finetuned: if dict_path: __lowerCamelCase = Dictionary.load(UpperCAmelCase__ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __lowerCamelCase = target_dict.pad_index __lowerCamelCase = target_dict.bos_index __lowerCamelCase = target_dict.eos_index __lowerCamelCase = len(target_dict.symbols ) __lowerCamelCase = os.path.join(UpperCAmelCase__ ,'''vocab.json''' ) if not os.path.isdir(UpperCAmelCase__ ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(UpperCAmelCase__ ) ) return os.makedirs(UpperCAmelCase__ ,exist_ok=UpperCAmelCase__ ) __lowerCamelCase = target_dict.indices # fairseq has the <pad> and <s> switched __lowerCamelCase = 0 __lowerCamelCase = 1 with open(UpperCAmelCase__ ,'''w''' ,encoding='''utf-8''' ) as vocab_handle: json.dump(UpperCAmelCase__ ,UpperCAmelCase__ ) __lowerCamelCase = WavaVecaCTCTokenizer( UpperCAmelCase__ ,unk_token=target_dict.unk_word ,pad_token=target_dict.pad_word ,bos_token=target_dict.bos_word ,eos_token=target_dict.eos_word ,word_delimiter_token='''|''' ,do_lower_case=UpperCAmelCase__ ,) __lowerCamelCase = True if config.feat_extract_norm == '''layer''' else False __lowerCamelCase = WavaVecaFeatureExtractor( feature_size=1 ,sampling_rate=1_60_00 ,padding_value=0 ,do_normalize=UpperCAmelCase__ ,return_attention_mask=UpperCAmelCase__ ,) __lowerCamelCase = WavaVecaProcessor(feature_extractor=UpperCAmelCase__ ,tokenizer=UpperCAmelCase__ ) processor.save_pretrained(UpperCAmelCase__ ) __lowerCamelCase = WavaVecaForCTC(UpperCAmelCase__ ) else: __lowerCamelCase = WavaVecaForPreTraining(UpperCAmelCase__ ) if is_finetuned or is_seq_class: __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] ,arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: __lowerCamelCase = argparse.Namespace(task='''audio_pretraining''' ) __lowerCamelCase = fairseq.tasks.setup_task(UpperCAmelCase__ ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ,task=UpperCAmelCase__ ) __lowerCamelCase = model[0].eval() recursively_load_weights(UpperCAmelCase__ ,UpperCAmelCase__ ,not is_finetuned ) hf_wavavec.save_pretrained(UpperCAmelCase__ ) if __name__ == "__main__": a_ = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--not_finetuned""", action="""store_true""", help="""Whether the model to convert is a fine-tuned model or not""" ) parser.add_argument( """--is_seq_class""", action="""store_true""", help="""Whether the model to convert is a fine-tuned sequence classification model or not""", ) a_ = parser.parse_args() a_ = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
715
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def a__ ( _UpperCamelCase : Tuple ,_UpperCamelCase : Any ,_UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Any ): if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = np.full((len(_UpperCamelCase ), sequence_length, 2) ,_UpperCamelCase ) else: __lowerCamelCase = np.full((len(_UpperCamelCase ), sequence_length) ,_UpperCamelCase ) for i, tensor in enumerate(_UpperCamelCase ): if padding_side == "right": if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = tensor[:sequence_length] else: __lowerCamelCase = tensor[:sequence_length] else: if isinstance(_UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase = tensor[:sequence_length] else: __lowerCamelCase = tensor[:sequence_length] return out_tensor.tolist() def a__ ( _UpperCamelCase : Dict ): __lowerCamelCase = ord(_UpperCamelCase ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 1_23 and cp <= 1_26): return True __lowerCamelCase = unicodedata.category(_UpperCamelCase ) if cat.startswith('''P''' ): return True return False @dataclass class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = 42 lowerCAmelCase__ = True lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = -1_0_0 lowerCAmelCase__ = "pt" def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' import torch __lowerCamelCase = '''label''' if '''label''' in features[0].keys() else '''labels''' __lowerCamelCase = [feature[label_name] for feature in features] if label_name in features[0].keys() else None __lowerCamelCase = self.tokenizer.pad( __UpperCAmelCase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch __lowerCamelCase = torch.tensor(batch['''entity_ids'''] ).shape[1] __lowerCamelCase = self.tokenizer.padding_side if padding_side == "right": __lowerCamelCase = [ list(__UpperCAmelCase ) + [self.label_pad_token_id] * (sequence_length - len(__UpperCAmelCase )) for label in labels ] else: __lowerCamelCase = [ [self.label_pad_token_id] * (sequence_length - len(__UpperCAmelCase )) + list(__UpperCAmelCase ) for label in labels ] __lowerCamelCase = [feature['''ner_tags'''] for feature in features] __lowerCamelCase = padding_tensor(__UpperCAmelCase , -1 , __UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = [feature['''original_entity_spans'''] for feature in features] __lowerCamelCase = padding_tensor(__UpperCAmelCase , (-1, -1) , __UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = {k: torch.tensor(__UpperCAmelCase , dtype=torch.intaa ) for k, v in batch.items()} return batch
622
0
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _interleave_iterable_datasets from .splits import NamedSplit from .utils import logging from .utils.py_utils import Literal a_ = logging.get_logger(__name__) a_ = TypeVar("""DatasetType""", Dataset, IterableDataset) def a__ ( _UpperCamelCase : List[DatasetType] ,_UpperCamelCase : Optional[List[float]] = None ,_UpperCamelCase : Optional[int] = None ,_UpperCamelCase : Optional[DatasetInfo] = None ,_UpperCamelCase : Optional[NamedSplit] = None ,_UpperCamelCase : Literal["first_exhausted", "all_exhausted"] = "first_exhausted" ,): from .arrow_dataset import Dataset from .iterable_dataset import IterableDataset if not datasets: raise ValueError('''Unable to interleave an empty list of datasets.''' ) for i, dataset in enumerate(_UpperCamelCase ): if not isinstance(_UpperCamelCase ,(Dataset, IterableDataset) ): if isinstance(_UpperCamelCase ,(DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} """ '''is an empty dataset dictionary.''' ) raise ValueError( F"""Dataset at position {i} has at least one split: {list(_UpperCamelCase )}\n""" F"""Please pick one to interleave with the other datasets, for example: dataset['{next(iter(_UpperCamelCase ) )}']""" ) raise ValueError( F"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(_UpperCamelCase ).__name__}.""" ) if i == 0: __lowerCamelCase ,__lowerCamelCase = ( (Dataset, IterableDataset) if isinstance(_UpperCamelCase ,_UpperCamelCase ) else (IterableDataset, Dataset) ) elif not isinstance(_UpperCamelCase ,_UpperCamelCase ): raise ValueError( F"""Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.""" ) if stopping_strategy not in ["first_exhausted", "all_exhausted"]: raise ValueError(F"""{stopping_strategy} is not supported. Please enter a valid stopping_strategy.""" ) if dataset_type is Dataset: return _interleave_map_style_datasets( _UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ,info=_UpperCamelCase ,split=_UpperCamelCase ,stopping_strategy=_UpperCamelCase ) else: return _interleave_iterable_datasets( _UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ,info=_UpperCamelCase ,split=_UpperCamelCase ,stopping_strategy=_UpperCamelCase ) def a__ ( _UpperCamelCase : List[DatasetType] ,_UpperCamelCase : Optional[DatasetInfo] = None ,_UpperCamelCase : Optional[NamedSplit] = None ,_UpperCamelCase : int = 0 ,): if not dsets: raise ValueError('''Unable to concatenate an empty list of datasets.''' ) for i, dataset in enumerate(_UpperCamelCase ): if not isinstance(_UpperCamelCase ,(Dataset, IterableDataset) ): if isinstance(_UpperCamelCase ,(DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} """ '''is an empty dataset dictionary.''' ) raise ValueError( F"""Dataset at position {i} has at least one split: {list(_UpperCamelCase )}\n""" F"""Please pick one to interleave with the other datasets, for example: dataset['{next(iter(_UpperCamelCase ) )}']""" ) raise ValueError( F"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(_UpperCamelCase ).__name__}.""" ) if i == 0: __lowerCamelCase ,__lowerCamelCase = ( (Dataset, IterableDataset) if isinstance(_UpperCamelCase ,_UpperCamelCase ) else (IterableDataset, Dataset) ) elif not isinstance(_UpperCamelCase ,_UpperCamelCase ): raise ValueError( F"""Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.""" ) if dataset_type is Dataset: return _concatenate_map_style_datasets(_UpperCamelCase ,info=_UpperCamelCase ,split=_UpperCamelCase ,axis=_UpperCamelCase ) else: return _concatenate_iterable_datasets(_UpperCamelCase ,info=_UpperCamelCase ,split=_UpperCamelCase ,axis=_UpperCamelCase )
716
from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=99 , __UpperCAmelCase=[1, 1, 2] , __UpperCAmelCase=1 , __UpperCAmelCase=32 , __UpperCAmelCase=4 , __UpperCAmelCase=8 , __UpperCAmelCase=37 , __UpperCAmelCase="gelu_new" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.0 , __UpperCAmelCase=512 , __UpperCAmelCase=3 , __UpperCAmelCase=0.02 , __UpperCAmelCase=3 , __UpperCAmelCase=4 , __UpperCAmelCase=None , __UpperCAmelCase=False , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_input_mask __lowerCamelCase = use_token_type_ids __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = block_sizes __lowerCamelCase = num_decoder_layers __lowerCamelCase = d_model __lowerCamelCase = n_head __lowerCamelCase = d_head __lowerCamelCase = d_inner __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout __lowerCamelCase = attention_dropout __lowerCamelCase = activation_dropout __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = 2 __lowerCamelCase = num_labels __lowerCamelCase = num_choices __lowerCamelCase = scope __lowerCamelCase = initializer_std # Used in the tests to check the size of the first attention layer __lowerCamelCase = n_head # Used in the tests to check the size of the first hidden state __lowerCamelCase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowerCamelCase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowerCamelCase = self.num_hidden_layers + 2 def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCamelCase = None if self.use_token_type_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCamelCase = FunnelConfig( vocab_size=self.vocab_size , block_sizes=self.block_sizes , num_decoder_layers=self.num_decoder_layers , d_model=self.d_model , n_head=self.n_head , d_head=self.d_head , d_inner=self.d_inner , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , activation_dropout=self.activation_dropout , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_std=self.initializer_std , ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = [input_ids, input_mask] __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = [input_ids, input_mask] __lowerCamelCase = model(__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 3, self.d_model) ) __lowerCamelCase = False __lowerCamelCase = TFFunnelBaseModel(config=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForPreTraining(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForMaskedLM(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = TFFunnelForSequenceClassification(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_choices __lowerCamelCase = TFFunnelForMultipleChoice(config=__UpperCAmelCase ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowerCamelCase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = TFFunnelForTokenClassification(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = TFFunnelForQuestionAnswering(config=__UpperCAmelCase ) __lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.prepare_config_and_inputs() ( ( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) , ) = config_and_inputs __lowerCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) lowerCAmelCase__ = ( { """feature-extraction""": (TFFunnelBaseModel, TFFunnelModel), """fill-mask""": TFFunnelForMaskedLM, """question-answering""": TFFunnelForQuestionAnswering, """text-classification""": TFFunnelForSequenceClassification, """token-classification""": TFFunnelForTokenClassification, """zero-shot""": TFFunnelForSequenceClassification, } if is_tf_available() else {} ) lowerCAmelCase__ = False lowerCAmelCase__ = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFFunnelModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCAmelCase ) @require_tf class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) lowerCAmelCase__ = False lowerCAmelCase__ = False def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = TFFunnelModelTester(self , base=__UpperCAmelCase ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCAmelCase )
622
0
import numpy as np from sklearn.datasets import fetch_california_housing from sklearn.metrics import mean_absolute_error, mean_squared_error from sklearn.model_selection import train_test_split from xgboost import XGBRegressor def a__ ( _UpperCamelCase : Optional[Any] ): return (data["data"], data["target"]) def a__ ( _UpperCamelCase : Union[str, Any] ,_UpperCamelCase : Tuple ,_UpperCamelCase : List[str] ): __lowerCamelCase = XGBRegressor(verbosity=0 ,random_state=42 ) xgb.fit(_UpperCamelCase ,_UpperCamelCase ) # Predict target for test data __lowerCamelCase = xgb.predict(_UpperCamelCase ) __lowerCamelCase = predictions.reshape(len(_UpperCamelCase ) ,1 ) return predictions def a__ ( ): __lowerCamelCase = fetch_california_housing() __lowerCamelCase ,__lowerCamelCase = data_handling(_UpperCamelCase ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = train_test_split( _UpperCamelCase ,_UpperCamelCase ,test_size=0.25 ,random_state=1 ) __lowerCamelCase = xgboost(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ) # Error printing print(F"""Mean Absolute Error : {mean_absolute_error(_UpperCamelCase ,_UpperCamelCase )}""" ) print(F"""Mean Square Error : {mean_squared_error(_UpperCamelCase ,_UpperCamelCase )}""" ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
717
from collections import namedtuple import requests from lxml import html # type: ignore a_ = namedtuple("""covid_data""", """cases deaths recovered""") def a__ ( _UpperCamelCase : str = "https://www.worldometers.info/coronavirus/" ): __lowerCamelCase = '''//div[@class = "maincounter-number"]/span/text()''' return covid_data(*html.fromstring(requests.get(_UpperCamelCase ).content ).xpath(_UpperCamelCase ) ) a_ = """Total COVID-19 cases in the world: {} Total deaths due to COVID-19 in the world: {} Total COVID-19 patients recovered in the world: {}""" print(fmt.format(*covid_stats()))
622
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import RoFormerTokenizer from .tokenization_utils import JiebaPreTokenizer a_ = logging.get_logger(__name__) a_ = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} a_ = { "vocab_file": { "junnyu/roformer_chinese_small": "https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/vocab.txt", "junnyu/roformer_chinese_base": "https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/vocab.txt", "junnyu/roformer_chinese_char_small": ( "https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/vocab.txt" ), "junnyu/roformer_chinese_char_base": ( "https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/vocab.txt" ), "junnyu/roformer_small_discriminator": ( "https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/vocab.txt" ), "junnyu/roformer_small_generator": ( "https://huggingface.co/junnyu/roformer_small_generator/resolve/main/vocab.txt" ), } } a_ = { "junnyu/roformer_chinese_small": 1_536, "junnyu/roformer_chinese_base": 1_536, "junnyu/roformer_chinese_char_small": 512, "junnyu/roformer_chinese_char_base": 512, "junnyu/roformer_small_discriminator": 128, "junnyu/roformer_small_generator": 128, } a_ = { "junnyu/roformer_chinese_small": {"do_lower_case": True}, "junnyu/roformer_chinese_base": {"do_lower_case": True}, "junnyu/roformer_chinese_char_small": {"do_lower_case": True}, "junnyu/roformer_chinese_char_base": {"do_lower_case": True}, "junnyu/roformer_small_discriminator": {"do_lower_case": True}, "junnyu/roformer_small_generator": {"do_lower_case": True}, } class __lowerCAmelCase ( __lowerCAmelCase ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase__ = RoFormerTokenizer def __init__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=True , __UpperCAmelCase="[UNK]" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="[PAD]" , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[MASK]" , __UpperCAmelCase=True , __UpperCAmelCase=None , **__UpperCAmelCase , ): '''simple docstring''' super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) __lowerCamelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( pre_tok_state.get('''lowercase''' , lowerCamelCase__ ) != do_lower_case or pre_tok_state.get('''strip_accents''' , lowerCamelCase__ ) != strip_accents ): __lowerCamelCase = getattr(lowerCamelCase__ , pre_tok_state.pop('''type''' ) ) __lowerCamelCase = do_lower_case __lowerCamelCase = strip_accents __lowerCamelCase = pre_tok_class(**lowerCamelCase__ ) __lowerCamelCase = do_lower_case def __getstate__( self ): '''simple docstring''' __lowerCamelCase = self.__dict__.copy() __lowerCamelCase = BertPreTokenizer() return state def __setstate__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = d __lowerCamelCase = self.__dict__['''_tokenizer'''].get_vocab() __lowerCamelCase = PreTokenizer.custom(JiebaPreTokenizer(lowerCamelCase__ ) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase=None ): '''simple docstring''' __lowerCamelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=False , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = BertPreTokenizer() return super().save_pretrained(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ )
718
def a__ ( _UpperCamelCase : str ,_UpperCamelCase : str = " " ): __lowerCamelCase = [] __lowerCamelCase = 0 for index, char in enumerate(_UpperCamelCase ): if char == separator: split_words.append(string[last_index:index] ) __lowerCamelCase = index + 1 elif index + 1 == len(_UpperCamelCase ): split_words.append(string[last_index : index + 1] ) return split_words if __name__ == "__main__": from doctest import testmod testmod()
622
0
from unittest.mock import patch import pyspark from datasets.packaged_modules.spark.spark import ( Spark, SparkExamplesIterable, _generate_iterable_examples, ) from ..utils import ( require_dill_gt_0_3_2, require_not_windows, ) def a__ ( _UpperCamelCase : Dict ,_UpperCamelCase : int ): __lowerCamelCase = [] for part_id in partition_order: __lowerCamelCase = df.where(F"""SPARK_PARTITION_ID() = {part_id}""" ).collect() for row_idx, row in enumerate(snake_case_ ): expected_row_ids_and_row_dicts.append((F"""{part_id}_{row_idx}""", row.asDict()) ) return expected_row_ids_and_row_dicts @require_not_windows @require_dill_gt_0_3_2 def a__ ( ): __lowerCamelCase = pyspark.sql.SparkSession.builder.master('''local[*]''' ).appName('''pyspark''' ).getOrCreate() __lowerCamelCase = spark.range(1_00 ).repartition(1 ) __lowerCamelCase = Spark(snake_case_ ) # The id ints will be converted to Pyarrow int64s, so each row will be 8 bytes. Setting a max_shard_size of 16 means # that each partition can hold 2 rows. spark_builder._repartition_df_if_needed(max_shard_size=16 ) # Given that the dataframe has 100 rows and each partition has 2 rows, we expect 50 partitions. assert spark_builder.df.rdd.getNumPartitions() == 50 @require_not_windows @require_dill_gt_0_3_2 def a__ ( ): __lowerCamelCase = pyspark.sql.SparkSession.builder.master('''local[*]''' ).appName('''pyspark''' ).getOrCreate() __lowerCamelCase = spark.range(10 ).repartition(2 ) __lowerCamelCase = [1, 0] __lowerCamelCase = _generate_iterable_examples(snake_case_ ,snake_case_ ) # Reverse the partitions. __lowerCamelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(snake_case_ ,snake_case_ ) for i, (row_id, row_dict) in enumerate(generate_fn() ): __lowerCamelCase = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def a__ ( ): __lowerCamelCase = pyspark.sql.SparkSession.builder.master('''local[*]''' ).appName('''pyspark''' ).getOrCreate() __lowerCamelCase = spark.range(10 ).repartition(1 ) __lowerCamelCase = SparkExamplesIterable(snake_case_ ) assert it.n_shards == 1 for i, (row_id, row_dict) in enumerate(snake_case_ ): assert row_id == F"""0_{i}""" assert row_dict == {"id": i} @require_not_windows @require_dill_gt_0_3_2 def a__ ( ): __lowerCamelCase = pyspark.sql.SparkSession.builder.master('''local[*]''' ).appName('''pyspark''' ).getOrCreate() __lowerCamelCase = spark.range(30 ).repartition(3 ) # Mock the generator so that shuffle reverses the partition indices. with patch('''numpy.random.Generator''' ) as generator_mock: __lowerCamelCase = lambda _UpperCamelCase : x.reverse() __lowerCamelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(snake_case_ ,[2, 1, 0] ) __lowerCamelCase = SparkExamplesIterable(snake_case_ ).shuffle_data_sources(snake_case_ ) assert shuffled_it.n_shards == 3 for i, (row_id, row_dict) in enumerate(snake_case_ ): __lowerCamelCase = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def a__ ( ): __lowerCamelCase = pyspark.sql.SparkSession.builder.master('''local[*]''' ).appName('''pyspark''' ).getOrCreate() __lowerCamelCase = spark.range(20 ).repartition(4 ) # Partitions 0 and 2 __lowerCamelCase = SparkExamplesIterable(snake_case_ ).shard_data_sources(worker_id=0 ,num_workers=2 ) assert shard_it_a.n_shards == 2 __lowerCamelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(snake_case_ ,[0, 2] ) for i, (row_id, row_dict) in enumerate(snake_case_ ): __lowerCamelCase = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict # Partitions 1 and 3 __lowerCamelCase = SparkExamplesIterable(snake_case_ ).shard_data_sources(worker_id=1 ,num_workers=2 ) assert shard_it_a.n_shards == 2 __lowerCamelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(snake_case_ ,[1, 3] ) for i, (row_id, row_dict) in enumerate(snake_case_ ): __lowerCamelCase = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def a__ ( ): __lowerCamelCase = pyspark.sql.SparkSession.builder.master('''local[*]''' ).appName('''pyspark''' ).getOrCreate() __lowerCamelCase = spark.range(1_00 ).repartition(1 ) __lowerCamelCase = Spark(snake_case_ ) # Choose a small max_shard_size for maximum partitioning. spark_builder._repartition_df_if_needed(max_shard_size=1 ) # The new number of partitions should not be greater than the number of rows. assert spark_builder.df.rdd.getNumPartitions() == 1_00
719
import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class __lowerCAmelCase ( lowerCAmelCase__ ): def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = tempfile.mkdtemp() __lowerCamelCase = 8 # DPR tok __lowerCamelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] __lowerCamelCase = os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) __lowerCamelCase = os.path.join(__UpperCAmelCase , DPR_VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) # BART tok __lowerCamelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] __lowerCamelCase = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) __lowerCamelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] __lowerCamelCase = {'''unk_token''': '''<unk>'''} __lowerCamelCase = os.path.join(self.tmpdirname , '''bart_tokenizer''' ) os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) __lowerCamelCase = os.path.join(__UpperCAmelCase , BART_VOCAB_FILES_NAMES['''vocab_file'''] ) __lowerCamelCase = os.path.join(__UpperCAmelCase , BART_VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(__UpperCAmelCase ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(__UpperCAmelCase ) ) def lowerCamelCase ( self ): '''simple docstring''' return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''dpr_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''bart_tokenizer''' ) ) def lowerCamelCase ( self ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_dataset() __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: __lowerCamelCase = dataset __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.get_dummy_dataset() __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''custom''' , ) if from_disk: __lowerCamelCase = os.path.join(self.tmpdirname , '''dataset''' ) __lowerCamelCase = os.path.join(self.tmpdirname , '''index.faiss''' ) dataset.get_index('''embeddings''' ).save(os.path.join(self.tmpdirname , '''index.faiss''' ) ) dataset.drop_index('''embeddings''' ) dataset.save_to_disk(os.path.join(self.tmpdirname , '''dataset''' ) ) del dataset __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , __UpperCAmelCase ) , ) return retriever def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''text''': ['''foo''', '''bar'''], '''title''': ['''Foo''', '''Bar'''], '''embeddings''': [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index('''embeddings''' , string_factory='''Flat''' , metric_type=faiss.METRIC_INNER_PRODUCT ) __lowerCamelCase = os.path.join(self.tmpdirname , '''hf_bert_base.hnswSQ8_correct_phi_128.c_index''' ) dataset.save_faiss_index('''embeddings''' , index_file_name + '''.index.dpr''' ) pickle.dump(dataset['''id'''] , open(index_file_name + '''.index_meta.dpr''' , '''wb''' ) ) __lowerCamelCase = os.path.join(self.tmpdirname , '''psgs_w100.tsv.pkl''' ) __lowerCamelCase = {sample['''id''']: [sample['''text'''], sample['''title''']] for sample in dataset} pickle.dump(__UpperCAmelCase , open(__UpperCAmelCase , '''wb''' ) ) __lowerCamelCase = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name='''legacy''' , index_path=self.tmpdirname , ) __lowerCamelCase = RagRetriever( __UpperCAmelCase , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch('''transformers.models.rag.retrieval_rag.load_dataset''' ) as mock_load_dataset: __lowerCamelCase = self.get_dummy_dataset() retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''embeddings''', '''id''', '''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''id'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''id'''][0] , '''1''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''id'''][0] , '''0''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_legacy_index_retriever() __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=__UpperCAmelCase ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__UpperCAmelCase ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ['''text''', '''title'''] ) self.assertEqual(len(doc_dicts[0]['''text'''] ) , __UpperCAmelCase ) self.assertEqual(doc_dicts[0]['''text'''][0] , '''bar''' ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]['''text'''][0] , '''foo''' ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__UpperCAmelCase ) __lowerCamelCase = RagRetriever.from_pretrained(__UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever.retrieve(__UpperCAmelCase , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def lowerCamelCase ( self ): '''simple docstring''' import torch __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_canonical_hf_index_retriever() __lowerCamelCase = [[5, 7], [10, 11]] __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever(__UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = ( out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , __UpperCAmelCase ) self.assertIsInstance(__UpperCAmelCase , np.ndarray ) __lowerCamelCase = retriever( __UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase , return_tensors='''pt''' , ) __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = ( # noqa: F841 out['''context_input_ids'''], out['''context_attention_mask'''], out['''retrieved_doc_embeds'''], out['''doc_ids'''], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_dpr_ctx_encoder_tokenizer() __lowerCamelCase = 1 __lowerCamelCase = self.get_dummy_custom_hf_index_retriever(from_disk=__UpperCAmelCase ) retriever.set_ctx_encoder_tokenizer(__UpperCAmelCase ) __lowerCamelCase = [[5, 7], [10, 11]] __lowerCamelCase = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) __lowerCamelCase = retriever(__UpperCAmelCase , __UpperCAmelCase , prefix=retriever.config.generator.prefix , n_docs=__UpperCAmelCase ) self.assertEqual( len(__UpperCAmelCase ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ('''tokenized_doc_ids''', '''tokenized_doc_attention_mask''') ) , __UpperCAmelCase ) # check for doc token related keys in dictionary.
622
0
from argparse import ArgumentParser from . import BaseTransformersCLICommand def a__ ( _UpperCamelCase : Union[str, Any] ): return DownloadCommand(args.model ,args.cache_dir ,args.force ,args.trust_remote_code ) class __lowerCAmelCase ( lowercase_ ): @staticmethod def lowerCamelCase ( __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' , type=__UpperCAmelCase , default=__UpperCAmelCase , help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' , action='''store_true''' , help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' , action='''store_true''' , help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' , ) download_parser.add_argument('''model''' , type=__UpperCAmelCase , help='''Name of the model to download''' ) download_parser.set_defaults(func=__UpperCAmelCase ) def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = model __lowerCamelCase = cache __lowerCamelCase = force __lowerCamelCase = trust_remote_code def lowerCamelCase ( self ): '''simple docstring''' from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
720
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """sail/poolformer_s12""": """https://huggingface.co/sail/poolformer_s12/resolve/main/config.json""", # See all PoolFormer models at https://huggingface.co/models?filter=poolformer } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """poolformer""" def __init__( self , __UpperCAmelCase=3 , __UpperCAmelCase=16 , __UpperCAmelCase=16 , __UpperCAmelCase=3 , __UpperCAmelCase=4.0 , __UpperCAmelCase=[2, 2, 6, 2] , __UpperCAmelCase=[64, 128, 320, 512] , __UpperCAmelCase=[7, 3, 3, 3] , __UpperCAmelCase=[4, 2, 2, 2] , __UpperCAmelCase=[2, 1, 1, 1] , __UpperCAmelCase=4 , __UpperCAmelCase=0.0 , __UpperCAmelCase="gelu" , __UpperCAmelCase=True , __UpperCAmelCase=1E-5 , __UpperCAmelCase=0.02 , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = num_channels __lowerCamelCase = patch_size __lowerCamelCase = stride __lowerCamelCase = padding __lowerCamelCase = pool_size __lowerCamelCase = hidden_sizes __lowerCamelCase = mlp_ratio __lowerCamelCase = depths __lowerCamelCase = patch_sizes __lowerCamelCase = strides __lowerCamelCase = num_encoder_blocks __lowerCamelCase = drop_path_rate __lowerCamelCase = hidden_act __lowerCamelCase = use_layer_scale __lowerCamelCase = layer_scale_init_value __lowerCamelCase = initializer_range super().__init__(**__UpperCAmelCase ) class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = version.parse("""1.11""" ) @property def lowerCamelCase ( self ): '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def lowerCamelCase ( self ): '''simple docstring''' return 2E-3
622
0
def a__ ( _UpperCamelCase : List[str] ): __lowerCamelCase = '''''' for ch in key: if ch == " " or ch not in key_no_dups and ch.isalpha(): key_no_dups += ch return key_no_dups def a__ ( _UpperCamelCase : Optional[Any] ): __lowerCamelCase = [chr(i + 65 ) for i in range(26 )] # Remove duplicate characters from key __lowerCamelCase = remove_duplicates(key.upper() ) __lowerCamelCase = len(lowerCamelCase__ ) # First fill cipher with key characters __lowerCamelCase = {alphabet[i]: char for i, char in enumerate(lowerCamelCase__ )} # Then map remaining characters in alphabet to # the alphabet from the beginning for i in range(len(lowerCamelCase__ ) ,26 ): __lowerCamelCase = alphabet[i - offset] # Ensure we are not mapping letters to letters previously mapped while char in key: offset -= 1 __lowerCamelCase = alphabet[i - offset] __lowerCamelCase = char return cipher_alphabet def a__ ( _UpperCamelCase : str ,_UpperCamelCase : Dict ): return "".join(cipher_map.get(lowerCamelCase__ ,lowerCamelCase__ ) for ch in message.upper() ) def a__ ( _UpperCamelCase : int ,_UpperCamelCase : int ): __lowerCamelCase = {v: k for k, v in cipher_map.items()} return "".join(rev_cipher_map.get(lowerCamelCase__ ,lowerCamelCase__ ) for ch in message.upper() ) def a__ ( ): __lowerCamelCase = input('''Enter message to encode or decode: ''' ).strip() __lowerCamelCase = input('''Enter keyword: ''' ).strip() __lowerCamelCase = input('''Encipher or decipher? E/D:''' ).strip()[0].lower() try: __lowerCamelCase = {'''e''': encipher, '''d''': decipher}[option] except KeyError: raise KeyError('''invalid input option''' ) __lowerCamelCase = create_cipher_map(lowerCamelCase__ ) print(func(lowerCamelCase__ ,lowerCamelCase__ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
721
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { """uclanlp/visualbert-vqa""": """https://huggingface.co/uclanlp/visualbert-vqa/resolve/main/config.json""", """uclanlp/visualbert-vqa-pre""": """https://huggingface.co/uclanlp/visualbert-vqa-pre/resolve/main/config.json""", """uclanlp/visualbert-vqa-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-vqa-coco-pre/resolve/main/config.json""" ), """uclanlp/visualbert-vcr""": """https://huggingface.co/uclanlp/visualbert-vcr/resolve/main/config.json""", """uclanlp/visualbert-vcr-pre""": """https://huggingface.co/uclanlp/visualbert-vcr-pre/resolve/main/config.json""", """uclanlp/visualbert-vcr-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-vcr-coco-pre/resolve/main/config.json""" ), """uclanlp/visualbert-nlvr2""": """https://huggingface.co/uclanlp/visualbert-nlvr2/resolve/main/config.json""", """uclanlp/visualbert-nlvr2-pre""": """https://huggingface.co/uclanlp/visualbert-nlvr2-pre/resolve/main/config.json""", """uclanlp/visualbert-nlvr2-coco-pre""": ( """https://huggingface.co/uclanlp/visualbert-nlvr2-coco-pre/resolve/main/config.json""" ) # See all VisualBERT models at https://huggingface.co/models?filter=visual_bert } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = """visual_bert""" def __init__( self , __UpperCAmelCase=30522 , __UpperCAmelCase=768 , __UpperCAmelCase=512 , __UpperCAmelCase=12 , __UpperCAmelCase=12 , __UpperCAmelCase=3072 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase=1 , __UpperCAmelCase=0 , __UpperCAmelCase=2 , **__UpperCAmelCase , ): '''simple docstring''' super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) __lowerCamelCase = vocab_size __lowerCamelCase = max_position_embeddings __lowerCamelCase = hidden_size __lowerCamelCase = visual_embedding_dim __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = initializer_range __lowerCamelCase = type_vocab_size __lowerCamelCase = layer_norm_eps __lowerCamelCase = bypass_transformer __lowerCamelCase = special_visual_initialize
622
0
import qiskit def a__ ( _UpperCamelCase : int ,_UpperCamelCase : int ): __lowerCamelCase = qiskit.Aer.get_backend('''aer_simulator''' ) __lowerCamelCase = qiskit.QuantumCircuit(4 ,2 ) # encode inputs in qubits 0 and 1 if bita == 1: qc_ha.x(0 ) if bita == 1: qc_ha.x(1 ) qc_ha.barrier() # use cnots to write XOR of the inputs on qubit2 qc_ha.cx(0 ,2 ) qc_ha.cx(1 ,2 ) # use ccx / toffoli gate to write AND of the inputs on qubit3 qc_ha.ccx(0 ,1 ,3 ) qc_ha.barrier() # extract outputs qc_ha.measure(2 ,0 ) # extract XOR value qc_ha.measure(3 ,1 ) # extract AND value # Execute the circuit on the qasm simulator __lowerCamelCase = qiskit.execute(_UpperCamelCase ,_UpperCamelCase ,shots=10_00 ) # Return the histogram data of the results of the experiment return job.result().get_counts(_UpperCamelCase ) if __name__ == "__main__": a_ = half_adder(1, 1) print(f"Half Adder Output Qubit Counts: {counts}")
700
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging a_ = logging.get_logger(__name__) a_ = {"""vocab_file""": """spiece.model"""} a_ = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", } } a_ = { """albert-base-v1""": 512, """albert-large-v1""": 512, """albert-xlarge-v1""": 512, """albert-xxlarge-v1""": 512, """albert-base-v2""": 512, """albert-large-v2""": 512, """albert-xlarge-v2""": 512, """albert-xxlarge-v2""": 512, } a_ = """▁""" class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , __UpperCAmelCase , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="<unk>" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="<pad>" , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[MASK]" , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. __lowerCamelCase = ( AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase , normalized=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token ) __lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__UpperCAmelCase , remove_space=__UpperCAmelCase , keep_accents=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) __lowerCamelCase = do_lower_case __lowerCamelCase = remove_space __lowerCamelCase = keep_accents __lowerCamelCase = vocab_file __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCAmelCase ) @property def lowerCamelCase ( self ): '''simple docstring''' return len(self.sp_model ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): '''simple docstring''' __lowerCamelCase = self.__dict__.copy() __lowerCamelCase = None return state def __setstate__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCamelCase = {} __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' if self.remove_space: __lowerCamelCase = ''' '''.join(inputs.strip().split() ) else: __lowerCamelCase = inputs __lowerCamelCase = outputs.replace('''``''' , '''"''' ).replace('''\'\'''' , '''"''' ) if not self.keep_accents: __lowerCamelCase = unicodedata.normalize('''NFKD''' , __UpperCAmelCase ) __lowerCamelCase = ''''''.join([c for c in outputs if not unicodedata.combining(__UpperCAmelCase )] ) if self.do_lower_case: __lowerCamelCase = outputs.lower() return outputs def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.preprocess_text(__UpperCAmelCase ) __lowerCamelCase = self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) __lowerCamelCase = [] for piece in pieces: if len(__UpperCAmelCase ) > 1 and piece[-1] == str(''',''' ) and piece[-2].isdigit(): __lowerCamelCase = self.sp_model.EncodeAsPieces(piece[:-1].replace(__UpperCAmelCase , '''''' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __lowerCamelCase = cur_pieces[1:] else: __lowerCamelCase = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__UpperCAmelCase ) else: new_pieces.append(__UpperCAmelCase ) return new_pieces def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.PieceToId(__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.IdToPiece(__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = '''''' __lowerCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__UpperCAmelCase ) + token __lowerCamelCase = True __lowerCamelCase = [] else: current_sub_tokens.append(__UpperCAmelCase ) __lowerCamelCase = False out_string += self.sp_model.decode(__UpperCAmelCase ) return out_string.strip() def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase ) if token_ids_a is not None: return [1] + ([0] * len(__UpperCAmelCase )) + [1] + ([0] * len(__UpperCAmelCase )) + [1] return [1] + ([0] * len(__UpperCAmelCase )) + [1] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __lowerCamelCase = [self.sep_token_id] __lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' if not os.path.isdir(__UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCamelCase = os.path.join( __UpperCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , '''wb''' ) as fi: __lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,)
622
0