code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
import re def a_ ( lowerCAmelCase_ : str ): return [char.split() for char in re.split(R'[^ a-z A-Z 0-9 \s]', str_ )] def a_ ( lowerCAmelCase_ : str ): __lowerCAmelCase = split_input(str_ ) return "".join( [''.join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : bool, lowerCAmelCase_ : str ): try: __lowerCAmelCase = split_input(lowerCAmelCase_ ) if upper: __lowerCAmelCase = ''.join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: __lowerCAmelCase = ''.join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def a_ ( lowerCAmelCase_ : str ): return to_simple_case(lowerCAmelCase_ ) def a_ ( lowerCAmelCase_ : str ): try: __lowerCAmelCase = to_simple_case(lowerCAmelCase_ ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : bool ): return to_complex_case(lowerCAmelCase_, lowerCAmelCase_, '_' ) def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : bool ): return to_complex_case(lowerCAmelCase_, lowerCAmelCase_, '-' ) if __name__ == "__main__": __import__('doctest').testmod()
284
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
284
1
import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings _snake_case : List[str] = logging.getLogger(__name__) @dataclass @add_start_docstrings(TrainingArguments.__doc__ ) class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = field(default=_UpperCamelCase , metadata={"""help""": """Whether to use SortishSampler or not."""} ) a_ = field( default=_UpperCamelCase , metadata={"""help""": """Whether to use generate to calculate generative metrics (ROUGE, BLEU)."""} ) a_ = field( default=_UpperCamelCase , metadata={ """help""": ( """The `max_length` to use on each evaluation loop when `predict_with_generate=True`. Will default """ """to the `max_length` value of the model configuration.""" ) } , ) a_ = field( default=_UpperCamelCase , metadata={ """help""": ( """The `num_beams` to use on each evaluation loop when `predict_with_generate=True`. Will default """ """to the `num_beams` value of the model configuration.""" ) } , ) a_ = field( default=_UpperCamelCase , metadata={ """help""": """Model id, file path or url pointing to a GenerationConfig json file, to use during prediction.""" } , ) def lowercase ( self : Tuple ) -> Dict: __lowerCAmelCase = super().to_dict() for k, v in d.items(): if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): __lowerCAmelCase = v.to_dict() return d
284
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def a_ ( lowerCAmelCase_ : List[str], lowerCAmelCase_ : Dict, lowerCAmelCase_ : Tuple=1024, lowerCAmelCase_ : Optional[Any]=1024, lowerCAmelCase_ : Tuple=False, **lowerCAmelCase_ : Union[str, Any] ): __lowerCAmelCase = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='train', **lowerCAmelCase_ ) __lowerCAmelCase = tok.pad_token_id def get_lens(lowerCAmelCase_ : Optional[Any] ): __lowerCAmelCase = tqdm( DataLoader(lowerCAmelCase_, batch_size=512, num_workers=8, shuffle=lowerCAmelCase_, collate_fn=ds.collate_fn ), desc=str(ds.len_file ), ) __lowerCAmelCase = [] for batch in dl: __lowerCAmelCase = batch['input_ids'].ne(lowerCAmelCase_ ).sum(1 ).tolist() __lowerCAmelCase = batch['labels'].ne(lowerCAmelCase_ ).sum(1 ).tolist() if consider_target: for src, tgt in zip(lowerCAmelCase_, lowerCAmelCase_ ): max_lens.append(max(lowerCAmelCase_, lowerCAmelCase_ ) ) else: max_lens.extend(lowerCAmelCase_ ) return max_lens __lowerCAmelCase = get_lens(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='val', **lowerCAmelCase_ ) __lowerCAmelCase = get_lens(lowerCAmelCase_ ) pickle_save(lowerCAmelCase_, train_ds.len_file ) pickle_save(lowerCAmelCase_, val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
284
1
import argparse from pathlib import Path import requests import torch from PIL import Image from transformers import ( RobertaTokenizer, TrOCRConfig, TrOCRForCausalLM, TrOCRProcessor, VisionEncoderDecoderModel, ViTConfig, ViTImageProcessor, ViTModel, ) from transformers.utils import logging logging.set_verbosity_info() _snake_case : str = logging.get_logger(__name__) def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : List[Any] ): __lowerCAmelCase = [] for i in range(encoder_config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F"""encoder.deit.blocks.{i}.norm1.weight""", F"""encoder.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""encoder.deit.blocks.{i}.norm1.bias""", F"""encoder.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append( (F"""encoder.deit.blocks.{i}.attn.proj.weight""", F"""encoder.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append( (F"""encoder.deit.blocks.{i}.attn.proj.bias""", F"""encoder.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append( (F"""encoder.deit.blocks.{i}.norm2.weight""", F"""encoder.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""encoder.deit.blocks.{i}.norm2.bias""", F"""encoder.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append( (F"""encoder.deit.blocks.{i}.mlp.fc1.weight""", F"""encoder.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append( (F"""encoder.deit.blocks.{i}.mlp.fc1.bias""", F"""encoder.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append( (F"""encoder.deit.blocks.{i}.mlp.fc2.weight""", F"""encoder.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""encoder.deit.blocks.{i}.mlp.fc2.bias""", F"""encoder.encoder.layer.{i}.output.dense.bias""") ) # cls token, position embeddings and patch embeddings of encoder rename_keys.extend( [ ('encoder.deit.cls_token', 'encoder.embeddings.cls_token'), ('encoder.deit.pos_embed', 'encoder.embeddings.position_embeddings'), ('encoder.deit.patch_embed.proj.weight', 'encoder.embeddings.patch_embeddings.projection.weight'), ('encoder.deit.patch_embed.proj.bias', 'encoder.embeddings.patch_embeddings.projection.bias'), ('encoder.deit.norm.weight', 'encoder.layernorm.weight'), ('encoder.deit.norm.bias', 'encoder.layernorm.bias'), ] ) return rename_keys def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : Optional[int] ): for i in range(encoder_config.num_hidden_layers ): # queries, keys and values (only weights, no biases) __lowerCAmelCase = state_dict.pop(F"""encoder.deit.blocks.{i}.attn.qkv.weight""" ) __lowerCAmelCase = in_proj_weight[ : encoder_config.hidden_size, : ] __lowerCAmelCase = in_proj_weight[ encoder_config.hidden_size : encoder_config.hidden_size * 2, : ] __lowerCAmelCase = in_proj_weight[ -encoder_config.hidden_size :, : ] def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : Union[str, Any] ): __lowerCAmelCase = dct.pop(lowerCAmelCase_ ) __lowerCAmelCase = val def a_ ( lowerCAmelCase_ : Union[str, Any] ): if "handwritten" in checkpoint_url: __lowerCAmelCase = 'https://fki.tic.heia-fr.ch/static/img/a01-122-02-00.jpg' # industry # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02-12.jpg" # have # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02-10.jpg" # let # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02.jpg" # # url = "https://fki.tic.heia-fr.ch/static/img/a01-122.jpg" elif "printed" in checkpoint_url or "stage1" in checkpoint_url: __lowerCAmelCase = 'https://www.researchgate.net/profile/Dinh-Sang/publication/338099565/figure/fig8/AS:840413229350922@1577381536857/An-receipt-example-in-the-SROIE-2019-dataset_Q640.jpg' __lowerCAmelCase = Image.open(requests.get(lowerCAmelCase_, stream=lowerCAmelCase_ ).raw ).convert('RGB' ) return im @torch.no_grad() def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : int ): __lowerCAmelCase = ViTConfig(image_size=384, qkv_bias=lowerCAmelCase_ ) __lowerCAmelCase = TrOCRConfig() # size of the architecture if "base" in checkpoint_url: __lowerCAmelCase = 768 elif "large" in checkpoint_url: # use ViT-large encoder __lowerCAmelCase = 1024 __lowerCAmelCase = 4096 __lowerCAmelCase = 24 __lowerCAmelCase = 16 __lowerCAmelCase = 1024 else: raise ValueError('Should either find \'base\' or \'large\' in checkpoint URL' ) # the large-printed + stage1 checkpoints uses sinusoidal position embeddings, no layernorm afterwards if "large-printed" in checkpoint_url or "stage1" in checkpoint_url: __lowerCAmelCase = False __lowerCAmelCase = 'relu' __lowerCAmelCase = 1024 __lowerCAmelCase = True __lowerCAmelCase = False __lowerCAmelCase = False # load HuggingFace model __lowerCAmelCase = ViTModel(lowerCAmelCase_, add_pooling_layer=lowerCAmelCase_ ) __lowerCAmelCase = TrOCRForCausalLM(lowerCAmelCase_ ) __lowerCAmelCase = VisionEncoderDecoderModel(encoder=lowerCAmelCase_, decoder=lowerCAmelCase_ ) model.eval() # load state_dict of original model, rename some keys __lowerCAmelCase = torch.hub.load_state_dict_from_url(lowerCAmelCase_, map_location='cpu', check_hash=lowerCAmelCase_ )['model'] __lowerCAmelCase = create_rename_keys(lowerCAmelCase_, lowerCAmelCase_ ) for src, dest in rename_keys: rename_key(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) read_in_q_k_v(lowerCAmelCase_, lowerCAmelCase_ ) # remove parameters we don't need del state_dict["encoder.deit.head.weight"] del state_dict["encoder.deit.head.bias"] del state_dict["decoder.version"] # add prefix to decoder keys for key, val in state_dict.copy().items(): __lowerCAmelCase = state_dict.pop(lowerCAmelCase_ ) if key.startswith('decoder' ) and "output_projection" not in key: __lowerCAmelCase = val else: __lowerCAmelCase = val # load state dict model.load_state_dict(lowerCAmelCase_ ) # Check outputs on an image __lowerCAmelCase = ViTImageProcessor(size=encoder_config.image_size ) __lowerCAmelCase = RobertaTokenizer.from_pretrained('roberta-large' ) __lowerCAmelCase = TrOCRProcessor(lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = processor(images=prepare_img(lowerCAmelCase_ ), return_tensors='pt' ).pixel_values # verify logits __lowerCAmelCase = torch.tensor([[model.config.decoder.decoder_start_token_id]] ) __lowerCAmelCase = model(pixel_values=lowerCAmelCase_, decoder_input_ids=lowerCAmelCase_ ) __lowerCAmelCase = outputs.logits __lowerCAmelCase = torch.Size([1, 1, 5_0265] ) if "trocr-base-handwritten" in checkpoint_url: __lowerCAmelCase = torch.tensor( [-1.4502, -4.6683, -0.5347, -2.9291, 9.1435, -3.0571, 8.9764, 1.7560, 8.7358, -1.5311] ) elif "trocr-large-handwritten" in checkpoint_url: __lowerCAmelCase = torch.tensor( [-2.6437, -1.3129, -2.2596, -5.3455, 6.3539, 1.7604, 5.4991, 1.4702, 5.6113, 2.0170] ) elif "trocr-base-printed" in checkpoint_url: __lowerCAmelCase = torch.tensor( [-5.6816, -5.8388, 1.1398, -6.9034, 6.8505, -2.4393, 1.2284, -1.0232, -1.9661, -3.9210] ) elif "trocr-large-printed" in checkpoint_url: __lowerCAmelCase = torch.tensor( [-6.0162, -7.0959, 4.4155, -5.1063, 7.0468, -3.1631, 2.6466, -0.3081, -0.8106, -1.7535] ) if "stage1" not in checkpoint_url: assert logits.shape == expected_shape, "Shape of logits not as expected" assert torch.allclose(logits[0, 0, :10], lowerCAmelCase_, atol=1E-3 ), "First elements of logits not as expected" Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowerCAmelCase_ ) print(F"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(lowerCAmelCase_ ) if __name__ == "__main__": _snake_case : Optional[Any] = argparse.ArgumentParser() parser.add_argument( '--checkpoint_url', default='https://layoutlm.blob.core.windows.net/trocr/model_zoo/fairseq/trocr-base-handwritten.pt', type=str, help='URL to the original PyTorch checkpoint (.pth file).', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) _snake_case : Tuple = parser.parse_args() convert_tr_ocr_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
284
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def a_ ( lowerCAmelCase_ : List[Any], lowerCAmelCase_ : str, lowerCAmelCase_ : Optional[int]=None, lowerCAmelCase_ : List[Any]=None ): if attention_mask is None: __lowerCAmelCase = tf.cast(tf.math.not_equal(lowerCAmelCase_, config.pad_token_id ), tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class _UpperCAmelCase : """simple docstring""" a_ = OPTConfig a_ = {} a_ = """gelu""" def __init__( self : Dict , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : List[str]=1_3 , lowerCAmelCase_ : Tuple=7 , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : Union[str, Any]=False , lowerCAmelCase_ : Any=9_9 , lowerCAmelCase_ : Any=1_6 , lowerCAmelCase_ : List[str]=2 , lowerCAmelCase_ : Dict=4 , lowerCAmelCase_ : str=4 , lowerCAmelCase_ : Any="gelu" , lowerCAmelCase_ : Optional[Any]=0.1 , lowerCAmelCase_ : List[Any]=0.1 , lowerCAmelCase_ : Tuple=2_0 , lowerCAmelCase_ : Optional[Any]=2 , lowerCAmelCase_ : Any=1 , lowerCAmelCase_ : List[Any]=0 , lowerCAmelCase_ : Optional[int]=1_6 , lowerCAmelCase_ : Dict=1_6 , ) -> int: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = eos_token_id __lowerCAmelCase = pad_token_id __lowerCAmelCase = bos_token_id __lowerCAmelCase = embed_dim __lowerCAmelCase = word_embed_proj_dim __lowerCAmelCase = False def lowercase ( self : List[str] ) -> Optional[Any]: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowerCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowerCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowerCAmelCase = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=lowerCAmelCase_ , **self.config_updates , ) __lowerCAmelCase = prepare_opt_inputs_dict(lowerCAmelCase_ , lowerCAmelCase_ ) return config, inputs_dict def lowercase ( self : Any , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Dict ) -> List[str]: __lowerCAmelCase = TFOPTModel(config=lowerCAmelCase_ ) __lowerCAmelCase = inputs_dict['input_ids'] __lowerCAmelCase = input_ids[:1, :] __lowerCAmelCase = inputs_dict['attention_mask'][:1, :] __lowerCAmelCase = 1 # first forward pass __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , use_cache=lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowerCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowerCAmelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowerCAmelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowerCAmelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , past_key_values=lowerCAmelCase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowerCAmelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowerCAmelCase = output_from_no_past[:, -3:, random_slice_idx] __lowerCAmelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowerCAmelCase_ , lowerCAmelCase_ , rtol=1e-3 ) @require_tf class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () a_ = (TFOPTForCausalLM,) if is_tf_available() else () a_ = ( {"""feature-extraction""": TFOPTModel, """text-generation""": TFOPTForCausalLM} if is_tf_available() else {} ) a_ = False a_ = False a_ = False a_ = 10 def lowercase ( self : List[str] ) -> Optional[int]: __lowerCAmelCase = TFOPTModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=lowerCAmelCase_ ) def lowercase ( self : Tuple ) -> Tuple: self.config_tester.run_common_tests() def lowercase ( self : Tuple ) -> Optional[Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCAmelCase_ ) def lowercase ( self : Union[str, Any] ) -> Dict: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(lowerCAmelCase_ : str , lowerCAmelCase_ : Union[str, Any] ): if hasattr(lowerCAmelCase_ , 'weight' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(lowerCAmelCase_ , 'weight' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 1_0, config.vocab_size + 1_0]: # build the embeddings __lowerCAmelCase = model_class(config=lowerCAmelCase_ ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_input_embeddings() ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(lowerCAmelCase_ ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_input_embeddings() ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. __lowerCAmelCase = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , lowerCAmelCase_ ) # check that weights remain the same after resizing __lowerCAmelCase = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCAmelCase = False self.assertTrue(lowerCAmelCase_ ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , lowerCAmelCase_ ) __lowerCAmelCase = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCAmelCase = False self.assertTrue(lowerCAmelCase_ ) def a_ ( lowerCAmelCase_ : Union[str, Any] ): return tf.constant(lowerCAmelCase_, dtype=tf.intaa ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" a_ = 99 def lowercase ( self : Optional[int] ) -> Any: __lowerCAmelCase = tf.ones((4, 1) , dtype=tf.intaa ) * 2 __lowerCAmelCase = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) __lowerCAmelCase = input_ids.shape[0] __lowerCAmelCase = OPTConfig( vocab_size=self.vocab_size , hidden_size=2_4 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=3_2 , max_position_embeddings=4_8 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def lowercase ( self : str ) -> List[str]: __lowerCAmelCase = TFOPTModel.from_pretrained('facebook/opt-350m' ) __lowerCAmelCase = _long_tensor([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] ) __lowerCAmelCase = tf.not_equal(lowerCAmelCase_ , model.config.pad_token_id ) with tf.GradientTape(): __lowerCAmelCase = model(input_ids=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ ).last_hidden_state __lowerCAmelCase = (1, 1_1, 5_1_2) self.assertEqual(output.shape , lowerCAmelCase_ ) __lowerCAmelCase = tf.constant( [[-0.28_73, -1.92_18, -0.30_33], [-1.27_10, -0.13_38, -0.19_02], [0.40_95, 0.12_14, -1.31_21]] ) self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase_ , atol=4e-3 ) ) __lowerCAmelCase = tf.function(lowerCAmelCase_ , jit_compile=lowerCAmelCase_ ) __lowerCAmelCase = xla_generate(lowerCAmelCase_ , lowerCAmelCase_ )[0] self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase_ , atol=4e-2 ) ) @require_tf @slow class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> Dict: super().setUp() __lowerCAmelCase = 'facebook/opt-350m' def lowercase ( self : Dict ) -> Any: __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(self.path_model ) __lowerCAmelCase = GPTaTokenizer.from_pretrained(self.path_model ) __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of', 'Paris is the capital of France and', 'Computers and mobile phones have taken', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' , padding=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) __lowerCAmelCase = tf.constant( [ [1.38_51, -13.89_23, -10.52_29, -10.75_33, -0.23_09, -10.23_84, -0.53_65, -9.09_47, -5.16_70], [-4.70_73, -10.62_76, -3.94_15, -21.52_42, -0.28_22, -0.28_22, -0.28_22, -0.28_22, -0.28_22], [0.62_47, -3.42_29, -8.91_79, -1.42_97, -14.16_50, 1.41_46, -9.02_18, -0.27_03, -0.27_03], [6.47_83, -1.99_13, -10.79_26, -2.33_36, 1.50_92, -0.99_74, -6.82_13, 1.34_77, 1.34_77], ] ) self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1e-4 ) ) __lowerCAmelCase = tf.function(lowerCAmelCase_ , jit_compile=lowerCAmelCase_ ) __lowerCAmelCase = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1e-4 ) ) @require_tf @slow class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @property def lowercase ( self : Optional[int] ) -> int: return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def lowercase ( self : int ) -> str: __lowerCAmelCase = 'facebook/opt-125m' __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of New York, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] __lowerCAmelCase = [] __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) for prompt in self.prompts: __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(lowerCAmelCase_ , max_length=1_0 ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Optional[Any] ) -> str: __lowerCAmelCase = 'facebook/opt-350m' __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = 'left' # use different length sentences to test batching __lowerCAmelCase = [ 'Hello, my dog is a little', 'Today, I', ] __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' , padding=lowerCAmelCase_ ) __lowerCAmelCase = inputs['input_ids'] __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ , attention_mask=inputs['attention_mask'] ) __lowerCAmelCase = tokenizer(sentences[0] , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ ) __lowerCAmelCase = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['attention_mask'][-1] , tf.intaa ) ) __lowerCAmelCase = tokenizer(sentences[1] , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ , max_length=model.config.max_length - num_paddings ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = [ 'Hello, my dog is a little bit of a dork.\nI\'m a little bit', 'Today, I was in the middle of a conversation with a friend about the', ] self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , [non_padded_sentence, padded_sentence] ) def lowercase ( self : List[Any] ) -> List[Any]: __lowerCAmelCase = 'facebook/opt-350m' __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of San Francisco, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] __lowerCAmelCase = [] __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) for prompt in self.prompts: __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(lowerCAmelCase_ , max_length=1_0 ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ )
284
1
def a_ ( lowerCAmelCase_ : int ): if not isinstance(lowerCAmelCase_, lowerCAmelCase_ ) or number < 0: raise ValueError('Input must be a non-negative integer' ) __lowerCAmelCase = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
284
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { 'configuration_layoutlmv3': [ 'LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv3Config', 'LayoutLMv3OnnxConfig', ], 'processing_layoutlmv3': ['LayoutLMv3Processor'], 'tokenization_layoutlmv3': ['LayoutLMv3Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Union[str, Any] = ['LayoutLMv3TokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[str] = [ 'LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv3ForQuestionAnswering', 'LayoutLMv3ForSequenceClassification', 'LayoutLMv3ForTokenClassification', 'LayoutLMv3Model', 'LayoutLMv3PreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[Any] = [ 'TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLayoutLMv3ForQuestionAnswering', 'TFLayoutLMv3ForSequenceClassification', 'TFLayoutLMv3ForTokenClassification', 'TFLayoutLMv3Model', 'TFLayoutLMv3PreTrainedModel', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Tuple = ['LayoutLMv3FeatureExtractor'] _snake_case : str = ['LayoutLMv3ImageProcessor'] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys _snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
1
def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : int ): return 1 if input_a == input_a else 0 def a_ ( ): assert xnor_gate(0, 0 ) == 1 assert xnor_gate(0, 1 ) == 0 assert xnor_gate(1, 0 ) == 0 assert xnor_gate(1, 1 ) == 1 if __name__ == "__main__": print(xnor_gate(0, 0)) print(xnor_gate(0, 1)) print(xnor_gate(1, 0)) print(xnor_gate(1, 1))
284
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case : Dict = {'configuration_mra': ['MRA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MraConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'MRA_PRETRAINED_MODEL_ARCHIVE_LIST', 'MraForMaskedLM', 'MraForMultipleChoice', 'MraForQuestionAnswering', 'MraForSequenceClassification', 'MraForTokenClassification', 'MraLayer', 'MraModel', 'MraPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys _snake_case : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure)
284
1
import argparse import OmegaConf import torch from diffusers import DDIMScheduler, LDMPipeline, UNetLDMModel, VQModel def a_ ( lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Union[str, Any] ): __lowerCAmelCase = OmegaConf.load(lowerCAmelCase_ ) __lowerCAmelCase = torch.load(lowerCAmelCase_, map_location='cpu' )['model'] __lowerCAmelCase = list(state_dict.keys() ) # extract state_dict for VQVAE __lowerCAmelCase = {} __lowerCAmelCase = 'first_stage_model.' for key in keys: if key.startswith(lowerCAmelCase_ ): __lowerCAmelCase = state_dict[key] # extract state_dict for UNetLDM __lowerCAmelCase = {} __lowerCAmelCase = 'model.diffusion_model.' for key in keys: if key.startswith(lowerCAmelCase_ ): __lowerCAmelCase = state_dict[key] __lowerCAmelCase = config.model.params.first_stage_config.params __lowerCAmelCase = config.model.params.unet_config.params __lowerCAmelCase = VQModel(**lowerCAmelCase_ ).eval() vqvae.load_state_dict(lowerCAmelCase_ ) __lowerCAmelCase = UNetLDMModel(**lowerCAmelCase_ ).eval() unet.load_state_dict(lowerCAmelCase_ ) __lowerCAmelCase = DDIMScheduler( timesteps=config.model.params.timesteps, beta_schedule='scaled_linear', beta_start=config.model.params.linear_start, beta_end=config.model.params.linear_end, clip_sample=lowerCAmelCase_, ) __lowerCAmelCase = LDMPipeline(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) pipeline.save_pretrained(lowerCAmelCase_ ) if __name__ == "__main__": _snake_case : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('--checkpoint_path', type=str, required=True) parser.add_argument('--config_path', type=str, required=True) parser.add_argument('--output_path', type=str, required=True) _snake_case : Optional[Any] = parser.parse_args() convert_ldm_original(args.checkpoint_path, args.config_path, args.output_path)
284
from __future__ import annotations import string from itertools import cycle, product from pathlib import Path _snake_case : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) _snake_case : list[int] = [ord(letter) for letter in string.ascii_lowercase] _snake_case : set[int] = {ord(char) for char in VALID_CHARS} _snake_case : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def a_ ( lowerCAmelCase_ : list[int], lowerCAmelCase_ : tuple[int, ...] ): __lowerCAmelCase = "" __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 for keychar, cipherchar in zip(cycle(lowerCAmelCase_ ), lowerCAmelCase_ ): __lowerCAmelCase = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(lowerCAmelCase_ ) return decoded def a_ ( lowerCAmelCase_ : list[int] ): __lowerCAmelCase = [] for key in product(lowerCAmelCase_, repeat=3 ): __lowerCAmelCase = try_key(lowerCAmelCase_, lowerCAmelCase_ ) if encoded is not None: possibles.append(lowerCAmelCase_ ) return possibles def a_ ( lowerCAmelCase_ : list[str], lowerCAmelCase_ : str ): return [possible for possible in possibles if common_word in possible.lower()] def a_ ( lowerCAmelCase_ : str = "p059_cipher.txt" ): __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = Path(lowerCAmelCase_ ).parent.joinpath(lowerCAmelCase_ ).read_text(encoding='utf-8' ) __lowerCAmelCase = [int(lowerCAmelCase_ ) for number in data.strip().split(',' )] __lowerCAmelCase = filter_valid_chars(lowerCAmelCase_ ) for common_word in COMMON_WORDS: __lowerCAmelCase = filter_common_word(lowerCAmelCase_, lowerCAmelCase_ ) if len(lowerCAmelCase_ ) == 1: break __lowerCAmelCase = possibles[0] return sum(ord(lowerCAmelCase_ ) for char in decoded_text ) if __name__ == "__main__": print(F"""{solution() = }""")
284
1
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : str , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : List[Any]=7 , lowerCAmelCase_ : Optional[Any]=3 , lowerCAmelCase_ : Dict=3_0 , lowerCAmelCase_ : Optional[int]=4_0_0 , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : str=None , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Any=1 / 2_5_5 , lowerCAmelCase_ : Optional[Any]=True , lowerCAmelCase_ : List[str]=[0.5, 0.5, 0.5] , lowerCAmelCase_ : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCAmelCase_ : Dict=True , ) -> Optional[int]: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __lowerCAmelCase = size if size is not None else {'shortest_edge': 1_8, 'longest_edge': 1_3_3_3} __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = num_channels __lowerCAmelCase = min_resolution __lowerCAmelCase = max_resolution __lowerCAmelCase = do_resize __lowerCAmelCase = size __lowerCAmelCase = do_rescale __lowerCAmelCase = rescale_factor __lowerCAmelCase = do_normalize __lowerCAmelCase = image_mean __lowerCAmelCase = image_std __lowerCAmelCase = do_pad def lowercase ( self : Union[str, Any] ) -> Optional[Any]: return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def lowercase ( self : str , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Dict=False ) -> Optional[int]: if not batched: __lowerCAmelCase = image_inputs[0] if isinstance(lowerCAmelCase_ , Image.Image ): __lowerCAmelCase , __lowerCAmelCase = image.size else: __lowerCAmelCase , __lowerCAmelCase = image.shape[1], image.shape[2] if w < h: __lowerCAmelCase = int(self.size['shortest_edge'] * h / w ) __lowerCAmelCase = self.size['shortest_edge'] elif w > h: __lowerCAmelCase = self.size['shortest_edge'] __lowerCAmelCase = int(self.size['shortest_edge'] * w / h ) else: __lowerCAmelCase = self.size['shortest_edge'] __lowerCAmelCase = self.size['shortest_edge'] else: __lowerCAmelCase = [] for image in image_inputs: __lowerCAmelCase , __lowerCAmelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __lowerCAmelCase = max(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : item[0] )[0] __lowerCAmelCase = max(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _UpperCAmelCase ( _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = DetrImageProcessor if is_vision_available() else None def lowercase ( self : List[Any] ) -> int: __lowerCAmelCase = DetrImageProcessingTester(self ) @property def lowercase ( self : Any ) -> List[str]: return self.image_processor_tester.prepare_image_processor_dict() def lowercase ( self : List[Any] ) -> int: __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCAmelCase_ , 'image_mean' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'image_std' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'do_normalize' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'do_rescale' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'rescale_factor' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'do_resize' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'size' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'do_pad' ) ) def lowercase ( self : str ) -> Optional[Any]: __lowerCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 1_8, 'longest_edge': 1_3_3_3} ) self.assertEqual(image_processor.do_pad , lowerCAmelCase_ ) __lowerCAmelCase = self.image_processing_class.from_dict( self.image_processor_dict , size=4_2 , max_size=8_4 , pad_and_return_pixel_mask=lowerCAmelCase_ ) self.assertEqual(image_processor.size , {'shortest_edge': 4_2, 'longest_edge': 8_4} ) self.assertEqual(image_processor.do_pad , lowerCAmelCase_ ) def lowercase ( self : Tuple ) -> Optional[Any]: pass def lowercase ( self : int ) -> Any: # Initialize image_processing __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase_ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase_ , Image.Image ) # Test not batched input __lowerCAmelCase = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processor_tester.get_expected_values(lowerCAmelCase_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase , __lowerCAmelCase = self.image_processor_tester.get_expected_values(lowerCAmelCase_ , batched=lowerCAmelCase_ ) __lowerCAmelCase = image_processing(lowerCAmelCase_ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowercase ( self : Dict ) -> List[Any]: # Initialize image_processing __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase_ , numpify=lowerCAmelCase_ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase_ , np.ndarray ) # Test not batched input __lowerCAmelCase = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processor_tester.get_expected_values(lowerCAmelCase_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase = image_processing(lowerCAmelCase_ , return_tensors='pt' ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processor_tester.get_expected_values(lowerCAmelCase_ , batched=lowerCAmelCase_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowercase ( self : Union[str, Any] ) -> Dict: # Initialize image_processing __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase_ , torchify=lowerCAmelCase_ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase_ , torch.Tensor ) # Test not batched input __lowerCAmelCase = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processor_tester.get_expected_values(lowerCAmelCase_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase = image_processing(lowerCAmelCase_ , return_tensors='pt' ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processor_tester.get_expected_values(lowerCAmelCase_ , batched=lowerCAmelCase_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def lowercase ( self : Tuple ) -> str: # prepare image and target __lowerCAmelCase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: __lowerCAmelCase = json.loads(f.read() ) __lowerCAmelCase = {'image_id': 3_9_7_6_9, 'annotations': target} # encode them __lowerCAmelCase = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50' ) __lowerCAmelCase = image_processing(images=lowerCAmelCase_ , annotations=lowerCAmelCase_ , return_tensors='pt' ) # verify pixel values __lowerCAmelCase = torch.Size([1, 3, 8_0_0, 1_0_6_6] ) self.assertEqual(encoding['pixel_values'].shape , lowerCAmelCase_ ) __lowerCAmelCase = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , lowerCAmelCase_ , atol=1e-4 ) ) # verify area __lowerCAmelCase = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , lowerCAmelCase_ ) ) # verify boxes __lowerCAmelCase = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , lowerCAmelCase_ ) __lowerCAmelCase = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , lowerCAmelCase_ , atol=1e-3 ) ) # verify image_id __lowerCAmelCase = torch.tensor([3_9_7_6_9] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , lowerCAmelCase_ ) ) # verify is_crowd __lowerCAmelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , lowerCAmelCase_ ) ) # verify class_labels __lowerCAmelCase = torch.tensor([7_5, 7_5, 6_3, 6_5, 1_7, 1_7] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , lowerCAmelCase_ ) ) # verify orig_size __lowerCAmelCase = torch.tensor([4_8_0, 6_4_0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , lowerCAmelCase_ ) ) # verify size __lowerCAmelCase = torch.tensor([8_0_0, 1_0_6_6] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , lowerCAmelCase_ ) ) @slow def lowercase ( self : Optional[int] ) -> str: # prepare image, target and masks_path __lowerCAmelCase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: __lowerCAmelCase = json.loads(f.read() ) __lowerCAmelCase = {'file_name': '000000039769.png', 'image_id': 3_9_7_6_9, 'segments_info': target} __lowerCAmelCase = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them __lowerCAmelCase = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50-panoptic' ) __lowerCAmelCase = image_processing(images=lowerCAmelCase_ , annotations=lowerCAmelCase_ , masks_path=lowerCAmelCase_ , return_tensors='pt' ) # verify pixel values __lowerCAmelCase = torch.Size([1, 3, 8_0_0, 1_0_6_6] ) self.assertEqual(encoding['pixel_values'].shape , lowerCAmelCase_ ) __lowerCAmelCase = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , lowerCAmelCase_ , atol=1e-4 ) ) # verify area __lowerCAmelCase = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , lowerCAmelCase_ ) ) # verify boxes __lowerCAmelCase = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , lowerCAmelCase_ ) __lowerCAmelCase = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , lowerCAmelCase_ , atol=1e-3 ) ) # verify image_id __lowerCAmelCase = torch.tensor([3_9_7_6_9] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , lowerCAmelCase_ ) ) # verify is_crowd __lowerCAmelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , lowerCAmelCase_ ) ) # verify class_labels __lowerCAmelCase = torch.tensor([1_7, 1_7, 6_3, 7_5, 7_5, 9_3] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , lowerCAmelCase_ ) ) # verify masks __lowerCAmelCase = 8_2_2_8_7_3 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , lowerCAmelCase_ ) # verify orig_size __lowerCAmelCase = torch.tensor([4_8_0, 6_4_0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , lowerCAmelCase_ ) ) # verify size __lowerCAmelCase = torch.tensor([8_0_0, 1_0_6_6] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , lowerCAmelCase_ ) )
284
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem _snake_case : Tuple = importlib.util.find_spec('s3fs') is not None if _has_safs: from .safilesystem import SaFileSystem # noqa: F401 _snake_case : List[compression.BaseCompressedFileFileSystem] = [ compression.BzaFileSystem, compression.GzipFileSystem, compression.LzaFileSystem, compression.XzFileSystem, compression.ZstdFileSystem, ] # Register custom filesystems for fs_class in COMPRESSION_FILESYSTEMS + [HfFileSystem]: if fs_class.protocol in fsspec.registry and fsspec.registry[fs_class.protocol] is not fs_class: warnings.warn(F"""A filesystem protocol was already set for {fs_class.protocol} and will be overwritten.""") fsspec.register_implementation(fs_class.protocol, fs_class, clobber=True) def a_ ( lowerCAmelCase_ : str ): if "://" in dataset_path: __lowerCAmelCase = dataset_path.split('://' )[1] return dataset_path def a_ ( lowerCAmelCase_ : fsspec.AbstractFileSystem ): if fs is not None and fs.protocol != "file": return True else: return False def a_ ( lowerCAmelCase_ : fsspec.AbstractFileSystem, lowerCAmelCase_ : str, lowerCAmelCase_ : str ): __lowerCAmelCase = not is_remote_filesystem(lowerCAmelCase_ ) if is_local: # LocalFileSystem.mv does copy + rm, it is more efficient to simply move a local directory shutil.move(fs._strip_protocol(lowerCAmelCase_ ), fs._strip_protocol(lowerCAmelCase_ ) ) else: fs.mv(lowerCAmelCase_, lowerCAmelCase_, recursive=lowerCAmelCase_ ) def a_ ( ): if hasattr(fsspec.asyn, 'reset_lock' ): # for future fsspec>2022.05.0 fsspec.asyn.reset_lock() else: __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = threading.Lock()
284
1
import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer _snake_case : Any = logging.get_logger(__name__) _snake_case : Any = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} _snake_case : List[str] = { 'vocab_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } _snake_case : List[str] = { 'vocab_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } _snake_case : int = { 'vocab_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json' ), }, } _snake_case : Dict = { 'facebook/dpr-ctx_encoder-single-nq-base': 512, 'facebook/dpr-ctx_encoder-multiset-base': 512, } _snake_case : List[Any] = { 'facebook/dpr-question_encoder-single-nq-base': 512, 'facebook/dpr-question_encoder-multiset-base': 512, } _snake_case : Tuple = { 'facebook/dpr-reader-single-nq-base': 512, 'facebook/dpr-reader-multiset-base': 512, } _snake_case : Tuple = { 'facebook/dpr-ctx_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-ctx_encoder-multiset-base': {'do_lower_case': True}, } _snake_case : Any = { 'facebook/dpr-question_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-question_encoder-multiset-base': {'do_lower_case': True}, } _snake_case : Optional[Any] = { 'facebook/dpr-reader-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-reader-multiset-base': {'do_lower_case': True}, } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = VOCAB_FILES_NAMES a_ = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP a_ = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION a_ = DPRContextEncoderTokenizer class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = VOCAB_FILES_NAMES a_ = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP a_ = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION a_ = DPRQuestionEncoderTokenizer _snake_case : str = collections.namedtuple( 'DPRSpanPrediction', ['span_score', 'relevance_score', 'doc_id', 'start_index', 'end_index', 'text'] ) _snake_case : List[Any] = collections.namedtuple('DPRReaderOutput', ['start_logits', 'end_logits', 'relevance_logits']) _snake_case : Tuple = R'\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `\'longest\'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `\'max_length\'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `\'do_not_pad\'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `\'longest_first\'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `\'only_first\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `\'only_second\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `\'do_not_truncate\'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `\'tf\'`: Return TensorFlow `tf.constant` objects.\n - `\'pt\'`: Return PyTorch `torch.Tensor` objects.\n - `\'np\'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer\'s default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Return:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n ' @add_start_docstrings(_UpperCamelCase ) class _UpperCAmelCase : """simple docstring""" def __call__( self : List[str] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Optional[str] = None , lowerCAmelCase_ : Optional[str] = None , lowerCAmelCase_ : Union[bool, str] = False , lowerCAmelCase_ : Union[bool, str] = False , lowerCAmelCase_ : Optional[int] = None , lowerCAmelCase_ : Optional[Union[str, TensorType]] = None , lowerCAmelCase_ : Optional[bool] = None , **lowerCAmelCase_ : Dict , ) -> BatchEncoding: if titles is None and texts is None: return super().__call__( lowerCAmelCase_ , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , **lowerCAmelCase_ , ) elif titles is None or texts is None: __lowerCAmelCase = titles if texts is None else texts return super().__call__( lowerCAmelCase_ , lowerCAmelCase_ , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , **lowerCAmelCase_ , ) __lowerCAmelCase = titles if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else [titles] __lowerCAmelCase = texts if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else [texts] __lowerCAmelCase = len(lowerCAmelCase_ ) __lowerCAmelCase = questions if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else [questions] * n_passages assert len(lowerCAmelCase_ ) == len( lowerCAmelCase_ ), f"""There should be as many titles than texts but got {len(lowerCAmelCase_ )} titles and {len(lowerCAmelCase_ )} texts.""" __lowerCAmelCase = super().__call__(lowerCAmelCase_ , lowerCAmelCase_ , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ )['input_ids'] __lowerCAmelCase = super().__call__(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ )['input_ids'] __lowerCAmelCase = { 'input_ids': [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(lowerCAmelCase_ , lowerCAmelCase_ ) ] } if return_attention_mask is not False: __lowerCAmelCase = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) __lowerCAmelCase = attention_mask return self.pad(lowerCAmelCase_ , padding=lowerCAmelCase_ , max_length=lowerCAmelCase_ , return_tensors=lowerCAmelCase_ ) def lowercase ( self : Optional[Any] , lowerCAmelCase_ : BatchEncoding , lowerCAmelCase_ : DPRReaderOutput , lowerCAmelCase_ : int = 1_6 , lowerCAmelCase_ : int = 6_4 , lowerCAmelCase_ : int = 4 , ) -> List[DPRSpanPrediction]: __lowerCAmelCase = reader_input['input_ids'] __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = reader_output[:3] __lowerCAmelCase = len(lowerCAmelCase_ ) __lowerCAmelCase = sorted(range(lowerCAmelCase_ ) , reverse=lowerCAmelCase_ , key=relevance_logits.__getitem__ ) __lowerCAmelCase = [] for doc_id in sorted_docs: __lowerCAmelCase = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence __lowerCAmelCase = sequence_ids.index(self.sep_token_id , 2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: __lowerCAmelCase = sequence_ids.index(self.pad_token_id ) else: __lowerCAmelCase = len(lowerCAmelCase_ ) __lowerCAmelCase = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=lowerCAmelCase_ , top_spans=lowerCAmelCase_ , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=lowerCAmelCase_ , start_index=lowerCAmelCase_ , end_index=lowerCAmelCase_ , text=self.decode(sequence_ids[start_index : end_index + 1] ) , ) ) if len(lowerCAmelCase_ ) >= num_spans: break return nbest_spans_predictions[:num_spans] def lowercase ( self : str , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : int , lowerCAmelCase_ : int , ) -> List[DPRSpanPrediction]: __lowerCAmelCase = [] for start_index, start_score in enumerate(lowerCAmelCase_ ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) __lowerCAmelCase = sorted(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : x[1] , reverse=lowerCAmelCase_ ) __lowerCAmelCase = [] for (start_index, end_index), score in scores: assert start_index <= end_index, f"""Wrong span indices: [{start_index}:{end_index}]""" __lowerCAmelCase = end_index - start_index + 1 assert length <= max_answer_length, f"""Span is too long: {length} > {max_answer_length}""" if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(lowerCAmelCase_ ) == top_spans: break return chosen_span_intervals @add_end_docstrings(_UpperCamelCase ) class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase ): """simple docstring""" a_ = VOCAB_FILES_NAMES a_ = READER_PRETRAINED_VOCAB_FILES_MAP a_ = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ = READER_PRETRAINED_INIT_CONFIGURATION a_ = ["""input_ids""", """attention_mask"""] a_ = DPRReaderTokenizer
284
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def a_ ( lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = filter(lambda lowerCAmelCase_ : p.requires_grad, model.parameters() ) __lowerCAmelCase = sum([np.prod(p.size() ) for p in model_parameters] ) return params _snake_case : Dict = logging.getLogger(__name__) def a_ ( lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Optional[int] ): if metric == "rouge2": __lowerCAmelCase = '{val_avg_rouge2:.4f}-{step_count}' elif metric == "bleu": __lowerCAmelCase = '{val_avg_bleu:.4f}-{step_count}' elif metric == "em": __lowerCAmelCase = '{val_avg_em:.4f}-{step_count}' else: raise NotImplementedError( F"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" ' function.' ) __lowerCAmelCase = ModelCheckpoint( dirpath=lowerCAmelCase_, filename=lowerCAmelCase_, monitor=F"""val_{metric}""", mode='max', save_top_k=3, every_n_epochs=1, ) return checkpoint_callback def a_ ( lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : Any ): return EarlyStopping( monitor=F"""val_{metric}""", mode='min' if 'loss' in metric else 'max', patience=lowerCAmelCase_, verbose=lowerCAmelCase_, ) class _UpperCAmelCase ( pl.Callback ): """simple docstring""" def lowercase ( self : Tuple , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : int ) -> Any: __lowerCAmelCase = {f"""lr_group_{i}""": param['lr'] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Optional[int] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule , lowerCAmelCase_ : str , lowerCAmelCase_ : List[Any]=True ) -> None: logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) __lowerCAmelCase = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['log', 'progress_bar', 'preds']} ) # Log results __lowerCAmelCase = Path(pl_module.hparams.output_dir ) if type_path == "test": __lowerCAmelCase = od / 'test_results.txt' __lowerCAmelCase = od / 'test_generations.txt' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __lowerCAmelCase = od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" __lowerCAmelCase = od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) generations_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'a+' ) as writer: for key in sorted(lowerCAmelCase_ ): if key in ["log", "progress_bar", "preds"]: continue __lowerCAmelCase = metrics[key] if isinstance(lowerCAmelCase_ , torch.Tensor ): __lowerCAmelCase = val.item() __lowerCAmelCase = f"""{key}: {val:.6f}\n""" writer.write(lowerCAmelCase_ ) if not save_generations: return if "preds" in metrics: __lowerCAmelCase = '\n'.join(metrics['preds'] ) generations_file.open('w+' ).write(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[str] ) -> Dict: try: __lowerCAmelCase = pl_module.model.model.num_parameters() except AttributeError: __lowerCAmelCase = pl_module.model.num_parameters() __lowerCAmelCase = count_trainable_parameters(lowerCAmelCase_ ) # mp stands for million parameters trainer.logger.log_metrics({'n_params': npars, 'mp': npars / 1e6, 'grad_mp': n_trainable_pars / 1e6} ) @rank_zero_only def lowercase ( self : int , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule ) -> Any: save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(lowerCAmelCase_ , lowerCAmelCase_ , 'test' ) @rank_zero_only def lowercase ( self : List[Any] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : Any ) -> int: save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
284
1
import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class _UpperCAmelCase : """simple docstring""" def __init__( self : Dict , lowerCAmelCase_ : List[str]=2 , lowerCAmelCase_ : Tuple=3 , lowerCAmelCase_ : int=6_4 , lowerCAmelCase_ : str=None ) -> Tuple: __lowerCAmelCase = np.random.default_rng(lowerCAmelCase_ ) __lowerCAmelCase = length __lowerCAmelCase = rng.normal(size=(length,) ).astype(np.floataa ) __lowerCAmelCase = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self : Optional[Any] ) -> Optional[int]: return self.length def __getitem__( self : List[Any] , lowerCAmelCase_ : Dict ) -> List[str]: return {"x": self.x[i], "y": self.y[i]} class _UpperCAmelCase ( torch.nn.Module ): """simple docstring""" def __init__( self : int , lowerCAmelCase_ : Any=0 , lowerCAmelCase_ : int=0 , lowerCAmelCase_ : Dict=False ) -> Dict: super().__init__() __lowerCAmelCase = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) __lowerCAmelCase = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) __lowerCAmelCase = True def lowercase ( self : int , lowerCAmelCase_ : Union[str, Any]=None ) -> Optional[Any]: if self.first_batch: print(f"""Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}""" ) __lowerCAmelCase = False return x * self.a[0] + self.b[0] class _UpperCAmelCase ( torch.nn.Module ): """simple docstring""" def __init__( self : Union[str, Any] , lowerCAmelCase_ : List[str]=0 , lowerCAmelCase_ : Union[str, Any]=0 , lowerCAmelCase_ : List[str]=False ) -> Optional[Any]: super().__init__() __lowerCAmelCase = torch.nn.Parameter(torch.tensor(lowerCAmelCase_ ).float() ) __lowerCAmelCase = torch.nn.Parameter(torch.tensor(lowerCAmelCase_ ).float() ) __lowerCAmelCase = True def lowercase ( self : Optional[int] , lowerCAmelCase_ : int=None ) -> int: if self.first_batch: print(f"""Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}""" ) __lowerCAmelCase = False return x * self.a + self.b def a_ ( lowerCAmelCase_ : Any, lowerCAmelCase_ : int = 16 ): from datasets import load_dataset from transformers import AutoTokenizer __lowerCAmelCase = AutoTokenizer.from_pretrained('bert-base-cased' ) __lowerCAmelCase = {'train': 'tests/test_samples/MRPC/train.csv', 'validation': 'tests/test_samples/MRPC/dev.csv'} __lowerCAmelCase = load_dataset('csv', data_files=lowerCAmelCase_ ) __lowerCAmelCase = datasets['train'].unique('label' ) __lowerCAmelCase = {v: i for i, v in enumerate(lowerCAmelCase_ )} def tokenize_function(lowerCAmelCase_ : str ): # max_length=None => use the model max length (it's actually the default) __lowerCAmelCase = tokenizer( examples['sentence1'], examples['sentence2'], truncation=lowerCAmelCase_, max_length=lowerCAmelCase_, padding='max_length' ) if "label" in examples: __lowerCAmelCase = [label_to_id[l] for l in examples['label']] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset __lowerCAmelCase = datasets.map( lowerCAmelCase_, batched=lowerCAmelCase_, remove_columns=['sentence1', 'sentence2', 'label'], ) def collate_fn(lowerCAmelCase_ : Tuple ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowerCAmelCase_, padding='max_length', max_length=128, return_tensors='pt' ) return tokenizer.pad(lowerCAmelCase_, padding='longest', return_tensors='pt' ) # Instantiate dataloaders. __lowerCAmelCase = DataLoader(tokenized_datasets['train'], shuffle=lowerCAmelCase_, collate_fn=lowerCAmelCase_, batch_size=2 ) __lowerCAmelCase = DataLoader(tokenized_datasets['validation'], shuffle=lowerCAmelCase_, collate_fn=lowerCAmelCase_, batch_size=1 ) return train_dataloader, eval_dataloader
284
import re from filelock import FileLock try: import nltk _snake_case : Any = True except (ImportError, ModuleNotFoundError): _snake_case : Union[str, Any] = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def a_ ( lowerCAmelCase_ : str ): re.sub('<n>', '', lowerCAmelCase_ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(lowerCAmelCase_ ) )
284
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: _snake_case : str = None _snake_case : Dict = logging.get_logger(__name__) _snake_case : List[str] = {'vocab_file': 'sentencepiece.model', 'tokenizer_file': 'tokenizer.json'} _snake_case : List[Any] = { 'vocab_file': { 'google/rembert': 'https://huggingface.co/google/rembert/resolve/main/sentencepiece.model', }, 'tokenizer_file': { 'google/rembert': 'https://huggingface.co/google/rembert/resolve/main/tokenizer.json', }, } _snake_case : str = { 'google/rembert': 256, } _snake_case : Optional[int] = '▁' class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = VOCAB_FILES_NAMES a_ = PRETRAINED_VOCAB_FILES_MAP a_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ = RemBertTokenizer def __init__( self : int , lowerCAmelCase_ : List[str]=None , lowerCAmelCase_ : int=None , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : Any=True , lowerCAmelCase_ : int=False , lowerCAmelCase_ : Any="[CLS]" , lowerCAmelCase_ : Dict="[SEP]" , lowerCAmelCase_ : Dict="<unk>" , lowerCAmelCase_ : str="[SEP]" , lowerCAmelCase_ : Tuple="<pad>" , lowerCAmelCase_ : Optional[int]="[CLS]" , lowerCAmelCase_ : int="[MASK]" , **lowerCAmelCase_ : Optional[Any] , ) -> Tuple: # Mask token behave like a normal word, i.e. include the space before it __lowerCAmelCase = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else mask_token super().__init__( lowerCAmelCase_ , tokenizer_file=lowerCAmelCase_ , do_lower_case=lowerCAmelCase_ , remove_space=lowerCAmelCase_ , keep_accents=lowerCAmelCase_ , bos_token=lowerCAmelCase_ , eos_token=lowerCAmelCase_ , unk_token=lowerCAmelCase_ , sep_token=lowerCAmelCase_ , pad_token=lowerCAmelCase_ , cls_token=lowerCAmelCase_ , mask_token=lowerCAmelCase_ , **lowerCAmelCase_ , ) __lowerCAmelCase = do_lower_case __lowerCAmelCase = remove_space __lowerCAmelCase = keep_accents __lowerCAmelCase = vocab_file __lowerCAmelCase = False if not self.vocab_file else True def lowercase ( self : Dict , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : Optional[List[int]] = None ) -> List[int]: __lowerCAmelCase = [self.sep_token_id] __lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def lowercase ( self : Tuple , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : Optional[List[int]] = None , lowerCAmelCase_ : bool = False ) -> List[int]: if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowerCAmelCase_ )) + [1] + ([0] * len(lowerCAmelCase_ )) + [1] return [1] + ([0] * len(lowerCAmelCase_ )) + [1] def lowercase ( self : Any , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : Optional[List[int]] = None ) -> List[int]: __lowerCAmelCase = [self.sep_token_id] __lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowercase ( self : int , lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase_ ): logger.error('Vocabulary path ({}) should be a directory'.format(lowerCAmelCase_ ) ) return __lowerCAmelCase = os.path.join( lowerCAmelCase_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase_ ): copyfile(self.vocab_file , lowerCAmelCase_ ) return (out_vocab_file,)
284
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case : List[Any] = { 'configuration_biogpt': ['BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BioGptConfig'], 'tokenization_biogpt': ['BioGptTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BioGptForCausalLM', 'BioGptForTokenClassification', 'BioGptForSequenceClassification', 'BioGptModel', 'BioGptPreTrainedModel', ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys _snake_case : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
1
def a_ ( lowerCAmelCase_ : list[list[float]] ): __lowerCAmelCase = [] for data in source_data: for i, el in enumerate(lowerCAmelCase_ ): if len(lowerCAmelCase_ ) < i + 1: data_lists.append([] ) data_lists[i].append(float(lowerCAmelCase_ ) ) return data_lists def a_ ( lowerCAmelCase_ : list[list[float]], lowerCAmelCase_ : list[int] ): __lowerCAmelCase = [] for dlist, weight in zip(lowerCAmelCase_, lowerCAmelCase_ ): __lowerCAmelCase = min(lowerCAmelCase_ ) __lowerCAmelCase = max(lowerCAmelCase_ ) __lowerCAmelCase = [] # for weight 0 score is 1 - actual score if weight == 0: for item in dlist: try: score.append(1 - ((item - mind) / (maxd - mind)) ) except ZeroDivisionError: score.append(1 ) elif weight == 1: for item in dlist: try: score.append((item - mind) / (maxd - mind) ) except ZeroDivisionError: score.append(0 ) # weight not 0 or 1 else: __lowerCAmelCase = F"""Invalid weight of {weight:f} provided""" raise ValueError(lowerCAmelCase_ ) score_lists.append(lowerCAmelCase_ ) return score_lists def a_ ( lowerCAmelCase_ : list[list[float]] ): __lowerCAmelCase = [0 for i in range(len(score_lists[0] ) )] for slist in score_lists: for j, ele in enumerate(lowerCAmelCase_ ): __lowerCAmelCase = final_scores[j] + ele return final_scores def a_ ( lowerCAmelCase_ : list[list[float]], lowerCAmelCase_ : list[int] ): __lowerCAmelCase = get_data(lowerCAmelCase_ ) __lowerCAmelCase = calculate_each_score(lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = generate_final_scores(lowerCAmelCase_ ) # append scores to source data for i, ele in enumerate(lowerCAmelCase_ ): source_data[i].append(lowerCAmelCase_ ) return source_data
284
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets _snake_case : Tuple = '\\n@article{hendrycksmath2021,\n title={Measuring Mathematical Problem Solving With the MATH Dataset},\n author={Dan Hendrycks\n and Collin Burns\n and Saurav Kadavath\n and Akul Arora\n and Steven Basart\n and Eric Tang\n and Dawn Song\n and Jacob Steinhardt},\n journal={arXiv preprint arXiv:2103.03874},\n year={2021}\n}\n' _snake_case : str = '\\nThis metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset.\nIt first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy.\n' _snake_case : List[str] = R'\nCalculates accuracy after canonicalizing inputs.\n\nArgs:\n predictions: list of predictions to score. Each prediction\n is a string that contains natural language and LaTex.\n references: list of reference for each prediction. Each\n reference is a string that contains natural language\n and LaTex.\nReturns:\n accuracy: accuracy after canonicalizing inputs\n (e.g., converting "1/2" to "\\frac{1}{2}")\n\nExamples:\n >>> metric = datasets.load_metric("competition_math")\n >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"])\n >>> print(results)\n {\'accuracy\': 1.0}\n' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): """simple docstring""" def lowercase ( self : str ) -> int: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' ), 'references': datasets.Value('string' ), } ) , homepage='https://github.com/hendrycks/math' , codebase_urls=['https://github.com/hendrycks/math'] , ) def lowercase ( self : str , lowerCAmelCase_ : Any , lowerCAmelCase_ : Optional[int] ) -> List[Any]: __lowerCAmelCase = 0.0 for i, j in zip(lowerCAmelCase_ , lowerCAmelCase_ ): n_correct += 1.0 if math_equivalence.is_equiv(lowerCAmelCase_ , lowerCAmelCase_ ) else 0.0 __lowerCAmelCase = n_correct / len(lowerCAmelCase_ ) return { "accuracy": accuracy, }
284
1
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : str = logging.get_logger(__name__) class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCAmelCase_ : Dict , **lowerCAmelCase_ : List[str] ) -> None: warnings.warn( 'The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use FlavaImageProcessor instead.' , lowerCAmelCase_ , ) super().__init__(*lowerCAmelCase_ , **lowerCAmelCase_ )
284
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = StableDiffusionInstructPixaPixPipeline a_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width""", """cross_attention_kwargs"""} a_ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase ( self : Optional[int] ) -> Optional[int]: torch.manual_seed(0 ) __lowerCAmelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=8 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , ) __lowerCAmelCase = PNDMScheduler(skip_prk_steps=lowerCAmelCase_ ) torch.manual_seed(0 ) __lowerCAmelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) __lowerCAmelCase = CLIPTextModel(lowerCAmelCase_ ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __lowerCAmelCase = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowercase ( self : Optional[Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Tuple=0 ) -> Dict: __lowerCAmelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) __lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCAmelCase = Image.fromarray(np.uinta(lowerCAmelCase_ ) ).convert('RGB' ) if str(lowerCAmelCase_ ).startswith('mps' ): __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) else: __lowerCAmelCase = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'image_guidance_scale': 1, 'output_type': 'numpy', } return inputs def lowercase ( self : Tuple ) -> List[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_26, 0.37_50, 0.45_47, 0.61_17, 0.58_66, 0.50_16, 0.43_27, 0.56_42, 0.48_15] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Dict: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = 'french fries' __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ , negative_prompt=lowerCAmelCase_ ) __lowerCAmelCase = output.images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_11, 0.36_42, 0.45_53, 0.62_36, 0.57_97, 0.50_13, 0.43_43, 0.56_11, 0.48_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Any: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = [inputs['prompt']] * 2 __lowerCAmelCase = np.array(inputs['image'] ).astype(np.floataa ) / 2_55.0 __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ).unsqueeze(0 ).to(lowerCAmelCase_ ) __lowerCAmelCase = image / 2 + 0.5 __lowerCAmelCase = image.permute(0 , 3 , 1 , 2 ) __lowerCAmelCase = image.repeat(2 , 1 , 1 , 1 ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[-1, -3:, -3:, -1] assert image.shape == (2, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.58_12, 0.57_48, 0.52_22, 0.59_08, 0.56_95, 0.71_74, 0.68_04, 0.55_23, 0.55_79] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Dict ) -> Optional[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = EulerAncestralDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' ) __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = [round(lowerCAmelCase_ , 4 ) for x in image_slice.flatten().tolist()] print(','.join([str(lowerCAmelCase_ ) for x in slice] ) ) assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.74_17, 0.38_42, 0.47_32, 0.57_76, 0.58_91, 0.51_39, 0.40_52, 0.56_73, 0.49_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> Optional[int]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def lowercase ( self : Optional[Any] ) -> Optional[Any]: __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = VaeImageProcessor(do_resize=lowerCAmelCase_ , do_normalize=lowerCAmelCase_ ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = pipe(**self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) )[0] __lowerCAmelCase = components['vae'] __lowerCAmelCase = self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): __lowerCAmelCase = vae.encode(inputs[image_param] ).latent_dist.mode() __lowerCAmelCase = pipe(**lowerCAmelCase_ )[0] __lowerCAmelCase = np.abs(out - out_latents_inputs ).max() self.assertLess(lowerCAmelCase_ , 1e-4 , 'passing latents as image input generate different result from passing image' ) @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> Optional[int]: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase ( self : List[str] , lowerCAmelCase_ : List[Any]=0 ) -> Any: __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = load_image( 'https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg' ) __lowerCAmelCase = { 'prompt': 'turn him into a cyborg', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'image_guidance_scale': 1.0, 'output_type': 'numpy', } return inputs def lowercase ( self : List[Any] ) -> str: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.59_02, 0.60_15, 0.60_27, 0.59_83, 0.60_92, 0.60_61, 0.57_65, 0.57_85, 0.55_55] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Tuple ) -> List[str]: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.65_78, 0.68_17, 0.69_72, 0.67_61, 0.68_56, 0.69_16, 0.64_28, 0.65_16, 0.63_01] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[Any] ) -> Dict: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.38_28, 0.38_34, 0.38_18, 0.37_92, 0.38_65, 0.37_52, 0.37_92, 0.38_47, 0.37_53] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> int: __lowerCAmelCase = 0 def callback_fn(lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : torch.FloatTensor ) -> None: __lowerCAmelCase = True nonlocal number_of_steps number_of_steps += 1 if step == 1: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.24_63, -0.46_44, -0.97_56, 1.51_76, 1.44_14, 0.78_66, 0.98_97, 0.85_21, 0.79_83] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.26_44, -0.46_26, -0.96_53, 1.51_76, 1.45_51, 0.76_86, 0.98_05, 0.84_52, 0.81_15] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 __lowerCAmelCase = False __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() pipe(**lowerCAmelCase_ , callback=lowerCAmelCase_ , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def lowercase ( self : Optional[int] ) -> Any: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 1_0**9 def lowercase ( self : List[Any] ) -> Any: __lowerCAmelCase = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 __lowerCAmelCase = inputs['image'].resize((5_0_4, 5_0_4) ) __lowerCAmelCase = 'timbrooks/instruct-pix2pix' __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( lowerCAmelCase_ , safety_checker=lowerCAmelCase_ , ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = output.images[0] __lowerCAmelCase = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 5_0_4, 3) __lowerCAmelCase = np.array([0.27_26, 0.25_29, 0.26_64, 0.26_55, 0.26_41, 0.26_42, 0.25_91, 0.26_49, 0.25_90] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3
284
1
import argparse import math import traceback import dateutil.parser as date_parser import requests def a_ ( lowerCAmelCase_ : str ): __lowerCAmelCase = {} __lowerCAmelCase = job['started_at'] __lowerCAmelCase = job['completed_at'] __lowerCAmelCase = date_parser.parse(lowerCAmelCase_ ) __lowerCAmelCase = date_parser.parse(lowerCAmelCase_ ) __lowerCAmelCase = round((end_datetime - start_datetime).total_seconds() / 60.0 ) __lowerCAmelCase = start __lowerCAmelCase = end __lowerCAmelCase = duration_in_min return job_info def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : int=None ): __lowerCAmelCase = None if token is not None: __lowerCAmelCase = {'Accept': 'application/vnd.github+json', 'Authorization': F"""Bearer {token}"""} __lowerCAmelCase = F"""https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100""" __lowerCAmelCase = requests.get(lowerCAmelCase_, headers=lowerCAmelCase_ ).json() __lowerCAmelCase = {} try: job_time.update({job['name']: extract_time_from_single_job(lowerCAmelCase_ ) for job in result['jobs']} ) __lowerCAmelCase = math.ceil((result['total_count'] - 100) / 100 ) for i in range(lowerCAmelCase_ ): __lowerCAmelCase = requests.get(url + F"""&page={i + 2}""", headers=lowerCAmelCase_ ).json() job_time.update({job['name']: extract_time_from_single_job(lowerCAmelCase_ ) for job in result['jobs']} ) return job_time except Exception: print(F"""Unknown error, could not fetch links:\n{traceback.format_exc()}""" ) return {} if __name__ == "__main__": _snake_case : int = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') _snake_case : Optional[int] = parser.parse_args() _snake_case : Any = get_job_time(args.workflow_run_id) _snake_case : Dict = dict(sorted(job_time.items(), key=lambda item: item[1]["duration"], reverse=True)) for k, v in job_time.items(): print(F"""{k}: {v["duration"]}""")
284
from timeit import timeit def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('the value of input must not be negative' ) __lowerCAmelCase = 0 while number: number &= number - 1 result += 1 return result def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('the value of input must not be negative' ) __lowerCAmelCase = 0 while number: if number % 2 == 1: result += 1 number >>= 1 return result def a_ ( ): def do_benchmark(lowerCAmelCase_ : int ) -> None: __lowerCAmelCase = 'import __main__ as z' print(F"""Benchmark when {number = }:""" ) print(F"""{get_set_bits_count_using_modulo_operator(lowerCAmelCase_ ) = }""" ) __lowerCAmelCase = timeit('z.get_set_bits_count_using_modulo_operator(25)', setup=lowerCAmelCase_ ) print(F"""timeit() runs in {timing} seconds""" ) print(F"""{get_set_bits_count_using_brian_kernighans_algorithm(lowerCAmelCase_ ) = }""" ) __lowerCAmelCase = timeit( 'z.get_set_bits_count_using_brian_kernighans_algorithm(25)', setup=lowerCAmelCase_, ) print(F"""timeit() runs in {timing} seconds""" ) for number in (25, 37, 58, 0): do_benchmark(lowerCAmelCase_ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
284
1
import os import sys _snake_case : str = os.path.join(os.path.dirname(__file__), 'src') sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) _snake_case : Union[str, Any] = [ 'torch', 'numpy', 'tokenizers', 'filelock', 'requests', 'tqdm', 'regex', 'sentencepiece', 'sacremoses', 'importlib_metadata', 'huggingface_hub', ] @add_start_docstrings(AutoConfig.__doc__ ) def a_ ( *lowerCAmelCase_ : Tuple, **lowerCAmelCase_ : List[Any] ): return AutoConfig.from_pretrained(*lowerCAmelCase_, **lowerCAmelCase_ ) @add_start_docstrings(AutoTokenizer.__doc__ ) def a_ ( *lowerCAmelCase_ : Union[str, Any], **lowerCAmelCase_ : Union[str, Any] ): return AutoTokenizer.from_pretrained(*lowerCAmelCase_, **lowerCAmelCase_ ) @add_start_docstrings(AutoModel.__doc__ ) def a_ ( *lowerCAmelCase_ : List[Any], **lowerCAmelCase_ : int ): return AutoModel.from_pretrained(*lowerCAmelCase_, **lowerCAmelCase_ ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def a_ ( *lowerCAmelCase_ : Dict, **lowerCAmelCase_ : Optional[int] ): return AutoModelForCausalLM.from_pretrained(*lowerCAmelCase_, **lowerCAmelCase_ ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def a_ ( *lowerCAmelCase_ : Dict, **lowerCAmelCase_ : List[Any] ): return AutoModelForMaskedLM.from_pretrained(*lowerCAmelCase_, **lowerCAmelCase_ ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def a_ ( *lowerCAmelCase_ : Optional[int], **lowerCAmelCase_ : Optional[int] ): return AutoModelForSequenceClassification.from_pretrained(*lowerCAmelCase_, **lowerCAmelCase_ ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def a_ ( *lowerCAmelCase_ : List[Any], **lowerCAmelCase_ : str ): return AutoModelForQuestionAnswering.from_pretrained(*lowerCAmelCase_, **lowerCAmelCase_ )
284
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss _snake_case : Dict = pytest.mark.integration @require_faiss class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : List[Any] ) -> Optional[Any]: __lowerCAmelCase = Dataset.from_dict({'filename': ['my_name-train' + '_' + str(lowerCAmelCase_ ) for x in np.arange(3_0 ).tolist()]} ) return dset def lowercase ( self : List[str] ) -> Tuple: import faiss __lowerCAmelCase = self._create_dummy_dataset() __lowerCAmelCase = dset.map( lambda lowerCAmelCase_ , lowerCAmelCase_ : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=lowerCAmelCase_ , keep_in_memory=lowerCAmelCase_ ) __lowerCAmelCase = dset.add_faiss_index('vecs' , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) dset.drop_index('vecs' ) def lowercase ( self : Optional[Any] ) -> str: import faiss __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT , ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def lowercase ( self : int ) -> Optional[Any]: import faiss __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCAmelCase_ ) as tmp_file: dset.save_faiss_index('vecs' , tmp_file.name ) dset.load_faiss_index('vecs2' , tmp_file.name ) os.unlink(tmp_file.name ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs2' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def lowercase ( self : Union[str, Any] ) -> List[Any]: __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' ) dset.drop_index('vecs' ) self.assertRaises(lowerCAmelCase_ , partial(dset.get_nearest_examples , 'vecs2' , np.ones(5 , dtype=np.floataa ) ) ) def lowercase ( self : Union[str, Any] ) -> Tuple: from elasticsearch import Elasticsearch __lowerCAmelCase = self._create_dummy_dataset() with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: __lowerCAmelCase = {'acknowledged': True} mocked_bulk.return_value([(True, None)] * 3_0 ) __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 2_9}]}} __lowerCAmelCase = Elasticsearch() dset.add_elasticsearch_index('filename' , es_client=lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('filename' , 'my_name-train_29' ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) @require_faiss class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : str ) -> int: import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 1_0 ) # single query __lowerCAmelCase = np.zeros(5 , dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertRaises(lowerCAmelCase_ , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries __lowerCAmelCase = np.eye(5 , dtype=np.floataa )[::-1] __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ ) self.assertRaises(lowerCAmelCase_ , index.search_batch , queries[0] ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , lowerCAmelCase_ ) def lowercase ( self : List[Any] ) -> List[str]: import faiss __lowerCAmelCase = FaissIndex(string_factory='Flat' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) __lowerCAmelCase = FaissIndex(string_factory='LSH' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(lowerCAmelCase_ ): __lowerCAmelCase = FaissIndex(string_factory='Flat' , custom_index=faiss.IndexFlat(5 ) ) def lowercase ( self : Union[str, Any] ) -> Dict: import faiss __lowerCAmelCase = faiss.IndexFlat(5 ) __lowerCAmelCase = FaissIndex(custom_index=lowerCAmelCase_ ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def lowercase ( self : str ) -> Any: import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCAmelCase_ ) as tmp_file: index.save(tmp_file.name ) __lowerCAmelCase = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) __lowerCAmelCase = np.zeros(5 , dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def a_ ( lowerCAmelCase_ : Union[str, Any] ): import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5, dtype=np.floataa ) ) __lowerCAmelCase = 'index.faiss' __lowerCAmelCase = F"""mock://{index_name}""" index.save(lowerCAmelCase_, storage_options=mockfs.storage_options ) __lowerCAmelCase = FaissIndex.load(lowerCAmelCase_, storage_options=mockfs.storage_options ) __lowerCAmelCase = np.zeros(5, dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : Any ) -> int: from elasticsearch import Elasticsearch with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: __lowerCAmelCase = Elasticsearch() __lowerCAmelCase = {'acknowledged': True} __lowerCAmelCase = ElasticSearchIndex(es_client=lowerCAmelCase_ ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(['foo', 'bar', 'foobar'] ) # single query __lowerCAmelCase = 'foo' __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout __lowerCAmelCase = 'foo' __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ , request_timeout=3_0 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries __lowerCAmelCase = ['foo', 'bar', 'foobar'] __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([1, 1, 1] , lowerCAmelCase_ ) # batched queries with timeout __lowerCAmelCase = ['foo', 'bar', 'foobar'] __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ , request_timeout=3_0 ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([1, 1, 1] , lowerCAmelCase_ )
284
1
import inspect from typing import Callable, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import DiffusionPipeline from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import logging _snake_case : List[str] = logging.get_logger(__name__) # pylint: disable=invalid-name class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def __init__( self : Optional[int] , lowerCAmelCase_ : AutoencoderKL , lowerCAmelCase_ : CLIPTextModel , lowerCAmelCase_ : CLIPTokenizer , lowerCAmelCase_ : UNetaDConditionModel , lowerCAmelCase_ : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCAmelCase_ : StableDiffusionSafetyChecker , lowerCAmelCase_ : CLIPImageProcessor , ) -> str: super().__init__() self.register_modules( vae=lowerCAmelCase_ , text_encoder=lowerCAmelCase_ , tokenizer=lowerCAmelCase_ , unet=lowerCAmelCase_ , scheduler=lowerCAmelCase_ , safety_checker=lowerCAmelCase_ , feature_extractor=lowerCAmelCase_ , ) def lowercase ( self : int , lowerCAmelCase_ : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __lowerCAmelCase = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCAmelCase_ ) def lowercase ( self : Tuple ) -> Any: self.enable_attention_slicing(lowerCAmelCase_ ) @torch.no_grad() def __call__( self : Tuple , lowerCAmelCase_ : Union[str, List[str]] , lowerCAmelCase_ : int = 5_1_2 , lowerCAmelCase_ : int = 5_1_2 , lowerCAmelCase_ : int = 5_0 , lowerCAmelCase_ : float = 7.5 , lowerCAmelCase_ : Optional[Union[str, List[str]]] = None , lowerCAmelCase_ : Optional[int] = 1 , lowerCAmelCase_ : float = 0.0 , lowerCAmelCase_ : Optional[torch.Generator] = None , lowerCAmelCase_ : Optional[torch.FloatTensor] = None , lowerCAmelCase_ : Optional[str] = "pil" , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCAmelCase_ : int = 1 , lowerCAmelCase_ : Optional[torch.FloatTensor] = None , **lowerCAmelCase_ : Optional[Any] , ) -> Any: if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): __lowerCAmelCase = 1 elif isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): __lowerCAmelCase = len(lowerCAmelCase_ ) else: raise ValueError(f"""`prompt` has to be of type `str` or `list` but is {type(lowerCAmelCase_ )}""" ) if height % 8 != 0 or width % 8 != 0: raise ValueError(f"""`height` and `width` have to be divisible by 8 but are {height} and {width}.""" ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or callback_steps <= 0) ): raise ValueError( f"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" f""" {type(lowerCAmelCase_ )}.""" ) # get prompt text embeddings __lowerCAmelCase = self.tokenizer( lowerCAmelCase_ , padding='max_length' , max_length=self.tokenizer.model_max_length , return_tensors='pt' , ) __lowerCAmelCase = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: __lowerCAmelCase = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( 'The following part of your input was truncated because CLIP can only handle sequences up to' f""" {self.tokenizer.model_max_length} tokens: {removed_text}""" ) __lowerCAmelCase = text_input_ids[:, : self.tokenizer.model_max_length] if text_embeddings is None: __lowerCAmelCase = self.text_encoder(text_input_ids.to(self.device ) )[0] # duplicate text embeddings for each generation per prompt, using mps friendly method __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = text_embeddings.shape __lowerCAmelCase = text_embeddings.repeat(1 , lowerCAmelCase_ , 1 ) __lowerCAmelCase = text_embeddings.view(bs_embed * num_images_per_prompt , lowerCAmelCase_ , -1 ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. __lowerCAmelCase = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: __lowerCAmelCase = 42 if negative_prompt is None: __lowerCAmelCase = [''] elif type(lowerCAmelCase_ ) is not type(lowerCAmelCase_ ): raise TypeError( f"""`negative_prompt` should be the same type to `prompt`, but got {type(lowerCAmelCase_ )} !=""" f""" {type(lowerCAmelCase_ )}.""" ) elif isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): __lowerCAmelCase = [negative_prompt] elif batch_size != len(lowerCAmelCase_ ): raise ValueError( f"""`negative_prompt`: {negative_prompt} has batch size {len(lowerCAmelCase_ )}, but `prompt`:""" f""" {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches""" ' the batch size of `prompt`.' ) else: __lowerCAmelCase = negative_prompt __lowerCAmelCase = text_input_ids.shape[-1] __lowerCAmelCase = self.tokenizer( lowerCAmelCase_ , padding='max_length' , max_length=lowerCAmelCase_ , truncation=lowerCAmelCase_ , return_tensors='pt' , ) __lowerCAmelCase = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt, using mps friendly method __lowerCAmelCase = uncond_embeddings.shape[1] __lowerCAmelCase = uncond_embeddings.repeat(lowerCAmelCase_ , lowerCAmelCase_ , 1 ) __lowerCAmelCase = uncond_embeddings.view(batch_size * num_images_per_prompt , lowerCAmelCase_ , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes __lowerCAmelCase = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. __lowerCAmelCase = (batch_size * num_images_per_prompt, self.unet.config.in_channels, height // 8, width // 8) __lowerCAmelCase = (batch_size * num_images_per_prompt, self.unet.config.in_channels, 6_4, 6_4) __lowerCAmelCase = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not exist on mps __lowerCAmelCase = torch.randn( lowerCAmelCase_ , generator=lowerCAmelCase_ , device='cpu' , dtype=lowerCAmelCase_ ).to(self.device ) __lowerCAmelCase = torch.randn(lowerCAmelCase_ , generator=lowerCAmelCase_ , device='cpu' , dtype=lowerCAmelCase_ ).to( self.device ) else: __lowerCAmelCase = torch.randn( lowerCAmelCase_ , generator=lowerCAmelCase_ , device=self.device , dtype=lowerCAmelCase_ ) __lowerCAmelCase = torch.randn(lowerCAmelCase_ , generator=lowerCAmelCase_ , device=self.device , dtype=lowerCAmelCase_ ) else: if latents_reference.shape != latents_shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {latents_shape}""" ) __lowerCAmelCase = latents_reference.to(self.device ) __lowerCAmelCase = latents.to(self.device ) # This is the key part of the pipeline where we # try to ensure that the generated images w/ the same seed # but different sizes actually result in similar images __lowerCAmelCase = (latents_shape[3] - latents_shape_reference[3]) // 2 __lowerCAmelCase = (latents_shape[2] - latents_shape_reference[2]) // 2 __lowerCAmelCase = latents_shape_reference[3] if dx >= 0 else latents_shape_reference[3] + 2 * dx __lowerCAmelCase = latents_shape_reference[2] if dy >= 0 else latents_shape_reference[2] + 2 * dy __lowerCAmelCase = 0 if dx < 0 else dx __lowerCAmelCase = 0 if dy < 0 else dy __lowerCAmelCase = max(-dx , 0 ) __lowerCAmelCase = max(-dy , 0 ) # import pdb # pdb.set_trace() __lowerCAmelCase = latents_reference[:, :, dy : dy + h, dx : dx + w] # set timesteps self.scheduler.set_timesteps(lowerCAmelCase_ ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand __lowerCAmelCase = self.scheduler.timesteps.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler __lowerCAmelCase = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __lowerCAmelCase = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __lowerCAmelCase = {} if accepts_eta: __lowerCAmelCase = eta for i, t in enumerate(self.progress_bar(lowerCAmelCase_ ) ): # expand the latents if we are doing classifier free guidance __lowerCAmelCase = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __lowerCAmelCase = self.scheduler.scale_model_input(lowerCAmelCase_ , lowerCAmelCase_ ) # predict the noise residual __lowerCAmelCase = self.unet(lowerCAmelCase_ , lowerCAmelCase_ , encoder_hidden_states=lowerCAmelCase_ ).sample # perform guidance if do_classifier_free_guidance: __lowerCAmelCase , __lowerCAmelCase = noise_pred.chunk(2 ) __lowerCAmelCase = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # compute the previous noisy sample x_t -> x_t-1 __lowerCAmelCase = self.scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) __lowerCAmelCase = 1 / 0.1_82_15 * latents __lowerCAmelCase = self.vae.decode(lowerCAmelCase_ ).sample __lowerCAmelCase = (image / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 __lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if self.safety_checker is not None: __lowerCAmelCase = self.feature_extractor(self.numpy_to_pil(lowerCAmelCase_ ) , return_tensors='pt' ).to( self.device ) __lowerCAmelCase , __lowerCAmelCase = self.safety_checker( images=lowerCAmelCase_ , clip_input=safety_checker_input.pixel_values.to(text_embeddings.dtype ) ) else: __lowerCAmelCase = None if output_type == "pil": __lowerCAmelCase = self.numpy_to_pil(lowerCAmelCase_ ) if not return_dict: return (image, has_nsfw_concept) return StableDiffusionPipelineOutput(images=lowerCAmelCase_ , nsfw_content_detected=lowerCAmelCase_ )
284
import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( 'kwargs, expected', [ ({'num_shards': 0, 'max_num_jobs': 1}, []), ({'num_shards': 10, 'max_num_jobs': 1}, [range(10 )]), ({'num_shards': 10, 'max_num_jobs': 10}, [range(lowerCAmelCase_, i + 1 ) for i in range(10 )]), ({'num_shards': 1, 'max_num_jobs': 10}, [range(1 )]), ({'num_shards': 10, 'max_num_jobs': 3}, [range(0, 4 ), range(4, 7 ), range(7, 10 )]), ({'num_shards': 3, 'max_num_jobs': 10}, [range(0, 1 ), range(1, 2 ), range(2, 3 )]), ], ) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : Any ): __lowerCAmelCase = _distribute_shards(**lowerCAmelCase_ ) assert out == expected @pytest.mark.parametrize( 'gen_kwargs, max_num_jobs, expected', [ ({'foo': 0}, 10, [{'foo': 0}]), ({'shards': [0, 1, 2, 3]}, 1, [{'shards': [0, 1, 2, 3]}]), ({'shards': [0, 1, 2, 3]}, 4, [{'shards': [0]}, {'shards': [1]}, {'shards': [2]}, {'shards': [3]}]), ({'shards': [0, 1]}, 4, [{'shards': [0]}, {'shards': [1]}]), ({'shards': [0, 1, 2, 3]}, 2, [{'shards': [0, 1]}, {'shards': [2, 3]}]), ], ) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : List[str], lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = _split_gen_kwargs(lowerCAmelCase_, lowerCAmelCase_ ) assert out == expected @pytest.mark.parametrize( 'gen_kwargs, expected', [ ({'foo': 0}, 1), ({'shards': [0]}, 1), ({'shards': [0, 1, 2, 3]}, 4), ({'shards': [0, 1, 2, 3], 'foo': 0}, 4), ({'shards': [0, 1, 2, 3], 'other': (0, 1)}, 4), ({'shards': [0, 1, 2, 3], 'shards2': [0, 1]}, RuntimeError), ], ) def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : Any ): if expected is RuntimeError: with pytest.raises(lowerCAmelCase_ ): _number_of_shards_in_gen_kwargs(lowerCAmelCase_ ) else: __lowerCAmelCase = _number_of_shards_in_gen_kwargs(lowerCAmelCase_ ) assert out == expected
284
1
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = StableUnCLIPImgaImgPipeline a_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS a_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a_ = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess a_ = frozenset([] ) def lowercase ( self : List[Any] ) -> Tuple: __lowerCAmelCase = 3_2 __lowerCAmelCase = embedder_hidden_size # image encoding components __lowerCAmelCase = CLIPImageProcessor(crop_size=3_2 , size=3_2 ) torch.manual_seed(0 ) __lowerCAmelCase = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=lowerCAmelCase_ , projection_dim=lowerCAmelCase_ , num_hidden_layers=5 , num_attention_heads=4 , image_size=3_2 , intermediate_size=3_7 , patch_size=1 , ) ) # regular denoising components torch.manual_seed(0 ) __lowerCAmelCase = StableUnCLIPImageNormalizer(embedding_dim=lowerCAmelCase_ ) __lowerCAmelCase = DDPMScheduler(beta_schedule='squaredcos_cap_v2' ) torch.manual_seed(0 ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) torch.manual_seed(0 ) __lowerCAmelCase = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=lowerCAmelCase_ , projection_dim=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) ) torch.manual_seed(0 ) __lowerCAmelCase = UNetaDConditionModel( sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('CrossAttnDownBlock2D', 'DownBlock2D') , up_block_types=('UpBlock2D', 'CrossAttnUpBlock2D') , block_out_channels=(3_2, 6_4) , attention_head_dim=(2, 4) , class_embed_type='projection' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=lowerCAmelCase_ , layers_per_block=1 , upcast_attention=lowerCAmelCase_ , use_linear_projection=lowerCAmelCase_ , ) torch.manual_seed(0 ) __lowerCAmelCase = DDIMScheduler( beta_schedule='scaled_linear' , beta_start=0.0_00_85 , beta_end=0.0_12 , prediction_type='v_prediction' , set_alpha_to_one=lowerCAmelCase_ , steps_offset=1 , ) torch.manual_seed(0 ) __lowerCAmelCase = AutoencoderKL() __lowerCAmelCase = { # image encoding components 'feature_extractor': feature_extractor, 'image_encoder': image_encoder.eval(), # image noising components 'image_normalizer': image_normalizer.eval(), 'image_noising_scheduler': image_noising_scheduler, # regular denoising components 'tokenizer': tokenizer, 'text_encoder': text_encoder.eval(), 'unet': unet.eval(), 'scheduler': scheduler, 'vae': vae.eval(), } return components def lowercase ( self : Optional[Any] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Dict=0 , lowerCAmelCase_ : List[str]=True ) -> Optional[Any]: if str(lowerCAmelCase_ ).startswith('mps' ): __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) else: __lowerCAmelCase = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) if pil_image: __lowerCAmelCase = input_image * 0.5 + 0.5 __lowerCAmelCase = input_image.clamp(0 , 1 ) __lowerCAmelCase = input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() __lowerCAmelCase = DiffusionPipeline.numpy_to_pil(lowerCAmelCase_ )[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def lowercase ( self : Dict ) -> str: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableUnCLIPImgaImgPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) inputs.update({'image_embeds': None} ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.38_72, 0.72_24, 0.56_01, 0.47_41, 0.68_72, 0.58_14, 0.46_36, 0.38_67, 0.50_78] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Optional[Any] ) -> Union[str, Any]: __lowerCAmelCase = torch_device in ['cpu', 'mps'] self._test_attention_slicing_forward_pass(test_max_difference=lowerCAmelCase_ ) def lowercase ( self : int ) -> int: __lowerCAmelCase = torch_device in ['cpu', 'mps'] self._test_inference_batch_single_identical(test_max_difference=lowerCAmelCase_ ) @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def lowercase ( self : List[Any] ) -> Union[str, Any]: self._test_xformers_attention_forwardGenerator_pass(test_max_difference=lowerCAmelCase_ ) @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase ( self : Optional[Any] ) -> Any: __lowerCAmelCase = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png' ) __lowerCAmelCase = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy' ) __lowerCAmelCase = StableUnCLIPImgaImgPipeline.from_pretrained( 'fusing/stable-unclip-2-1-l-img2img' , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __lowerCAmelCase = torch.Generator(device='cpu' ).manual_seed(0 ) __lowerCAmelCase = pipe(lowerCAmelCase_ , 'anime turle' , generator=lowerCAmelCase_ , output_type='np' ) __lowerCAmelCase = output.images[0] assert image.shape == (7_6_8, 7_6_8, 3) assert_mean_pixel_difference(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Optional[Any] ) -> List[Any]: __lowerCAmelCase = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png' ) __lowerCAmelCase = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy' ) __lowerCAmelCase = StableUnCLIPImgaImgPipeline.from_pretrained( 'fusing/stable-unclip-2-1-h-img2img' , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __lowerCAmelCase = torch.Generator(device='cpu' ).manual_seed(0 ) __lowerCAmelCase = pipe(lowerCAmelCase_ , 'anime turle' , generator=lowerCAmelCase_ , output_type='np' ) __lowerCAmelCase = output.images[0] assert image.shape == (7_6_8, 7_6_8, 3) assert_mean_pixel_difference(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Optional[int] ) -> Optional[Any]: __lowerCAmelCase = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png' ) torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCAmelCase = StableUnCLIPImgaImgPipeline.from_pretrained( 'fusing/stable-unclip-2-1-h-img2img' , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __lowerCAmelCase = pipe( lowerCAmelCase_ , 'anime turtle' , num_inference_steps=2 , output_type='np' , ) __lowerCAmelCase = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 1_0**9
284
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class _UpperCAmelCase ( _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = BertJapaneseTokenizer a_ = False a_ = True def lowercase ( self : Optional[Any] ) -> List[str]: super().setUp() __lowerCAmelCase = [ '[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは', '世界', '##世界', '、', '##、', '。', '##。', ] __lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def lowercase ( self : List[Any] , lowerCAmelCase_ : Tuple ) -> str: __lowerCAmelCase = 'こんにちは、世界。 \nこんばんは、世界。' __lowerCAmelCase = 'こんにちは 、 世界 。 こんばんは 、 世界 。' return input_text, output_text def lowercase ( self : List[Any] , lowerCAmelCase_ : str ) -> Dict: __lowerCAmelCase , __lowerCAmelCase = self.get_input_output_texts(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(lowerCAmelCase_ , clean_up_tokenization_spaces=lowerCAmelCase_ ) return text, ids def lowercase ( self : List[str] ) -> Optional[int]: pass # TODO add if relevant def lowercase ( self : Optional[Any] ) -> Optional[Any]: pass # TODO add if relevant def lowercase ( self : Union[str, Any] ) -> Any: pass # TODO add if relevant def lowercase ( self : Dict ) -> Tuple: __lowerCAmelCase = self.tokenizer_class(self.vocab_file ) __lowerCAmelCase = tokenizer.tokenize('こんにちは、世界。\nこんばんは、世界。' ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) def lowercase ( self : List[str] ) -> List[str]: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='mecab' ) self.assertIsNotNone(lowerCAmelCase_ ) __lowerCAmelCase = 'こんにちは、世界。\nこんばんは、世界。' __lowerCAmelCase = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __lowerCAmelCase = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowerCAmelCase_ , 'wb' ) as handle: pickle.dump(lowerCAmelCase_ , lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'rb' ) as handle: __lowerCAmelCase = pickle.load(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer_new.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Dict ) -> Tuple: __lowerCAmelCase = MecabTokenizer(mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : List[Any] ) -> int: try: __lowerCAmelCase = MecabTokenizer(mecab_dic='unidic_lite' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : Tuple ) -> Optional[Any]: try: __lowerCAmelCase = MecabTokenizer(mecab_dic='unidic' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : Tuple ) -> Union[str, Any]: __lowerCAmelCase = MecabTokenizer(do_lower_case=lowerCAmelCase_ , mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iphone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : Union[str, Any] ) -> Optional[Any]: try: __lowerCAmelCase = MecabTokenizer( do_lower_case=lowerCAmelCase_ , normalize_text=lowerCAmelCase_ , mecab_option='-d /usr/local/lib/mecab/dic/jumandic' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れた', '\u3000', '。'] , ) def lowercase ( self : Any ) -> Union[str, Any]: __lowerCAmelCase = MecabTokenizer(normalize_text=lowerCAmelCase_ , mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', ' ', '。'] , ) @require_sudachi def lowercase ( self : List[str] ) -> List[str]: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='sudachi' ) self.assertIsNotNone(lowerCAmelCase_ ) __lowerCAmelCase = 'こんにちは、世界。\nこんばんは、世界。' __lowerCAmelCase = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __lowerCAmelCase = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowerCAmelCase_ , 'wb' ) as handle: pickle.dump(lowerCAmelCase_ , lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'rb' ) as handle: __lowerCAmelCase = pickle.load(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer_new.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) @require_sudachi def lowercase ( self : Union[str, Any] ) -> List[str]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iPhone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', ' ', '。', ' ', ' '] , ) @require_sudachi def lowercase ( self : Tuple ) -> Optional[Any]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='A' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国', '人', '参政', '権'] ) @require_sudachi def lowercase ( self : Tuple ) -> List[Any]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='B' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国人', '参政権'] ) @require_sudachi def lowercase ( self : List[str] ) -> Union[str, Any]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='C' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国人参政権'] ) @require_sudachi def lowercase ( self : Dict ) -> List[str]: __lowerCAmelCase = SudachiTokenizer(do_lower_case=lowerCAmelCase_ , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iphone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', ' ', '。', ' ', ' '] , ) @require_sudachi def lowercase ( self : Union[str, Any] ) -> List[Any]: __lowerCAmelCase = SudachiTokenizer(normalize_text=lowerCAmelCase_ , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iPhone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', '\u3000', '。', ' ', ' '] , ) @require_sudachi def lowercase ( self : int ) -> str: __lowerCAmelCase = SudachiTokenizer(trim_whitespace=lowerCAmelCase_ , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) @require_jumanpp def lowercase ( self : Union[str, Any] ) -> Any: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='jumanpp' ) self.assertIsNotNone(lowerCAmelCase_ ) __lowerCAmelCase = 'こんにちは、世界。\nこんばんは、世界。' __lowerCAmelCase = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __lowerCAmelCase = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowerCAmelCase_ , 'wb' ) as handle: pickle.dump(lowerCAmelCase_ , lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'rb' ) as handle: __lowerCAmelCase = pickle.load(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer_new.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) @require_jumanpp def lowercase ( self : List[Any] ) -> Optional[Any]: __lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def lowercase ( self : Any ) -> Union[str, Any]: __lowerCAmelCase = JumanppTokenizer(do_lower_case=lowerCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iphone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def lowercase ( self : Dict ) -> Dict: __lowerCAmelCase = JumanppTokenizer(normalize_text=lowerCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['ア', 'ッ', 'フ', '゚', 'ル', 'ストア', 'で', 'iPhone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def lowercase ( self : List[str] ) -> List[str]: __lowerCAmelCase = JumanppTokenizer(trim_whitespace=lowerCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れた', '。'] , ) @require_jumanpp def lowercase ( self : Any ) -> Any: __lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('ありがとうございますm(_ _)m見つけるのが大変です。' ) , ['ありがとう', 'ございます', 'm(_ _)m', '見つける', 'の', 'が', '大変です', '。'] , ) def lowercase ( self : Any ) -> str: __lowerCAmelCase = ['[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは'] __lowerCAmelCase = {} for i, token in enumerate(lowerCAmelCase_ ): __lowerCAmelCase = i __lowerCAmelCase = WordpieceTokenizer(vocab=lowerCAmelCase_ , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('こんにちは' ) , ['こんにちは'] ) self.assertListEqual(tokenizer.tokenize('こんばんは' ) , ['こん', '##ばんは'] ) self.assertListEqual(tokenizer.tokenize('こんばんは こんばんにちは こんにちは' ) , ['こん', '##ばんは', '[UNK]', 'こんにちは'] ) def lowercase ( self : List[Any] ) -> Tuple: __lowerCAmelCase = BertJapaneseTokenizer.from_pretrained('nlp-waseda/roberta-base-japanese-with-auto-jumanpp' ) __lowerCAmelCase = tokenizer.subword_tokenizer __lowerCAmelCase = subword_tokenizer.tokenize('国境 の 長い トンネル を 抜ける と 雪国 であった 。' ) self.assertListEqual(lowerCAmelCase_ , ['▁国境', '▁の', '▁長い', '▁トンネル', '▁を', '▁抜ける', '▁と', '▁雪', '国', '▁であった', '▁。'] ) __lowerCAmelCase = subword_tokenizer.tokenize('こんばんは こんばん にち は こんにちは' ) self.assertListEqual(lowerCAmelCase_ , ['▁こん', 'ばん', 'は', '▁こん', 'ばん', '▁に', 'ち', '▁は', '▁こんにちは'] ) def lowercase ( self : int ) -> str: __lowerCAmelCase = self.tokenizer_class.from_pretrained('cl-tohoku/bert-base-japanese' ) __lowerCAmelCase = tokenizer.encode('ありがとう。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.encode('どういたしまして。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ , lowerCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class _UpperCAmelCase ( _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = BertJapaneseTokenizer a_ = False def lowercase ( self : Optional[Any] ) -> Tuple: super().setUp() __lowerCAmelCase = ['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] __lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def lowercase ( self : str , **lowerCAmelCase_ : Tuple ) -> Union[str, Any]: return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='character' , **lowerCAmelCase_ ) def lowercase ( self : Tuple , lowerCAmelCase_ : Tuple ) -> Optional[int]: __lowerCAmelCase = 'こんにちは、世界。 \nこんばんは、世界。' __lowerCAmelCase = 'こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。' return input_text, output_text def lowercase ( self : Dict ) -> str: pass # TODO add if relevant def lowercase ( self : Any ) -> str: pass # TODO add if relevant def lowercase ( self : List[Any] ) -> int: pass # TODO add if relevant def lowercase ( self : str ) -> str: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , subword_tokenizer_type='character' ) __lowerCAmelCase = tokenizer.tokenize('こんにちは、世界。 \nこんばんは、世界。' ) self.assertListEqual( lowerCAmelCase_ , ['こ', 'ん', 'に', 'ち', 'は', '、', '世', '界', '。', 'こ', 'ん', 'ば', 'ん', 'は', '、', '世', '界', '。'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 4, 5, 6, 7, 1_1, 9, 1_0, 1_2, 3, 4, 8, 4, 7, 1_1, 9, 1_0, 1_2] ) def lowercase ( self : str ) -> Optional[int]: __lowerCAmelCase = ['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] __lowerCAmelCase = {} for i, token in enumerate(lowerCAmelCase_ ): __lowerCAmelCase = i __lowerCAmelCase = CharacterTokenizer(vocab=lowerCAmelCase_ , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('こんにちは' ) , ['こ', 'ん', 'に', 'ち', 'は'] ) self.assertListEqual(tokenizer.tokenize('こんにちほ' ) , ['こ', 'ん', 'に', 'ち', '[UNK]'] ) def lowercase ( self : int ) -> str: __lowerCAmelCase = self.tokenizer_class.from_pretrained('cl-tohoku/bert-base-japanese-char' ) __lowerCAmelCase = tokenizer.encode('ありがとう。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.encode('どういたしまして。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ , lowerCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : str ) -> Union[str, Any]: __lowerCAmelCase = 'cl-tohoku/bert-base-japanese' __lowerCAmelCase = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : List[str] ) -> Optional[int]: __lowerCAmelCase = 'cl-tohoku/bert-base-japanese' with self.assertLogs('transformers' , level='WARNING' ) as cm: BertTokenizer.from_pretrained(lowerCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( 'The tokenizer class you load from this checkpoint is not the same type as the class this function' ' is called from.' ) ) __lowerCAmelCase = 'bert-base-cased' with self.assertLogs('transformers' , level='WARNING' ) as cm: BertJapaneseTokenizer.from_pretrained(lowerCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( 'The tokenizer class you load from this checkpoint is not the same type as the class this function' ' is called from.' ) )
284
1
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets _snake_case : Tuple = '\\n@inproceedings{pillutla-etal:mauve:neurips2021,\n title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers},\n author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid},\n booktitle = {NeurIPS},\n year = {2021}\n}\n\n' _snake_case : str = '\\nMAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure.\n\nMAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences.\n\nFor details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021).\n\nThis metrics is a wrapper around the official implementation of MAUVE:\nhttps://github.com/krishnap25/mauve\n' _snake_case : List[str] = '\nCalculates MAUVE scores between two lists of generated text and reference text.\nArgs:\n predictions: list of generated text to score. Each predictions\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\nOptional Args:\n num_buckets: the size of the histogram to quantize P and Q. Options: \'auto\' (default) or an integer\n pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1\n kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9\n kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5\n kmeans_max_iter: maximum number of k-means iterations. Default 500\n featurize_model_name: name of the model from which features are obtained. Default \'gpt2-large\' Use one of [\'gpt2\', \'gpt2-medium\', \'gpt2-large\', \'gpt2-xl\'].\n device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU\n max_text_length: maximum number of tokens to consider. Default 1024\n divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25\n mauve_scaling_factor: "c" from the paper. Default 5.\n verbose: If True (default), print running time updates\n seed: random seed to initialize k-means cluster assignments.\nReturns:\n mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer,\n frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer,\n divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve,\n p_hist: a discrete distribution, which is a quantized version of the text distribution p_text,\n q_hist: same as above, but with q_text.\nExamples:\n\n >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest\n >>> import datasets\n >>> mauve = datasets.load_metric(\'mauve\')\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP\n >>> print(out.mauve) # doctest: +SKIP\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): """simple docstring""" def lowercase ( self : Any ) -> Dict: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='https://github.com/krishnap25/mauve' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/krishnap25/mauve'] , reference_urls=[ 'https://arxiv.org/abs/2102.01454', 'https://github.com/krishnap25/mauve', ] , ) def lowercase ( self : int , lowerCAmelCase_ : int , lowerCAmelCase_ : Any , lowerCAmelCase_ : Union[str, Any]=None , lowerCAmelCase_ : List[Any]=None , lowerCAmelCase_ : int=None , lowerCAmelCase_ : List[str]=None , lowerCAmelCase_ : Tuple="auto" , lowerCAmelCase_ : int=-1 , lowerCAmelCase_ : Optional[Any]=0.9 , lowerCAmelCase_ : List[Any]=5 , lowerCAmelCase_ : int=5_0_0 , lowerCAmelCase_ : Any="gpt2-large" , lowerCAmelCase_ : Any=-1 , lowerCAmelCase_ : Dict=1_0_2_4 , lowerCAmelCase_ : Optional[Any]=2_5 , lowerCAmelCase_ : List[Any]=5 , lowerCAmelCase_ : int=True , lowerCAmelCase_ : Dict=2_5 , ) -> Tuple: __lowerCAmelCase = compute_mauve( p_text=lowerCAmelCase_ , q_text=lowerCAmelCase_ , p_features=lowerCAmelCase_ , q_features=lowerCAmelCase_ , p_tokens=lowerCAmelCase_ , q_tokens=lowerCAmelCase_ , num_buckets=lowerCAmelCase_ , pca_max_data=lowerCAmelCase_ , kmeans_explained_var=lowerCAmelCase_ , kmeans_num_redo=lowerCAmelCase_ , kmeans_max_iter=lowerCAmelCase_ , featurize_model_name=lowerCAmelCase_ , device_id=lowerCAmelCase_ , max_text_length=lowerCAmelCase_ , divergence_curve_discretization_size=lowerCAmelCase_ , mauve_scaling_factor=lowerCAmelCase_ , verbose=lowerCAmelCase_ , seed=lowerCAmelCase_ , ) return out
284
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : List[Any] = { 'microsoft/beit-base-patch16-224-pt22k': ( 'https://huggingface.co/microsoft/beit-base-patch16-224-pt22k/resolve/main/config.json' ), # See all BEiT models at https://huggingface.co/models?filter=beit } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """beit""" def __init__( self : List[Any] , lowerCAmelCase_ : Tuple=8_1_9_2 , lowerCAmelCase_ : Optional[int]=7_6_8 , lowerCAmelCase_ : int=1_2 , lowerCAmelCase_ : Optional[int]=1_2 , lowerCAmelCase_ : Any=3_0_7_2 , lowerCAmelCase_ : Optional[int]="gelu" , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Any=0.02 , lowerCAmelCase_ : int=1e-12 , lowerCAmelCase_ : int=2_2_4 , lowerCAmelCase_ : str=1_6 , lowerCAmelCase_ : int=3 , lowerCAmelCase_ : Dict=False , lowerCAmelCase_ : int=False , lowerCAmelCase_ : List[Any]=False , lowerCAmelCase_ : int=False , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : Union[str, Any]=0.1 , lowerCAmelCase_ : List[str]=True , lowerCAmelCase_ : List[Any]=[3, 5, 7, 1_1] , lowerCAmelCase_ : Optional[Any]=[1, 2, 3, 6] , lowerCAmelCase_ : Tuple=True , lowerCAmelCase_ : Dict=0.4 , lowerCAmelCase_ : Tuple=2_5_6 , lowerCAmelCase_ : Any=1 , lowerCAmelCase_ : Any=False , lowerCAmelCase_ : Optional[int]=2_5_5 , **lowerCAmelCase_ : Any , ) -> Dict: super().__init__(**lowerCAmelCase_ ) __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = initializer_range __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = image_size __lowerCAmelCase = patch_size __lowerCAmelCase = num_channels __lowerCAmelCase = use_mask_token __lowerCAmelCase = use_absolute_position_embeddings __lowerCAmelCase = use_relative_position_bias __lowerCAmelCase = use_shared_relative_position_bias __lowerCAmelCase = layer_scale_init_value __lowerCAmelCase = drop_path_rate __lowerCAmelCase = use_mean_pooling # decode head attributes (semantic segmentation) __lowerCAmelCase = out_indices __lowerCAmelCase = pool_scales # auxiliary head attributes (semantic segmentation) __lowerCAmelCase = use_auxiliary_head __lowerCAmelCase = auxiliary_loss_weight __lowerCAmelCase = auxiliary_channels __lowerCAmelCase = auxiliary_num_convs __lowerCAmelCase = auxiliary_concat_input __lowerCAmelCase = semantic_loss_ignore_index class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = version.parse("""1.11""" ) @property def lowercase ( self : Tuple ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def lowercase ( self : Optional[Any] ) -> float: return 1e-4
284
1
import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoImageProcessor, ViTImageProcessor from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test sys.path.append(str(Path(__file__).parent.parent / 'utils')) from test_module.custom_image_processing import CustomImageProcessor # noqa E402 _snake_case : Optional[Any] = get_tests_dir('fixtures') class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : Optional[int] ) -> int: # A mock response for an HTTP head request to emulate server down __lowerCAmelCase = mock.Mock() __lowerCAmelCase = 5_0_0 __lowerCAmelCase = {} __lowerCAmelCase = HTTPError __lowerCAmelCase = {} # Download this model to make sure it's in the cache. __lowerCAmelCase = ViTImageProcessor.from_pretrained('hf-internal-testing/tiny-random-vit' ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch('requests.Session.request' , return_value=lowerCAmelCase_ ) as mock_head: __lowerCAmelCase = ViTImageProcessor.from_pretrained('hf-internal-testing/tiny-random-vit' ) # This check we did call the fake head request mock_head.assert_called() def lowercase ( self : str ) -> List[Any]: # This test is for deprecated behavior and can be removed in v5 __lowerCAmelCase = ViTImageProcessor.from_pretrained( 'https://huggingface.co/hf-internal-testing/tiny-random-vit/resolve/main/preprocessor_config.json' ) def lowercase ( self : Dict ) -> int: with self.assertRaises(lowerCAmelCase_ ): # config is in subfolder, the following should not work without specifying the subfolder __lowerCAmelCase = AutoImageProcessor.from_pretrained('hf-internal-testing/stable-diffusion-all-variants' ) __lowerCAmelCase = AutoImageProcessor.from_pretrained( 'hf-internal-testing/stable-diffusion-all-variants' , subfolder='feature_extractor' ) self.assertIsNotNone(lowerCAmelCase_ ) @is_staging_test class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @classmethod def lowercase ( cls : Union[str, Any] ) -> List[Any]: __lowerCAmelCase = TOKEN HfFolder.save_token(lowerCAmelCase_ ) @classmethod def lowercase ( cls : Dict ) -> Dict: try: delete_repo(token=cls._token , repo_id='test-image-processor' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-image-processor-org' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-image-processor' ) except HTTPError: pass def lowercase ( self : Any ) -> int: __lowerCAmelCase = ViTImageProcessor.from_pretrained(lowerCAmelCase_ ) image_processor.push_to_hub('test-image-processor' , use_auth_token=self._token ) __lowerCAmelCase = ViTImageProcessor.from_pretrained(f"""{USER}/test-image-processor""" ) for k, v in image_processor.__dict__.items(): self.assertEqual(lowerCAmelCase_ , getattr(lowerCAmelCase_ , lowerCAmelCase_ ) ) # Reset repo delete_repo(token=self._token , repo_id='test-image-processor' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( lowerCAmelCase_ , repo_id='test-image-processor' , push_to_hub=lowerCAmelCase_ , use_auth_token=self._token ) __lowerCAmelCase = ViTImageProcessor.from_pretrained(f"""{USER}/test-image-processor""" ) for k, v in image_processor.__dict__.items(): self.assertEqual(lowerCAmelCase_ , getattr(lowerCAmelCase_ , lowerCAmelCase_ ) ) def lowercase ( self : List[str] ) -> Tuple: __lowerCAmelCase = ViTImageProcessor.from_pretrained(lowerCAmelCase_ ) image_processor.push_to_hub('valid_org/test-image-processor' , use_auth_token=self._token ) __lowerCAmelCase = ViTImageProcessor.from_pretrained('valid_org/test-image-processor' ) for k, v in image_processor.__dict__.items(): self.assertEqual(lowerCAmelCase_ , getattr(lowerCAmelCase_ , lowerCAmelCase_ ) ) # Reset repo delete_repo(token=self._token , repo_id='valid_org/test-image-processor' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( lowerCAmelCase_ , repo_id='valid_org/test-image-processor-org' , push_to_hub=lowerCAmelCase_ , use_auth_token=self._token ) __lowerCAmelCase = ViTImageProcessor.from_pretrained('valid_org/test-image-processor-org' ) for k, v in image_processor.__dict__.items(): self.assertEqual(lowerCAmelCase_ , getattr(lowerCAmelCase_ , lowerCAmelCase_ ) ) def lowercase ( self : Tuple ) -> List[str]: CustomImageProcessor.register_for_auto_class() __lowerCAmelCase = CustomImageProcessor.from_pretrained(lowerCAmelCase_ ) image_processor.push_to_hub('test-dynamic-image-processor' , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual( image_processor.auto_map , {'AutoImageProcessor': 'custom_image_processing.CustomImageProcessor'} , ) __lowerCAmelCase = AutoImageProcessor.from_pretrained( f"""{USER}/test-dynamic-image-processor""" , trust_remote_code=lowerCAmelCase_ ) # Can't make an isinstance check because the new_image_processor is from the CustomImageProcessor class of a dynamic module self.assertEqual(new_image_processor.__class__.__name__ , 'CustomImageProcessor' )
284
def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : int ): return [sentence[i : i + ngram_size] for i in range(len(lowerCAmelCase_ ) - ngram_size + 1 )] if __name__ == "__main__": from doctest import testmod testmod()
284
1
import cva import numpy as np class _UpperCAmelCase : """simple docstring""" def __init__( self : Union[str, Any] , lowerCAmelCase_ : float , lowerCAmelCase_ : int ) -> Optional[Any]: if k in (0.04, 0.06): __lowerCAmelCase = k __lowerCAmelCase = window_size else: raise ValueError('invalid k value' ) def __str__( self : int ) -> str: return str(self.k ) def lowercase ( self : Dict , lowerCAmelCase_ : str ) -> tuple[cva.Mat, list[list[int]]]: __lowerCAmelCase = cva.imread(lowerCAmelCase_ , 0 ) __lowerCAmelCase , __lowerCAmelCase = img.shape __lowerCAmelCase = [] __lowerCAmelCase = img.copy() __lowerCAmelCase = cva.cvtColor(lowerCAmelCase_ , cva.COLOR_GRAY2RGB ) __lowerCAmelCase , __lowerCAmelCase = np.gradient(lowerCAmelCase_ ) __lowerCAmelCase = dx**2 __lowerCAmelCase = dy**2 __lowerCAmelCase = dx * dy __lowerCAmelCase = 0.04 __lowerCAmelCase = self.window_size // 2 for y in range(lowerCAmelCase_ , h - offset ): for x in range(lowerCAmelCase_ , w - offset ): __lowerCAmelCase = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() __lowerCAmelCase = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() __lowerCAmelCase = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() __lowerCAmelCase = (wxx * wyy) - (wxy**2) __lowerCAmelCase = wxx + wyy __lowerCAmelCase = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 2_5_5 ) return color_img, corner_list if __name__ == "__main__": _snake_case : List[str] = HarrisCorner(0.04, 3) _snake_case , _snake_case : str = edge_detect.detect('path_to_image') cva.imwrite('detect.png', color_img)
284
from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : Any = { 'google/pegasus-large': 'https://huggingface.co/google/pegasus-large/resolve/main/config.json', # See all PEGASUS models at https://huggingface.co/models?filter=pegasus } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """pegasus""" a_ = ["""past_key_values"""] a_ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[str] , lowerCAmelCase_ : Union[str, Any]=5_0_2_6_5 , lowerCAmelCase_ : Union[str, Any]=1_0_2_4 , lowerCAmelCase_ : Union[str, Any]=1_2 , lowerCAmelCase_ : Dict=4_0_9_6 , lowerCAmelCase_ : str=1_6 , lowerCAmelCase_ : List[Any]=1_2 , lowerCAmelCase_ : Union[str, Any]=4_0_9_6 , lowerCAmelCase_ : Union[str, Any]=1_6 , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Optional[int]=0.0 , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Any=True , lowerCAmelCase_ : Dict="gelu" , lowerCAmelCase_ : Optional[int]=1_0_2_4 , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : str=0.0 , lowerCAmelCase_ : Union[str, Any]=0.0 , lowerCAmelCase_ : Optional[int]=0.02 , lowerCAmelCase_ : Tuple=0 , lowerCAmelCase_ : Tuple=False , lowerCAmelCase_ : Union[str, Any]=0 , lowerCAmelCase_ : int=1 , lowerCAmelCase_ : Tuple=1 , **lowerCAmelCase_ : Tuple , ) -> List[str]: __lowerCAmelCase = vocab_size __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = d_model __lowerCAmelCase = encoder_ffn_dim __lowerCAmelCase = encoder_layers __lowerCAmelCase = encoder_attention_heads __lowerCAmelCase = decoder_ffn_dim __lowerCAmelCase = decoder_layers __lowerCAmelCase = decoder_attention_heads __lowerCAmelCase = dropout __lowerCAmelCase = attention_dropout __lowerCAmelCase = activation_dropout __lowerCAmelCase = activation_function __lowerCAmelCase = init_std __lowerCAmelCase = encoder_layerdrop __lowerCAmelCase = decoder_layerdrop __lowerCAmelCase = use_cache __lowerCAmelCase = encoder_layers __lowerCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , is_encoder_decoder=lowerCAmelCase_ , decoder_start_token_id=lowerCAmelCase_ , forced_eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ , ) @property def lowercase ( self : List[Any] ) -> int: return self.encoder_attention_heads @property def lowercase ( self : Optional[Any] ) -> int: return self.d_model
284
1
from timeit import timeit def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('the value of input must not be negative' ) __lowerCAmelCase = 0 while number: number &= number - 1 result += 1 return result def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('the value of input must not be negative' ) __lowerCAmelCase = 0 while number: if number % 2 == 1: result += 1 number >>= 1 return result def a_ ( ): def do_benchmark(lowerCAmelCase_ : int ) -> None: __lowerCAmelCase = 'import __main__ as z' print(F"""Benchmark when {number = }:""" ) print(F"""{get_set_bits_count_using_modulo_operator(lowerCAmelCase_ ) = }""" ) __lowerCAmelCase = timeit('z.get_set_bits_count_using_modulo_operator(25)', setup=lowerCAmelCase_ ) print(F"""timeit() runs in {timing} seconds""" ) print(F"""{get_set_bits_count_using_brian_kernighans_algorithm(lowerCAmelCase_ ) = }""" ) __lowerCAmelCase = timeit( 'z.get_set_bits_count_using_brian_kernighans_algorithm(25)', setup=lowerCAmelCase_, ) print(F"""timeit() runs in {timing} seconds""" ) for number in (25, 37, 58, 0): do_benchmark(lowerCAmelCase_ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
284
def a_ ( lowerCAmelCase_ : int ): if p < 2: raise ValueError('p should not be less than 2!' ) elif p == 2: return True __lowerCAmelCase = 4 __lowerCAmelCase = (1 << p) - 1 for _ in range(p - 2 ): __lowerCAmelCase = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
284
1
def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : bool = False ): if not isinstance(lowerCAmelCase_, lowerCAmelCase_ ): __lowerCAmelCase = F"""Expected string as input, found {type(lowerCAmelCase_ )}""" raise ValueError(lowerCAmelCase_ ) if not isinstance(lowerCAmelCase_, lowerCAmelCase_ ): __lowerCAmelCase = F"""Expected boolean as use_pascal parameter, found {type(lowerCAmelCase_ )}""" raise ValueError(lowerCAmelCase_ ) __lowerCAmelCase = input_str.split('_' ) __lowerCAmelCase = 0 if use_pascal else 1 __lowerCAmelCase = words[start_index:] __lowerCAmelCase = [word[0].upper() + word[1:] for word in words_to_capitalize] __lowerCAmelCase = '' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
284
from __future__ import annotations import math def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : int, lowerCAmelCase_ : bool, lowerCAmelCase_ : list[int], lowerCAmelCase_ : float ): if depth < 0: raise ValueError('Depth cannot be less than 0' ) if len(lowerCAmelCase_ ) == 0: raise ValueError('Scores cannot be empty' ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1, node_index * 2, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), minimax(depth + 1, node_index * 2 + 1, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), ) return min( minimax(depth + 1, node_index * 2, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), minimax(depth + 1, node_index * 2 + 1, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), ) def a_ ( ): __lowerCAmelCase = [90, 23, 6, 33, 21, 65, 123, 3_4423] __lowerCAmelCase = math.log(len(lowerCAmelCase_ ), 2 ) print('Optimal value : ', end='' ) print(minimax(0, 0, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
284
1
_snake_case : int = frozenset( [ 'prompt', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) _snake_case : Tuple = frozenset(['prompt', 'negative_prompt']) _snake_case : Optional[int] = frozenset([]) _snake_case : Tuple = frozenset(['image']) _snake_case : Any = frozenset( [ 'image', 'height', 'width', 'guidance_scale', ] ) _snake_case : Optional[Any] = frozenset(['image']) _snake_case : Union[str, Any] = frozenset( [ 'prompt', 'image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) _snake_case : Tuple = frozenset(['prompt', 'image', 'negative_prompt']) _snake_case : Union[str, Any] = frozenset( [ # Text guided image variation with an image mask 'prompt', 'image', 'mask_image', 'height', 'width', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', ] ) _snake_case : Optional[int] = frozenset(['prompt', 'image', 'mask_image', 'negative_prompt']) _snake_case : str = frozenset( [ # image variation with an image mask 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) _snake_case : str = frozenset(['image', 'mask_image']) _snake_case : Union[str, Any] = frozenset( [ 'example_image', 'image', 'mask_image', 'height', 'width', 'guidance_scale', ] ) _snake_case : List[str] = frozenset(['example_image', 'image', 'mask_image']) _snake_case : Union[str, Any] = frozenset(['class_labels']) _snake_case : int = frozenset(['class_labels']) _snake_case : List[Any] = frozenset(['batch_size']) _snake_case : Any = frozenset([]) _snake_case : List[str] = frozenset(['batch_size']) _snake_case : Tuple = frozenset([]) _snake_case : Dict = frozenset( [ 'prompt', 'audio_length_in_s', 'guidance_scale', 'negative_prompt', 'prompt_embeds', 'negative_prompt_embeds', 'cross_attention_kwargs', ] ) _snake_case : List[str] = frozenset(['prompt', 'negative_prompt']) _snake_case : List[Any] = frozenset(['input_tokens']) _snake_case : Any = frozenset(['input_tokens'])
284
def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('number must not be negative' ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
284
1
import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCAmelCase : """simple docstring""" def __init__( self : Dict , lowerCAmelCase_ : Dict , lowerCAmelCase_ : int=1_3 , lowerCAmelCase_ : Any=7 , lowerCAmelCase_ : Optional[Any]=True , lowerCAmelCase_ : Tuple=True , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : Tuple=9_9 , lowerCAmelCase_ : int=1_6 , lowerCAmelCase_ : List[str]=3_6 , lowerCAmelCase_ : Optional[int]=6 , lowerCAmelCase_ : Tuple=6 , lowerCAmelCase_ : Dict=6 , lowerCAmelCase_ : List[str]=3_7 , lowerCAmelCase_ : Dict="gelu" , lowerCAmelCase_ : Any=0.1 , lowerCAmelCase_ : str=0.1 , lowerCAmelCase_ : List[str]=5_1_2 , lowerCAmelCase_ : Any=1_6 , lowerCAmelCase_ : List[Any]=2 , lowerCAmelCase_ : List[Any]=0.02 , lowerCAmelCase_ : str=3 , lowerCAmelCase_ : Dict=4 , lowerCAmelCase_ : Union[str, Any]=None , ) -> Union[str, Any]: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_input_mask __lowerCAmelCase = use_token_type_ids __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = embedding_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_hidden_groups __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = type_vocab_size __lowerCAmelCase = type_sequence_label_size __lowerCAmelCase = initializer_range __lowerCAmelCase = num_labels __lowerCAmelCase = num_choices __lowerCAmelCase = scope def lowercase ( self : Any ) -> Dict: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase = None if self.use_input_mask: __lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCAmelCase = None if self.use_token_type_ids: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = None if self.use_labels: __lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCAmelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowercase ( self : List[Any] ) -> Optional[Any]: return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def lowercase ( self : Tuple , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : str , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Optional[int] ) -> Optional[int]: __lowerCAmelCase = AlbertModel(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ ) __lowerCAmelCase = model(lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ ) __lowerCAmelCase = model(lowerCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowercase ( self : List[Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Any , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : int , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : Tuple ) -> Optional[Any]: __lowerCAmelCase = AlbertForPreTraining(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() __lowerCAmelCase = model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ , sentence_order_label=lowerCAmelCase_ , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels) ) def lowercase ( self : str , lowerCAmelCase_ : str , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : str , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : int ) -> Any: __lowerCAmelCase = AlbertForMaskedLM(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase ( self : str , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : int , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : int , lowerCAmelCase_ : Any ) -> Tuple: __lowerCAmelCase = AlbertForQuestionAnswering(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() __lowerCAmelCase = model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , start_positions=lowerCAmelCase_ , end_positions=lowerCAmelCase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowercase ( self : Optional[Any] , lowerCAmelCase_ : str , lowerCAmelCase_ : Dict , lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Any ) -> Union[str, Any]: __lowerCAmelCase = self.num_labels __lowerCAmelCase = AlbertForSequenceClassification(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase ( self : Optional[int] , lowerCAmelCase_ : str , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : str , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : int ) -> str: __lowerCAmelCase = self.num_labels __lowerCAmelCase = AlbertForTokenClassification(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowercase ( self : str , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[int] ) -> Tuple: __lowerCAmelCase = self.num_choices __lowerCAmelCase = AlbertForMultipleChoice(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() __lowerCAmelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCAmelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCAmelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCAmelCase = model( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , labels=lowerCAmelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowercase ( self : Optional[int] ) -> str: __lowerCAmelCase = self.prepare_config_and_inputs() ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = config_and_inputs __lowerCAmelCase = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = ( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) a_ = ( { """feature-extraction""": AlbertModel, """fill-mask""": AlbertForMaskedLM, """question-answering""": AlbertForQuestionAnswering, """text-classification""": AlbertForSequenceClassification, """token-classification""": AlbertForTokenClassification, """zero-shot""": AlbertForSequenceClassification, } if is_torch_available() else {} ) a_ = True def lowercase ( self : Optional[int] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Dict=False ) -> Dict: __lowerCAmelCase = super()._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ , return_labels=lowerCAmelCase_ ) if return_labels: if model_class in get_values(lowerCAmelCase_ ): __lowerCAmelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=lowerCAmelCase_ ) __lowerCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCAmelCase_ ) return inputs_dict def lowercase ( self : Optional[int] ) -> Any: __lowerCAmelCase = AlbertModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=lowerCAmelCase_ , hidden_size=3_7 ) def lowercase ( self : List[str] ) -> Tuple: self.config_tester.run_common_tests() def lowercase ( self : int ) -> Tuple: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase_ ) def lowercase ( self : int ) -> str: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCAmelCase_ ) def lowercase ( self : Any ) -> Optional[Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCAmelCase_ ) def lowercase ( self : List[Any] ) -> Optional[Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowerCAmelCase_ ) def lowercase ( self : Optional[int] ) -> Union[str, Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCAmelCase_ ) def lowercase ( self : Any ) -> str: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCAmelCase_ ) def lowercase ( self : Union[str, Any] ) -> Optional[Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __lowerCAmelCase = type self.model_tester.create_and_check_model(*lowerCAmelCase_ ) @slow def lowercase ( self : Dict ) -> int: for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = AlbertModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) @require_torch class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def lowercase ( self : Tuple ) -> int: __lowerCAmelCase = AlbertModel.from_pretrained('albert-base-v2' ) __lowerCAmelCase = torch.tensor([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) __lowerCAmelCase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] __lowerCAmelCase = torch.Size((1, 1_1, 7_6_8) ) self.assertEqual(output.shape , lowerCAmelCase_ ) __lowerCAmelCase = torch.tensor( [[[-0.65_13, 1.50_35, -0.27_66], [-0.65_15, 1.50_46, -0.27_80], [-0.65_12, 1.50_49, -0.27_84]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowerCAmelCase_ , atol=1e-4 ) )
284
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
284
1
from __future__ import annotations import math def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : int, lowerCAmelCase_ : bool, lowerCAmelCase_ : list[int], lowerCAmelCase_ : float ): if depth < 0: raise ValueError('Depth cannot be less than 0' ) if len(lowerCAmelCase_ ) == 0: raise ValueError('Scores cannot be empty' ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1, node_index * 2, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), minimax(depth + 1, node_index * 2 + 1, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), ) return min( minimax(depth + 1, node_index * 2, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), minimax(depth + 1, node_index * 2 + 1, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), ) def a_ ( ): __lowerCAmelCase = [90, 23, 6, 33, 21, 65, 123, 3_4423] __lowerCAmelCase = math.log(len(lowerCAmelCase_ ), 2 ) print('Optimal value : ', end='' ) print(minimax(0, 0, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
284
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def a_ ( lowerCAmelCase_ : List[str], lowerCAmelCase_ : Dict, lowerCAmelCase_ : Tuple=1024, lowerCAmelCase_ : Optional[Any]=1024, lowerCAmelCase_ : Tuple=False, **lowerCAmelCase_ : Union[str, Any] ): __lowerCAmelCase = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='train', **lowerCAmelCase_ ) __lowerCAmelCase = tok.pad_token_id def get_lens(lowerCAmelCase_ : Optional[Any] ): __lowerCAmelCase = tqdm( DataLoader(lowerCAmelCase_, batch_size=512, num_workers=8, shuffle=lowerCAmelCase_, collate_fn=ds.collate_fn ), desc=str(ds.len_file ), ) __lowerCAmelCase = [] for batch in dl: __lowerCAmelCase = batch['input_ids'].ne(lowerCAmelCase_ ).sum(1 ).tolist() __lowerCAmelCase = batch['labels'].ne(lowerCAmelCase_ ).sum(1 ).tolist() if consider_target: for src, tgt in zip(lowerCAmelCase_, lowerCAmelCase_ ): max_lens.append(max(lowerCAmelCase_, lowerCAmelCase_ ) ) else: max_lens.extend(lowerCAmelCase_ ) return max_lens __lowerCAmelCase = get_lens(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='val', **lowerCAmelCase_ ) __lowerCAmelCase = get_lens(lowerCAmelCase_ ) pickle_save(lowerCAmelCase_, train_ds.len_file ) pickle_save(lowerCAmelCase_, val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
284
1
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def a_ ( lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = filter(lambda lowerCAmelCase_ : p.requires_grad, model.parameters() ) __lowerCAmelCase = sum([np.prod(p.size() ) for p in model_parameters] ) return params _snake_case : Dict = logging.getLogger(__name__) def a_ ( lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Optional[int] ): if metric == "rouge2": __lowerCAmelCase = '{val_avg_rouge2:.4f}-{step_count}' elif metric == "bleu": __lowerCAmelCase = '{val_avg_bleu:.4f}-{step_count}' elif metric == "em": __lowerCAmelCase = '{val_avg_em:.4f}-{step_count}' else: raise NotImplementedError( F"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" ' function.' ) __lowerCAmelCase = ModelCheckpoint( dirpath=lowerCAmelCase_, filename=lowerCAmelCase_, monitor=F"""val_{metric}""", mode='max', save_top_k=3, every_n_epochs=1, ) return checkpoint_callback def a_ ( lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : Any ): return EarlyStopping( monitor=F"""val_{metric}""", mode='min' if 'loss' in metric else 'max', patience=lowerCAmelCase_, verbose=lowerCAmelCase_, ) class _UpperCAmelCase ( pl.Callback ): """simple docstring""" def lowercase ( self : Tuple , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : int ) -> Any: __lowerCAmelCase = {f"""lr_group_{i}""": param['lr'] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Optional[int] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule , lowerCAmelCase_ : str , lowerCAmelCase_ : List[Any]=True ) -> None: logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) __lowerCAmelCase = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['log', 'progress_bar', 'preds']} ) # Log results __lowerCAmelCase = Path(pl_module.hparams.output_dir ) if type_path == "test": __lowerCAmelCase = od / 'test_results.txt' __lowerCAmelCase = od / 'test_generations.txt' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __lowerCAmelCase = od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" __lowerCAmelCase = od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) generations_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'a+' ) as writer: for key in sorted(lowerCAmelCase_ ): if key in ["log", "progress_bar", "preds"]: continue __lowerCAmelCase = metrics[key] if isinstance(lowerCAmelCase_ , torch.Tensor ): __lowerCAmelCase = val.item() __lowerCAmelCase = f"""{key}: {val:.6f}\n""" writer.write(lowerCAmelCase_ ) if not save_generations: return if "preds" in metrics: __lowerCAmelCase = '\n'.join(metrics['preds'] ) generations_file.open('w+' ).write(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[str] ) -> Dict: try: __lowerCAmelCase = pl_module.model.model.num_parameters() except AttributeError: __lowerCAmelCase = pl_module.model.num_parameters() __lowerCAmelCase = count_trainable_parameters(lowerCAmelCase_ ) # mp stands for million parameters trainer.logger.log_metrics({'n_params': npars, 'mp': npars / 1e6, 'grad_mp': n_trainable_pars / 1e6} ) @rank_zero_only def lowercase ( self : int , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule ) -> Any: save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(lowerCAmelCase_ , lowerCAmelCase_ , 'test' ) @rank_zero_only def lowercase ( self : List[Any] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : Any ) -> int: save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
284
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def a_ ( lowerCAmelCase_ : List[Any], lowerCAmelCase_ : str, lowerCAmelCase_ : Optional[int]=None, lowerCAmelCase_ : List[Any]=None ): if attention_mask is None: __lowerCAmelCase = tf.cast(tf.math.not_equal(lowerCAmelCase_, config.pad_token_id ), tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class _UpperCAmelCase : """simple docstring""" a_ = OPTConfig a_ = {} a_ = """gelu""" def __init__( self : Dict , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : List[str]=1_3 , lowerCAmelCase_ : Tuple=7 , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : Union[str, Any]=False , lowerCAmelCase_ : Any=9_9 , lowerCAmelCase_ : Any=1_6 , lowerCAmelCase_ : List[str]=2 , lowerCAmelCase_ : Dict=4 , lowerCAmelCase_ : str=4 , lowerCAmelCase_ : Any="gelu" , lowerCAmelCase_ : Optional[Any]=0.1 , lowerCAmelCase_ : List[Any]=0.1 , lowerCAmelCase_ : Tuple=2_0 , lowerCAmelCase_ : Optional[Any]=2 , lowerCAmelCase_ : Any=1 , lowerCAmelCase_ : List[Any]=0 , lowerCAmelCase_ : Optional[int]=1_6 , lowerCAmelCase_ : Dict=1_6 , ) -> int: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = eos_token_id __lowerCAmelCase = pad_token_id __lowerCAmelCase = bos_token_id __lowerCAmelCase = embed_dim __lowerCAmelCase = word_embed_proj_dim __lowerCAmelCase = False def lowercase ( self : List[str] ) -> Optional[Any]: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowerCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowerCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowerCAmelCase = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=lowerCAmelCase_ , **self.config_updates , ) __lowerCAmelCase = prepare_opt_inputs_dict(lowerCAmelCase_ , lowerCAmelCase_ ) return config, inputs_dict def lowercase ( self : Any , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Dict ) -> List[str]: __lowerCAmelCase = TFOPTModel(config=lowerCAmelCase_ ) __lowerCAmelCase = inputs_dict['input_ids'] __lowerCAmelCase = input_ids[:1, :] __lowerCAmelCase = inputs_dict['attention_mask'][:1, :] __lowerCAmelCase = 1 # first forward pass __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , use_cache=lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowerCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowerCAmelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowerCAmelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowerCAmelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , past_key_values=lowerCAmelCase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowerCAmelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowerCAmelCase = output_from_no_past[:, -3:, random_slice_idx] __lowerCAmelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowerCAmelCase_ , lowerCAmelCase_ , rtol=1e-3 ) @require_tf class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () a_ = (TFOPTForCausalLM,) if is_tf_available() else () a_ = ( {"""feature-extraction""": TFOPTModel, """text-generation""": TFOPTForCausalLM} if is_tf_available() else {} ) a_ = False a_ = False a_ = False a_ = 10 def lowercase ( self : List[str] ) -> Optional[int]: __lowerCAmelCase = TFOPTModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=lowerCAmelCase_ ) def lowercase ( self : Tuple ) -> Tuple: self.config_tester.run_common_tests() def lowercase ( self : Tuple ) -> Optional[Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCAmelCase_ ) def lowercase ( self : Union[str, Any] ) -> Dict: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(lowerCAmelCase_ : str , lowerCAmelCase_ : Union[str, Any] ): if hasattr(lowerCAmelCase_ , 'weight' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(lowerCAmelCase_ , 'weight' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 1_0, config.vocab_size + 1_0]: # build the embeddings __lowerCAmelCase = model_class(config=lowerCAmelCase_ ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_input_embeddings() ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(lowerCAmelCase_ ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_input_embeddings() ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. __lowerCAmelCase = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , lowerCAmelCase_ ) # check that weights remain the same after resizing __lowerCAmelCase = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCAmelCase = False self.assertTrue(lowerCAmelCase_ ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , lowerCAmelCase_ ) __lowerCAmelCase = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCAmelCase = False self.assertTrue(lowerCAmelCase_ ) def a_ ( lowerCAmelCase_ : Union[str, Any] ): return tf.constant(lowerCAmelCase_, dtype=tf.intaa ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" a_ = 99 def lowercase ( self : Optional[int] ) -> Any: __lowerCAmelCase = tf.ones((4, 1) , dtype=tf.intaa ) * 2 __lowerCAmelCase = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) __lowerCAmelCase = input_ids.shape[0] __lowerCAmelCase = OPTConfig( vocab_size=self.vocab_size , hidden_size=2_4 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=3_2 , max_position_embeddings=4_8 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def lowercase ( self : str ) -> List[str]: __lowerCAmelCase = TFOPTModel.from_pretrained('facebook/opt-350m' ) __lowerCAmelCase = _long_tensor([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] ) __lowerCAmelCase = tf.not_equal(lowerCAmelCase_ , model.config.pad_token_id ) with tf.GradientTape(): __lowerCAmelCase = model(input_ids=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ ).last_hidden_state __lowerCAmelCase = (1, 1_1, 5_1_2) self.assertEqual(output.shape , lowerCAmelCase_ ) __lowerCAmelCase = tf.constant( [[-0.28_73, -1.92_18, -0.30_33], [-1.27_10, -0.13_38, -0.19_02], [0.40_95, 0.12_14, -1.31_21]] ) self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase_ , atol=4e-3 ) ) __lowerCAmelCase = tf.function(lowerCAmelCase_ , jit_compile=lowerCAmelCase_ ) __lowerCAmelCase = xla_generate(lowerCAmelCase_ , lowerCAmelCase_ )[0] self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase_ , atol=4e-2 ) ) @require_tf @slow class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> Dict: super().setUp() __lowerCAmelCase = 'facebook/opt-350m' def lowercase ( self : Dict ) -> Any: __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(self.path_model ) __lowerCAmelCase = GPTaTokenizer.from_pretrained(self.path_model ) __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of', 'Paris is the capital of France and', 'Computers and mobile phones have taken', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' , padding=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) __lowerCAmelCase = tf.constant( [ [1.38_51, -13.89_23, -10.52_29, -10.75_33, -0.23_09, -10.23_84, -0.53_65, -9.09_47, -5.16_70], [-4.70_73, -10.62_76, -3.94_15, -21.52_42, -0.28_22, -0.28_22, -0.28_22, -0.28_22, -0.28_22], [0.62_47, -3.42_29, -8.91_79, -1.42_97, -14.16_50, 1.41_46, -9.02_18, -0.27_03, -0.27_03], [6.47_83, -1.99_13, -10.79_26, -2.33_36, 1.50_92, -0.99_74, -6.82_13, 1.34_77, 1.34_77], ] ) self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1e-4 ) ) __lowerCAmelCase = tf.function(lowerCAmelCase_ , jit_compile=lowerCAmelCase_ ) __lowerCAmelCase = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1e-4 ) ) @require_tf @slow class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @property def lowercase ( self : Optional[int] ) -> int: return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def lowercase ( self : int ) -> str: __lowerCAmelCase = 'facebook/opt-125m' __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of New York, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] __lowerCAmelCase = [] __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) for prompt in self.prompts: __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(lowerCAmelCase_ , max_length=1_0 ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Optional[Any] ) -> str: __lowerCAmelCase = 'facebook/opt-350m' __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = 'left' # use different length sentences to test batching __lowerCAmelCase = [ 'Hello, my dog is a little', 'Today, I', ] __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' , padding=lowerCAmelCase_ ) __lowerCAmelCase = inputs['input_ids'] __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ , attention_mask=inputs['attention_mask'] ) __lowerCAmelCase = tokenizer(sentences[0] , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ ) __lowerCAmelCase = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['attention_mask'][-1] , tf.intaa ) ) __lowerCAmelCase = tokenizer(sentences[1] , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ , max_length=model.config.max_length - num_paddings ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = [ 'Hello, my dog is a little bit of a dork.\nI\'m a little bit', 'Today, I was in the middle of a conversation with a friend about the', ] self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , [non_padded_sentence, padded_sentence] ) def lowercase ( self : List[Any] ) -> List[Any]: __lowerCAmelCase = 'facebook/opt-350m' __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of San Francisco, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] __lowerCAmelCase = [] __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) for prompt in self.prompts: __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(lowerCAmelCase_ , max_length=1_0 ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ )
284
1
def a_ ( lowerCAmelCase_ : int ): # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError('The given input must be positive' ) # get the generated string sequence __lowerCAmelCase = gray_code_sequence_string(lowerCAmelCase_ ) # # convert them to integers for i in range(len(lowerCAmelCase_ ) ): __lowerCAmelCase = int(sequence[i], 2 ) return sequence def a_ ( lowerCAmelCase_ : int ): # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] __lowerCAmelCase = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits __lowerCAmelCase = gray_code_sequence_string(bit_count - 1 ) __lowerCAmelCase = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): __lowerCAmelCase = '0' + smaller_sequence[i] sequence.append(lowerCAmelCase_ ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): __lowerCAmelCase = '1' + smaller_sequence[i] sequence.append(lowerCAmelCase_ ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
284
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { 'configuration_layoutlmv3': [ 'LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv3Config', 'LayoutLMv3OnnxConfig', ], 'processing_layoutlmv3': ['LayoutLMv3Processor'], 'tokenization_layoutlmv3': ['LayoutLMv3Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Union[str, Any] = ['LayoutLMv3TokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[str] = [ 'LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv3ForQuestionAnswering', 'LayoutLMv3ForSequenceClassification', 'LayoutLMv3ForTokenClassification', 'LayoutLMv3Model', 'LayoutLMv3PreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[Any] = [ 'TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLayoutLMv3ForQuestionAnswering', 'TFLayoutLMv3ForSequenceClassification', 'TFLayoutLMv3ForTokenClassification', 'TFLayoutLMv3Model', 'TFLayoutLMv3PreTrainedModel', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Tuple = ['LayoutLMv3FeatureExtractor'] _snake_case : str = ['LayoutLMv3ImageProcessor'] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys _snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
1
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss _snake_case : Dict = pytest.mark.integration @require_faiss class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : List[Any] ) -> Optional[Any]: __lowerCAmelCase = Dataset.from_dict({'filename': ['my_name-train' + '_' + str(lowerCAmelCase_ ) for x in np.arange(3_0 ).tolist()]} ) return dset def lowercase ( self : List[str] ) -> Tuple: import faiss __lowerCAmelCase = self._create_dummy_dataset() __lowerCAmelCase = dset.map( lambda lowerCAmelCase_ , lowerCAmelCase_ : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=lowerCAmelCase_ , keep_in_memory=lowerCAmelCase_ ) __lowerCAmelCase = dset.add_faiss_index('vecs' , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) dset.drop_index('vecs' ) def lowercase ( self : Optional[Any] ) -> str: import faiss __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT , ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def lowercase ( self : int ) -> Optional[Any]: import faiss __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCAmelCase_ ) as tmp_file: dset.save_faiss_index('vecs' , tmp_file.name ) dset.load_faiss_index('vecs2' , tmp_file.name ) os.unlink(tmp_file.name ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs2' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def lowercase ( self : Union[str, Any] ) -> List[Any]: __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' ) dset.drop_index('vecs' ) self.assertRaises(lowerCAmelCase_ , partial(dset.get_nearest_examples , 'vecs2' , np.ones(5 , dtype=np.floataa ) ) ) def lowercase ( self : Union[str, Any] ) -> Tuple: from elasticsearch import Elasticsearch __lowerCAmelCase = self._create_dummy_dataset() with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: __lowerCAmelCase = {'acknowledged': True} mocked_bulk.return_value([(True, None)] * 3_0 ) __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 2_9}]}} __lowerCAmelCase = Elasticsearch() dset.add_elasticsearch_index('filename' , es_client=lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('filename' , 'my_name-train_29' ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) @require_faiss class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : str ) -> int: import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 1_0 ) # single query __lowerCAmelCase = np.zeros(5 , dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertRaises(lowerCAmelCase_ , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries __lowerCAmelCase = np.eye(5 , dtype=np.floataa )[::-1] __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ ) self.assertRaises(lowerCAmelCase_ , index.search_batch , queries[0] ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , lowerCAmelCase_ ) def lowercase ( self : List[Any] ) -> List[str]: import faiss __lowerCAmelCase = FaissIndex(string_factory='Flat' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) __lowerCAmelCase = FaissIndex(string_factory='LSH' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(lowerCAmelCase_ ): __lowerCAmelCase = FaissIndex(string_factory='Flat' , custom_index=faiss.IndexFlat(5 ) ) def lowercase ( self : Union[str, Any] ) -> Dict: import faiss __lowerCAmelCase = faiss.IndexFlat(5 ) __lowerCAmelCase = FaissIndex(custom_index=lowerCAmelCase_ ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def lowercase ( self : str ) -> Any: import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCAmelCase_ ) as tmp_file: index.save(tmp_file.name ) __lowerCAmelCase = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) __lowerCAmelCase = np.zeros(5 , dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def a_ ( lowerCAmelCase_ : Union[str, Any] ): import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5, dtype=np.floataa ) ) __lowerCAmelCase = 'index.faiss' __lowerCAmelCase = F"""mock://{index_name}""" index.save(lowerCAmelCase_, storage_options=mockfs.storage_options ) __lowerCAmelCase = FaissIndex.load(lowerCAmelCase_, storage_options=mockfs.storage_options ) __lowerCAmelCase = np.zeros(5, dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : Any ) -> int: from elasticsearch import Elasticsearch with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: __lowerCAmelCase = Elasticsearch() __lowerCAmelCase = {'acknowledged': True} __lowerCAmelCase = ElasticSearchIndex(es_client=lowerCAmelCase_ ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(['foo', 'bar', 'foobar'] ) # single query __lowerCAmelCase = 'foo' __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout __lowerCAmelCase = 'foo' __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ , request_timeout=3_0 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries __lowerCAmelCase = ['foo', 'bar', 'foobar'] __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([1, 1, 1] , lowerCAmelCase_ ) # batched queries with timeout __lowerCAmelCase = ['foo', 'bar', 'foobar'] __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ , request_timeout=3_0 ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([1, 1, 1] , lowerCAmelCase_ )
284
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case : Dict = {'configuration_mra': ['MRA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MraConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'MRA_PRETRAINED_MODEL_ARCHIVE_LIST', 'MraForMaskedLM', 'MraForMultipleChoice', 'MraForQuestionAnswering', 'MraForSequenceClassification', 'MraForTokenClassification', 'MraLayer', 'MraModel', 'MraPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys _snake_case : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure)
284
1
import math import sys def a_ ( lowerCAmelCase_ : int ): if number != int(lowerCAmelCase_ ): raise ValueError('the value of input must be a natural number' ) if number < 0: raise ValueError('the value of input must not be a negative number' ) if number == 0: return 1 __lowerCAmelCase = [-1] * (number + 1) __lowerCAmelCase = 0 for i in range(1, number + 1 ): __lowerCAmelCase = sys.maxsize __lowerCAmelCase = int(math.sqrt(lowerCAmelCase_ ) ) for j in range(1, root + 1 ): __lowerCAmelCase = 1 + answers[i - (j**2)] __lowerCAmelCase = min(lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = answer return answers[number] if __name__ == "__main__": import doctest doctest.testmod()
284
from __future__ import annotations import string from itertools import cycle, product from pathlib import Path _snake_case : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) _snake_case : list[int] = [ord(letter) for letter in string.ascii_lowercase] _snake_case : set[int] = {ord(char) for char in VALID_CHARS} _snake_case : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def a_ ( lowerCAmelCase_ : list[int], lowerCAmelCase_ : tuple[int, ...] ): __lowerCAmelCase = "" __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 for keychar, cipherchar in zip(cycle(lowerCAmelCase_ ), lowerCAmelCase_ ): __lowerCAmelCase = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(lowerCAmelCase_ ) return decoded def a_ ( lowerCAmelCase_ : list[int] ): __lowerCAmelCase = [] for key in product(lowerCAmelCase_, repeat=3 ): __lowerCAmelCase = try_key(lowerCAmelCase_, lowerCAmelCase_ ) if encoded is not None: possibles.append(lowerCAmelCase_ ) return possibles def a_ ( lowerCAmelCase_ : list[str], lowerCAmelCase_ : str ): return [possible for possible in possibles if common_word in possible.lower()] def a_ ( lowerCAmelCase_ : str = "p059_cipher.txt" ): __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = Path(lowerCAmelCase_ ).parent.joinpath(lowerCAmelCase_ ).read_text(encoding='utf-8' ) __lowerCAmelCase = [int(lowerCAmelCase_ ) for number in data.strip().split(',' )] __lowerCAmelCase = filter_valid_chars(lowerCAmelCase_ ) for common_word in COMMON_WORDS: __lowerCAmelCase = filter_common_word(lowerCAmelCase_, lowerCAmelCase_ ) if len(lowerCAmelCase_ ) == 1: break __lowerCAmelCase = possibles[0] return sum(ord(lowerCAmelCase_ ) for char in decoded_text ) if __name__ == "__main__": print(F"""{solution() = }""")
284
1
import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """M-CLIP""" def __init__( self : Optional[Any] , lowerCAmelCase_ : Union[str, Any]=1_0_2_4 , lowerCAmelCase_ : Any=7_6_8 , **lowerCAmelCase_ : List[Any] ) -> Union[str, Any]: __lowerCAmelCase = transformerDimSize __lowerCAmelCase = imageDimSize super().__init__(**lowerCAmelCase_ ) class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = MCLIPConfig def __init__( self : str , lowerCAmelCase_ : List[str] , *lowerCAmelCase_ : str , **lowerCAmelCase_ : Dict ) -> Optional[Any]: super().__init__(lowerCAmelCase_ , *lowerCAmelCase_ , **lowerCAmelCase_ ) __lowerCAmelCase = XLMRobertaModel(lowerCAmelCase_ ) __lowerCAmelCase = torch.nn.Linear( in_features=config.transformerDimensions , out_features=config.numDims ) def lowercase ( self : Optional[int] , lowerCAmelCase_ : Any , lowerCAmelCase_ : Tuple ) -> Tuple: __lowerCAmelCase = self.transformer(input_ids=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] __lowerCAmelCase = (embs * attention_mask.unsqueeze(2 )).sum(dim=1 ) / attention_mask.sum(dim=1 )[:, None] return self.LinearTransformation(lowerCAmelCase_ ), embs
284
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem _snake_case : Tuple = importlib.util.find_spec('s3fs') is not None if _has_safs: from .safilesystem import SaFileSystem # noqa: F401 _snake_case : List[compression.BaseCompressedFileFileSystem] = [ compression.BzaFileSystem, compression.GzipFileSystem, compression.LzaFileSystem, compression.XzFileSystem, compression.ZstdFileSystem, ] # Register custom filesystems for fs_class in COMPRESSION_FILESYSTEMS + [HfFileSystem]: if fs_class.protocol in fsspec.registry and fsspec.registry[fs_class.protocol] is not fs_class: warnings.warn(F"""A filesystem protocol was already set for {fs_class.protocol} and will be overwritten.""") fsspec.register_implementation(fs_class.protocol, fs_class, clobber=True) def a_ ( lowerCAmelCase_ : str ): if "://" in dataset_path: __lowerCAmelCase = dataset_path.split('://' )[1] return dataset_path def a_ ( lowerCAmelCase_ : fsspec.AbstractFileSystem ): if fs is not None and fs.protocol != "file": return True else: return False def a_ ( lowerCAmelCase_ : fsspec.AbstractFileSystem, lowerCAmelCase_ : str, lowerCAmelCase_ : str ): __lowerCAmelCase = not is_remote_filesystem(lowerCAmelCase_ ) if is_local: # LocalFileSystem.mv does copy + rm, it is more efficient to simply move a local directory shutil.move(fs._strip_protocol(lowerCAmelCase_ ), fs._strip_protocol(lowerCAmelCase_ ) ) else: fs.mv(lowerCAmelCase_, lowerCAmelCase_, recursive=lowerCAmelCase_ ) def a_ ( ): if hasattr(fsspec.asyn, 'reset_lock' ): # for future fsspec>2022.05.0 fsspec.asyn.reset_lock() else: __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = threading.Lock()
284
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : List[Any] = { 'configuration_owlvit': [ 'OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'OwlViTConfig', 'OwlViTOnnxConfig', 'OwlViTTextConfig', 'OwlViTVisionConfig', ], 'processing_owlvit': ['OwlViTProcessor'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Dict = ['OwlViTFeatureExtractor'] _snake_case : int = ['OwlViTImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[str] = [ 'OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'OwlViTModel', 'OwlViTPreTrainedModel', 'OwlViTTextModel', 'OwlViTVisionModel', 'OwlViTForObjectDetection', ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _snake_case : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def a_ ( lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = filter(lambda lowerCAmelCase_ : p.requires_grad, model.parameters() ) __lowerCAmelCase = sum([np.prod(p.size() ) for p in model_parameters] ) return params _snake_case : Dict = logging.getLogger(__name__) def a_ ( lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Optional[int] ): if metric == "rouge2": __lowerCAmelCase = '{val_avg_rouge2:.4f}-{step_count}' elif metric == "bleu": __lowerCAmelCase = '{val_avg_bleu:.4f}-{step_count}' elif metric == "em": __lowerCAmelCase = '{val_avg_em:.4f}-{step_count}' else: raise NotImplementedError( F"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" ' function.' ) __lowerCAmelCase = ModelCheckpoint( dirpath=lowerCAmelCase_, filename=lowerCAmelCase_, monitor=F"""val_{metric}""", mode='max', save_top_k=3, every_n_epochs=1, ) return checkpoint_callback def a_ ( lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : Any ): return EarlyStopping( monitor=F"""val_{metric}""", mode='min' if 'loss' in metric else 'max', patience=lowerCAmelCase_, verbose=lowerCAmelCase_, ) class _UpperCAmelCase ( pl.Callback ): """simple docstring""" def lowercase ( self : Tuple , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : int ) -> Any: __lowerCAmelCase = {f"""lr_group_{i}""": param['lr'] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Optional[int] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule , lowerCAmelCase_ : str , lowerCAmelCase_ : List[Any]=True ) -> None: logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) __lowerCAmelCase = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['log', 'progress_bar', 'preds']} ) # Log results __lowerCAmelCase = Path(pl_module.hparams.output_dir ) if type_path == "test": __lowerCAmelCase = od / 'test_results.txt' __lowerCAmelCase = od / 'test_generations.txt' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __lowerCAmelCase = od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" __lowerCAmelCase = od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) generations_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'a+' ) as writer: for key in sorted(lowerCAmelCase_ ): if key in ["log", "progress_bar", "preds"]: continue __lowerCAmelCase = metrics[key] if isinstance(lowerCAmelCase_ , torch.Tensor ): __lowerCAmelCase = val.item() __lowerCAmelCase = f"""{key}: {val:.6f}\n""" writer.write(lowerCAmelCase_ ) if not save_generations: return if "preds" in metrics: __lowerCAmelCase = '\n'.join(metrics['preds'] ) generations_file.open('w+' ).write(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[str] ) -> Dict: try: __lowerCAmelCase = pl_module.model.model.num_parameters() except AttributeError: __lowerCAmelCase = pl_module.model.num_parameters() __lowerCAmelCase = count_trainable_parameters(lowerCAmelCase_ ) # mp stands for million parameters trainer.logger.log_metrics({'n_params': npars, 'mp': npars / 1e6, 'grad_mp': n_trainable_pars / 1e6} ) @rank_zero_only def lowercase ( self : int , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule ) -> Any: save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(lowerCAmelCase_ , lowerCAmelCase_ , 'test' ) @rank_zero_only def lowercase ( self : List[Any] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : Any ) -> int: save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
284
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import GLPNImageProcessor class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : List[str] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : List[Any]=7 , lowerCAmelCase_ : Any=3 , lowerCAmelCase_ : List[Any]=1_8 , lowerCAmelCase_ : str=3_0 , lowerCAmelCase_ : Tuple=4_0_0 , lowerCAmelCase_ : Optional[int]=True , lowerCAmelCase_ : Tuple=3_2 , lowerCAmelCase_ : List[Any]=True , ) -> Dict: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = num_channels __lowerCAmelCase = image_size __lowerCAmelCase = min_resolution __lowerCAmelCase = max_resolution __lowerCAmelCase = do_resize __lowerCAmelCase = size_divisor __lowerCAmelCase = do_rescale def lowercase ( self : Optional[int] ) -> List[str]: return { "do_resize": self.do_resize, "size_divisor": self.size_divisor, "do_rescale": self.do_rescale, } @require_torch @require_vision class _UpperCAmelCase ( _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = GLPNImageProcessor if is_vision_available() else None def lowercase ( self : Optional[Any] ) -> Union[str, Any]: __lowerCAmelCase = GLPNImageProcessingTester(self ) @property def lowercase ( self : List[Any] ) -> Union[str, Any]: return self.image_processor_tester.prepare_image_processor_dict() def lowercase ( self : List[Any] ) -> List[Any]: __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCAmelCase_ , 'do_resize' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'size_divisor' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'resample' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'do_rescale' ) ) def lowercase ( self : int ) -> str: pass def lowercase ( self : Optional[int] ) -> Optional[int]: # Initialize image_processing __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase_ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase_ , Image.Image ) # Test not batched input (GLPNImageProcessor doesn't support batching) __lowerCAmelCase = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def lowercase ( self : List[str] ) -> Optional[Any]: # Initialize image_processing __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase_ , numpify=lowerCAmelCase_ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase_ , np.ndarray ) # Test not batched input (GLPNImageProcessor doesn't support batching) __lowerCAmelCase = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def lowercase ( self : Tuple ) -> Optional[Any]: # Initialize image_processing __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase_ , torchify=lowerCAmelCase_ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase_ , torch.Tensor ) # Test not batched input (GLPNImageProcessor doesn't support batching) __lowerCAmelCase = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 )
284
import re from filelock import FileLock try: import nltk _snake_case : Any = True except (ImportError, ModuleNotFoundError): _snake_case : Union[str, Any] = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def a_ ( lowerCAmelCase_ : str ): re.sub('<n>', '', lowerCAmelCase_ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(lowerCAmelCase_ ) )
284
1
import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def a_ ( lowerCAmelCase_ : int = 8 ): __lowerCAmelCase = ascii_letters + digits + punctuation return "".join(secrets.choice(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ) ) def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : int ): # Password Generator = full boot with random_number, random_letters, and # random_character FUNCTIONS # Put your code here... i -= len(lowerCAmelCase_ ) __lowerCAmelCase = i // 3 __lowerCAmelCase = i % 3 # chars = chars_incl + random_letters(ascii_letters, i / 3 + remainder) + # random_number(digits, i / 3) + random_characters(punctuation, i / 3) __lowerCAmelCase = ( chars_incl + random(lowerCAmelCase_, quotient + remainder ) + random(lowerCAmelCase_, lowerCAmelCase_ ) + random(lowerCAmelCase_, lowerCAmelCase_ ) ) __lowerCAmelCase = list(lowerCAmelCase_ ) shuffle(lowerCAmelCase_ ) return "".join(lowerCAmelCase_ ) # random is a generalised function for letters, characters and numbers def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : int ): return "".join(secrets.choice(lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ ) ) def a_ ( lowerCAmelCase_ : List[Any], lowerCAmelCase_ : List[str] ): pass # Put your code here... def a_ ( lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : Dict ): pass # Put your code here... def a_ ( lowerCAmelCase_ : Tuple, lowerCAmelCase_ : List[Any] ): pass # Put your code here... def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : int = 8 ): if len(lowerCAmelCase_ ) < min_length: # Your Password must be at least 8 characters long return False __lowerCAmelCase = any(char in ascii_uppercase for char in password ) __lowerCAmelCase = any(char in ascii_lowercase for char in password ) __lowerCAmelCase = any(char in digits for char in password ) __lowerCAmelCase = any(char in punctuation for char in password ) return upper and lower and num and spec_char # Passwords should contain UPPERCASE, lowerase # numbers, and special characters def a_ ( ): __lowerCAmelCase = int(input('Please indicate the max length of your password: ' ).strip() ) __lowerCAmelCase = input( 'Please indicate the characters that must be in your password: ' ).strip() print('Password generated:', password_generator(lowerCAmelCase_ ) ) print( 'Alternative Password generated:', alternative_password_generator(lowerCAmelCase_, lowerCAmelCase_ ), ) print('[If you are thinking of using this passsword, You better save it.]' ) if __name__ == "__main__": main()
284
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case : List[Any] = { 'configuration_biogpt': ['BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BioGptConfig'], 'tokenization_biogpt': ['BioGptTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BioGptForCausalLM', 'BioGptForTokenClassification', 'BioGptForSequenceClassification', 'BioGptModel', 'BioGptPreTrainedModel', ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys _snake_case : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.generation import DisjunctiveConstraint @require_torch class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : Dict ) -> List[str]: # For consistency across different places the DisjunctiveConstraint is called, # dc.token_ids is a list of integers. It is also initialized only by integers. __lowerCAmelCase = [[1, 2, 4], [1, 2, 3, 4]] __lowerCAmelCase = DisjunctiveConstraint(lowerCAmelCase_ ) self.assertTrue(isinstance(dc.token_ids , lowerCAmelCase_ ) ) with self.assertRaises(lowerCAmelCase_ ): DisjunctiveConstraint(torch.LongTensor([[1, 2, 4], [1, 2, 3]] ) ) with self.assertRaises(lowerCAmelCase_ ): DisjunctiveConstraint([torch.LongTensor([1, 2, 4] ), torch.LongTensor([1, 2, 3, 4, 5] )] ) def lowercase ( self : List[str] ) -> Optional[int]: # We can't have constraints that are complete subsets of another. This leads to a preverse # interpretation of "constraint fulfillment": does generating [1,2,3] fulfill the constraint? # It would mean that it generated [1,2] which fulfills it, but it's in the middle of potentially # fulfilling [1,2,3,4]. If we believe that [1,2,3] does fulfill the constraint, then the algorithm # will necessarily never reach [1,2,3,4], giving users a false sense of control (better to just not allow it). __lowerCAmelCase = [[1, 2], [1, 2, 3, 4]] with self.assertRaises(lowerCAmelCase_ ): DisjunctiveConstraint(lowerCAmelCase_ ) # fails here def lowercase ( self : List[str] ) -> Dict: __lowerCAmelCase = [[1, 2, 3], [1, 2, 4]] __lowerCAmelCase = DisjunctiveConstraint(lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = dc.update(1 ) __lowerCAmelCase = stepped is True and completed is False and reset is False self.assertTrue(lowerCAmelCase_ ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = dc.update(2 ) __lowerCAmelCase = stepped is True and completed is False and reset is False self.assertTrue(lowerCAmelCase_ ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = dc.update(3 ) __lowerCAmelCase = stepped is True and completed is True and reset is False self.assertTrue(lowerCAmelCase_ ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 3] ) def lowercase ( self : Tuple ) -> Dict: __lowerCAmelCase = [[1, 2, 3], [1, 2, 4, 5], [1, 2, 5]] __lowerCAmelCase = DisjunctiveConstraint(lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = dc.update(4 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2, 4] ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 4, 5] ) dc.reset() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 3 ) self.assertTrue(dc.current_seq == [1] ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 2 ) self.assertTrue(dc.current_seq == [1, 2] ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.remaining() == 0 ) self.assertTrue(dc.current_seq == [1, 2, 5] )
284
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets _snake_case : Tuple = '\\n@article{hendrycksmath2021,\n title={Measuring Mathematical Problem Solving With the MATH Dataset},\n author={Dan Hendrycks\n and Collin Burns\n and Saurav Kadavath\n and Akul Arora\n and Steven Basart\n and Eric Tang\n and Dawn Song\n and Jacob Steinhardt},\n journal={arXiv preprint arXiv:2103.03874},\n year={2021}\n}\n' _snake_case : str = '\\nThis metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset.\nIt first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy.\n' _snake_case : List[str] = R'\nCalculates accuracy after canonicalizing inputs.\n\nArgs:\n predictions: list of predictions to score. Each prediction\n is a string that contains natural language and LaTex.\n references: list of reference for each prediction. Each\n reference is a string that contains natural language\n and LaTex.\nReturns:\n accuracy: accuracy after canonicalizing inputs\n (e.g., converting "1/2" to "\\frac{1}{2}")\n\nExamples:\n >>> metric = datasets.load_metric("competition_math")\n >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"])\n >>> print(results)\n {\'accuracy\': 1.0}\n' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): """simple docstring""" def lowercase ( self : str ) -> int: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' ), 'references': datasets.Value('string' ), } ) , homepage='https://github.com/hendrycks/math' , codebase_urls=['https://github.com/hendrycks/math'] , ) def lowercase ( self : str , lowerCAmelCase_ : Any , lowerCAmelCase_ : Optional[int] ) -> List[Any]: __lowerCAmelCase = 0.0 for i, j in zip(lowerCAmelCase_ , lowerCAmelCase_ ): n_correct += 1.0 if math_equivalence.is_equiv(lowerCAmelCase_ , lowerCAmelCase_ ) else 0.0 __lowerCAmelCase = n_correct / len(lowerCAmelCase_ ) return { "accuracy": accuracy, }
284
1
# HF Trainer benchmarking tool # # This tool can be used to run and compare multiple dimensions of the HF Trainers args. # # It then prints a report once in github format with all the information that needs to be shared # with others and second time in a console-friendly format, so it's easier to use for tuning things up. # # The main idea is: # # ./trainer-benchmark.py --base-cmd '<cmd args that don't change>' \ # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' \ # --target-metric-key train_samples_per_second # # The variations can be any command line argument that you want to compare and not just dtype as in # the example. # # --variations allows you to compare variations in multiple dimensions. # # as the first dimention has 2 options and the second 3 in our example, this will run the trainer 6 # times adding one of: # # 1. --tf32 0 --fp16 0 # 2. --tf32 0 --fp16 1 # 3. --tf32 0 --bf16 1 # 4. --tf32 1 --fp16 0 # 5. --tf32 1 --fp16 1 # 6. --tf32 1 --bf16 1 # # and print the results. This is just a cartesian product - and more than 2 dimensions can be used. # # If you want to rely on defaults, this: # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' # is identical to this: # --variations '--tf32 0|--tf32 1' '|--fp16|--bf16' # # the leading empty variation in the 2nd dimension is a valid variation. # # So here we get the following 6 variations: # # 1. --tf32 0 # 2. --tf32 0 --fp16 # 3. --tf32 0 --bf16 # 4. --tf32 1 # 5. --tf32 1 --fp16 # 6. --tf32 1 --bf16 # # In this particular case we don't know what the default tf32 setting is as it's normally # pytorch-version dependent). That's why it's best to do an explicit setting of each variation: # `--tf32 0|--tf32 1` # # Here is a full example of a train: # # CUDA_VISIBLE_DEVICES=0 python ./scripts/benchmark/trainer-benchmark.py \ # --base-cmd \ # ' examples/pytorch/translation/run_translation.py --model_name_or_path t5-small \ # --output_dir output_dir --do_train --label_smoothing 0.1 --logging_strategy no \ # --save_strategy no --per_device_train_batch_size 32 --max_source_length 512 \ # --max_target_length 512 --num_train_epochs 1 --overwrite_output_dir \ # --source_lang en --target_lang ro --dataset_name wmt16 --dataset_config "ro-en" \ # --source_prefix "translate English to Romanian: " --warmup_steps 50 \ # --max_train_samples 20000 --dataloader_num_workers 2 ' \ # --target-metric-key train_samples_per_second --repeat-times 1 --variations \ # '|--fp16|--bf16' '--tf32 0|--tf32 1' --report-metric-keys train_loss \ # --repeat-times 1 --base-variation '--tf32 0' # # and here is a possible output: # # # | Variation | Train | Diff | Train | # | | samples | % | loss | # | | per | | | # | | second | | | # |:----------------|----------:|-------:|--------:| # | --tf32 0 | 285.11 | 0 | 2.51 | # | --tf32 1 | 342.09 | 20 | 2.51 | # | --fp16 --tf32 0 | 423.49 | 49 | 2.51 | # | --fp16 --tf32 1 | 423.13 | 48 | 2.51 | # | --bf16 --tf32 0 | 416.80 | 46 | 2.52 | # | --bf16 --tf32 1 | 415.87 | 46 | 2.52 | # # # So you can quickly compare the different outcomes. # # Typically running each experiment once is enough, but if the environment is unstable you can # re-run each multiple times, e.g., 3 using --repeat-times 3 and it will report the averaged results. # # By default it'll use the lowest result as the base line to use as 100% and then compare the rest to # it as can be seen from the table above, but you can also specify which combination is the one to use as # the baseline, e.g., to change to another entry use: --base-variation '--tf32 1 --fp16 0' # # --target-metric-key is there to tell the program which metrics to compare - the different metric keys are # inside output_dir/all_results.json. e.g., to measure eval performance instead of train use: # --target-metric-key eval_samples_per_second # but of course you will need to adjust the --base-cmd value in the example to perform evaluation as # well (as currently it doesn't) # import argparse import datetime import io import itertools import json import math import os import platform import re import shlex import subprocess import sys from pathlib import Path from statistics import fmean import pandas as pd import torch from tqdm import tqdm import transformers _snake_case : Any = float('nan') class _UpperCAmelCase : """simple docstring""" def __init__( self : Optional[int] , lowerCAmelCase_ : Tuple ) -> Optional[int]: __lowerCAmelCase = sys.stdout __lowerCAmelCase = open(lowerCAmelCase_ , 'a' ) def __getattr__( self : Optional[Any] , lowerCAmelCase_ : Union[str, Any] ) -> Optional[Any]: return getattr(self.stdout , lowerCAmelCase_ ) def lowercase ( self : List[Any] , lowerCAmelCase_ : Optional[int] ) -> str: self.stdout.write(lowerCAmelCase_ ) # strip tqdm codes self.file.write(re.sub(R'^.*\r' , '' , lowerCAmelCase_ , 0 , re.M ) ) def a_ ( lowerCAmelCase_ : Optional[Any]=80, lowerCAmelCase_ : Optional[int]=False ): __lowerCAmelCase = [] # deal with critical env vars __lowerCAmelCase = ['CUDA_VISIBLE_DEVICES'] for key in env_keys: __lowerCAmelCase = os.environ.get(lowerCAmelCase_, lowerCAmelCase_ ) if val is not None: cmd.append(F"""{key}={val}""" ) # python executable (not always needed if the script is executable) __lowerCAmelCase = sys.executable if full_python_path else sys.executable.split('/' )[-1] cmd.append(lowerCAmelCase_ ) # now the normal args cmd += list(map(shlex.quote, sys.argv ) ) # split up into up to MAX_WIDTH lines with shell multi-line escapes __lowerCAmelCase = [] __lowerCAmelCase = '' while len(lowerCAmelCase_ ) > 0: current_line += F"""{cmd.pop(0 )} """ if len(lowerCAmelCase_ ) == 0 or len(lowerCAmelCase_ ) + len(cmd[0] ) + 1 > max_width - 1: lines.append(lowerCAmelCase_ ) __lowerCAmelCase = '' return "\\\n".join(lowerCAmelCase_ ) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : int ): # unwrap multi-line input __lowerCAmelCase = re.sub(R'[\\\n]+', ' ', args.base_cmd ) # remove --output_dir if any and set our own __lowerCAmelCase = re.sub('--output_dir\s+[^\s]+', '', args.base_cmd ) args.base_cmd += F""" --output_dir {output_dir}""" # ensure we have --overwrite_output_dir __lowerCAmelCase = re.sub('--overwrite_output_dir\s+', '', args.base_cmd ) args.base_cmd += " --overwrite_output_dir" return [sys.executable] + shlex.split(args.base_cmd ) def a_ ( lowerCAmelCase_ : List[str], lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : List[Any], lowerCAmelCase_ : Tuple, lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : List[Any], lowerCAmelCase_ : int ): # Enable to debug everything but the run itself, to do it fast and see the progress. # This is useful for debugging the output formatting quickly - we can remove it later once # everybody is happy with the output if 0: import random from time import sleep sleep(0 ) return dict( {k: random.uniform(0, 100 ) for k in metric_keys}, **{target_metric_key: random.choice([nan, 10.31, 100.2, 55.6666, 222.2222_2222] )}, ) __lowerCAmelCase = subprocess.run(lowerCAmelCase_, capture_output=lowerCAmelCase_, text=lowerCAmelCase_ ) if verbose: print('STDOUT', result.stdout ) print('STDERR', result.stderr ) # save the streams __lowerCAmelCase = variation.replace(' ', '-' ) with open(Path(lowerCAmelCase_ ) / F"""log.{prefix}.stdout.txt""", 'w' ) as f: f.write(result.stdout ) with open(Path(lowerCAmelCase_ ) / F"""log.{prefix}.stderr.txt""", 'w' ) as f: f.write(result.stderr ) if result.returncode != 0: if verbose: print('failed' ) return {target_metric_key: nan} with io.open(F"""{output_dir}/all_results.json""", 'r', encoding='utf-8' ) as f: __lowerCAmelCase = json.load(lowerCAmelCase_ ) # filter out just the keys we want return {k: v for k, v in metrics.items() if k in metric_keys} def a_ ( lowerCAmelCase_ : List[Any], lowerCAmelCase_ : int, lowerCAmelCase_ : str, lowerCAmelCase_ : List[Any], lowerCAmelCase_ : str, lowerCAmelCase_ : Dict, lowerCAmelCase_ : str, lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Tuple, lowerCAmelCase_ : Tuple, ): __lowerCAmelCase = [] __lowerCAmelCase = [] __lowerCAmelCase = F"""{id}: {variation:<{longest_variation_len}}""" __lowerCAmelCase = F"""{preamble}: """ __lowerCAmelCase = set(report_metric_keys + [target_metric_key] ) for i in tqdm(range(lowerCAmelCase_ ), desc=lowerCAmelCase_, leave=lowerCAmelCase_ ): __lowerCAmelCase = process_run_single( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = single_run_metrics[target_metric_key] if not math.isnan(lowerCAmelCase_ ): metrics.append(lowerCAmelCase_ ) results.append(lowerCAmelCase_ ) outcome += "✓" else: outcome += "✘" __lowerCAmelCase = F"""\33[2K\r{outcome}""" if len(lowerCAmelCase_ ) > 0: __lowerCAmelCase = {k: fmean([x[k] for x in metrics] ) for k in metrics[0].keys()} __lowerCAmelCase = round(mean_metrics[target_metric_key], 2 ) __lowerCAmelCase = F"""{outcome} {mean_target}""" if len(lowerCAmelCase_ ) > 1: results_str += F""" {tuple(round(lowerCAmelCase_, 2 ) for x in results )}""" print(lowerCAmelCase_ ) __lowerCAmelCase = variation return mean_metrics else: print(lowerCAmelCase_ ) return {variation_key: variation, target_metric_key: nan} def a_ ( ): __lowerCAmelCase = torch.cuda.get_device_properties(torch.device('cuda' ) ) return F""" Datetime : {datetime.datetime.now().strftime("%Y-%m-%d %H:%M:%S" )} Software: transformers: {transformers.__version__} torch : {torch.__version__} cuda : {torch.version.cuda} python : {platform.python_version()} Hardware: {torch.cuda.device_count()} GPUs : {properties.name}, {properties.total_memory/2**30:0.2f}GB """ def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : List[str], lowerCAmelCase_ : Tuple, lowerCAmelCase_ : Any, lowerCAmelCase_ : List[str] ): __lowerCAmelCase = pd.DataFrame(lowerCAmelCase_ ) __lowerCAmelCase = 'variation' __lowerCAmelCase = 'diff_%' __lowerCAmelCase = nan if base_variation is not None and len(df[df[variation_key] == base_variation] ): # this may still return nan __lowerCAmelCase = df.loc[df[variation_key] == base_variation][target_metric_key].item() if math.isnan(lowerCAmelCase_ ): # as a fallback, use the minimal value as the sentinel __lowerCAmelCase = df.loc[df[target_metric_key] != nan][target_metric_key].min() # create diff column if possible if not math.isnan(lowerCAmelCase_ ): __lowerCAmelCase = df.apply( lambda lowerCAmelCase_ : round(100 * (r[target_metric_key] - sentinel_value) / sentinel_value ) if not math.isnan(r[target_metric_key] ) else 0, axis='columns', ) # re-order columns __lowerCAmelCase = [variation_key, target_metric_key, diff_key, *report_metric_keys] __lowerCAmelCase = df.reindex(lowerCAmelCase_, axis='columns' ) # reorder cols # capitalize __lowerCAmelCase = df.rename(str.capitalize, axis='columns' ) # make the cols as narrow as possible __lowerCAmelCase = df.rename(lambda lowerCAmelCase_ : c.replace('_', '<br>' ), axis='columns' ) __lowerCAmelCase = df.rename(lambda lowerCAmelCase_ : c.replace('_', '\n' ), axis='columns' ) __lowerCAmelCase = ['', 'Copy between the cut-here-lines and paste as is to github or a forum'] report += ["----------8<-----------------8<--------"] report += ["*** Results:", df_github.to_markdown(index=lowerCAmelCase_, floatfmt='.2f' )] report += ["```"] report += ["*** Setup:", get_versions()] report += ["*** The benchmark command line was:", get_original_command()] report += ["```"] report += ["----------8<-----------------8<--------"] report += ["*** Results (console):", df_console.to_markdown(index=lowerCAmelCase_, floatfmt='.2f' )] print('\n\n'.join(lowerCAmelCase_ ) ) def a_ ( ): __lowerCAmelCase = argparse.ArgumentParser() parser.add_argument( '--base-cmd', default=lowerCAmelCase_, type=lowerCAmelCase_, required=lowerCAmelCase_, help='Base cmd', ) parser.add_argument( '--variations', default=lowerCAmelCase_, type=lowerCAmelCase_, nargs='+', required=lowerCAmelCase_, help='Multi-dimensional variations, example: \'|--fp16|--bf16\' \'|--tf32\'', ) parser.add_argument( '--base-variation', default=lowerCAmelCase_, type=lowerCAmelCase_, help='Baseline variation to compare to. if None the minimal target value will be used to compare against', ) parser.add_argument( '--target-metric-key', default=lowerCAmelCase_, type=lowerCAmelCase_, required=lowerCAmelCase_, help='Target metric key in output_dir/all_results.json, e.g., train_samples_per_second', ) parser.add_argument( '--report-metric-keys', default='', type=lowerCAmelCase_, help='Report metric keys - other metric keys from output_dir/all_results.json to report, e.g., train_loss. Use a single argument e.g., \'train_loss train_samples', ) parser.add_argument( '--repeat-times', default=1, type=lowerCAmelCase_, help='How many times to re-run each variation - an average will be reported', ) parser.add_argument( '--output_dir', default='output_benchmark', type=lowerCAmelCase_, help='The output directory where all the benchmark reports will go to and additionally this directory will be used to override --output_dir in the script that is being benchmarked', ) parser.add_argument( '--verbose', default=lowerCAmelCase_, action='store_true', help='Whether to show the outputs of each run or just the benchmark progress', ) __lowerCAmelCase = parser.parse_args() __lowerCAmelCase = args.output_dir Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) __lowerCAmelCase = get_base_command(lowerCAmelCase_, lowerCAmelCase_ ) # split each dimension into its --foo variations __lowerCAmelCase = [list(map(str.strip, re.split(R'\|', lowerCAmelCase_ ) ) ) for x in args.variations] # build a cartesian product of dimensions and convert those back into cmd-line arg strings, # while stripping white space for inputs that were empty __lowerCAmelCase = list(map(str.strip, map(' '.join, itertools.product(*lowerCAmelCase_ ) ) ) ) __lowerCAmelCase = max(len(lowerCAmelCase_ ) for x in variations ) # split wanted keys __lowerCAmelCase = args.report_metric_keys.split() # capture prints into a log file for convenience __lowerCAmelCase = F"""benchmark-report-{datetime.datetime.now().strftime("%Y-%m-%d-%H-%M-%S" )}.txt""" print(F"""\nNote: each run's output is also logged under {output_dir}/log.*.std*.txt""" ) print(F"""and this script's output is also piped into {report_fn}""" ) __lowerCAmelCase = Tee(lowerCAmelCase_ ) print(F"""\n*** Running {len(lowerCAmelCase_ )} benchmarks:""" ) print(F"""Base command: {" ".join(lowerCAmelCase_ )}""" ) __lowerCAmelCase = 'variation' __lowerCAmelCase = [] for id, variation in enumerate(tqdm(lowerCAmelCase_, desc='Total completion: ', leave=lowerCAmelCase_ ) ): __lowerCAmelCase = base_cmd + variation.split() results.append( process_run( id + 1, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, args.target_metric_key, lowerCAmelCase_, args.repeat_times, lowerCAmelCase_, args.verbose, ) ) process_results(lowerCAmelCase_, args.target_metric_key, lowerCAmelCase_, args.base_variation, lowerCAmelCase_ ) if __name__ == "__main__": main()
284
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = StableDiffusionInstructPixaPixPipeline a_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width""", """cross_attention_kwargs"""} a_ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase ( self : Optional[int] ) -> Optional[int]: torch.manual_seed(0 ) __lowerCAmelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=8 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , ) __lowerCAmelCase = PNDMScheduler(skip_prk_steps=lowerCAmelCase_ ) torch.manual_seed(0 ) __lowerCAmelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) __lowerCAmelCase = CLIPTextModel(lowerCAmelCase_ ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __lowerCAmelCase = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowercase ( self : Optional[Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Tuple=0 ) -> Dict: __lowerCAmelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) __lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCAmelCase = Image.fromarray(np.uinta(lowerCAmelCase_ ) ).convert('RGB' ) if str(lowerCAmelCase_ ).startswith('mps' ): __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) else: __lowerCAmelCase = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'image_guidance_scale': 1, 'output_type': 'numpy', } return inputs def lowercase ( self : Tuple ) -> List[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_26, 0.37_50, 0.45_47, 0.61_17, 0.58_66, 0.50_16, 0.43_27, 0.56_42, 0.48_15] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Dict: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = 'french fries' __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ , negative_prompt=lowerCAmelCase_ ) __lowerCAmelCase = output.images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_11, 0.36_42, 0.45_53, 0.62_36, 0.57_97, 0.50_13, 0.43_43, 0.56_11, 0.48_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Any: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = [inputs['prompt']] * 2 __lowerCAmelCase = np.array(inputs['image'] ).astype(np.floataa ) / 2_55.0 __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ).unsqueeze(0 ).to(lowerCAmelCase_ ) __lowerCAmelCase = image / 2 + 0.5 __lowerCAmelCase = image.permute(0 , 3 , 1 , 2 ) __lowerCAmelCase = image.repeat(2 , 1 , 1 , 1 ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[-1, -3:, -3:, -1] assert image.shape == (2, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.58_12, 0.57_48, 0.52_22, 0.59_08, 0.56_95, 0.71_74, 0.68_04, 0.55_23, 0.55_79] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Dict ) -> Optional[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = EulerAncestralDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' ) __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = [round(lowerCAmelCase_ , 4 ) for x in image_slice.flatten().tolist()] print(','.join([str(lowerCAmelCase_ ) for x in slice] ) ) assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.74_17, 0.38_42, 0.47_32, 0.57_76, 0.58_91, 0.51_39, 0.40_52, 0.56_73, 0.49_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> Optional[int]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def lowercase ( self : Optional[Any] ) -> Optional[Any]: __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = VaeImageProcessor(do_resize=lowerCAmelCase_ , do_normalize=lowerCAmelCase_ ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = pipe(**self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) )[0] __lowerCAmelCase = components['vae'] __lowerCAmelCase = self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): __lowerCAmelCase = vae.encode(inputs[image_param] ).latent_dist.mode() __lowerCAmelCase = pipe(**lowerCAmelCase_ )[0] __lowerCAmelCase = np.abs(out - out_latents_inputs ).max() self.assertLess(lowerCAmelCase_ , 1e-4 , 'passing latents as image input generate different result from passing image' ) @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> Optional[int]: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase ( self : List[str] , lowerCAmelCase_ : List[Any]=0 ) -> Any: __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = load_image( 'https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg' ) __lowerCAmelCase = { 'prompt': 'turn him into a cyborg', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'image_guidance_scale': 1.0, 'output_type': 'numpy', } return inputs def lowercase ( self : List[Any] ) -> str: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.59_02, 0.60_15, 0.60_27, 0.59_83, 0.60_92, 0.60_61, 0.57_65, 0.57_85, 0.55_55] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Tuple ) -> List[str]: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.65_78, 0.68_17, 0.69_72, 0.67_61, 0.68_56, 0.69_16, 0.64_28, 0.65_16, 0.63_01] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[Any] ) -> Dict: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.38_28, 0.38_34, 0.38_18, 0.37_92, 0.38_65, 0.37_52, 0.37_92, 0.38_47, 0.37_53] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> int: __lowerCAmelCase = 0 def callback_fn(lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : torch.FloatTensor ) -> None: __lowerCAmelCase = True nonlocal number_of_steps number_of_steps += 1 if step == 1: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.24_63, -0.46_44, -0.97_56, 1.51_76, 1.44_14, 0.78_66, 0.98_97, 0.85_21, 0.79_83] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.26_44, -0.46_26, -0.96_53, 1.51_76, 1.45_51, 0.76_86, 0.98_05, 0.84_52, 0.81_15] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 __lowerCAmelCase = False __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() pipe(**lowerCAmelCase_ , callback=lowerCAmelCase_ , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def lowercase ( self : Optional[int] ) -> Any: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 1_0**9 def lowercase ( self : List[Any] ) -> Any: __lowerCAmelCase = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 __lowerCAmelCase = inputs['image'].resize((5_0_4, 5_0_4) ) __lowerCAmelCase = 'timbrooks/instruct-pix2pix' __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( lowerCAmelCase_ , safety_checker=lowerCAmelCase_ , ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = output.images[0] __lowerCAmelCase = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 5_0_4, 3) __lowerCAmelCase = np.array([0.27_26, 0.25_29, 0.26_64, 0.26_55, 0.26_41, 0.26_42, 0.25_91, 0.26_49, 0.25_90] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3
284
1
_snake_case : Union[str, Any] = '\n# Transformers 설치 방법\n! pip install transformers datasets\n# 마지막 릴리스 대신 소스에서 설치하려면, 위 명령을 주석으로 바꾸고 아래 명령을 해제하세요.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' _snake_case : Union[str, Any] = [{'type': 'code', 'content': INSTALL_CONTENT}] _snake_case : str = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
284
from timeit import timeit def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('the value of input must not be negative' ) __lowerCAmelCase = 0 while number: number &= number - 1 result += 1 return result def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('the value of input must not be negative' ) __lowerCAmelCase = 0 while number: if number % 2 == 1: result += 1 number >>= 1 return result def a_ ( ): def do_benchmark(lowerCAmelCase_ : int ) -> None: __lowerCAmelCase = 'import __main__ as z' print(F"""Benchmark when {number = }:""" ) print(F"""{get_set_bits_count_using_modulo_operator(lowerCAmelCase_ ) = }""" ) __lowerCAmelCase = timeit('z.get_set_bits_count_using_modulo_operator(25)', setup=lowerCAmelCase_ ) print(F"""timeit() runs in {timing} seconds""" ) print(F"""{get_set_bits_count_using_brian_kernighans_algorithm(lowerCAmelCase_ ) = }""" ) __lowerCAmelCase = timeit( 'z.get_set_bits_count_using_brian_kernighans_algorithm(25)', setup=lowerCAmelCase_, ) print(F"""timeit() runs in {timing} seconds""" ) for number in (25, 37, 58, 0): do_benchmark(lowerCAmelCase_ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
284
1
import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : Any , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Tuple=7 , lowerCAmelCase_ : Union[str, Any]=3 , lowerCAmelCase_ : List[Any]=1_8 , lowerCAmelCase_ : Dict=3_0 , lowerCAmelCase_ : Tuple=4_0_0 , lowerCAmelCase_ : Optional[int]=True , lowerCAmelCase_ : Dict=None , lowerCAmelCase_ : Any=True , ) -> Tuple: __lowerCAmelCase = size if size is not None else {'height': 1_8, 'width': 1_8} __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = num_channels __lowerCAmelCase = image_size __lowerCAmelCase = min_resolution __lowerCAmelCase = max_resolution __lowerCAmelCase = do_resize __lowerCAmelCase = size __lowerCAmelCase = apply_ocr def lowercase ( self : Any ) -> str: return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class _UpperCAmelCase ( _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = LayoutLMvaImageProcessor if is_pytesseract_available() else None def lowercase ( self : Dict ) -> Tuple: __lowerCAmelCase = LayoutLMvaImageProcessingTester(self ) @property def lowercase ( self : Optional[int] ) -> Optional[Any]: return self.image_processor_tester.prepare_image_processor_dict() def lowercase ( self : Optional[int] ) -> str: __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCAmelCase_ , 'do_resize' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'size' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'apply_ocr' ) ) def lowercase ( self : List[Any] ) -> Union[str, Any]: __lowerCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'height': 1_8, 'width': 1_8} ) __lowerCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 ) self.assertEqual(image_processor.size , {'height': 4_2, 'width': 4_2} ) def lowercase ( self : Optional[int] ) -> str: pass def lowercase ( self : List[Any] ) -> str: # Initialize image_processing __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase_ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase_ , Image.Image ) # Test not batched input __lowerCAmelCase = image_processing(image_inputs[0] , return_tensors='pt' ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) self.assertIsInstance(encoding.words , lowerCAmelCase_ ) self.assertIsInstance(encoding.boxes , lowerCAmelCase_ ) # Test batched __lowerCAmelCase = image_processing(lowerCAmelCase_ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) def lowercase ( self : str ) -> Optional[int]: # Initialize image_processing __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase_ , numpify=lowerCAmelCase_ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase_ , np.ndarray ) # Test not batched input __lowerCAmelCase = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) # Test batched __lowerCAmelCase = image_processing(lowerCAmelCase_ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) def lowercase ( self : Any ) -> Tuple: # Initialize image_processing __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase_ , torchify=lowerCAmelCase_ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase_ , torch.Tensor ) # Test not batched input __lowerCAmelCase = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) # Test batched __lowerCAmelCase = image_processing(lowerCAmelCase_ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) def lowercase ( self : List[str] ) -> int: # with apply_OCR = True __lowerCAmelCase = LayoutLMvaImageProcessor() from datasets import load_dataset __lowerCAmelCase = load_dataset('hf-internal-testing/fixtures_docvqa' , split='test' ) __lowerCAmelCase = Image.open(ds[0]['file'] ).convert('RGB' ) __lowerCAmelCase = image_processing(lowerCAmelCase_ , return_tensors='pt' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 2_2_4, 2_2_4) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 __lowerCAmelCase = [['11:14', 'to', '11:39', 'a.m', '11:39', 'to', '11:44', 'a.m.', '11:44', 'a.m.', 'to', '12:25', 'p.m.', '12:25', 'to', '12:58', 'p.m.', '12:58', 'to', '4:00', 'p.m.', '2:00', 'to', '5:00', 'p.m.', 'Coffee', 'Break', 'Coffee', 'will', 'be', 'served', 'for', 'men', 'and', 'women', 'in', 'the', 'lobby', 'adjacent', 'to', 'exhibit', 'area.', 'Please', 'move', 'into', 'exhibit', 'area.', '(Exhibits', 'Open)', 'TRRF', 'GENERAL', 'SESSION', '(PART', '|)', 'Presiding:', 'Lee', 'A.', 'Waller', 'TRRF', 'Vice', 'President', '“Introductory', 'Remarks”', 'Lee', 'A.', 'Waller,', 'TRRF', 'Vice', 'Presi-', 'dent', 'Individual', 'Interviews', 'with', 'TRRF', 'Public', 'Board', 'Members', 'and', 'Sci-', 'entific', 'Advisory', 'Council', 'Mem-', 'bers', 'Conducted', 'by', 'TRRF', 'Treasurer', 'Philip', 'G.', 'Kuehn', 'to', 'get', 'answers', 'which', 'the', 'public', 'refrigerated', 'warehousing', 'industry', 'is', 'looking', 'for.', 'Plus', 'questions', 'from', 'the', 'floor.', 'Dr.', 'Emil', 'M.', 'Mrak,', 'University', 'of', 'Cal-', 'ifornia,', 'Chairman,', 'TRRF', 'Board;', 'Sam', 'R.', 'Cecil,', 'University', 'of', 'Georgia', 'College', 'of', 'Agriculture;', 'Dr.', 'Stanley', 'Charm,', 'Tufts', 'University', 'School', 'of', 'Medicine;', 'Dr.', 'Robert', 'H.', 'Cotton,', 'ITT', 'Continental', 'Baking', 'Company;', 'Dr.', 'Owen', 'Fennema,', 'University', 'of', 'Wis-', 'consin;', 'Dr.', 'Robert', 'E.', 'Hardenburg,', 'USDA.', 'Questions', 'and', 'Answers', 'Exhibits', 'Open', 'Capt.', 'Jack', 'Stoney', 'Room', 'TRRF', 'Scientific', 'Advisory', 'Council', 'Meeting', 'Ballroom', 'Foyer']] # noqa: E231 __lowerCAmelCase = [[[1_4_1, 5_7, 2_1_4, 6_9], [2_2_8, 5_8, 2_5_2, 6_9], [1_4_1, 7_5, 2_1_6, 8_8], [2_3_0, 7_9, 2_8_0, 8_8], [1_4_2, 2_6_0, 2_1_8, 2_7_3], [2_3_0, 2_6_1, 2_5_5, 2_7_3], [1_4_3, 2_7_9, 2_1_8, 2_9_0], [2_3_1, 2_8_2, 2_9_0, 2_9_1], [1_4_3, 3_4_2, 2_1_8, 3_5_4], [2_3_1, 3_4_5, 2_8_9, 3_5_5], [2_0_2, 3_6_2, 2_2_7, 3_7_3], [1_4_3, 3_7_9, 2_2_0, 3_9_2], [2_3_1, 3_8_2, 2_9_1, 3_9_4], [1_4_4, 7_1_4, 2_2_0, 7_2_6], [2_3_1, 7_1_5, 2_5_6, 7_2_6], [1_4_4, 7_3_2, 2_2_0, 7_4_5], [2_3_2, 7_3_6, 2_9_1, 7_4_7], [1_4_4, 7_6_9, 2_1_8, 7_8_2], [2_3_1, 7_7_0, 2_5_6, 7_8_2], [1_4_1, 7_8_8, 2_0_2, 8_0_1], [2_1_5, 7_9_1, 2_7_4, 8_0_4], [1_4_3, 8_2_6, 2_0_4, 8_3_8], [2_1_5, 8_2_6, 2_4_0, 8_3_8], [1_4_2, 8_4_4, 2_0_2, 8_5_7], [2_1_5, 8_4_7, 2_7_4, 8_5_9], [3_3_4, 5_7, 4_2_7, 6_9], [4_4_0, 5_7, 5_2_2, 6_9], [3_6_9, 7_5, 4_6_1, 8_8], [4_6_9, 7_5, 5_1_6, 8_8], [5_2_8, 7_6, 5_6_2, 8_8], [5_7_0, 7_6, 6_6_7, 8_8], [6_7_5, 7_5, 7_1_1, 8_7], [7_2_1, 7_9, 7_7_8, 8_8], [7_8_9, 7_5, 8_4_0, 8_8], [3_6_9, 9_7, 4_7_0, 1_0_7], [4_8_4, 9_4, 5_0_7, 1_0_6], [5_1_8, 9_4, 5_6_2, 1_0_7], [5_7_6, 9_4, 6_5_5, 1_1_0], [6_6_8, 9_4, 7_9_2, 1_0_9], [8_0_4, 9_5, 8_2_9, 1_0_7], [3_6_9, 1_1_3, 4_6_5, 1_2_5], [4_7_7, 1_1_6, 5_4_7, 1_2_5], [5_6_2, 1_1_3, 6_5_8, 1_2_5], [6_7_1, 1_1_6, 7_4_8, 1_2_5], [7_6_1, 1_1_3, 8_1_1, 1_2_5], [3_6_9, 1_3_1, 4_6_5, 1_4_3], [4_7_7, 1_3_3, 5_4_8, 1_4_3], [5_6_3, 1_3_0, 6_9_8, 1_4_5], [7_1_0, 1_3_0, 8_0_2, 1_4_6], [3_3_6, 1_7_1, 4_1_2, 1_8_3], [4_2_3, 1_7_1, 5_7_2, 1_8_3], [5_8_2, 1_7_0, 7_1_6, 1_8_4], [7_2_8, 1_7_1, 8_1_7, 1_8_7], [8_2_9, 1_7_1, 8_4_4, 1_8_6], [3_3_8, 1_9_7, 4_8_2, 2_1_2], [5_0_7, 1_9_6, 5_5_7, 2_0_9], [5_6_9, 1_9_6, 5_9_5, 2_0_8], [6_1_0, 1_9_6, 7_0_2, 2_0_9], [5_0_5, 2_1_4, 5_8_3, 2_2_6], [5_9_5, 2_1_4, 6_5_6, 2_2_7], [6_7_0, 2_1_5, 8_0_7, 2_2_7], [3_3_5, 2_5_9, 5_4_3, 2_7_4], [5_5_6, 2_5_9, 7_0_8, 2_7_2], [3_7_2, 2_7_9, 4_2_2, 2_9_1], [4_3_5, 2_7_9, 4_6_0, 2_9_1], [4_7_4, 2_7_9, 5_7_4, 2_9_2], [5_8_7, 2_7_8, 6_6_4, 2_9_1], [6_7_6, 2_7_8, 7_3_8, 2_9_1], [7_5_1, 2_7_9, 8_3_4, 2_9_1], [3_7_2, 2_9_8, 4_3_4, 3_1_0], [3_3_5, 3_4_1, 4_8_3, 3_5_4], [4_9_7, 3_4_1, 6_5_5, 3_5_4], [6_6_7, 3_4_1, 7_2_8, 3_5_4], [7_4_0, 3_4_1, 8_2_5, 3_5_4], [3_3_5, 3_6_0, 4_3_0, 3_7_2], [4_4_2, 3_6_0, 5_3_4, 3_7_2], [5_4_5, 3_5_9, 6_8_7, 3_7_2], [6_9_7, 3_6_0, 7_5_4, 3_7_2], [7_6_5, 3_6_0, 8_2_3, 3_7_3], [3_3_4, 3_7_8, 4_2_8, 3_9_1], [4_4_0, 3_7_8, 5_7_7, 3_9_4], [5_9_0, 3_7_8, 7_0_5, 3_9_1], [7_2_0, 3_7_8, 8_0_1, 3_9_1], [3_3_4, 3_9_7, 4_0_0, 4_0_9], [3_7_0, 4_1_6, 5_2_9, 4_2_9], [5_4_4, 4_1_6, 5_7_6, 4_3_2], [5_8_7, 4_1_6, 6_6_5, 4_2_8], [6_7_7, 4_1_6, 8_1_4, 4_2_9], [3_7_2, 4_3_5, 4_5_2, 4_5_0], [4_6_5, 4_3_4, 4_9_5, 4_4_7], [5_1_1, 4_3_4, 6_0_0, 4_4_7], [6_1_1, 4_3_6, 6_3_7, 4_4_7], [6_4_9, 4_3_6, 6_9_4, 4_5_1], [7_0_5, 4_3_8, 8_2_4, 4_4_7], [3_6_9, 4_5_3, 4_5_2, 4_6_6], [4_6_4, 4_5_4, 5_0_9, 4_6_6], [5_2_2, 4_5_3, 6_1_1, 4_6_9], [6_2_5, 4_5_3, 7_9_2, 4_6_9], [3_7_0, 4_7_2, 5_5_6, 4_8_8], [5_7_0, 4_7_2, 6_8_4, 4_8_7], [6_9_7, 4_7_2, 7_1_8, 4_8_5], [7_3_2, 4_7_2, 8_3_5, 4_8_8], [3_6_9, 4_9_0, 4_1_1, 5_0_3], [4_2_5, 4_9_0, 4_8_4, 5_0_3], [4_9_6, 4_9_0, 6_3_5, 5_0_6], [6_4_5, 4_9_0, 7_0_7, 5_0_3], [7_1_8, 4_9_1, 7_6_1, 5_0_3], [7_7_1, 4_9_0, 8_4_0, 5_0_3], [3_3_6, 5_1_0, 3_7_4, 5_2_1], [3_8_8, 5_1_0, 4_4_7, 5_2_2], [4_6_0, 5_1_0, 4_8_9, 5_2_1], [5_0_3, 5_1_0, 5_8_0, 5_2_2], [5_9_2, 5_0_9, 7_3_6, 5_2_5], [7_4_5, 5_0_9, 7_7_0, 5_2_2], [7_8_1, 5_0_9, 8_4_0, 5_2_2], [3_3_8, 5_2_8, 4_3_4, 5_4_1], [4_4_8, 5_2_8, 5_9_6, 5_4_1], [6_0_9, 5_2_7, 6_8_7, 5_4_0], [7_0_0, 5_2_8, 7_9_2, 5_4_1], [3_3_6, 5_4_6, 3_9_7, 5_5_9], [4_0_7, 5_4_6, 4_3_1, 5_5_9], [4_4_3, 5_4_6, 5_2_5, 5_6_0], [5_3_7, 5_4_6, 6_8_0, 5_6_2], [6_8_8, 5_4_6, 7_1_4, 5_5_9], [7_2_2, 5_4_6, 8_3_7, 5_6_2], [3_3_6, 5_6_5, 4_4_9, 5_8_1], [4_6_1, 5_6_5, 4_8_5, 5_7_7], [4_9_7, 5_6_5, 6_6_5, 5_8_1], [6_8_1, 5_6_5, 7_1_8, 5_7_7], [7_3_2, 5_6_5, 8_3_7, 5_8_0], [3_3_7, 5_8_4, 4_3_8, 5_9_7], [4_5_2, 5_8_3, 5_2_1, 5_9_6], [5_3_5, 5_8_4, 6_7_7, 5_9_9], [6_9_0, 5_8_3, 7_8_7, 5_9_6], [8_0_1, 5_8_3, 8_2_5, 5_9_6], [3_3_8, 6_0_2, 4_7_8, 6_1_5], [4_9_2, 6_0_2, 5_3_0, 6_1_4], [5_4_3, 6_0_2, 6_3_8, 6_1_5], [6_5_0, 6_0_2, 6_7_6, 6_1_4], [6_8_8, 6_0_2, 7_8_8, 6_1_5], [8_0_2, 6_0_2, 8_4_3, 6_1_4], [3_3_7, 6_2_1, 5_0_2, 6_3_3], [5_1_6, 6_2_1, 6_1_5, 6_3_7], [6_2_9, 6_2_1, 7_7_4, 6_3_6], [7_8_9, 6_2_1, 8_2_7, 6_3_3], [3_3_7, 6_3_9, 4_1_8, 6_5_2], [4_3_2, 6_4_0, 5_7_1, 6_5_3], [5_8_7, 6_3_9, 7_3_1, 6_5_5], [7_4_3, 6_3_9, 7_6_9, 6_5_2], [7_8_0, 6_3_9, 8_4_1, 6_5_2], [3_3_8, 6_5_8, 4_4_0, 6_7_3], [4_5_5, 6_5_8, 4_9_1, 6_7_0], [5_0_8, 6_5_8, 6_0_2, 6_7_1], [6_1_6, 6_5_8, 6_3_8, 6_7_0], [6_5_4, 6_5_8, 8_3_5, 6_7_4], [3_3_7, 6_7_7, 4_2_9, 6_8_9], [3_3_7, 7_1_4, 4_8_2, 7_2_6], [4_9_5, 7_1_4, 5_4_8, 7_2_6], [5_6_1, 7_1_4, 6_8_3, 7_2_6], [3_3_8, 7_7_0, 4_6_1, 7_8_2], [4_7_4, 7_6_9, 5_5_4, 7_8_5], [4_8_9, 7_8_8, 5_6_2, 8_0_3], [5_7_6, 7_8_8, 6_4_3, 8_0_1], [6_5_6, 7_8_7, 7_5_1, 8_0_4], [7_6_4, 7_8_8, 8_4_4, 8_0_1], [3_3_4, 8_2_5, 4_2_1, 8_3_8], [4_3_0, 8_2_4, 5_7_4, 8_3_8], [5_8_4, 8_2_4, 7_2_3, 8_4_1], [3_3_5, 8_4_4, 4_5_0, 8_5_7], [4_6_4, 8_4_3, 5_8_3, 8_6_0], [6_2_8, 8_6_2, 7_5_5, 8_7_5], [7_6_9, 8_6_1, 8_4_8, 8_7_8]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , lowerCAmelCase_ ) self.assertListEqual(encoding.boxes , lowerCAmelCase_ ) # with apply_OCR = False __lowerCAmelCase = LayoutLMvaImageProcessor(apply_ocr=lowerCAmelCase_ ) __lowerCAmelCase = image_processing(lowerCAmelCase_ , return_tensors='pt' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 2_2_4, 2_2_4) )
284
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss _snake_case : Dict = pytest.mark.integration @require_faiss class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : List[Any] ) -> Optional[Any]: __lowerCAmelCase = Dataset.from_dict({'filename': ['my_name-train' + '_' + str(lowerCAmelCase_ ) for x in np.arange(3_0 ).tolist()]} ) return dset def lowercase ( self : List[str] ) -> Tuple: import faiss __lowerCAmelCase = self._create_dummy_dataset() __lowerCAmelCase = dset.map( lambda lowerCAmelCase_ , lowerCAmelCase_ : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=lowerCAmelCase_ , keep_in_memory=lowerCAmelCase_ ) __lowerCAmelCase = dset.add_faiss_index('vecs' , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) dset.drop_index('vecs' ) def lowercase ( self : Optional[Any] ) -> str: import faiss __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT , ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def lowercase ( self : int ) -> Optional[Any]: import faiss __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCAmelCase_ ) as tmp_file: dset.save_faiss_index('vecs' , tmp_file.name ) dset.load_faiss_index('vecs2' , tmp_file.name ) os.unlink(tmp_file.name ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs2' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def lowercase ( self : Union[str, Any] ) -> List[Any]: __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' ) dset.drop_index('vecs' ) self.assertRaises(lowerCAmelCase_ , partial(dset.get_nearest_examples , 'vecs2' , np.ones(5 , dtype=np.floataa ) ) ) def lowercase ( self : Union[str, Any] ) -> Tuple: from elasticsearch import Elasticsearch __lowerCAmelCase = self._create_dummy_dataset() with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: __lowerCAmelCase = {'acknowledged': True} mocked_bulk.return_value([(True, None)] * 3_0 ) __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 2_9}]}} __lowerCAmelCase = Elasticsearch() dset.add_elasticsearch_index('filename' , es_client=lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('filename' , 'my_name-train_29' ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) @require_faiss class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : str ) -> int: import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 1_0 ) # single query __lowerCAmelCase = np.zeros(5 , dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertRaises(lowerCAmelCase_ , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries __lowerCAmelCase = np.eye(5 , dtype=np.floataa )[::-1] __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ ) self.assertRaises(lowerCAmelCase_ , index.search_batch , queries[0] ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , lowerCAmelCase_ ) def lowercase ( self : List[Any] ) -> List[str]: import faiss __lowerCAmelCase = FaissIndex(string_factory='Flat' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) __lowerCAmelCase = FaissIndex(string_factory='LSH' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(lowerCAmelCase_ ): __lowerCAmelCase = FaissIndex(string_factory='Flat' , custom_index=faiss.IndexFlat(5 ) ) def lowercase ( self : Union[str, Any] ) -> Dict: import faiss __lowerCAmelCase = faiss.IndexFlat(5 ) __lowerCAmelCase = FaissIndex(custom_index=lowerCAmelCase_ ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def lowercase ( self : str ) -> Any: import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCAmelCase_ ) as tmp_file: index.save(tmp_file.name ) __lowerCAmelCase = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) __lowerCAmelCase = np.zeros(5 , dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def a_ ( lowerCAmelCase_ : Union[str, Any] ): import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5, dtype=np.floataa ) ) __lowerCAmelCase = 'index.faiss' __lowerCAmelCase = F"""mock://{index_name}""" index.save(lowerCAmelCase_, storage_options=mockfs.storage_options ) __lowerCAmelCase = FaissIndex.load(lowerCAmelCase_, storage_options=mockfs.storage_options ) __lowerCAmelCase = np.zeros(5, dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : Any ) -> int: from elasticsearch import Elasticsearch with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: __lowerCAmelCase = Elasticsearch() __lowerCAmelCase = {'acknowledged': True} __lowerCAmelCase = ElasticSearchIndex(es_client=lowerCAmelCase_ ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(['foo', 'bar', 'foobar'] ) # single query __lowerCAmelCase = 'foo' __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout __lowerCAmelCase = 'foo' __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ , request_timeout=3_0 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries __lowerCAmelCase = ['foo', 'bar', 'foobar'] __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([1, 1, 1] , lowerCAmelCase_ ) # batched queries with timeout __lowerCAmelCase = ['foo', 'bar', 'foobar'] __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ , request_timeout=3_0 ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([1, 1, 1] , lowerCAmelCase_ )
284
1
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _snake_case : Dict = datasets.logging.get_logger(__name__) _snake_case : List[Any] = '\\n@InProceedings{moosavi2019minimum,\n author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube},\n title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection},\n year = {2019},\n booktitle = {Proceedings of the 57th Annual Meeting of\n the Association for Computational Linguistics (Volume 1: Long Papers)},\n publisher = {Association for Computational Linguistics},\n address = {Florence, Italy},\n}\n\n@inproceedings{10.3115/1072399.1072405,\nauthor = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette},\ntitle = {A Model-Theoretic Coreference Scoring Scheme},\nyear = {1995},\nisbn = {1558604022},\npublisher = {Association for Computational Linguistics},\naddress = {USA},\nurl = {https://doi.org/10.3115/1072399.1072405},\ndoi = {10.3115/1072399.1072405},\nbooktitle = {Proceedings of the 6th Conference on Message Understanding},\npages = {45–52},\nnumpages = {8},\nlocation = {Columbia, Maryland},\nseries = {MUC6 ’95}\n}\n\n@INPROCEEDINGS{Bagga98algorithmsfor,\n author = {Amit Bagga and Breck Baldwin},\n title = {Algorithms for Scoring Coreference Chains},\n booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference},\n year = {1998},\n pages = {563--566}\n}\n\n@INPROCEEDINGS{Luo05oncoreference,\n author = {Xiaoqiang Luo},\n title = {On coreference resolution performance metrics},\n booktitle = {In Proc. of HLT/EMNLP},\n year = {2005},\n pages = {25--32},\n publisher = {URL}\n}\n\n@inproceedings{moosavi-strube-2016-coreference,\n title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric",\n author = "Moosavi, Nafise Sadat and\n Strube, Michael",\n booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",\n month = aug,\n year = "2016",\n address = "Berlin, Germany",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/P16-1060",\n doi = "10.18653/v1/P16-1060",\n pages = "632--642",\n}\n\n' _snake_case : Tuple = '\\nCoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which\nimplements of the common evaluation metrics including MUC [Vilain et al, 1995],\nB-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005],\nLEA [Moosavi and Strube, 2016] and the averaged CoNLL score\n(the average of the F1 values of MUC, B-cubed and CEAFe)\n[Denis and Baldridge, 2009a; Pradhan et al., 2011].\n\nThis wrapper of CoVal currently only work with CoNLL line format:\nThe CoNLL format has one word per line with all the annotation for this word in column separated by spaces:\nColumn Type Description\n1 Document ID This is a variation on the document filename\n2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n3 Word number\n4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release.\n5 Part-of-Speech\n6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column.\n7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-"\n8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7.\n9 Word sense This is the word sense of the word in Column 3.\n10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data.\n11 Named Entities These columns identifies the spans representing various named entities.\n12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7.\nN Coreference Coreference chain information encoded in a parenthesis structure.\nMore informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html\n\nDetails on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md\n\nCoVal code was written by @ns-moosavi.\nSome parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py\nThe test suite is taken from https://github.com/conll/reference-coreference-scorers/\nMention evaluation and the test suite are added by @andreasvc.\nParsing CoNLL files is developed by Leo Born.\n' _snake_case : List[str] = '\nCalculates coreference evaluation metrics.\nArgs:\n predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format.\n Each prediction is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format.\n Each reference is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n keep_singletons: After extracting all mentions of key or system files,\n mentions whose corresponding coreference chain is of size one,\n are considered as singletons. The default evaluation mode will include\n singletons in evaluations if they are included in the key or the system files.\n By setting \'keep_singletons=False\', all singletons in the key and system files\n will be excluded from the evaluation.\n NP_only: Most of the recent coreference resolvers only resolve NP mentions and\n leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs.\n min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans.\n Minimum spans are determined using the MINA algorithm.\n\nReturns:\n \'mentions\': mentions\n \'muc\': MUC metric [Vilain et al, 1995]\n \'bcub\': B-cubed [Bagga and Baldwin, 1998]\n \'ceafe\': CEAFe [Luo et al., 2005]\n \'lea\': LEA [Moosavi and Strube, 2016]\n \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe)\n\nExamples:\n\n >>> coval = datasets.load_metric(\'coval\')\n >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\',\n ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\',\n ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\',\n ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\',\n ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\',\n ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\']\n >>> references = [words]\n >>> predictions = [words]\n >>> results = coval.compute(predictions=predictions, references=references)\n >>> print(results) # doctest:+ELLIPSIS\n {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0}\n' def a_ ( lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : List[Any], lowerCAmelCase_ : Union[str, Any]=False, lowerCAmelCase_ : Tuple=False, lowerCAmelCase_ : int=True, lowerCAmelCase_ : Tuple=False, lowerCAmelCase_ : Tuple="dummy_doc" ): __lowerCAmelCase = {doc: key_lines} __lowerCAmelCase = {doc: sys_lines} __lowerCAmelCase = {} __lowerCAmelCase = 0 __lowerCAmelCase = 0 __lowerCAmelCase = 0 __lowerCAmelCase = 0 __lowerCAmelCase = 0 __lowerCAmelCase = 0 __lowerCAmelCase , __lowerCAmelCase = reader.get_doc_mentions(lowerCAmelCase_, key_doc_lines[doc], lowerCAmelCase_ ) key_singletons_num += singletons_num if NP_only or min_span: __lowerCAmelCase = reader.set_annotated_parse_trees(lowerCAmelCase_, key_doc_lines[doc], lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase = reader.get_doc_mentions(lowerCAmelCase_, sys_doc_lines[doc], lowerCAmelCase_ ) sys_singletons_num += singletons_num if NP_only or min_span: __lowerCAmelCase = reader.set_annotated_parse_trees(lowerCAmelCase_, key_doc_lines[doc], lowerCAmelCase_, lowerCAmelCase_ ) if remove_nested: __lowerCAmelCase , __lowerCAmelCase = reader.remove_nested_coref_mentions(lowerCAmelCase_, lowerCAmelCase_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters __lowerCAmelCase , __lowerCAmelCase = reader.remove_nested_coref_mentions(lowerCAmelCase_, lowerCAmelCase_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters __lowerCAmelCase = reader.get_mention_assignments(lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = reader.get_mention_assignments(lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( 'Number of removed nested coreferring mentions in the key ' F"""annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}""" ) logger.info( 'Number of resulting singleton clusters in the key ' F"""annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}""" ) if not keep_singletons: logger.info( F"""{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system """ 'files, respectively' ) return doc_coref_infos def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : List[str], lowerCAmelCase_ : List[str], lowerCAmelCase_ : Tuple, lowerCAmelCase_ : List[str], lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = get_coref_infos(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = {} __lowerCAmelCase = 0 __lowerCAmelCase = 0 for name, metric in metrics: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = evaluator.evaluate_documents(lowerCAmelCase_, lowerCAmelCase_, beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F"""{name}/recall""": recall, F"""{name}/precision""": precision, F"""{name}/f1""": fa} ) logger.info( name.ljust(10 ), F"""Recall: {recall * 100:.2f}""", F""" Precision: {precision * 100:.2f}""", F""" F1: {fa * 100:.2f}""", ) if conll_subparts_num == 3: __lowerCAmelCase = (conll / 3) * 100 logger.info(F"""CoNLL score: {conll:.2f}""" ) output_scores.update({'conll_score': conll} ) return output_scores def a_ ( lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = False for line in key_lines: if not line.startswith('#' ): if len(line.split() ) > 6: __lowerCAmelCase = line.split()[5] if not parse_col == "-": __lowerCAmelCase = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): """simple docstring""" def lowercase ( self : Any ) -> Optional[int]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('string' ) ), 'references': datasets.Sequence(datasets.Value('string' ) ), } ) , codebase_urls=['https://github.com/ns-moosavi/coval'] , reference_urls=[ 'https://github.com/ns-moosavi/coval', 'https://www.aclweb.org/anthology/P16-1060', 'http://www.conll.cemantix.org/2012/data.html', ] , ) def lowercase ( self : Tuple , lowerCAmelCase_ : int , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Tuple=False , lowerCAmelCase_ : Union[str, Any]=False , lowerCAmelCase_ : int=False ) -> Dict: __lowerCAmelCase = [ ('mentions', evaluator.mentions), ('muc', evaluator.muc), ('bcub', evaluator.b_cubed), ('ceafe', evaluator.ceafe), ('lea', evaluator.lea), ] if min_span: __lowerCAmelCase = util.check_gold_parse_annotation(lowerCAmelCase_ ) if not has_gold_parse: raise NotImplementedError('References should have gold parse annotation to use \'min_span\'.' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" __lowerCAmelCase = evaluate( key_lines=lowerCAmelCase_ , sys_lines=lowerCAmelCase_ , metrics=lowerCAmelCase_ , NP_only=lowerCAmelCase_ , remove_nested=lowerCAmelCase_ , keep_singletons=lowerCAmelCase_ , min_span=lowerCAmelCase_ , ) return score
284
import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( 'kwargs, expected', [ ({'num_shards': 0, 'max_num_jobs': 1}, []), ({'num_shards': 10, 'max_num_jobs': 1}, [range(10 )]), ({'num_shards': 10, 'max_num_jobs': 10}, [range(lowerCAmelCase_, i + 1 ) for i in range(10 )]), ({'num_shards': 1, 'max_num_jobs': 10}, [range(1 )]), ({'num_shards': 10, 'max_num_jobs': 3}, [range(0, 4 ), range(4, 7 ), range(7, 10 )]), ({'num_shards': 3, 'max_num_jobs': 10}, [range(0, 1 ), range(1, 2 ), range(2, 3 )]), ], ) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : Any ): __lowerCAmelCase = _distribute_shards(**lowerCAmelCase_ ) assert out == expected @pytest.mark.parametrize( 'gen_kwargs, max_num_jobs, expected', [ ({'foo': 0}, 10, [{'foo': 0}]), ({'shards': [0, 1, 2, 3]}, 1, [{'shards': [0, 1, 2, 3]}]), ({'shards': [0, 1, 2, 3]}, 4, [{'shards': [0]}, {'shards': [1]}, {'shards': [2]}, {'shards': [3]}]), ({'shards': [0, 1]}, 4, [{'shards': [0]}, {'shards': [1]}]), ({'shards': [0, 1, 2, 3]}, 2, [{'shards': [0, 1]}, {'shards': [2, 3]}]), ], ) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : List[str], lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = _split_gen_kwargs(lowerCAmelCase_, lowerCAmelCase_ ) assert out == expected @pytest.mark.parametrize( 'gen_kwargs, expected', [ ({'foo': 0}, 1), ({'shards': [0]}, 1), ({'shards': [0, 1, 2, 3]}, 4), ({'shards': [0, 1, 2, 3], 'foo': 0}, 4), ({'shards': [0, 1, 2, 3], 'other': (0, 1)}, 4), ({'shards': [0, 1, 2, 3], 'shards2': [0, 1]}, RuntimeError), ], ) def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : Any ): if expected is RuntimeError: with pytest.raises(lowerCAmelCase_ ): _number_of_shards_in_gen_kwargs(lowerCAmelCase_ ) else: __lowerCAmelCase = _number_of_shards_in_gen_kwargs(lowerCAmelCase_ ) assert out == expected
284
1
import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class _UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : Tuple ) -> List[str]: super().__init__() __lowerCAmelCase = nn.Linear(3 , 4 ) __lowerCAmelCase = nn.BatchNormad(4 ) __lowerCAmelCase = nn.Linear(4 , 5 ) def lowercase ( self : int , lowerCAmelCase_ : Any ) -> Tuple: return self.lineara(self.batchnorm(self.lineara(lowerCAmelCase_ ) ) ) class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : List[str] ) -> str: __lowerCAmelCase = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(lowerCAmelCase_ , model.state_dict() ) __lowerCAmelCase = os.path.join(lowerCAmelCase_ , 'index.json' ) self.assertTrue(os.path.isfile(lowerCAmelCase_ ) ) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: __lowerCAmelCase = os.path.join(lowerCAmelCase_ , f"""{key}.dat""" ) self.assertTrue(os.path.isfile(lowerCAmelCase_ ) ) # TODO: add tests on the fact weights are properly loaded def lowercase ( self : Any ) -> Optional[int]: __lowerCAmelCase = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: __lowerCAmelCase = torch.randn(2 , 3 , dtype=lowerCAmelCase_ ) with TemporaryDirectory() as tmp_dir: __lowerCAmelCase = offload_weight(lowerCAmelCase_ , 'weight' , lowerCAmelCase_ , {} ) __lowerCAmelCase = os.path.join(lowerCAmelCase_ , 'weight.dat' ) self.assertTrue(os.path.isfile(lowerCAmelCase_ ) ) self.assertDictEqual(lowerCAmelCase_ , {'weight': {'shape': [2, 3], 'dtype': str(lowerCAmelCase_ ).split('.' )[1]}} ) __lowerCAmelCase = load_offloaded_weight(lowerCAmelCase_ , index['weight'] ) self.assertTrue(torch.equal(lowerCAmelCase_ , lowerCAmelCase_ ) ) def lowercase ( self : Any ) -> Tuple: __lowerCAmelCase = ModelForTest() __lowerCAmelCase = model.state_dict() __lowerCAmelCase = {k: v for k, v in state_dict.items() if 'linear2' not in k} __lowerCAmelCase = {k: v for k, v in state_dict.items() if 'linear2' in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(lowerCAmelCase_ , lowerCAmelCase_ ) __lowerCAmelCase = OffloadedWeightsLoader(state_dict=lowerCAmelCase_ , save_folder=lowerCAmelCase_ ) # Every key is there with the right value self.assertEqual(sorted(lowerCAmelCase_ ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(lowerCAmelCase_ , weight_map[key] ) ) __lowerCAmelCase = {k: v for k, v in state_dict.items() if 'weight' in k} __lowerCAmelCase = {k: v for k, v in state_dict.items() if 'weight' not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(lowerCAmelCase_ , lowerCAmelCase_ ) __lowerCAmelCase = OffloadedWeightsLoader(state_dict=lowerCAmelCase_ , save_folder=lowerCAmelCase_ ) # Every key is there with the right value self.assertEqual(sorted(lowerCAmelCase_ ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(lowerCAmelCase_ , weight_map[key] ) ) with TemporaryDirectory() as tmp_dir: offload_state_dict(lowerCAmelCase_ , lowerCAmelCase_ ) # Duplicates are removed __lowerCAmelCase = OffloadedWeightsLoader(state_dict=lowerCAmelCase_ , save_folder=lowerCAmelCase_ ) # Every key is there with the right value self.assertEqual(sorted(lowerCAmelCase_ ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(lowerCAmelCase_ , weight_map[key] ) ) def lowercase ( self : Dict ) -> List[str]: __lowerCAmelCase = {'a.1': 0, 'a.10': 1, 'a.2': 2} __lowerCAmelCase = extract_submodules_state_dict(lowerCAmelCase_ , ['a.1', 'a.2'] ) self.assertDictEqual(lowerCAmelCase_ , {'a.1': 0, 'a.2': 2} ) __lowerCAmelCase = {'a.1.a': 0, 'a.10.a': 1, 'a.2.a': 2} __lowerCAmelCase = extract_submodules_state_dict(lowerCAmelCase_ , ['a.1', 'a.2'] ) self.assertDictEqual(lowerCAmelCase_ , {'a.1.a': 0, 'a.2.a': 2} )
284
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class _UpperCAmelCase ( _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = BertJapaneseTokenizer a_ = False a_ = True def lowercase ( self : Optional[Any] ) -> List[str]: super().setUp() __lowerCAmelCase = [ '[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは', '世界', '##世界', '、', '##、', '。', '##。', ] __lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def lowercase ( self : List[Any] , lowerCAmelCase_ : Tuple ) -> str: __lowerCAmelCase = 'こんにちは、世界。 \nこんばんは、世界。' __lowerCAmelCase = 'こんにちは 、 世界 。 こんばんは 、 世界 。' return input_text, output_text def lowercase ( self : List[Any] , lowerCAmelCase_ : str ) -> Dict: __lowerCAmelCase , __lowerCAmelCase = self.get_input_output_texts(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(lowerCAmelCase_ , clean_up_tokenization_spaces=lowerCAmelCase_ ) return text, ids def lowercase ( self : List[str] ) -> Optional[int]: pass # TODO add if relevant def lowercase ( self : Optional[Any] ) -> Optional[Any]: pass # TODO add if relevant def lowercase ( self : Union[str, Any] ) -> Any: pass # TODO add if relevant def lowercase ( self : Dict ) -> Tuple: __lowerCAmelCase = self.tokenizer_class(self.vocab_file ) __lowerCAmelCase = tokenizer.tokenize('こんにちは、世界。\nこんばんは、世界。' ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) def lowercase ( self : List[str] ) -> List[str]: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='mecab' ) self.assertIsNotNone(lowerCAmelCase_ ) __lowerCAmelCase = 'こんにちは、世界。\nこんばんは、世界。' __lowerCAmelCase = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __lowerCAmelCase = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowerCAmelCase_ , 'wb' ) as handle: pickle.dump(lowerCAmelCase_ , lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'rb' ) as handle: __lowerCAmelCase = pickle.load(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer_new.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Dict ) -> Tuple: __lowerCAmelCase = MecabTokenizer(mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : List[Any] ) -> int: try: __lowerCAmelCase = MecabTokenizer(mecab_dic='unidic_lite' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : Tuple ) -> Optional[Any]: try: __lowerCAmelCase = MecabTokenizer(mecab_dic='unidic' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : Tuple ) -> Union[str, Any]: __lowerCAmelCase = MecabTokenizer(do_lower_case=lowerCAmelCase_ , mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iphone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : Union[str, Any] ) -> Optional[Any]: try: __lowerCAmelCase = MecabTokenizer( do_lower_case=lowerCAmelCase_ , normalize_text=lowerCAmelCase_ , mecab_option='-d /usr/local/lib/mecab/dic/jumandic' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れた', '\u3000', '。'] , ) def lowercase ( self : Any ) -> Union[str, Any]: __lowerCAmelCase = MecabTokenizer(normalize_text=lowerCAmelCase_ , mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', ' ', '。'] , ) @require_sudachi def lowercase ( self : List[str] ) -> List[str]: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='sudachi' ) self.assertIsNotNone(lowerCAmelCase_ ) __lowerCAmelCase = 'こんにちは、世界。\nこんばんは、世界。' __lowerCAmelCase = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __lowerCAmelCase = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowerCAmelCase_ , 'wb' ) as handle: pickle.dump(lowerCAmelCase_ , lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'rb' ) as handle: __lowerCAmelCase = pickle.load(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer_new.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) @require_sudachi def lowercase ( self : Union[str, Any] ) -> List[str]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iPhone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', ' ', '。', ' ', ' '] , ) @require_sudachi def lowercase ( self : Tuple ) -> Optional[Any]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='A' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国', '人', '参政', '権'] ) @require_sudachi def lowercase ( self : Tuple ) -> List[Any]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='B' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国人', '参政権'] ) @require_sudachi def lowercase ( self : List[str] ) -> Union[str, Any]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='C' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国人参政権'] ) @require_sudachi def lowercase ( self : Dict ) -> List[str]: __lowerCAmelCase = SudachiTokenizer(do_lower_case=lowerCAmelCase_ , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iphone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', ' ', '。', ' ', ' '] , ) @require_sudachi def lowercase ( self : Union[str, Any] ) -> List[Any]: __lowerCAmelCase = SudachiTokenizer(normalize_text=lowerCAmelCase_ , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iPhone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', '\u3000', '。', ' ', ' '] , ) @require_sudachi def lowercase ( self : int ) -> str: __lowerCAmelCase = SudachiTokenizer(trim_whitespace=lowerCAmelCase_ , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) @require_jumanpp def lowercase ( self : Union[str, Any] ) -> Any: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='jumanpp' ) self.assertIsNotNone(lowerCAmelCase_ ) __lowerCAmelCase = 'こんにちは、世界。\nこんばんは、世界。' __lowerCAmelCase = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __lowerCAmelCase = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowerCAmelCase_ , 'wb' ) as handle: pickle.dump(lowerCAmelCase_ , lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'rb' ) as handle: __lowerCAmelCase = pickle.load(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer_new.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) @require_jumanpp def lowercase ( self : List[Any] ) -> Optional[Any]: __lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def lowercase ( self : Any ) -> Union[str, Any]: __lowerCAmelCase = JumanppTokenizer(do_lower_case=lowerCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iphone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def lowercase ( self : Dict ) -> Dict: __lowerCAmelCase = JumanppTokenizer(normalize_text=lowerCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['ア', 'ッ', 'フ', '゚', 'ル', 'ストア', 'で', 'iPhone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def lowercase ( self : List[str] ) -> List[str]: __lowerCAmelCase = JumanppTokenizer(trim_whitespace=lowerCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れた', '。'] , ) @require_jumanpp def lowercase ( self : Any ) -> Any: __lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('ありがとうございますm(_ _)m見つけるのが大変です。' ) , ['ありがとう', 'ございます', 'm(_ _)m', '見つける', 'の', 'が', '大変です', '。'] , ) def lowercase ( self : Any ) -> str: __lowerCAmelCase = ['[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは'] __lowerCAmelCase = {} for i, token in enumerate(lowerCAmelCase_ ): __lowerCAmelCase = i __lowerCAmelCase = WordpieceTokenizer(vocab=lowerCAmelCase_ , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('こんにちは' ) , ['こんにちは'] ) self.assertListEqual(tokenizer.tokenize('こんばんは' ) , ['こん', '##ばんは'] ) self.assertListEqual(tokenizer.tokenize('こんばんは こんばんにちは こんにちは' ) , ['こん', '##ばんは', '[UNK]', 'こんにちは'] ) def lowercase ( self : List[Any] ) -> Tuple: __lowerCAmelCase = BertJapaneseTokenizer.from_pretrained('nlp-waseda/roberta-base-japanese-with-auto-jumanpp' ) __lowerCAmelCase = tokenizer.subword_tokenizer __lowerCAmelCase = subword_tokenizer.tokenize('国境 の 長い トンネル を 抜ける と 雪国 であった 。' ) self.assertListEqual(lowerCAmelCase_ , ['▁国境', '▁の', '▁長い', '▁トンネル', '▁を', '▁抜ける', '▁と', '▁雪', '国', '▁であった', '▁。'] ) __lowerCAmelCase = subword_tokenizer.tokenize('こんばんは こんばん にち は こんにちは' ) self.assertListEqual(lowerCAmelCase_ , ['▁こん', 'ばん', 'は', '▁こん', 'ばん', '▁に', 'ち', '▁は', '▁こんにちは'] ) def lowercase ( self : int ) -> str: __lowerCAmelCase = self.tokenizer_class.from_pretrained('cl-tohoku/bert-base-japanese' ) __lowerCAmelCase = tokenizer.encode('ありがとう。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.encode('どういたしまして。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ , lowerCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class _UpperCAmelCase ( _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = BertJapaneseTokenizer a_ = False def lowercase ( self : Optional[Any] ) -> Tuple: super().setUp() __lowerCAmelCase = ['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] __lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def lowercase ( self : str , **lowerCAmelCase_ : Tuple ) -> Union[str, Any]: return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='character' , **lowerCAmelCase_ ) def lowercase ( self : Tuple , lowerCAmelCase_ : Tuple ) -> Optional[int]: __lowerCAmelCase = 'こんにちは、世界。 \nこんばんは、世界。' __lowerCAmelCase = 'こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。' return input_text, output_text def lowercase ( self : Dict ) -> str: pass # TODO add if relevant def lowercase ( self : Any ) -> str: pass # TODO add if relevant def lowercase ( self : List[Any] ) -> int: pass # TODO add if relevant def lowercase ( self : str ) -> str: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , subword_tokenizer_type='character' ) __lowerCAmelCase = tokenizer.tokenize('こんにちは、世界。 \nこんばんは、世界。' ) self.assertListEqual( lowerCAmelCase_ , ['こ', 'ん', 'に', 'ち', 'は', '、', '世', '界', '。', 'こ', 'ん', 'ば', 'ん', 'は', '、', '世', '界', '。'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 4, 5, 6, 7, 1_1, 9, 1_0, 1_2, 3, 4, 8, 4, 7, 1_1, 9, 1_0, 1_2] ) def lowercase ( self : str ) -> Optional[int]: __lowerCAmelCase = ['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] __lowerCAmelCase = {} for i, token in enumerate(lowerCAmelCase_ ): __lowerCAmelCase = i __lowerCAmelCase = CharacterTokenizer(vocab=lowerCAmelCase_ , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('こんにちは' ) , ['こ', 'ん', 'に', 'ち', 'は'] ) self.assertListEqual(tokenizer.tokenize('こんにちほ' ) , ['こ', 'ん', 'に', 'ち', '[UNK]'] ) def lowercase ( self : int ) -> str: __lowerCAmelCase = self.tokenizer_class.from_pretrained('cl-tohoku/bert-base-japanese-char' ) __lowerCAmelCase = tokenizer.encode('ありがとう。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.encode('どういたしまして。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ , lowerCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : str ) -> Union[str, Any]: __lowerCAmelCase = 'cl-tohoku/bert-base-japanese' __lowerCAmelCase = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : List[str] ) -> Optional[int]: __lowerCAmelCase = 'cl-tohoku/bert-base-japanese' with self.assertLogs('transformers' , level='WARNING' ) as cm: BertTokenizer.from_pretrained(lowerCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( 'The tokenizer class you load from this checkpoint is not the same type as the class this function' ' is called from.' ) ) __lowerCAmelCase = 'bert-base-cased' with self.assertLogs('transformers' , level='WARNING' ) as cm: BertJapaneseTokenizer.from_pretrained(lowerCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( 'The tokenizer class you load from this checkpoint is not the same type as the class this function' ' is called from.' ) )
284
1
# This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny model through reduction of a normal pre-trained model, but keeping the # full vocab, merges file, and thus also resulting in a larger model due to a large vocab size. # This gives ~3MB in total for all files. # # If you want a 50 times smaller than this see `fsmt-make-super-tiny-model.py`, which is slightly more complicated # # # It will be used then as "stas/tiny-wmt19-en-de" # Build from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration _snake_case : Tuple = 'facebook/wmt19-en-de' _snake_case : List[str] = FSMTTokenizer.from_pretrained(mname) # get the correct vocab sizes, etc. from the master model _snake_case : Tuple = FSMTConfig.from_pretrained(mname) config.update( dict( d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) ) _snake_case : Optional[Any] = FSMTForConditionalGeneration(config) print(F"""num of params {tiny_model.num_parameters()}""") # Test _snake_case : int = tokenizer(['Making tiny model'], return_tensors='pt') _snake_case : str = tiny_model(**batch) print('test output:', len(outputs.logits[0])) # Save _snake_case : Dict = 'tiny-wmt19-en-de' tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F"""Generated {mname_tiny}""") # Upload # transformers-cli upload tiny-wmt19-en-de
284
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : List[Any] = { 'microsoft/beit-base-patch16-224-pt22k': ( 'https://huggingface.co/microsoft/beit-base-patch16-224-pt22k/resolve/main/config.json' ), # See all BEiT models at https://huggingface.co/models?filter=beit } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """beit""" def __init__( self : List[Any] , lowerCAmelCase_ : Tuple=8_1_9_2 , lowerCAmelCase_ : Optional[int]=7_6_8 , lowerCAmelCase_ : int=1_2 , lowerCAmelCase_ : Optional[int]=1_2 , lowerCAmelCase_ : Any=3_0_7_2 , lowerCAmelCase_ : Optional[int]="gelu" , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Any=0.02 , lowerCAmelCase_ : int=1e-12 , lowerCAmelCase_ : int=2_2_4 , lowerCAmelCase_ : str=1_6 , lowerCAmelCase_ : int=3 , lowerCAmelCase_ : Dict=False , lowerCAmelCase_ : int=False , lowerCAmelCase_ : List[Any]=False , lowerCAmelCase_ : int=False , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : Union[str, Any]=0.1 , lowerCAmelCase_ : List[str]=True , lowerCAmelCase_ : List[Any]=[3, 5, 7, 1_1] , lowerCAmelCase_ : Optional[Any]=[1, 2, 3, 6] , lowerCAmelCase_ : Tuple=True , lowerCAmelCase_ : Dict=0.4 , lowerCAmelCase_ : Tuple=2_5_6 , lowerCAmelCase_ : Any=1 , lowerCAmelCase_ : Any=False , lowerCAmelCase_ : Optional[int]=2_5_5 , **lowerCAmelCase_ : Any , ) -> Dict: super().__init__(**lowerCAmelCase_ ) __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = initializer_range __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = image_size __lowerCAmelCase = patch_size __lowerCAmelCase = num_channels __lowerCAmelCase = use_mask_token __lowerCAmelCase = use_absolute_position_embeddings __lowerCAmelCase = use_relative_position_bias __lowerCAmelCase = use_shared_relative_position_bias __lowerCAmelCase = layer_scale_init_value __lowerCAmelCase = drop_path_rate __lowerCAmelCase = use_mean_pooling # decode head attributes (semantic segmentation) __lowerCAmelCase = out_indices __lowerCAmelCase = pool_scales # auxiliary head attributes (semantic segmentation) __lowerCAmelCase = use_auxiliary_head __lowerCAmelCase = auxiliary_loss_weight __lowerCAmelCase = auxiliary_channels __lowerCAmelCase = auxiliary_num_convs __lowerCAmelCase = auxiliary_concat_input __lowerCAmelCase = semantic_loss_ignore_index class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = version.parse("""1.11""" ) @property def lowercase ( self : Tuple ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def lowercase ( self : Optional[Any] ) -> float: return 1e-4
284
1
_snake_case : List[str] = 65521 def a_ ( lowerCAmelCase_ : str ): __lowerCAmelCase = 1 __lowerCAmelCase = 0 for plain_chr in plain_text: __lowerCAmelCase = (a + ord(lowerCAmelCase_ )) % MOD_ADLER __lowerCAmelCase = (b + a) % MOD_ADLER return (b << 16) | a
284
def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : int ): return [sentence[i : i + ngram_size] for i in range(len(lowerCAmelCase_ ) - ngram_size + 1 )] if __name__ == "__main__": from doctest import testmod testmod()
284
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _snake_case : Tuple = { 'configuration_gpt_bigcode': ['GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GPTBigCodeConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST', 'GPTBigCodeForSequenceClassification', 'GPTBigCodeForTokenClassification', 'GPTBigCodeForCausalLM', 'GPTBigCodeModel', 'GPTBigCodePreTrainedModel', ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : Any = { 'google/pegasus-large': 'https://huggingface.co/google/pegasus-large/resolve/main/config.json', # See all PEGASUS models at https://huggingface.co/models?filter=pegasus } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """pegasus""" a_ = ["""past_key_values"""] a_ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[str] , lowerCAmelCase_ : Union[str, Any]=5_0_2_6_5 , lowerCAmelCase_ : Union[str, Any]=1_0_2_4 , lowerCAmelCase_ : Union[str, Any]=1_2 , lowerCAmelCase_ : Dict=4_0_9_6 , lowerCAmelCase_ : str=1_6 , lowerCAmelCase_ : List[Any]=1_2 , lowerCAmelCase_ : Union[str, Any]=4_0_9_6 , lowerCAmelCase_ : Union[str, Any]=1_6 , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Optional[int]=0.0 , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Any=True , lowerCAmelCase_ : Dict="gelu" , lowerCAmelCase_ : Optional[int]=1_0_2_4 , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : str=0.0 , lowerCAmelCase_ : Union[str, Any]=0.0 , lowerCAmelCase_ : Optional[int]=0.02 , lowerCAmelCase_ : Tuple=0 , lowerCAmelCase_ : Tuple=False , lowerCAmelCase_ : Union[str, Any]=0 , lowerCAmelCase_ : int=1 , lowerCAmelCase_ : Tuple=1 , **lowerCAmelCase_ : Tuple , ) -> List[str]: __lowerCAmelCase = vocab_size __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = d_model __lowerCAmelCase = encoder_ffn_dim __lowerCAmelCase = encoder_layers __lowerCAmelCase = encoder_attention_heads __lowerCAmelCase = decoder_ffn_dim __lowerCAmelCase = decoder_layers __lowerCAmelCase = decoder_attention_heads __lowerCAmelCase = dropout __lowerCAmelCase = attention_dropout __lowerCAmelCase = activation_dropout __lowerCAmelCase = activation_function __lowerCAmelCase = init_std __lowerCAmelCase = encoder_layerdrop __lowerCAmelCase = decoder_layerdrop __lowerCAmelCase = use_cache __lowerCAmelCase = encoder_layers __lowerCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , is_encoder_decoder=lowerCAmelCase_ , decoder_start_token_id=lowerCAmelCase_ , forced_eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ , ) @property def lowercase ( self : List[Any] ) -> int: return self.encoder_attention_heads @property def lowercase ( self : Optional[Any] ) -> int: return self.d_model
284
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType _snake_case : int = logging.get_logger(__name__) _snake_case : Any = { 'microsoft/deberta-v2-xlarge': 'https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json', 'microsoft/deberta-v2-xxlarge': 'https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json', 'microsoft/deberta-v2-xlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json' ), 'microsoft/deberta-v2-xxlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json' ), } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """deberta-v2""" def __init__( self : Optional[Any] , lowerCAmelCase_ : Tuple=1_2_8_1_0_0 , lowerCAmelCase_ : Any=1_5_3_6 , lowerCAmelCase_ : Optional[Any]=2_4 , lowerCAmelCase_ : str=2_4 , lowerCAmelCase_ : Dict=6_1_4_4 , lowerCAmelCase_ : List[Any]="gelu" , lowerCAmelCase_ : str=0.1 , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : List[Any]=5_1_2 , lowerCAmelCase_ : str=0 , lowerCAmelCase_ : List[Any]=0.02 , lowerCAmelCase_ : Any=1e-7 , lowerCAmelCase_ : List[str]=False , lowerCAmelCase_ : int=-1 , lowerCAmelCase_ : Any=0 , lowerCAmelCase_ : Optional[Any]=True , lowerCAmelCase_ : Any=None , lowerCAmelCase_ : Tuple=0 , lowerCAmelCase_ : Union[str, Any]="gelu" , **lowerCAmelCase_ : int , ) -> str: super().__init__(**lowerCAmelCase_ ) __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = type_vocab_size __lowerCAmelCase = initializer_range __lowerCAmelCase = relative_attention __lowerCAmelCase = max_relative_positions __lowerCAmelCase = pad_token_id __lowerCAmelCase = position_biased_input # Backwards compatibility if type(lowerCAmelCase_ ) == str: __lowerCAmelCase = [x.strip() for x in pos_att_type.lower().split('|' )] __lowerCAmelCase = pos_att_type __lowerCAmelCase = vocab_size __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = kwargs.get('pooler_hidden_size' , lowerCAmelCase_ ) __lowerCAmelCase = pooler_dropout __lowerCAmelCase = pooler_hidden_act class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" @property def lowercase ( self : Optional[Any] ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": __lowerCAmelCase = {0: 'batch', 1: 'choice', 2: 'sequence'} else: __lowerCAmelCase = {0: 'batch', 1: 'sequence'} if self._config.type_vocab_size > 0: return OrderedDict( [('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('token_type_ids', dynamic_axis)] ) else: return OrderedDict([('input_ids', dynamic_axis), ('attention_mask', dynamic_axis)] ) @property def lowercase ( self : List[str] ) -> int: return 1_2 def lowercase ( self : int , lowerCAmelCase_ : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , lowerCAmelCase_ : int = -1 , lowerCAmelCase_ : int = -1 , lowerCAmelCase_ : int = -1 , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : Optional["TensorType"] = None , lowerCAmelCase_ : int = 3 , lowerCAmelCase_ : int = 4_0 , lowerCAmelCase_ : int = 4_0 , lowerCAmelCase_ : "PreTrainedTokenizerBase" = None , ) -> Mapping[str, Any]: __lowerCAmelCase = super().generate_dummy_inputs(preprocessor=lowerCAmelCase_ , framework=lowerCAmelCase_ ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
284
def a_ ( lowerCAmelCase_ : int ): if p < 2: raise ValueError('p should not be less than 2!' ) elif p == 2: return True __lowerCAmelCase = 4 __lowerCAmelCase = (1 << p) - 1 for _ in range(p - 2 ): __lowerCAmelCase = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
284
1
import doctest import glob import importlib import inspect import os import re from contextlib import contextmanager from functools import wraps from unittest.mock import patch import numpy as np import pytest from absl.testing import parameterized import datasets from datasets import load_metric from .utils import for_all_test_methods, local, slow # mark all tests as integration _snake_case : List[Any] = pytest.mark.integration _snake_case : Dict = {'comet'} _snake_case : Union[str, Any] = importlib.util.find_spec('fairseq') is not None _snake_case : Optional[int] = {'code_eval'} _snake_case : int = os.name == 'nt' _snake_case : Optional[int] = {'bertscore', 'frugalscore', 'perplexity'} _snake_case : int = importlib.util.find_spec('transformers') is not None def a_ ( lowerCAmelCase_ : Optional[int] ): @wraps(lowerCAmelCase_ ) def wrapper(self : List[Any], lowerCAmelCase_ : Union[str, Any] ): if not _has_fairseq and metric_name in REQUIRE_FAIRSEQ: self.skipTest('"test requires Fairseq"' ) else: test_case(self, lowerCAmelCase_ ) return wrapper def a_ ( lowerCAmelCase_ : Optional[Any] ): @wraps(lowerCAmelCase_ ) def wrapper(self : List[str], lowerCAmelCase_ : Tuple ): if not _has_transformers and metric_name in REQUIRE_TRANSFORMERS: self.skipTest('"test requires transformers"' ) else: test_case(self, lowerCAmelCase_ ) return wrapper def a_ ( lowerCAmelCase_ : Dict ): @wraps(lowerCAmelCase_ ) def wrapper(self : int, lowerCAmelCase_ : Optional[int] ): if _on_windows and metric_name in UNSUPPORTED_ON_WINDOWS: self.skipTest('"test not supported on Windows"' ) else: test_case(self, lowerCAmelCase_ ) return wrapper def a_ ( ): __lowerCAmelCase = [metric_dir.split(os.sep )[-2] for metric_dir in glob.glob('./metrics/*/' )] return [{"testcase_name": x, "metric_name": x} for x in metrics if x != "gleu"] # gleu is unfinished @parameterized.named_parameters(get_local_metric_names() ) @for_all_test_methods( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) @local class _UpperCAmelCase ( parameterized.TestCase ): """simple docstring""" a_ = {} a_ = None @pytest.mark.filterwarnings('ignore:metric_module_factory is deprecated:FutureWarning' ) @pytest.mark.filterwarnings('ignore:load_metric is deprecated:FutureWarning' ) def lowercase ( self : Optional[Any] , lowerCAmelCase_ : str ) -> Optional[int]: __lowerCAmelCase = '[...]' __lowerCAmelCase = importlib.import_module( datasets.load.metric_module_factory(os.path.join('metrics' , lowerCAmelCase_ ) ).module_path ) __lowerCAmelCase = datasets.load.import_main_class(metric_module.__name__ , dataset=lowerCAmelCase_ ) # check parameters __lowerCAmelCase = inspect.signature(metric._compute ).parameters self.assertTrue(all(p.kind != p.VAR_KEYWORD for p in parameters.values() ) ) # no **kwargs # run doctest with self.patch_intensive_calls(lowerCAmelCase_ , metric_module.__name__ ): with self.use_local_metrics(): try: __lowerCAmelCase = doctest.testmod(lowerCAmelCase_ , verbose=lowerCAmelCase_ , raise_on_error=lowerCAmelCase_ ) except doctest.UnexpectedException as e: raise e.exc_info[1] # raise the exception that doctest caught self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @slow def lowercase ( self : Any , lowerCAmelCase_ : Tuple ) -> List[str]: __lowerCAmelCase = '[...]' __lowerCAmelCase = importlib.import_module( datasets.load.metric_module_factory(os.path.join('metrics' , lowerCAmelCase_ ) ).module_path ) # run doctest with self.use_local_metrics(): __lowerCAmelCase = doctest.testmod(lowerCAmelCase_ , verbose=lowerCAmelCase_ , raise_on_error=lowerCAmelCase_ ) self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @contextmanager def lowercase ( self : str , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : List[str] ) -> int: if metric_name in self.INTENSIVE_CALLS_PATCHER: with self.INTENSIVE_CALLS_PATCHER[metric_name](lowerCAmelCase_ ): yield else: yield @contextmanager def lowercase ( self : Union[str, Any] ) -> List[str]: def load_local_metric(lowerCAmelCase_ : Any , *lowerCAmelCase_ : str , **lowerCAmelCase_ : List[str] ): return load_metric(os.path.join('metrics' , lowerCAmelCase_ ) , *lowerCAmelCase_ , **lowerCAmelCase_ ) with patch('datasets.load_metric' ) as mock_load_metric: __lowerCAmelCase = load_local_metric yield @classmethod def lowercase ( cls : List[str] , lowerCAmelCase_ : List[Any] ) -> List[str]: def wrapper(lowerCAmelCase_ : str ): __lowerCAmelCase = contextmanager(lowerCAmelCase_ ) __lowerCAmelCase = patcher return patcher return wrapper @LocalMetricTest.register_intensive_calls_patcher('bleurt' ) def a_ ( lowerCAmelCase_ : Optional[int] ): import tensorflow.compat.va as tf from bleurt.score import Predictor tf.flags.DEFINE_string('sv', '', '' ) # handle pytest cli flags class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : List[str] , lowerCAmelCase_ : Optional[Any] ) -> Dict: assert len(input_dict['input_ids'] ) == 2 return np.array([1.03, 1.04] ) # mock predict_fn which is supposed to do a forward pass with a bleurt model with patch('bleurt.score._create_predictor' ) as mock_create_predictor: __lowerCAmelCase = MockedPredictor() yield @LocalMetricTest.register_intensive_calls_patcher('bertscore' ) def a_ ( lowerCAmelCase_ : Dict ): import torch def bert_cos_score_idf(lowerCAmelCase_ : Dict, lowerCAmelCase_ : List[Any], *lowerCAmelCase_ : List[str], **lowerCAmelCase_ : Union[str, Any] ): return torch.tensor([[1.0, 1.0, 1.0]] * len(lowerCAmelCase_ ) ) # mock get_model which is supposed to do download a bert model # mock bert_cos_score_idf which is supposed to do a forward pass with a bert model with patch('bert_score.scorer.get_model' ), patch( 'bert_score.scorer.bert_cos_score_idf' ) as mock_bert_cos_score_idf: __lowerCAmelCase = bert_cos_score_idf yield @LocalMetricTest.register_intensive_calls_patcher('comet' ) def a_ ( lowerCAmelCase_ : Dict ): def load_from_checkpoint(lowerCAmelCase_ : Dict ): class _UpperCAmelCase : """simple docstring""" def lowercase ( self : Any , lowerCAmelCase_ : Any , *lowerCAmelCase_ : Any , **lowerCAmelCase_ : List[Any] ) -> Optional[Any]: assert len(lowerCAmelCase_ ) == 2 __lowerCAmelCase = [0.19, 0.92] return scores, sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) return Model() # mock load_from_checkpoint which is supposed to do download a bert model # mock load_from_checkpoint which is supposed to do download a bert model with patch('comet.download_model' ) as mock_download_model: __lowerCAmelCase = None with patch('comet.load_from_checkpoint' ) as mock_load_from_checkpoint: __lowerCAmelCase = load_from_checkpoint yield def a_ ( ): __lowerCAmelCase = load_metric(os.path.join('metrics', 'seqeval' ) ) __lowerCAmelCase = 'ERROR' __lowerCAmelCase = F"""Scheme should be one of [IOB1, IOB2, IOE1, IOE2, IOBES, BILOU], got {wrong_scheme}""" with pytest.raises(lowerCAmelCase_, match=re.escape(lowerCAmelCase_ ) ): metric.compute(predictions=[], references=[], scheme=lowerCAmelCase_ )
284
from __future__ import annotations import math def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : int, lowerCAmelCase_ : bool, lowerCAmelCase_ : list[int], lowerCAmelCase_ : float ): if depth < 0: raise ValueError('Depth cannot be less than 0' ) if len(lowerCAmelCase_ ) == 0: raise ValueError('Scores cannot be empty' ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1, node_index * 2, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), minimax(depth + 1, node_index * 2 + 1, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), ) return min( minimax(depth + 1, node_index * 2, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), minimax(depth + 1, node_index * 2 + 1, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), ) def a_ ( ): __lowerCAmelCase = [90, 23, 6, 33, 21, 65, 123, 3_4423] __lowerCAmelCase = math.log(len(lowerCAmelCase_ ), 2 ) print('Optimal value : ', end='' ) print(minimax(0, 0, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
284
1
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def a_ ( lowerCAmelCase_ : List[str], lowerCAmelCase_ : Dict, lowerCAmelCase_ : Tuple=1024, lowerCAmelCase_ : Optional[Any]=1024, lowerCAmelCase_ : Tuple=False, **lowerCAmelCase_ : Union[str, Any] ): __lowerCAmelCase = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='train', **lowerCAmelCase_ ) __lowerCAmelCase = tok.pad_token_id def get_lens(lowerCAmelCase_ : Optional[Any] ): __lowerCAmelCase = tqdm( DataLoader(lowerCAmelCase_, batch_size=512, num_workers=8, shuffle=lowerCAmelCase_, collate_fn=ds.collate_fn ), desc=str(ds.len_file ), ) __lowerCAmelCase = [] for batch in dl: __lowerCAmelCase = batch['input_ids'].ne(lowerCAmelCase_ ).sum(1 ).tolist() __lowerCAmelCase = batch['labels'].ne(lowerCAmelCase_ ).sum(1 ).tolist() if consider_target: for src, tgt in zip(lowerCAmelCase_, lowerCAmelCase_ ): max_lens.append(max(lowerCAmelCase_, lowerCAmelCase_ ) ) else: max_lens.extend(lowerCAmelCase_ ) return max_lens __lowerCAmelCase = get_lens(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='val', **lowerCAmelCase_ ) __lowerCAmelCase = get_lens(lowerCAmelCase_ ) pickle_save(lowerCAmelCase_, train_ds.len_file ) pickle_save(lowerCAmelCase_, val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
284
def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('number must not be negative' ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
284
1
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Tuple = { 'configuration_autoformer': [ 'AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'AutoformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Dict = [ 'AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'AutoformerForPrediction', 'AutoformerModel', 'AutoformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_autoformer import ( AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_autoformer import ( AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, AutoformerForPrediction, AutoformerModel, AutoformerPreTrainedModel, ) else: import sys _snake_case : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
284
1
import warnings from ...utils import logging from .image_processing_segformer import SegformerImageProcessor _snake_case : Any = logging.get_logger(__name__) class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCAmelCase_ : Union[str, Any] , **lowerCAmelCase_ : List[str] ) -> None: warnings.warn( 'The class SegformerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use SegformerImageProcessor instead.' , lowerCAmelCase_ , ) super().__init__(*lowerCAmelCase_ , **lowerCAmelCase_ )
284
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def a_ ( lowerCAmelCase_ : List[str], lowerCAmelCase_ : Dict, lowerCAmelCase_ : Tuple=1024, lowerCAmelCase_ : Optional[Any]=1024, lowerCAmelCase_ : Tuple=False, **lowerCAmelCase_ : Union[str, Any] ): __lowerCAmelCase = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='train', **lowerCAmelCase_ ) __lowerCAmelCase = tok.pad_token_id def get_lens(lowerCAmelCase_ : Optional[Any] ): __lowerCAmelCase = tqdm( DataLoader(lowerCAmelCase_, batch_size=512, num_workers=8, shuffle=lowerCAmelCase_, collate_fn=ds.collate_fn ), desc=str(ds.len_file ), ) __lowerCAmelCase = [] for batch in dl: __lowerCAmelCase = batch['input_ids'].ne(lowerCAmelCase_ ).sum(1 ).tolist() __lowerCAmelCase = batch['labels'].ne(lowerCAmelCase_ ).sum(1 ).tolist() if consider_target: for src, tgt in zip(lowerCAmelCase_, lowerCAmelCase_ ): max_lens.append(max(lowerCAmelCase_, lowerCAmelCase_ ) ) else: max_lens.extend(lowerCAmelCase_ ) return max_lens __lowerCAmelCase = get_lens(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='val', **lowerCAmelCase_ ) __lowerCAmelCase = get_lens(lowerCAmelCase_ ) pickle_save(lowerCAmelCase_, train_ds.len_file ) pickle_save(lowerCAmelCase_, val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
284
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _snake_case : Tuple = logging.get_logger(__name__) _snake_case : str = { 'hustvl/yolos-small': 'https://huggingface.co/hustvl/yolos-small/resolve/main/config.json', # See all YOLOS models at https://huggingface.co/models?filter=yolos } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """yolos""" def __init__( self : Any , lowerCAmelCase_ : Optional[int]=7_6_8 , lowerCAmelCase_ : int=1_2 , lowerCAmelCase_ : Tuple=1_2 , lowerCAmelCase_ : Dict=3_0_7_2 , lowerCAmelCase_ : Optional[int]="gelu" , lowerCAmelCase_ : Union[str, Any]=0.0 , lowerCAmelCase_ : Tuple=0.0 , lowerCAmelCase_ : Optional[int]=0.02 , lowerCAmelCase_ : int=1e-12 , lowerCAmelCase_ : Any=[5_1_2, 8_6_4] , lowerCAmelCase_ : List[Any]=1_6 , lowerCAmelCase_ : int=3 , lowerCAmelCase_ : Union[str, Any]=True , lowerCAmelCase_ : Optional[Any]=1_0_0 , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : List[str]=False , lowerCAmelCase_ : Any=1 , lowerCAmelCase_ : str=5 , lowerCAmelCase_ : Optional[int]=2 , lowerCAmelCase_ : List[Any]=5 , lowerCAmelCase_ : List[Any]=2 , lowerCAmelCase_ : Any=0.1 , **lowerCAmelCase_ : Tuple , ) -> Any: super().__init__(**lowerCAmelCase_ ) __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = initializer_range __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = image_size __lowerCAmelCase = patch_size __lowerCAmelCase = num_channels __lowerCAmelCase = qkv_bias __lowerCAmelCase = num_detection_tokens __lowerCAmelCase = use_mid_position_embeddings __lowerCAmelCase = auxiliary_loss # Hungarian matcher __lowerCAmelCase = class_cost __lowerCAmelCase = bbox_cost __lowerCAmelCase = giou_cost # Loss coefficients __lowerCAmelCase = bbox_loss_coefficient __lowerCAmelCase = giou_loss_coefficient __lowerCAmelCase = eos_coefficient class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = version.parse("""1.11""" ) @property def lowercase ( self : Dict ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def lowercase ( self : Union[str, Any] ) -> float: return 1e-4 @property def lowercase ( self : Tuple ) -> int: return 1_2
284
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def a_ ( lowerCAmelCase_ : List[Any], lowerCAmelCase_ : str, lowerCAmelCase_ : Optional[int]=None, lowerCAmelCase_ : List[Any]=None ): if attention_mask is None: __lowerCAmelCase = tf.cast(tf.math.not_equal(lowerCAmelCase_, config.pad_token_id ), tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class _UpperCAmelCase : """simple docstring""" a_ = OPTConfig a_ = {} a_ = """gelu""" def __init__( self : Dict , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : List[str]=1_3 , lowerCAmelCase_ : Tuple=7 , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : Union[str, Any]=False , lowerCAmelCase_ : Any=9_9 , lowerCAmelCase_ : Any=1_6 , lowerCAmelCase_ : List[str]=2 , lowerCAmelCase_ : Dict=4 , lowerCAmelCase_ : str=4 , lowerCAmelCase_ : Any="gelu" , lowerCAmelCase_ : Optional[Any]=0.1 , lowerCAmelCase_ : List[Any]=0.1 , lowerCAmelCase_ : Tuple=2_0 , lowerCAmelCase_ : Optional[Any]=2 , lowerCAmelCase_ : Any=1 , lowerCAmelCase_ : List[Any]=0 , lowerCAmelCase_ : Optional[int]=1_6 , lowerCAmelCase_ : Dict=1_6 , ) -> int: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = eos_token_id __lowerCAmelCase = pad_token_id __lowerCAmelCase = bos_token_id __lowerCAmelCase = embed_dim __lowerCAmelCase = word_embed_proj_dim __lowerCAmelCase = False def lowercase ( self : List[str] ) -> Optional[Any]: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowerCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowerCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowerCAmelCase = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=lowerCAmelCase_ , **self.config_updates , ) __lowerCAmelCase = prepare_opt_inputs_dict(lowerCAmelCase_ , lowerCAmelCase_ ) return config, inputs_dict def lowercase ( self : Any , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Dict ) -> List[str]: __lowerCAmelCase = TFOPTModel(config=lowerCAmelCase_ ) __lowerCAmelCase = inputs_dict['input_ids'] __lowerCAmelCase = input_ids[:1, :] __lowerCAmelCase = inputs_dict['attention_mask'][:1, :] __lowerCAmelCase = 1 # first forward pass __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , use_cache=lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowerCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowerCAmelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowerCAmelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowerCAmelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , past_key_values=lowerCAmelCase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowerCAmelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowerCAmelCase = output_from_no_past[:, -3:, random_slice_idx] __lowerCAmelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowerCAmelCase_ , lowerCAmelCase_ , rtol=1e-3 ) @require_tf class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () a_ = (TFOPTForCausalLM,) if is_tf_available() else () a_ = ( {"""feature-extraction""": TFOPTModel, """text-generation""": TFOPTForCausalLM} if is_tf_available() else {} ) a_ = False a_ = False a_ = False a_ = 10 def lowercase ( self : List[str] ) -> Optional[int]: __lowerCAmelCase = TFOPTModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=lowerCAmelCase_ ) def lowercase ( self : Tuple ) -> Tuple: self.config_tester.run_common_tests() def lowercase ( self : Tuple ) -> Optional[Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCAmelCase_ ) def lowercase ( self : Union[str, Any] ) -> Dict: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(lowerCAmelCase_ : str , lowerCAmelCase_ : Union[str, Any] ): if hasattr(lowerCAmelCase_ , 'weight' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(lowerCAmelCase_ , 'weight' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 1_0, config.vocab_size + 1_0]: # build the embeddings __lowerCAmelCase = model_class(config=lowerCAmelCase_ ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_input_embeddings() ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(lowerCAmelCase_ ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_input_embeddings() ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. __lowerCAmelCase = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , lowerCAmelCase_ ) # check that weights remain the same after resizing __lowerCAmelCase = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCAmelCase = False self.assertTrue(lowerCAmelCase_ ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , lowerCAmelCase_ ) __lowerCAmelCase = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCAmelCase = False self.assertTrue(lowerCAmelCase_ ) def a_ ( lowerCAmelCase_ : Union[str, Any] ): return tf.constant(lowerCAmelCase_, dtype=tf.intaa ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" a_ = 99 def lowercase ( self : Optional[int] ) -> Any: __lowerCAmelCase = tf.ones((4, 1) , dtype=tf.intaa ) * 2 __lowerCAmelCase = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) __lowerCAmelCase = input_ids.shape[0] __lowerCAmelCase = OPTConfig( vocab_size=self.vocab_size , hidden_size=2_4 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=3_2 , max_position_embeddings=4_8 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def lowercase ( self : str ) -> List[str]: __lowerCAmelCase = TFOPTModel.from_pretrained('facebook/opt-350m' ) __lowerCAmelCase = _long_tensor([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] ) __lowerCAmelCase = tf.not_equal(lowerCAmelCase_ , model.config.pad_token_id ) with tf.GradientTape(): __lowerCAmelCase = model(input_ids=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ ).last_hidden_state __lowerCAmelCase = (1, 1_1, 5_1_2) self.assertEqual(output.shape , lowerCAmelCase_ ) __lowerCAmelCase = tf.constant( [[-0.28_73, -1.92_18, -0.30_33], [-1.27_10, -0.13_38, -0.19_02], [0.40_95, 0.12_14, -1.31_21]] ) self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase_ , atol=4e-3 ) ) __lowerCAmelCase = tf.function(lowerCAmelCase_ , jit_compile=lowerCAmelCase_ ) __lowerCAmelCase = xla_generate(lowerCAmelCase_ , lowerCAmelCase_ )[0] self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase_ , atol=4e-2 ) ) @require_tf @slow class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> Dict: super().setUp() __lowerCAmelCase = 'facebook/opt-350m' def lowercase ( self : Dict ) -> Any: __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(self.path_model ) __lowerCAmelCase = GPTaTokenizer.from_pretrained(self.path_model ) __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of', 'Paris is the capital of France and', 'Computers and mobile phones have taken', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' , padding=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) __lowerCAmelCase = tf.constant( [ [1.38_51, -13.89_23, -10.52_29, -10.75_33, -0.23_09, -10.23_84, -0.53_65, -9.09_47, -5.16_70], [-4.70_73, -10.62_76, -3.94_15, -21.52_42, -0.28_22, -0.28_22, -0.28_22, -0.28_22, -0.28_22], [0.62_47, -3.42_29, -8.91_79, -1.42_97, -14.16_50, 1.41_46, -9.02_18, -0.27_03, -0.27_03], [6.47_83, -1.99_13, -10.79_26, -2.33_36, 1.50_92, -0.99_74, -6.82_13, 1.34_77, 1.34_77], ] ) self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1e-4 ) ) __lowerCAmelCase = tf.function(lowerCAmelCase_ , jit_compile=lowerCAmelCase_ ) __lowerCAmelCase = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1e-4 ) ) @require_tf @slow class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @property def lowercase ( self : Optional[int] ) -> int: return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def lowercase ( self : int ) -> str: __lowerCAmelCase = 'facebook/opt-125m' __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of New York, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] __lowerCAmelCase = [] __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) for prompt in self.prompts: __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(lowerCAmelCase_ , max_length=1_0 ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Optional[Any] ) -> str: __lowerCAmelCase = 'facebook/opt-350m' __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = 'left' # use different length sentences to test batching __lowerCAmelCase = [ 'Hello, my dog is a little', 'Today, I', ] __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' , padding=lowerCAmelCase_ ) __lowerCAmelCase = inputs['input_ids'] __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ , attention_mask=inputs['attention_mask'] ) __lowerCAmelCase = tokenizer(sentences[0] , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ ) __lowerCAmelCase = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['attention_mask'][-1] , tf.intaa ) ) __lowerCAmelCase = tokenizer(sentences[1] , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ , max_length=model.config.max_length - num_paddings ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = [ 'Hello, my dog is a little bit of a dork.\nI\'m a little bit', 'Today, I was in the middle of a conversation with a friend about the', ] self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , [non_padded_sentence, padded_sentence] ) def lowercase ( self : List[Any] ) -> List[Any]: __lowerCAmelCase = 'facebook/opt-350m' __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of San Francisco, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] __lowerCAmelCase = [] __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) for prompt in self.prompts: __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(lowerCAmelCase_ , max_length=1_0 ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ )
284
1
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available from .timesteps import ( fastaa_timesteps, smartaa_timesteps, smartaa_timesteps, smartaaa_timesteps, smartaaa_timesteps, superaa_timesteps, superaa_timesteps, superaaa_timesteps, ) @dataclass class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = 42 a_ = 42 a_ = 42 try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_if import IFPipeline from .pipeline_if_imgaimg import IFImgaImgPipeline from .pipeline_if_imgaimg_superresolution import IFImgaImgSuperResolutionPipeline from .pipeline_if_inpainting import IFInpaintingPipeline from .pipeline_if_inpainting_superresolution import IFInpaintingSuperResolutionPipeline from .pipeline_if_superresolution import IFSuperResolutionPipeline from .safety_checker import IFSafetyChecker from .watermark import IFWatermarker
284
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { 'configuration_layoutlmv3': [ 'LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv3Config', 'LayoutLMv3OnnxConfig', ], 'processing_layoutlmv3': ['LayoutLMv3Processor'], 'tokenization_layoutlmv3': ['LayoutLMv3Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Union[str, Any] = ['LayoutLMv3TokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[str] = [ 'LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv3ForQuestionAnswering', 'LayoutLMv3ForSequenceClassification', 'LayoutLMv3ForTokenClassification', 'LayoutLMv3Model', 'LayoutLMv3PreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[Any] = [ 'TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLayoutLMv3ForQuestionAnswering', 'TFLayoutLMv3ForSequenceClassification', 'TFLayoutLMv3ForTokenClassification', 'TFLayoutLMv3Model', 'TFLayoutLMv3PreTrainedModel', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Tuple = ['LayoutLMv3FeatureExtractor'] _snake_case : str = ['LayoutLMv3ImageProcessor'] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys _snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
1
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, logging if is_torch_available(): import torch _snake_case : Any = logging.get_logger(__name__) class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = ["""pixel_values"""] def __init__( self : List[str] , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Optional[Dict[str, int]] = None , lowerCAmelCase_ : PILImageResampling = PILImageResampling.BILINEAR , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Dict[str, int] = None , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Union[int, float] = 1 / 2_5_5 , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Optional[Union[float, List[float]]] = None , lowerCAmelCase_ : Optional[Union[float, List[float]]] = None , **lowerCAmelCase_ : str , ) -> None: super().__init__(**lowerCAmelCase_ ) __lowerCAmelCase = size if size is not None else {'shortest_edge': 2_5_6} __lowerCAmelCase = get_size_dict(lowerCAmelCase_ , default_to_square=lowerCAmelCase_ ) __lowerCAmelCase = crop_size if crop_size is not None else {'height': 2_2_4, 'width': 2_2_4} __lowerCAmelCase = get_size_dict(lowerCAmelCase_ , param_name='crop_size' ) __lowerCAmelCase = do_resize __lowerCAmelCase = size __lowerCAmelCase = resample __lowerCAmelCase = do_center_crop __lowerCAmelCase = crop_size __lowerCAmelCase = do_rescale __lowerCAmelCase = rescale_factor __lowerCAmelCase = do_normalize __lowerCAmelCase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __lowerCAmelCase = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : np.ndarray , lowerCAmelCase_ : Dict[str, int] , lowerCAmelCase_ : PILImageResampling = PILImageResampling.BICUBIC , lowerCAmelCase_ : Optional[Union[str, ChannelDimension]] = None , **lowerCAmelCase_ : Any , ) -> np.ndarray: __lowerCAmelCase = get_size_dict(lowerCAmelCase_ , default_to_square=lowerCAmelCase_ ) if "shortest_edge" not in size: raise ValueError(f"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""" ) __lowerCAmelCase = get_resize_output_image_size(lowerCAmelCase_ , size=size['shortest_edge'] , default_to_square=lowerCAmelCase_ ) return resize(lowerCAmelCase_ , size=lowerCAmelCase_ , resample=lowerCAmelCase_ , data_format=lowerCAmelCase_ , **lowerCAmelCase_ ) def lowercase ( self : Optional[Any] , lowerCAmelCase_ : np.ndarray , lowerCAmelCase_ : Dict[str, int] , lowerCAmelCase_ : Optional[Union[str, ChannelDimension]] = None , **lowerCAmelCase_ : Optional[Any] , ) -> np.ndarray: __lowerCAmelCase = get_size_dict(lowerCAmelCase_ ) if "height" not in size or "width" not in size: raise ValueError(f"""The `size` parameter must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(lowerCAmelCase_ , size=(size['height'], size['width']) , data_format=lowerCAmelCase_ , **lowerCAmelCase_ ) def lowercase ( self : List[str] , lowerCAmelCase_ : np.ndarray , lowerCAmelCase_ : float , lowerCAmelCase_ : Optional[Union[str, ChannelDimension]] = None , **lowerCAmelCase_ : Optional[Any] ) -> np.ndarray: return rescale(lowerCAmelCase_ , scale=lowerCAmelCase_ , data_format=lowerCAmelCase_ , **lowerCAmelCase_ ) def lowercase ( self : Any , lowerCAmelCase_ : np.ndarray , lowerCAmelCase_ : Union[float, List[float]] , lowerCAmelCase_ : Union[float, List[float]] , lowerCAmelCase_ : Optional[Union[str, ChannelDimension]] = None , **lowerCAmelCase_ : Optional[Any] , ) -> np.ndarray: return normalize(lowerCAmelCase_ , mean=lowerCAmelCase_ , std=lowerCAmelCase_ , data_format=lowerCAmelCase_ , **lowerCAmelCase_ ) def lowercase ( self : Tuple , lowerCAmelCase_ : ImageInput , lowerCAmelCase_ : Optional[bool] = None , lowerCAmelCase_ : Dict[str, int] = None , lowerCAmelCase_ : PILImageResampling = None , lowerCAmelCase_ : bool = None , lowerCAmelCase_ : Dict[str, int] = None , lowerCAmelCase_ : Optional[bool] = None , lowerCAmelCase_ : Optional[float] = None , lowerCAmelCase_ : Optional[bool] = None , lowerCAmelCase_ : Optional[Union[float, List[float]]] = None , lowerCAmelCase_ : Optional[Union[float, List[float]]] = None , lowerCAmelCase_ : Optional[Union[str, TensorType]] = None , lowerCAmelCase_ : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCAmelCase_ : int , ) -> List[str]: __lowerCAmelCase = do_resize if do_resize is not None else self.do_resize __lowerCAmelCase = size if size is not None else self.size __lowerCAmelCase = get_size_dict(lowerCAmelCase_ , default_to_square=lowerCAmelCase_ ) __lowerCAmelCase = resample if resample is not None else self.resample __lowerCAmelCase = do_center_crop if do_center_crop is not None else self.do_center_crop __lowerCAmelCase = crop_size if crop_size is not None else self.crop_size __lowerCAmelCase = get_size_dict(lowerCAmelCase_ , param_name='crop_size' ) __lowerCAmelCase = do_rescale if do_rescale is not None else self.do_rescale __lowerCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCAmelCase = do_normalize if do_normalize is not None else self.do_normalize __lowerCAmelCase = image_mean if image_mean is not None else self.image_mean __lowerCAmelCase = image_std if image_std is not None else self.image_std __lowerCAmelCase = make_list_of_images(lowerCAmelCase_ ) if not valid_images(lowerCAmelCase_ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. __lowerCAmelCase = [to_numpy_array(lowerCAmelCase_ ) for image in images] if do_resize: __lowerCAmelCase = [self.resize(image=lowerCAmelCase_ , size=lowerCAmelCase_ , resample=lowerCAmelCase_ ) for image in images] if do_center_crop: __lowerCAmelCase = [self.center_crop(image=lowerCAmelCase_ , size=lowerCAmelCase_ ) for image in images] if do_rescale: __lowerCAmelCase = [self.rescale(image=lowerCAmelCase_ , scale=lowerCAmelCase_ ) for image in images] if do_normalize: __lowerCAmelCase = [self.normalize(image=lowerCAmelCase_ , mean=lowerCAmelCase_ , std=lowerCAmelCase_ ) for image in images] __lowerCAmelCase = [to_channel_dimension_format(lowerCAmelCase_ , lowerCAmelCase_ ) for image in images] __lowerCAmelCase = {'pixel_values': images} return BatchFeature(data=lowerCAmelCase_ , tensor_type=lowerCAmelCase_ ) def lowercase ( self : Optional[int] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : List[Tuple] = None ) -> Union[str, Any]: __lowerCAmelCase = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(lowerCAmelCase_ ) != len(lowerCAmelCase_ ): raise ValueError( 'Make sure that you pass in as many target sizes as the batch dimension of the logits' ) if is_torch_tensor(lowerCAmelCase_ ): __lowerCAmelCase = target_sizes.numpy() __lowerCAmelCase = [] for idx in range(len(lowerCAmelCase_ ) ): __lowerCAmelCase = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='bilinear' , align_corners=lowerCAmelCase_ ) __lowerCAmelCase = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(lowerCAmelCase_ ) else: __lowerCAmelCase = logits.argmax(dim=1 ) __lowerCAmelCase = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
284
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case : Dict = {'configuration_mra': ['MRA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MraConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'MRA_PRETRAINED_MODEL_ARCHIVE_LIST', 'MraForMaskedLM', 'MraForMultipleChoice', 'MraForQuestionAnswering', 'MraForSequenceClassification', 'MraForTokenClassification', 'MraLayer', 'MraModel', 'MraPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys _snake_case : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure)
284
1
def a_ ( ): __lowerCAmelCase = 0 for i in range(1, 1001 ): total += i**i return str(lowerCAmelCase_ )[-10:] if __name__ == "__main__": print(solution())
284
from __future__ import annotations import string from itertools import cycle, product from pathlib import Path _snake_case : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) _snake_case : list[int] = [ord(letter) for letter in string.ascii_lowercase] _snake_case : set[int] = {ord(char) for char in VALID_CHARS} _snake_case : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def a_ ( lowerCAmelCase_ : list[int], lowerCAmelCase_ : tuple[int, ...] ): __lowerCAmelCase = "" __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 for keychar, cipherchar in zip(cycle(lowerCAmelCase_ ), lowerCAmelCase_ ): __lowerCAmelCase = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(lowerCAmelCase_ ) return decoded def a_ ( lowerCAmelCase_ : list[int] ): __lowerCAmelCase = [] for key in product(lowerCAmelCase_, repeat=3 ): __lowerCAmelCase = try_key(lowerCAmelCase_, lowerCAmelCase_ ) if encoded is not None: possibles.append(lowerCAmelCase_ ) return possibles def a_ ( lowerCAmelCase_ : list[str], lowerCAmelCase_ : str ): return [possible for possible in possibles if common_word in possible.lower()] def a_ ( lowerCAmelCase_ : str = "p059_cipher.txt" ): __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = Path(lowerCAmelCase_ ).parent.joinpath(lowerCAmelCase_ ).read_text(encoding='utf-8' ) __lowerCAmelCase = [int(lowerCAmelCase_ ) for number in data.strip().split(',' )] __lowerCAmelCase = filter_valid_chars(lowerCAmelCase_ ) for common_word in COMMON_WORDS: __lowerCAmelCase = filter_common_word(lowerCAmelCase_, lowerCAmelCase_ ) if len(lowerCAmelCase_ ) == 1: break __lowerCAmelCase = possibles[0] return sum(ord(lowerCAmelCase_ ) for char in decoded_text ) if __name__ == "__main__": print(F"""{solution() = }""")
284
1
import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging _snake_case : Tuple = logging.get_logger(__name__) def a_ ( lowerCAmelCase_ : Dict, lowerCAmelCase_ : Dict ): __lowerCAmelCase = set() __lowerCAmelCase = [] def parse_line(lowerCAmelCase_ : Optional[Any] ): for line in fp: if isinstance(lowerCAmelCase_, lowerCAmelCase_ ): __lowerCAmelCase = line.decode('UTF-8' ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(' ' ): # process a single warning and move it to `selected_warnings`. if len(lowerCAmelCase_ ) > 0: __lowerCAmelCase = '\n'.join(lowerCAmelCase_ ) # Only keep the warnings specified in `targets` if any(F""": {x}: """ in warning for x in targets ): selected_warnings.add(lowerCAmelCase_ ) buffer.clear() continue else: __lowerCAmelCase = line.strip() buffer.append(lowerCAmelCase_ ) if from_gh: for filename in os.listdir(lowerCAmelCase_ ): __lowerCAmelCase = os.path.join(lowerCAmelCase_, lowerCAmelCase_ ) if not os.path.isdir(lowerCAmelCase_ ): # read the file if filename != "warnings.txt": continue with open(lowerCAmelCase_ ) as fp: parse_line(lowerCAmelCase_ ) else: try: with zipfile.ZipFile(lowerCAmelCase_ ) as z: for filename in z.namelist(): if not os.path.isdir(lowerCAmelCase_ ): # read the file if filename != "warnings.txt": continue with z.open(lowerCAmelCase_ ) as fp: parse_line(lowerCAmelCase_ ) except Exception: logger.warning( F"""{artifact_path} is either an invalid zip file or something else wrong. This file is skipped.""" ) return selected_warnings def a_ ( lowerCAmelCase_ : Tuple, lowerCAmelCase_ : Dict ): __lowerCAmelCase = set() __lowerCAmelCase = [os.path.join(lowerCAmelCase_, lowerCAmelCase_ ) for p in os.listdir(lowerCAmelCase_ ) if (p.endswith('.zip' ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(lowerCAmelCase_, lowerCAmelCase_ ) ) return selected_warnings if __name__ == "__main__": def a_ ( lowerCAmelCase_ : Dict ): return values.split(',' ) _snake_case : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') parser.add_argument( '--output_dir', type=str, required=True, help='Where to store the downloaded artifacts and other result files.', ) parser.add_argument('--token', default=None, type=str, help='A token that has actions:read permission.') # optional parameters parser.add_argument( '--targets', default='DeprecationWarning,UserWarning,FutureWarning', type=list_str, help='Comma-separated list of target warning(s) which we want to extract.', ) parser.add_argument( '--from_gh', action='store_true', help='If running from a GitHub action workflow and collecting warnings from its artifacts.', ) _snake_case : int = parser.parse_args() _snake_case : List[Any] = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links _snake_case : Dict = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, 'artifacts.json'), 'w', encoding='UTF-8') as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print('=' * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts _snake_case : Optional[Any] = extract_warnings(args.output_dir, args.targets) _snake_case : int = sorted(selected_warnings) with open(os.path.join(args.output_dir, 'selected_warnings.json'), 'w', encoding='UTF-8') as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
284
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem _snake_case : Tuple = importlib.util.find_spec('s3fs') is not None if _has_safs: from .safilesystem import SaFileSystem # noqa: F401 _snake_case : List[compression.BaseCompressedFileFileSystem] = [ compression.BzaFileSystem, compression.GzipFileSystem, compression.LzaFileSystem, compression.XzFileSystem, compression.ZstdFileSystem, ] # Register custom filesystems for fs_class in COMPRESSION_FILESYSTEMS + [HfFileSystem]: if fs_class.protocol in fsspec.registry and fsspec.registry[fs_class.protocol] is not fs_class: warnings.warn(F"""A filesystem protocol was already set for {fs_class.protocol} and will be overwritten.""") fsspec.register_implementation(fs_class.protocol, fs_class, clobber=True) def a_ ( lowerCAmelCase_ : str ): if "://" in dataset_path: __lowerCAmelCase = dataset_path.split('://' )[1] return dataset_path def a_ ( lowerCAmelCase_ : fsspec.AbstractFileSystem ): if fs is not None and fs.protocol != "file": return True else: return False def a_ ( lowerCAmelCase_ : fsspec.AbstractFileSystem, lowerCAmelCase_ : str, lowerCAmelCase_ : str ): __lowerCAmelCase = not is_remote_filesystem(lowerCAmelCase_ ) if is_local: # LocalFileSystem.mv does copy + rm, it is more efficient to simply move a local directory shutil.move(fs._strip_protocol(lowerCAmelCase_ ), fs._strip_protocol(lowerCAmelCase_ ) ) else: fs.mv(lowerCAmelCase_, lowerCAmelCase_, recursive=lowerCAmelCase_ ) def a_ ( ): if hasattr(fsspec.asyn, 'reset_lock' ): # for future fsspec>2022.05.0 fsspec.asyn.reset_lock() else: __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = threading.Lock()
284
1
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def a_ ( lowerCAmelCase_ : BertModel, lowerCAmelCase_ : str, lowerCAmelCase_ : str ): __lowerCAmelCase = ('dense.weight', 'attention.self.query', 'attention.self.key', 'attention.self.value') __lowerCAmelCase = ( ('layer.', 'layer_'), ('word_embeddings.weight', 'word_embeddings'), ('position_embeddings.weight', 'position_embeddings'), ('token_type_embeddings.weight', 'token_type_embeddings'), ('.', '/'), ('LayerNorm/weight', 'LayerNorm/gamma'), ('LayerNorm/bias', 'LayerNorm/beta'), ('weight', 'kernel'), ) if not os.path.isdir(lowerCAmelCase_ ): os.makedirs(lowerCAmelCase_ ) __lowerCAmelCase = model.state_dict() def to_tf_var_name(lowerCAmelCase_ : str ): for patt, repl in iter(lowerCAmelCase_ ): __lowerCAmelCase = name.replace(lowerCAmelCase_, lowerCAmelCase_ ) return F"""bert/{name}""" def create_tf_var(lowerCAmelCase_ : np.ndarray, lowerCAmelCase_ : str, lowerCAmelCase_ : tf.Session ): __lowerCAmelCase = tf.dtypes.as_dtype(tensor.dtype ) __lowerCAmelCase = tf.get_variable(dtype=lowerCAmelCase_, shape=tensor.shape, name=lowerCAmelCase_, initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(lowerCAmelCase_ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: __lowerCAmelCase = to_tf_var_name(lowerCAmelCase_ ) __lowerCAmelCase = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): __lowerCAmelCase = torch_tensor.T __lowerCAmelCase = create_tf_var(tensor=lowerCAmelCase_, name=lowerCAmelCase_, session=lowerCAmelCase_ ) tf.keras.backend.set_value(lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = session.run(lowerCAmelCase_ ) print(F"""Successfully created {tf_name}: {np.allclose(lowerCAmelCase_, lowerCAmelCase_ )}""" ) __lowerCAmelCase = tf.train.Saver(tf.trainable_variables() ) saver.save(lowerCAmelCase_, os.path.join(lowerCAmelCase_, model_name.replace('-', '_' ) + '.ckpt' ) ) def a_ ( lowerCAmelCase_ : str=None ): __lowerCAmelCase = argparse.ArgumentParser() parser.add_argument('--model_name', type=lowerCAmelCase_, required=lowerCAmelCase_, help='model name e.g. bert-base-uncased' ) parser.add_argument( '--cache_dir', type=lowerCAmelCase_, default=lowerCAmelCase_, required=lowerCAmelCase_, help='Directory containing pytorch model' ) parser.add_argument('--pytorch_model_path', type=lowerCAmelCase_, required=lowerCAmelCase_, help='/path/to/<pytorch-model-name>.bin' ) parser.add_argument('--tf_cache_dir', type=lowerCAmelCase_, required=lowerCAmelCase_, help='Directory in which to save tensorflow model' ) __lowerCAmelCase = parser.parse_args(lowerCAmelCase_ ) __lowerCAmelCase = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name, state_dict=torch.load(args.pytorch_model_path ), cache_dir=args.cache_dir, ) convert_pytorch_checkpoint_to_tf(model=lowerCAmelCase_, ckpt_dir=args.tf_cache_dir, model_name=args.model_name ) if __name__ == "__main__": main()
284
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def a_ ( lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = filter(lambda lowerCAmelCase_ : p.requires_grad, model.parameters() ) __lowerCAmelCase = sum([np.prod(p.size() ) for p in model_parameters] ) return params _snake_case : Dict = logging.getLogger(__name__) def a_ ( lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Optional[int] ): if metric == "rouge2": __lowerCAmelCase = '{val_avg_rouge2:.4f}-{step_count}' elif metric == "bleu": __lowerCAmelCase = '{val_avg_bleu:.4f}-{step_count}' elif metric == "em": __lowerCAmelCase = '{val_avg_em:.4f}-{step_count}' else: raise NotImplementedError( F"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" ' function.' ) __lowerCAmelCase = ModelCheckpoint( dirpath=lowerCAmelCase_, filename=lowerCAmelCase_, monitor=F"""val_{metric}""", mode='max', save_top_k=3, every_n_epochs=1, ) return checkpoint_callback def a_ ( lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : Any ): return EarlyStopping( monitor=F"""val_{metric}""", mode='min' if 'loss' in metric else 'max', patience=lowerCAmelCase_, verbose=lowerCAmelCase_, ) class _UpperCAmelCase ( pl.Callback ): """simple docstring""" def lowercase ( self : Tuple , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : int ) -> Any: __lowerCAmelCase = {f"""lr_group_{i}""": param['lr'] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Optional[int] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule , lowerCAmelCase_ : str , lowerCAmelCase_ : List[Any]=True ) -> None: logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) __lowerCAmelCase = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['log', 'progress_bar', 'preds']} ) # Log results __lowerCAmelCase = Path(pl_module.hparams.output_dir ) if type_path == "test": __lowerCAmelCase = od / 'test_results.txt' __lowerCAmelCase = od / 'test_generations.txt' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __lowerCAmelCase = od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" __lowerCAmelCase = od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) generations_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'a+' ) as writer: for key in sorted(lowerCAmelCase_ ): if key in ["log", "progress_bar", "preds"]: continue __lowerCAmelCase = metrics[key] if isinstance(lowerCAmelCase_ , torch.Tensor ): __lowerCAmelCase = val.item() __lowerCAmelCase = f"""{key}: {val:.6f}\n""" writer.write(lowerCAmelCase_ ) if not save_generations: return if "preds" in metrics: __lowerCAmelCase = '\n'.join(metrics['preds'] ) generations_file.open('w+' ).write(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[str] ) -> Dict: try: __lowerCAmelCase = pl_module.model.model.num_parameters() except AttributeError: __lowerCAmelCase = pl_module.model.num_parameters() __lowerCAmelCase = count_trainable_parameters(lowerCAmelCase_ ) # mp stands for million parameters trainer.logger.log_metrics({'n_params': npars, 'mp': npars / 1e6, 'grad_mp': n_trainable_pars / 1e6} ) @rank_zero_only def lowercase ( self : int , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule ) -> Any: save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(lowerCAmelCase_ , lowerCAmelCase_ , 'test' ) @rank_zero_only def lowercase ( self : List[Any] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : Any ) -> int: save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
284
1
from __future__ import annotations _snake_case : Any = { 'A': ['B', 'C', 'E'], 'B': ['A', 'D', 'E'], 'C': ['A', 'F', 'G'], 'D': ['B'], 'E': ['A', 'B', 'D'], 'F': ['C'], 'G': ['C'], } class _UpperCAmelCase : """simple docstring""" def __init__( self : Optional[Any] , lowerCAmelCase_ : dict[str, list[str]] , lowerCAmelCase_ : str ) -> None: __lowerCAmelCase = graph # mapping node to its parent in resulting breadth first tree __lowerCAmelCase = {} __lowerCAmelCase = source_vertex def lowercase ( self : Optional[Any] ) -> None: __lowerCAmelCase = {self.source_vertex} __lowerCAmelCase = None __lowerCAmelCase = [self.source_vertex] # first in first out queue while queue: __lowerCAmelCase = queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(lowerCAmelCase_ ) __lowerCAmelCase = vertex queue.append(lowerCAmelCase_ ) def lowercase ( self : int , lowerCAmelCase_ : str ) -> str: if target_vertex == self.source_vertex: return self.source_vertex __lowerCAmelCase = self.parent.get(lowerCAmelCase_ ) if target_vertex_parent is None: __lowerCAmelCase = ( f"""No path from vertex: {self.source_vertex} to vertex: {target_vertex}""" ) raise ValueError(lowerCAmelCase_ ) return self.shortest_path(lowerCAmelCase_ ) + f"""->{target_vertex}""" if __name__ == "__main__": _snake_case : Dict = Graph(graph, 'G') g.breath_first_search() print(g.shortest_path('D')) print(g.shortest_path('G')) print(g.shortest_path('Foo'))
284
import re from filelock import FileLock try: import nltk _snake_case : Any = True except (ImportError, ModuleNotFoundError): _snake_case : Union[str, Any] = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def a_ ( lowerCAmelCase_ : str ): re.sub('<n>', '', lowerCAmelCase_ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(lowerCAmelCase_ ) )
284
1
def a_ ( lowerCAmelCase_ : list[int], lowerCAmelCase_ : list[int] ): # Check if the input is valid if not len(lowerCAmelCase_ ) == len(lowerCAmelCase_ ) == 3: raise ValueError('Please enter a valid equation.' ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError('Both a & b of two equations can\'t be zero.' ) # Extract the coefficients __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = equationa __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = equationa # Calculate the determinants of the matrices __lowerCAmelCase = aa * ba - aa * ba __lowerCAmelCase = ca * ba - ca * ba __lowerCAmelCase = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError('Infinite solutions. (Consistent system)' ) else: raise ValueError('No solution. (Inconsistent system)' ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: __lowerCAmelCase = determinant_x / determinant __lowerCAmelCase = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
284
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case : List[Any] = { 'configuration_biogpt': ['BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BioGptConfig'], 'tokenization_biogpt': ['BioGptTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BioGptForCausalLM', 'BioGptForTokenClassification', 'BioGptForSequenceClassification', 'BioGptModel', 'BioGptPreTrainedModel', ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys _snake_case : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
1
import os import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers.models.realm.configuration_realm import RealmConfig from transformers.models.realm.retrieval_realm import _REALM_BLOCK_RECORDS_FILENAME, RealmRetriever from transformers.models.realm.tokenization_realm import VOCAB_FILES_NAMES, RealmTokenizer class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : int ) -> Union[str, Any]: __lowerCAmelCase = tempfile.mkdtemp() __lowerCAmelCase = 5 # Realm tok __lowerCAmelCase = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'test', 'question', 'this', 'is', 'the', 'first', 'second', 'third', 'fourth', 'fifth', 'record', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __lowerCAmelCase = os.path.join(self.tmpdirname , 'realm_tokenizer' ) os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) __lowerCAmelCase = os.path.join(lowerCAmelCase_ , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) __lowerCAmelCase = os.path.join(self.tmpdirname , 'realm_block_records' ) os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) def lowercase ( self : Tuple ) -> RealmTokenizer: return RealmTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'realm_tokenizer' ) ) def lowercase ( self : int ) -> Optional[Any]: shutil.rmtree(self.tmpdirname ) def lowercase ( self : Optional[Any] ) -> List[Any]: __lowerCAmelCase = RealmConfig(num_block_records=self.num_block_records ) return config def lowercase ( self : int ) -> List[Any]: __lowerCAmelCase = Dataset.from_dict( { 'id': ['0', '1'], 'question': ['foo', 'bar'], 'answers': [['Foo', 'Bar'], ['Bar']], } ) return dataset def lowercase ( self : Optional[int] ) -> Optional[int]: __lowerCAmelCase = np.array( [ B'This is the first record', B'This is the second record', B'This is the third record', B'This is the fourth record', B'This is the fifth record', B'This is a longer longer longer record', ] , dtype=lowerCAmelCase_ , ) return block_records def lowercase ( self : str ) -> int: __lowerCAmelCase = RealmRetriever( block_records=self.get_dummy_block_records() , tokenizer=self.get_tokenizer() , ) return retriever def lowercase ( self : Union[str, Any] ) -> List[Any]: __lowerCAmelCase = self.get_config() __lowerCAmelCase = self.get_dummy_retriever() __lowerCAmelCase = retriever.tokenizer __lowerCAmelCase = np.array([0, 3] , dtype='long' ) __lowerCAmelCase = tokenizer(['Test question'] ).input_ids __lowerCAmelCase = tokenizer( ['the fourth'] , add_special_tokens=lowerCAmelCase_ , return_token_type_ids=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , ).input_ids __lowerCAmelCase = config.reader_seq_len __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = retriever( lowerCAmelCase_ , lowerCAmelCase_ , answer_ids=lowerCAmelCase_ , max_length=lowerCAmelCase_ , return_tensors='np' ) self.assertEqual(len(lowerCAmelCase_ ) , 2 ) self.assertEqual(len(lowerCAmelCase_ ) , 2 ) self.assertEqual(len(lowerCAmelCase_ ) , 2 ) self.assertEqual(concat_inputs.input_ids.shape , (2, 1_0) ) self.assertEqual(concat_inputs.attention_mask.shape , (2, 1_0) ) self.assertEqual(concat_inputs.token_type_ids.shape , (2, 1_0) ) self.assertEqual(concat_inputs.special_tokens_mask.shape , (2, 1_0) ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[0] ) , ['[CLS]', 'test', 'question', '[SEP]', 'this', 'is', 'the', 'first', 'record', '[SEP]'] , ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[1] ) , ['[CLS]', 'test', 'question', '[SEP]', 'this', 'is', 'the', 'fourth', 'record', '[SEP]'] , ) def lowercase ( self : Tuple ) -> List[Any]: __lowerCAmelCase = self.get_config() __lowerCAmelCase = self.get_dummy_retriever() __lowerCAmelCase = retriever.tokenizer __lowerCAmelCase = np.array([0, 3, 5] , dtype='long' ) __lowerCAmelCase = tokenizer(['Test question'] ).input_ids __lowerCAmelCase = tokenizer( ['the fourth', 'longer longer'] , add_special_tokens=lowerCAmelCase_ , return_token_type_ids=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , ).input_ids __lowerCAmelCase = config.reader_seq_len __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = retriever( lowerCAmelCase_ , lowerCAmelCase_ , answer_ids=lowerCAmelCase_ , max_length=lowerCAmelCase_ , return_tensors='np' ) self.assertEqual([False, True, True] , lowerCAmelCase_ ) self.assertEqual([[-1, -1, -1], [6, -1, -1], [6, 7, 8]] , lowerCAmelCase_ ) self.assertEqual([[-1, -1, -1], [7, -1, -1], [7, 8, 9]] , lowerCAmelCase_ ) def lowercase ( self : str ) -> Dict: __lowerCAmelCase = self.get_dummy_retriever() retriever.save_pretrained(os.path.join(self.tmpdirname , 'realm_block_records' ) ) # Test local path __lowerCAmelCase = retriever.from_pretrained(os.path.join(self.tmpdirname , 'realm_block_records' ) ) self.assertEqual(retriever.block_records[0] , B'This is the first record' ) # Test mocked remote path with patch('transformers.models.realm.retrieval_realm.hf_hub_download' ) as mock_hf_hub_download: __lowerCAmelCase = os.path.join( os.path.join(self.tmpdirname , 'realm_block_records' ) , _REALM_BLOCK_RECORDS_FILENAME ) __lowerCAmelCase = RealmRetriever.from_pretrained('google/realm-cc-news-pretrained-openqa' ) self.assertEqual(retriever.block_records[0] , B'This is the first record' )
284
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets _snake_case : Tuple = '\\n@article{hendrycksmath2021,\n title={Measuring Mathematical Problem Solving With the MATH Dataset},\n author={Dan Hendrycks\n and Collin Burns\n and Saurav Kadavath\n and Akul Arora\n and Steven Basart\n and Eric Tang\n and Dawn Song\n and Jacob Steinhardt},\n journal={arXiv preprint arXiv:2103.03874},\n year={2021}\n}\n' _snake_case : str = '\\nThis metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset.\nIt first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy.\n' _snake_case : List[str] = R'\nCalculates accuracy after canonicalizing inputs.\n\nArgs:\n predictions: list of predictions to score. Each prediction\n is a string that contains natural language and LaTex.\n references: list of reference for each prediction. Each\n reference is a string that contains natural language\n and LaTex.\nReturns:\n accuracy: accuracy after canonicalizing inputs\n (e.g., converting "1/2" to "\\frac{1}{2}")\n\nExamples:\n >>> metric = datasets.load_metric("competition_math")\n >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"])\n >>> print(results)\n {\'accuracy\': 1.0}\n' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): """simple docstring""" def lowercase ( self : str ) -> int: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' ), 'references': datasets.Value('string' ), } ) , homepage='https://github.com/hendrycks/math' , codebase_urls=['https://github.com/hendrycks/math'] , ) def lowercase ( self : str , lowerCAmelCase_ : Any , lowerCAmelCase_ : Optional[int] ) -> List[Any]: __lowerCAmelCase = 0.0 for i, j in zip(lowerCAmelCase_ , lowerCAmelCase_ ): n_correct += 1.0 if math_equivalence.is_equiv(lowerCAmelCase_ , lowerCAmelCase_ ) else 0.0 __lowerCAmelCase = n_correct / len(lowerCAmelCase_ ) return { "accuracy": accuracy, }
284
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case : List[Any] = { 'configuration_biogpt': ['BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BioGptConfig'], 'tokenization_biogpt': ['BioGptTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BioGptForCausalLM', 'BioGptForTokenClassification', 'BioGptForSequenceClassification', 'BioGptModel', 'BioGptPreTrainedModel', ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys _snake_case : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = StableDiffusionInstructPixaPixPipeline a_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width""", """cross_attention_kwargs"""} a_ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase ( self : Optional[int] ) -> Optional[int]: torch.manual_seed(0 ) __lowerCAmelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=8 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , ) __lowerCAmelCase = PNDMScheduler(skip_prk_steps=lowerCAmelCase_ ) torch.manual_seed(0 ) __lowerCAmelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) __lowerCAmelCase = CLIPTextModel(lowerCAmelCase_ ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __lowerCAmelCase = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowercase ( self : Optional[Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Tuple=0 ) -> Dict: __lowerCAmelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) __lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCAmelCase = Image.fromarray(np.uinta(lowerCAmelCase_ ) ).convert('RGB' ) if str(lowerCAmelCase_ ).startswith('mps' ): __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) else: __lowerCAmelCase = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'image_guidance_scale': 1, 'output_type': 'numpy', } return inputs def lowercase ( self : Tuple ) -> List[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_26, 0.37_50, 0.45_47, 0.61_17, 0.58_66, 0.50_16, 0.43_27, 0.56_42, 0.48_15] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Dict: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = 'french fries' __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ , negative_prompt=lowerCAmelCase_ ) __lowerCAmelCase = output.images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_11, 0.36_42, 0.45_53, 0.62_36, 0.57_97, 0.50_13, 0.43_43, 0.56_11, 0.48_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Any: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = [inputs['prompt']] * 2 __lowerCAmelCase = np.array(inputs['image'] ).astype(np.floataa ) / 2_55.0 __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ).unsqueeze(0 ).to(lowerCAmelCase_ ) __lowerCAmelCase = image / 2 + 0.5 __lowerCAmelCase = image.permute(0 , 3 , 1 , 2 ) __lowerCAmelCase = image.repeat(2 , 1 , 1 , 1 ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[-1, -3:, -3:, -1] assert image.shape == (2, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.58_12, 0.57_48, 0.52_22, 0.59_08, 0.56_95, 0.71_74, 0.68_04, 0.55_23, 0.55_79] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Dict ) -> Optional[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = EulerAncestralDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' ) __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = [round(lowerCAmelCase_ , 4 ) for x in image_slice.flatten().tolist()] print(','.join([str(lowerCAmelCase_ ) for x in slice] ) ) assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.74_17, 0.38_42, 0.47_32, 0.57_76, 0.58_91, 0.51_39, 0.40_52, 0.56_73, 0.49_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> Optional[int]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def lowercase ( self : Optional[Any] ) -> Optional[Any]: __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = VaeImageProcessor(do_resize=lowerCAmelCase_ , do_normalize=lowerCAmelCase_ ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = pipe(**self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) )[0] __lowerCAmelCase = components['vae'] __lowerCAmelCase = self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): __lowerCAmelCase = vae.encode(inputs[image_param] ).latent_dist.mode() __lowerCAmelCase = pipe(**lowerCAmelCase_ )[0] __lowerCAmelCase = np.abs(out - out_latents_inputs ).max() self.assertLess(lowerCAmelCase_ , 1e-4 , 'passing latents as image input generate different result from passing image' ) @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> Optional[int]: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase ( self : List[str] , lowerCAmelCase_ : List[Any]=0 ) -> Any: __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = load_image( 'https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg' ) __lowerCAmelCase = { 'prompt': 'turn him into a cyborg', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'image_guidance_scale': 1.0, 'output_type': 'numpy', } return inputs def lowercase ( self : List[Any] ) -> str: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.59_02, 0.60_15, 0.60_27, 0.59_83, 0.60_92, 0.60_61, 0.57_65, 0.57_85, 0.55_55] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Tuple ) -> List[str]: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.65_78, 0.68_17, 0.69_72, 0.67_61, 0.68_56, 0.69_16, 0.64_28, 0.65_16, 0.63_01] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[Any] ) -> Dict: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.38_28, 0.38_34, 0.38_18, 0.37_92, 0.38_65, 0.37_52, 0.37_92, 0.38_47, 0.37_53] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> int: __lowerCAmelCase = 0 def callback_fn(lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : torch.FloatTensor ) -> None: __lowerCAmelCase = True nonlocal number_of_steps number_of_steps += 1 if step == 1: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.24_63, -0.46_44, -0.97_56, 1.51_76, 1.44_14, 0.78_66, 0.98_97, 0.85_21, 0.79_83] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.26_44, -0.46_26, -0.96_53, 1.51_76, 1.45_51, 0.76_86, 0.98_05, 0.84_52, 0.81_15] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 __lowerCAmelCase = False __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() pipe(**lowerCAmelCase_ , callback=lowerCAmelCase_ , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def lowercase ( self : Optional[int] ) -> Any: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 1_0**9 def lowercase ( self : List[Any] ) -> Any: __lowerCAmelCase = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 __lowerCAmelCase = inputs['image'].resize((5_0_4, 5_0_4) ) __lowerCAmelCase = 'timbrooks/instruct-pix2pix' __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( lowerCAmelCase_ , safety_checker=lowerCAmelCase_ , ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = output.images[0] __lowerCAmelCase = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 5_0_4, 3) __lowerCAmelCase = np.array([0.27_26, 0.25_29, 0.26_64, 0.26_55, 0.26_41, 0.26_42, 0.25_91, 0.26_49, 0.25_90] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3
284
1
import os from pickle import UnpicklingError from typing import Dict, Tuple import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict, unflatten_dict import transformers from .utils import logging _snake_case : Optional[int] = logging.get_logger(__name__) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : Any, lowerCAmelCase_ : Any, lowerCAmelCase_ : Any=False ): try: import torch # noqa: F401 except ImportError: logger.error( 'Loading a PyTorch model in Flax, requires both PyTorch and Flax to be installed. Please see' ' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation' ' instructions.' ) raise if not is_sharded: __lowerCAmelCase = os.path.abspath(lowerCAmelCase_ ) logger.info(F"""Loading PyTorch weights from {pt_path}""" ) __lowerCAmelCase = torch.load(lowerCAmelCase_, map_location='cpu' ) logger.info(F"""PyTorch checkpoint contains {sum(t.numel() for t in pt_state_dict.values() ):,} parameters.""" ) __lowerCAmelCase = convert_pytorch_state_dict_to_flax(lowerCAmelCase_, lowerCAmelCase_ ) else: # model is sharded and pytorch_checkpoint_path already contains the list of .pt shard files __lowerCAmelCase = convert_pytorch_sharded_state_dict_to_flax(lowerCAmelCase_, lowerCAmelCase_ ) return flax_state_dict def a_ ( lowerCAmelCase_ : Tuple[str], lowerCAmelCase_ : np.ndarray, lowerCAmelCase_ : Dict[str, jnp.ndarray], lowerCAmelCase_ : str, ): def is_key_or_prefix_key_in_dict(lowerCAmelCase_ : Tuple[str] ) -> bool: return len(set(lowerCAmelCase_ ) & {key, (model_prefix,) + key} ) > 0 # layer norm __lowerCAmelCase = pt_tuple_key[:-1] + ('scale',) if pt_tuple_key[-1] in ["weight", "gamma"] and is_key_or_prefix_key_in_dict(lowerCAmelCase_ ): return renamed_pt_tuple_key, pt_tensor # batch norm layer mean __lowerCAmelCase = pt_tuple_key[:-1] + ('mean',) if pt_tuple_key[-1] == "running_mean" and not is_key_or_prefix_key_in_dict(lowerCAmelCase_ ): return renamed_pt_tuple_key, pt_tensor # batch norm layer var __lowerCAmelCase = pt_tuple_key[:-1] + ('var',) if pt_tuple_key[-1] == "running_var" and not is_key_or_prefix_key_in_dict(lowerCAmelCase_ ): return renamed_pt_tuple_key, pt_tensor # embedding __lowerCAmelCase = pt_tuple_key[:-1] + ('embedding',) if pt_tuple_key[-1] == "weight" and is_key_or_prefix_key_in_dict(lowerCAmelCase_ ): return renamed_pt_tuple_key, pt_tensor # conv layer __lowerCAmelCase = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4 and not is_key_or_prefix_key_in_dict(lowerCAmelCase_ ): __lowerCAmelCase = pt_tensor.transpose(2, 3, 1, 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer __lowerCAmelCase = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and not is_key_or_prefix_key_in_dict(lowerCAmelCase_ ): __lowerCAmelCase = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight __lowerCAmelCase = pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias __lowerCAmelCase = pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 __lowerCAmelCase = None if pt_tuple_key[-3::2] == ("parametrizations", "original0"): __lowerCAmelCase = pt_tuple_key[-2] + '_g' elif pt_tuple_key[-3::2] == ("parametrizations", "original1"): __lowerCAmelCase = pt_tuple_key[-2] + '_v' if name is not None: __lowerCAmelCase = pt_tuple_key[:-3] + (name,) return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def a_ ( lowerCAmelCase_ : Tuple, lowerCAmelCase_ : Tuple ): # convert pytorch tensor to numpy __lowerCAmelCase = {k: v.numpy() for k, v in pt_state_dict.items()} __lowerCAmelCase = flax_model.base_model_prefix # use params dict if the model contains batch norm layers if "params" in flax_model.params: __lowerCAmelCase = flax_model.params['params'] else: __lowerCAmelCase = flax_model.params __lowerCAmelCase = flatten_dict(lowerCAmelCase_ ) # add batch_stats keys,values to dict if "batch_stats" in flax_model.params: __lowerCAmelCase = flatten_dict(flax_model.params['batch_stats'] ) random_flax_state_dict.update(lowerCAmelCase_ ) __lowerCAmelCase = {} __lowerCAmelCase = (model_prefix not in flax_model_params) and ( model_prefix in {k.split('.' )[0] for k in pt_state_dict.keys()} ) __lowerCAmelCase = (model_prefix in flax_model_params) and ( model_prefix not in {k.split('.' )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): __lowerCAmelCase = tuple(pt_key.split('.' ) ) # remove base model prefix if necessary __lowerCAmelCase = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: __lowerCAmelCase = pt_tuple_key[1:] # Correctly rename weight parameters __lowerCAmelCase , __lowerCAmelCase = rename_key_and_reshape_tensor( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) # add model prefix if necessary __lowerCAmelCase = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: __lowerCAmelCase = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ F"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1] or "var" in flax_key[-1]: __lowerCAmelCase = jnp.asarray(lowerCAmelCase_ ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(lowerCAmelCase_, lowerCAmelCase_ ) continue # also add unexpected weight so that warning is thrown __lowerCAmelCase = jnp.asarray(lowerCAmelCase_ ) else: # also add unexpected weight so that warning is thrown __lowerCAmelCase = jnp.asarray(lowerCAmelCase_ ) return unflatten_dict(lowerCAmelCase_ ) def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : int ): import torch # Load the index __lowerCAmelCase = {} for shard_file in shard_filenames: # load using msgpack utils __lowerCAmelCase = torch.load(lowerCAmelCase_ ) __lowerCAmelCase = {k: v.numpy() for k, v in pt_state_dict.items()} __lowerCAmelCase = flax_model.base_model_prefix # use params dict if the model contains batch norm layers and then add batch_stats keys,values to dict if "batch_stats" in flax_model.params: __lowerCAmelCase = flax_model.params['params'] __lowerCAmelCase = flatten_dict(lowerCAmelCase_ ) random_flax_state_dict.update(flatten_dict(flax_model.params['batch_stats'] ) ) else: __lowerCAmelCase = flax_model.params __lowerCAmelCase = flatten_dict(lowerCAmelCase_ ) __lowerCAmelCase = (model_prefix not in flax_model_params) and ( model_prefix in {k.split('.' )[0] for k in pt_state_dict.keys()} ) __lowerCAmelCase = (model_prefix in flax_model_params) and ( model_prefix not in {k.split('.' )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): __lowerCAmelCase = tuple(pt_key.split('.' ) ) # remove base model prefix if necessary __lowerCAmelCase = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: __lowerCAmelCase = pt_tuple_key[1:] # Correctly rename weight parameters __lowerCAmelCase , __lowerCAmelCase = rename_key_and_reshape_tensor( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) # add model prefix if necessary __lowerCAmelCase = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: __lowerCAmelCase = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ F"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1]: __lowerCAmelCase = jnp.asarray(lowerCAmelCase_ ) continue if "var" in flax_key[-1]: __lowerCAmelCase = jnp.asarray(lowerCAmelCase_ ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(lowerCAmelCase_, lowerCAmelCase_ ) continue # also add unexpected weight so that warning is thrown __lowerCAmelCase = jnp.asarray(lowerCAmelCase_ ) else: # also add unexpected weight so that warning is thrown __lowerCAmelCase = jnp.asarray(lowerCAmelCase_ ) return unflatten_dict(lowerCAmelCase_ ) def a_ ( lowerCAmelCase_ : Dict, lowerCAmelCase_ : List[Any] ): __lowerCAmelCase = os.path.abspath(lowerCAmelCase_ ) logger.info(F"""Loading Flax weights from {flax_checkpoint_path}""" ) # import correct flax class __lowerCAmelCase = getattr(lowerCAmelCase_, 'Flax' + model.__class__.__name__ ) # load flax weight dict with open(lowerCAmelCase_, 'rb' ) as state_f: try: __lowerCAmelCase = from_bytes(lowerCAmelCase_, state_f.read() ) except UnpicklingError: raise EnvironmentError(F"""Unable to convert {flax_checkpoint_path} to Flax deserializable object. """ ) return load_flax_weights_in_pytorch_model(lowerCAmelCase_, lowerCAmelCase_ ) def a_ ( lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Optional[Any] ): try: import torch # noqa: F401 except ImportError: logger.error( 'Loading a Flax weights in PyTorch, requires both PyTorch and Flax to be installed. Please see' ' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation' ' instructions.' ) raise # check if we have bf16 weights __lowerCAmelCase = flatten_dict(jax.tree_util.tree_map(lambda lowerCAmelCase_ : x.dtype == jnp.bfloataa, lowerCAmelCase_ ) ).values() if any(lowerCAmelCase_ ): # convert all weights to fp32 if the are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( 'Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` ' 'before loading those in PyTorch model.' ) __lowerCAmelCase = jax.tree_util.tree_map( lambda lowerCAmelCase_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params, lowerCAmelCase_ ) __lowerCAmelCase = flatten_dict(lowerCAmelCase_ ) __lowerCAmelCase = pt_model.state_dict() __lowerCAmelCase = (pt_model.base_model_prefix in flax_state) and ( pt_model.base_model_prefix not in {k.split('.' )[0] for k in pt_model_dict.keys()} ) __lowerCAmelCase = (pt_model.base_model_prefix not in flax_state) and ( pt_model.base_model_prefix in {k.split('.' )[0] for k in pt_model_dict.keys()} ) # keep track of unexpected & missing keys __lowerCAmelCase = [] __lowerCAmelCase = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): __lowerCAmelCase = flax_key_tuple[0] == pt_model.base_model_prefix __lowerCAmelCase = '.'.join((pt_model.base_model_prefix,) + flax_key_tuple ) in pt_model_dict # adapt flax_key to prepare for loading from/to base model only if load_model_with_head_into_base_model and has_base_model_prefix: __lowerCAmelCase = flax_key_tuple[1:] elif load_base_model_into_model_with_head and require_base_model_prefix: __lowerCAmelCase = (pt_model.base_model_prefix,) + flax_key_tuple # rename flax weights to PyTorch format if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 4 and ".".join(lowerCAmelCase_ ) not in pt_model_dict: # conv layer __lowerCAmelCase = flax_key_tuple[:-1] + ('weight',) __lowerCAmelCase = jnp.transpose(lowerCAmelCase_, (3, 2, 0, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(lowerCAmelCase_ ) not in pt_model_dict: # linear layer __lowerCAmelCase = flax_key_tuple[:-1] + ('weight',) __lowerCAmelCase = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: __lowerCAmelCase = flax_key_tuple[:-1] + ('weight',) # adding batch stats from flax batch norm to pt elif "mean" in flax_key_tuple[-1]: __lowerCAmelCase = flax_key_tuple[:-1] + ('running_mean',) elif "var" in flax_key_tuple[-1]: __lowerCAmelCase = flax_key_tuple[:-1] + ('running_var',) if "batch_stats" in flax_state: __lowerCAmelCase = '.'.join(flax_key_tuple[1:] ) # Remove the params/batch_stats header else: __lowerCAmelCase = '.'.join(lowerCAmelCase_ ) # We also need to look at `pt_model_dict` and see if there are keys requiring further transformation. __lowerCAmelCase = {} # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 for key in pt_model_dict: __lowerCAmelCase = key.split('.' ) __lowerCAmelCase = None if key_components[-3::2] == ["parametrizations", "original0"]: __lowerCAmelCase = key_components[-2] + '_g' elif key_components[-3::2] == ["parametrizations", "original1"]: __lowerCAmelCase = key_components[-2] + '_v' if name is not None: __lowerCAmelCase = key_components[:-3] + [name] __lowerCAmelCase = '.'.join(lowerCAmelCase_ ) __lowerCAmelCase = key if flax_key in special_pt_names: __lowerCAmelCase = special_pt_names[flax_key] if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"""Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected """ F"""to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) else: # add weight to pytorch dict __lowerCAmelCase = np.asarray(lowerCAmelCase_ ) if not isinstance(lowerCAmelCase_, np.ndarray ) else flax_tensor __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ) # remove from missing keys missing_keys.remove(lowerCAmelCase_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(lowerCAmelCase_ ) pt_model.load_state_dict(lowerCAmelCase_ ) # re-transform missing_keys to list __lowerCAmelCase = list(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: logger.warning( 'Some weights of the Flax model were not used when initializing the PyTorch model' F""" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing""" F""" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture""" ' (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This' F""" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect""" ' to be exactly identical (e.g. initializing a BertForSequenceClassification model from a' ' FlaxBertForSequenceClassification model).' ) else: logger.warning(F"""All Flax model weights were used when initializing {pt_model.__class__.__name__}.\n""" ) if len(lowerCAmelCase_ ) > 0: logger.warning( F"""Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly""" F""" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to""" ' use it for predictions and inference.' ) else: logger.warning( F"""All the weights of {pt_model.__class__.__name__} were initialized from the Flax model.\n""" 'If your task is similar to the task the model of the checkpoint was trained on, ' F"""you can already use {pt_model.__class__.__name__} for predictions without further training.""" ) return pt_model
284
from timeit import timeit def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('the value of input must not be negative' ) __lowerCAmelCase = 0 while number: number &= number - 1 result += 1 return result def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('the value of input must not be negative' ) __lowerCAmelCase = 0 while number: if number % 2 == 1: result += 1 number >>= 1 return result def a_ ( ): def do_benchmark(lowerCAmelCase_ : int ) -> None: __lowerCAmelCase = 'import __main__ as z' print(F"""Benchmark when {number = }:""" ) print(F"""{get_set_bits_count_using_modulo_operator(lowerCAmelCase_ ) = }""" ) __lowerCAmelCase = timeit('z.get_set_bits_count_using_modulo_operator(25)', setup=lowerCAmelCase_ ) print(F"""timeit() runs in {timing} seconds""" ) print(F"""{get_set_bits_count_using_brian_kernighans_algorithm(lowerCAmelCase_ ) = }""" ) __lowerCAmelCase = timeit( 'z.get_set_bits_count_using_brian_kernighans_algorithm(25)', setup=lowerCAmelCase_, ) print(F"""timeit() runs in {timing} seconds""" ) for number in (25, 37, 58, 0): do_benchmark(lowerCAmelCase_ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
284
1
from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL _snake_case : int = logging.get_logger(__name__) def a_ ( lowerCAmelCase_ : List[str] ): if isinstance(lowerCAmelCase_, (list, tuple) ) and isinstance(videos[0], (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(lowerCAmelCase_, (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(lowerCAmelCase_ ): return [[videos]] raise ValueError(F"""Could not make batched video from {videos}""" ) class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = ["""pixel_values"""] def __init__( self : int , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Dict[str, int] = None , lowerCAmelCase_ : PILImageResampling = PILImageResampling.BILINEAR , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Dict[str, int] = None , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Union[int, float] = 1 / 2_5_5 , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Optional[Union[float, List[float]]] = None , lowerCAmelCase_ : Optional[Union[float, List[float]]] = None , **lowerCAmelCase_ : int , ) -> None: super().__init__(**lowerCAmelCase_ ) __lowerCAmelCase = size if size is not None else {'shortest_edge': 2_5_6} __lowerCAmelCase = get_size_dict(lowerCAmelCase_ , default_to_square=lowerCAmelCase_ ) __lowerCAmelCase = crop_size if crop_size is not None else {'height': 2_2_4, 'width': 2_2_4} __lowerCAmelCase = get_size_dict(lowerCAmelCase_ , param_name='crop_size' ) __lowerCAmelCase = do_resize __lowerCAmelCase = size __lowerCAmelCase = do_center_crop __lowerCAmelCase = crop_size __lowerCAmelCase = resample __lowerCAmelCase = do_rescale __lowerCAmelCase = rescale_factor __lowerCAmelCase = offset __lowerCAmelCase = do_normalize __lowerCAmelCase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __lowerCAmelCase = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowercase ( self : str , lowerCAmelCase_ : np.ndarray , lowerCAmelCase_ : Dict[str, int] , lowerCAmelCase_ : PILImageResampling = PILImageResampling.BILINEAR , lowerCAmelCase_ : Optional[Union[str, ChannelDimension]] = None , **lowerCAmelCase_ : List[str] , ) -> np.ndarray: __lowerCAmelCase = get_size_dict(lowerCAmelCase_ , default_to_square=lowerCAmelCase_ ) if "shortest_edge" in size: __lowerCAmelCase = get_resize_output_image_size(lowerCAmelCase_ , size['shortest_edge'] , default_to_square=lowerCAmelCase_ ) elif "height" in size and "width" in size: __lowerCAmelCase = (size['height'], size['width']) else: raise ValueError(f"""Size must have 'height' and 'width' or 'shortest_edge' as keys. Got {size.keys()}""" ) return resize(lowerCAmelCase_ , size=lowerCAmelCase_ , resample=lowerCAmelCase_ , data_format=lowerCAmelCase_ , **lowerCAmelCase_ ) def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : np.ndarray , lowerCAmelCase_ : Dict[str, int] , lowerCAmelCase_ : Optional[Union[str, ChannelDimension]] = None , **lowerCAmelCase_ : Dict , ) -> np.ndarray: __lowerCAmelCase = get_size_dict(lowerCAmelCase_ ) if "height" not in size or "width" not in size: raise ValueError(f"""Size must have 'height' and 'width' as keys. Got {size.keys()}""" ) return center_crop(lowerCAmelCase_ , size=(size['height'], size['width']) , data_format=lowerCAmelCase_ , **lowerCAmelCase_ ) def lowercase ( self : Optional[Any] , lowerCAmelCase_ : np.ndarray , lowerCAmelCase_ : Union[int, float] , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Optional[Union[str, ChannelDimension]] = None , **lowerCAmelCase_ : Tuple , ) -> Dict: __lowerCAmelCase = image.astype(np.floataa ) if offset: __lowerCAmelCase = image - (scale / 2) return rescale(lowerCAmelCase_ , scale=lowerCAmelCase_ , data_format=lowerCAmelCase_ , **lowerCAmelCase_ ) def lowercase ( self : Any , lowerCAmelCase_ : np.ndarray , lowerCAmelCase_ : Union[float, List[float]] , lowerCAmelCase_ : Union[float, List[float]] , lowerCAmelCase_ : Optional[Union[str, ChannelDimension]] = None , **lowerCAmelCase_ : List[Any] , ) -> np.ndarray: return normalize(lowerCAmelCase_ , mean=lowerCAmelCase_ , std=lowerCAmelCase_ , data_format=lowerCAmelCase_ , **lowerCAmelCase_ ) def lowercase ( self : int , lowerCAmelCase_ : ImageInput , lowerCAmelCase_ : bool = None , lowerCAmelCase_ : Dict[str, int] = None , lowerCAmelCase_ : PILImageResampling = None , lowerCAmelCase_ : bool = None , lowerCAmelCase_ : Dict[str, int] = None , lowerCAmelCase_ : bool = None , lowerCAmelCase_ : float = None , lowerCAmelCase_ : bool = None , lowerCAmelCase_ : bool = None , lowerCAmelCase_ : Optional[Union[float, List[float]]] = None , lowerCAmelCase_ : Optional[Union[float, List[float]]] = None , lowerCAmelCase_ : Optional[ChannelDimension] = ChannelDimension.FIRST , ) -> np.ndarray: if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) if offset and not do_rescale: raise ValueError('For offset, do_rescale must also be set to True.' ) # All transformations expect numpy arrays. __lowerCAmelCase = to_numpy_array(lowerCAmelCase_ ) if do_resize: __lowerCAmelCase = self.resize(image=lowerCAmelCase_ , size=lowerCAmelCase_ , resample=lowerCAmelCase_ ) if do_center_crop: __lowerCAmelCase = self.center_crop(lowerCAmelCase_ , size=lowerCAmelCase_ ) if do_rescale: __lowerCAmelCase = self.rescale(image=lowerCAmelCase_ , scale=lowerCAmelCase_ , offset=lowerCAmelCase_ ) if do_normalize: __lowerCAmelCase = self.normalize(image=lowerCAmelCase_ , mean=lowerCAmelCase_ , std=lowerCAmelCase_ ) __lowerCAmelCase = to_channel_dimension_format(lowerCAmelCase_ , lowerCAmelCase_ ) return image def lowercase ( self : Optional[int] , lowerCAmelCase_ : ImageInput , lowerCAmelCase_ : bool = None , lowerCAmelCase_ : Dict[str, int] = None , lowerCAmelCase_ : PILImageResampling = None , lowerCAmelCase_ : bool = None , lowerCAmelCase_ : Dict[str, int] = None , lowerCAmelCase_ : bool = None , lowerCAmelCase_ : float = None , lowerCAmelCase_ : bool = None , lowerCAmelCase_ : bool = None , lowerCAmelCase_ : Optional[Union[float, List[float]]] = None , lowerCAmelCase_ : Optional[Union[float, List[float]]] = None , lowerCAmelCase_ : Optional[Union[str, TensorType]] = None , lowerCAmelCase_ : ChannelDimension = ChannelDimension.FIRST , **lowerCAmelCase_ : Any , ) -> PIL.Image.Image: __lowerCAmelCase = do_resize if do_resize is not None else self.do_resize __lowerCAmelCase = resample if resample is not None else self.resample __lowerCAmelCase = do_center_crop if do_center_crop is not None else self.do_center_crop __lowerCAmelCase = do_rescale if do_rescale is not None else self.do_rescale __lowerCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCAmelCase = offset if offset is not None else self.offset __lowerCAmelCase = do_normalize if do_normalize is not None else self.do_normalize __lowerCAmelCase = image_mean if image_mean is not None else self.image_mean __lowerCAmelCase = image_std if image_std is not None else self.image_std __lowerCAmelCase = size if size is not None else self.size __lowerCAmelCase = get_size_dict(lowerCAmelCase_ , default_to_square=lowerCAmelCase_ ) __lowerCAmelCase = crop_size if crop_size is not None else self.crop_size __lowerCAmelCase = get_size_dict(lowerCAmelCase_ , param_name='crop_size' ) if not valid_images(lowerCAmelCase_ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) __lowerCAmelCase = make_batched(lowerCAmelCase_ ) __lowerCAmelCase = [ [ self._preprocess_image( image=lowerCAmelCase_ , do_resize=lowerCAmelCase_ , size=lowerCAmelCase_ , resample=lowerCAmelCase_ , do_center_crop=lowerCAmelCase_ , crop_size=lowerCAmelCase_ , do_rescale=lowerCAmelCase_ , rescale_factor=lowerCAmelCase_ , offset=lowerCAmelCase_ , do_normalize=lowerCAmelCase_ , image_mean=lowerCAmelCase_ , image_std=lowerCAmelCase_ , data_format=lowerCAmelCase_ , ) for img in video ] for video in videos ] __lowerCAmelCase = {'pixel_values': videos} return BatchFeature(data=lowerCAmelCase_ , tensor_type=lowerCAmelCase_ )
284
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss _snake_case : Dict = pytest.mark.integration @require_faiss class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : List[Any] ) -> Optional[Any]: __lowerCAmelCase = Dataset.from_dict({'filename': ['my_name-train' + '_' + str(lowerCAmelCase_ ) for x in np.arange(3_0 ).tolist()]} ) return dset def lowercase ( self : List[str] ) -> Tuple: import faiss __lowerCAmelCase = self._create_dummy_dataset() __lowerCAmelCase = dset.map( lambda lowerCAmelCase_ , lowerCAmelCase_ : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=lowerCAmelCase_ , keep_in_memory=lowerCAmelCase_ ) __lowerCAmelCase = dset.add_faiss_index('vecs' , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) dset.drop_index('vecs' ) def lowercase ( self : Optional[Any] ) -> str: import faiss __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT , ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def lowercase ( self : int ) -> Optional[Any]: import faiss __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCAmelCase_ ) as tmp_file: dset.save_faiss_index('vecs' , tmp_file.name ) dset.load_faiss_index('vecs2' , tmp_file.name ) os.unlink(tmp_file.name ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs2' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def lowercase ( self : Union[str, Any] ) -> List[Any]: __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' ) dset.drop_index('vecs' ) self.assertRaises(lowerCAmelCase_ , partial(dset.get_nearest_examples , 'vecs2' , np.ones(5 , dtype=np.floataa ) ) ) def lowercase ( self : Union[str, Any] ) -> Tuple: from elasticsearch import Elasticsearch __lowerCAmelCase = self._create_dummy_dataset() with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: __lowerCAmelCase = {'acknowledged': True} mocked_bulk.return_value([(True, None)] * 3_0 ) __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 2_9}]}} __lowerCAmelCase = Elasticsearch() dset.add_elasticsearch_index('filename' , es_client=lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('filename' , 'my_name-train_29' ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) @require_faiss class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : str ) -> int: import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 1_0 ) # single query __lowerCAmelCase = np.zeros(5 , dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertRaises(lowerCAmelCase_ , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries __lowerCAmelCase = np.eye(5 , dtype=np.floataa )[::-1] __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ ) self.assertRaises(lowerCAmelCase_ , index.search_batch , queries[0] ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , lowerCAmelCase_ ) def lowercase ( self : List[Any] ) -> List[str]: import faiss __lowerCAmelCase = FaissIndex(string_factory='Flat' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) __lowerCAmelCase = FaissIndex(string_factory='LSH' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(lowerCAmelCase_ ): __lowerCAmelCase = FaissIndex(string_factory='Flat' , custom_index=faiss.IndexFlat(5 ) ) def lowercase ( self : Union[str, Any] ) -> Dict: import faiss __lowerCAmelCase = faiss.IndexFlat(5 ) __lowerCAmelCase = FaissIndex(custom_index=lowerCAmelCase_ ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def lowercase ( self : str ) -> Any: import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCAmelCase_ ) as tmp_file: index.save(tmp_file.name ) __lowerCAmelCase = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) __lowerCAmelCase = np.zeros(5 , dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def a_ ( lowerCAmelCase_ : Union[str, Any] ): import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5, dtype=np.floataa ) ) __lowerCAmelCase = 'index.faiss' __lowerCAmelCase = F"""mock://{index_name}""" index.save(lowerCAmelCase_, storage_options=mockfs.storage_options ) __lowerCAmelCase = FaissIndex.load(lowerCAmelCase_, storage_options=mockfs.storage_options ) __lowerCAmelCase = np.zeros(5, dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : Any ) -> int: from elasticsearch import Elasticsearch with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: __lowerCAmelCase = Elasticsearch() __lowerCAmelCase = {'acknowledged': True} __lowerCAmelCase = ElasticSearchIndex(es_client=lowerCAmelCase_ ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(['foo', 'bar', 'foobar'] ) # single query __lowerCAmelCase = 'foo' __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout __lowerCAmelCase = 'foo' __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ , request_timeout=3_0 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries __lowerCAmelCase = ['foo', 'bar', 'foobar'] __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([1, 1, 1] , lowerCAmelCase_ ) # batched queries with timeout __lowerCAmelCase = ['foo', 'bar', 'foobar'] __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ , request_timeout=3_0 ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([1, 1, 1] , lowerCAmelCase_ )
284
1
from __future__ import annotations import csv import requests from bsa import BeautifulSoup def a_ ( lowerCAmelCase_ : str = "" ): __lowerCAmelCase = url or 'https://www.imdb.com/chart/top/?ref_=nv_mv_250' __lowerCAmelCase = BeautifulSoup(requests.get(lowerCAmelCase_ ).text, 'html.parser' ) __lowerCAmelCase = soup.find_all('td', attrs='titleColumn' ) __lowerCAmelCase = soup.find_all('td', class_='ratingColumn imdbRating' ) return { title.a.text: float(rating.strong.text ) for title, rating in zip(lowerCAmelCase_, lowerCAmelCase_ ) } def a_ ( lowerCAmelCase_ : str = "IMDb_Top_250_Movies.csv" ): __lowerCAmelCase = get_imdb_top_aaa_movies() with open(lowerCAmelCase_, 'w', newline='' ) as out_file: __lowerCAmelCase = csv.writer(lowerCAmelCase_ ) writer.writerow(['Movie title', 'IMDb rating'] ) for title, rating in movies.items(): writer.writerow([title, rating] ) if __name__ == "__main__": write_movies()
284
import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( 'kwargs, expected', [ ({'num_shards': 0, 'max_num_jobs': 1}, []), ({'num_shards': 10, 'max_num_jobs': 1}, [range(10 )]), ({'num_shards': 10, 'max_num_jobs': 10}, [range(lowerCAmelCase_, i + 1 ) for i in range(10 )]), ({'num_shards': 1, 'max_num_jobs': 10}, [range(1 )]), ({'num_shards': 10, 'max_num_jobs': 3}, [range(0, 4 ), range(4, 7 ), range(7, 10 )]), ({'num_shards': 3, 'max_num_jobs': 10}, [range(0, 1 ), range(1, 2 ), range(2, 3 )]), ], ) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : Any ): __lowerCAmelCase = _distribute_shards(**lowerCAmelCase_ ) assert out == expected @pytest.mark.parametrize( 'gen_kwargs, max_num_jobs, expected', [ ({'foo': 0}, 10, [{'foo': 0}]), ({'shards': [0, 1, 2, 3]}, 1, [{'shards': [0, 1, 2, 3]}]), ({'shards': [0, 1, 2, 3]}, 4, [{'shards': [0]}, {'shards': [1]}, {'shards': [2]}, {'shards': [3]}]), ({'shards': [0, 1]}, 4, [{'shards': [0]}, {'shards': [1]}]), ({'shards': [0, 1, 2, 3]}, 2, [{'shards': [0, 1]}, {'shards': [2, 3]}]), ], ) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : List[str], lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = _split_gen_kwargs(lowerCAmelCase_, lowerCAmelCase_ ) assert out == expected @pytest.mark.parametrize( 'gen_kwargs, expected', [ ({'foo': 0}, 1), ({'shards': [0]}, 1), ({'shards': [0, 1, 2, 3]}, 4), ({'shards': [0, 1, 2, 3], 'foo': 0}, 4), ({'shards': [0, 1, 2, 3], 'other': (0, 1)}, 4), ({'shards': [0, 1, 2, 3], 'shards2': [0, 1]}, RuntimeError), ], ) def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : Any ): if expected is RuntimeError: with pytest.raises(lowerCAmelCase_ ): _number_of_shards_in_gen_kwargs(lowerCAmelCase_ ) else: __lowerCAmelCase = _number_of_shards_in_gen_kwargs(lowerCAmelCase_ ) assert out == expected
284
1
import math from datetime import datetime, timedelta def a_ ( lowerCAmelCase_ : int ): __lowerCAmelCase = year % 19 __lowerCAmelCase = year % 4 __lowerCAmelCase = year % 7 __lowerCAmelCase = math.floor(year / 100 ) __lowerCAmelCase = math.floor((13 + 8 * leap_day_inhibits) / 25 ) __lowerCAmelCase = leap_day_inhibits / 4 __lowerCAmelCase = ( 15 - lunar_orbit_correction + leap_day_inhibits - leap_day_reinstall_number ) % 30 __lowerCAmelCase = (4 + leap_day_inhibits - leap_day_reinstall_number) % 7 # days to be added to March 21 __lowerCAmelCase = (19 * metonic_cycle + secular_moon_shift) % 30 # PHM -> Paschal Full Moon __lowerCAmelCase = ( 2 * julian_leap_year + 4 * non_leap_year + 6 * days_to_add + century_starting_point ) % 7 if days_to_add == 29 and days_from_phm_to_sunday == 6: return datetime(lowerCAmelCase_, 4, 19 ) elif days_to_add == 28 and days_from_phm_to_sunday == 6: return datetime(lowerCAmelCase_, 4, 18 ) else: return datetime(lowerCAmelCase_, 3, 22 ) + timedelta( days=int(days_to_add + days_from_phm_to_sunday ) ) if __name__ == "__main__": for year in (1994, 2000, 2010, 2021, 2023): _snake_case : int = 'will be' if year > datetime.now().year else 'was' print(F"""Easter in {year} {tense} {gauss_easter(year)}""")
284
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class _UpperCAmelCase ( _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = BertJapaneseTokenizer a_ = False a_ = True def lowercase ( self : Optional[Any] ) -> List[str]: super().setUp() __lowerCAmelCase = [ '[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは', '世界', '##世界', '、', '##、', '。', '##。', ] __lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def lowercase ( self : List[Any] , lowerCAmelCase_ : Tuple ) -> str: __lowerCAmelCase = 'こんにちは、世界。 \nこんばんは、世界。' __lowerCAmelCase = 'こんにちは 、 世界 。 こんばんは 、 世界 。' return input_text, output_text def lowercase ( self : List[Any] , lowerCAmelCase_ : str ) -> Dict: __lowerCAmelCase , __lowerCAmelCase = self.get_input_output_texts(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(lowerCAmelCase_ , clean_up_tokenization_spaces=lowerCAmelCase_ ) return text, ids def lowercase ( self : List[str] ) -> Optional[int]: pass # TODO add if relevant def lowercase ( self : Optional[Any] ) -> Optional[Any]: pass # TODO add if relevant def lowercase ( self : Union[str, Any] ) -> Any: pass # TODO add if relevant def lowercase ( self : Dict ) -> Tuple: __lowerCAmelCase = self.tokenizer_class(self.vocab_file ) __lowerCAmelCase = tokenizer.tokenize('こんにちは、世界。\nこんばんは、世界。' ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) def lowercase ( self : List[str] ) -> List[str]: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='mecab' ) self.assertIsNotNone(lowerCAmelCase_ ) __lowerCAmelCase = 'こんにちは、世界。\nこんばんは、世界。' __lowerCAmelCase = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __lowerCAmelCase = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowerCAmelCase_ , 'wb' ) as handle: pickle.dump(lowerCAmelCase_ , lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'rb' ) as handle: __lowerCAmelCase = pickle.load(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer_new.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Dict ) -> Tuple: __lowerCAmelCase = MecabTokenizer(mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : List[Any] ) -> int: try: __lowerCAmelCase = MecabTokenizer(mecab_dic='unidic_lite' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : Tuple ) -> Optional[Any]: try: __lowerCAmelCase = MecabTokenizer(mecab_dic='unidic' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : Tuple ) -> Union[str, Any]: __lowerCAmelCase = MecabTokenizer(do_lower_case=lowerCAmelCase_ , mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iphone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : Union[str, Any] ) -> Optional[Any]: try: __lowerCAmelCase = MecabTokenizer( do_lower_case=lowerCAmelCase_ , normalize_text=lowerCAmelCase_ , mecab_option='-d /usr/local/lib/mecab/dic/jumandic' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れた', '\u3000', '。'] , ) def lowercase ( self : Any ) -> Union[str, Any]: __lowerCAmelCase = MecabTokenizer(normalize_text=lowerCAmelCase_ , mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', ' ', '。'] , ) @require_sudachi def lowercase ( self : List[str] ) -> List[str]: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='sudachi' ) self.assertIsNotNone(lowerCAmelCase_ ) __lowerCAmelCase = 'こんにちは、世界。\nこんばんは、世界。' __lowerCAmelCase = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __lowerCAmelCase = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowerCAmelCase_ , 'wb' ) as handle: pickle.dump(lowerCAmelCase_ , lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'rb' ) as handle: __lowerCAmelCase = pickle.load(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer_new.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) @require_sudachi def lowercase ( self : Union[str, Any] ) -> List[str]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iPhone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', ' ', '。', ' ', ' '] , ) @require_sudachi def lowercase ( self : Tuple ) -> Optional[Any]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='A' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国', '人', '参政', '権'] ) @require_sudachi def lowercase ( self : Tuple ) -> List[Any]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='B' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国人', '参政権'] ) @require_sudachi def lowercase ( self : List[str] ) -> Union[str, Any]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='C' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国人参政権'] ) @require_sudachi def lowercase ( self : Dict ) -> List[str]: __lowerCAmelCase = SudachiTokenizer(do_lower_case=lowerCAmelCase_ , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iphone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', ' ', '。', ' ', ' '] , ) @require_sudachi def lowercase ( self : Union[str, Any] ) -> List[Any]: __lowerCAmelCase = SudachiTokenizer(normalize_text=lowerCAmelCase_ , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iPhone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', '\u3000', '。', ' ', ' '] , ) @require_sudachi def lowercase ( self : int ) -> str: __lowerCAmelCase = SudachiTokenizer(trim_whitespace=lowerCAmelCase_ , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) @require_jumanpp def lowercase ( self : Union[str, Any] ) -> Any: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='jumanpp' ) self.assertIsNotNone(lowerCAmelCase_ ) __lowerCAmelCase = 'こんにちは、世界。\nこんばんは、世界。' __lowerCAmelCase = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __lowerCAmelCase = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowerCAmelCase_ , 'wb' ) as handle: pickle.dump(lowerCAmelCase_ , lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'rb' ) as handle: __lowerCAmelCase = pickle.load(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer_new.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) @require_jumanpp def lowercase ( self : List[Any] ) -> Optional[Any]: __lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def lowercase ( self : Any ) -> Union[str, Any]: __lowerCAmelCase = JumanppTokenizer(do_lower_case=lowerCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iphone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def lowercase ( self : Dict ) -> Dict: __lowerCAmelCase = JumanppTokenizer(normalize_text=lowerCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['ア', 'ッ', 'フ', '゚', 'ル', 'ストア', 'で', 'iPhone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def lowercase ( self : List[str] ) -> List[str]: __lowerCAmelCase = JumanppTokenizer(trim_whitespace=lowerCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れた', '。'] , ) @require_jumanpp def lowercase ( self : Any ) -> Any: __lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('ありがとうございますm(_ _)m見つけるのが大変です。' ) , ['ありがとう', 'ございます', 'm(_ _)m', '見つける', 'の', 'が', '大変です', '。'] , ) def lowercase ( self : Any ) -> str: __lowerCAmelCase = ['[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは'] __lowerCAmelCase = {} for i, token in enumerate(lowerCAmelCase_ ): __lowerCAmelCase = i __lowerCAmelCase = WordpieceTokenizer(vocab=lowerCAmelCase_ , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('こんにちは' ) , ['こんにちは'] ) self.assertListEqual(tokenizer.tokenize('こんばんは' ) , ['こん', '##ばんは'] ) self.assertListEqual(tokenizer.tokenize('こんばんは こんばんにちは こんにちは' ) , ['こん', '##ばんは', '[UNK]', 'こんにちは'] ) def lowercase ( self : List[Any] ) -> Tuple: __lowerCAmelCase = BertJapaneseTokenizer.from_pretrained('nlp-waseda/roberta-base-japanese-with-auto-jumanpp' ) __lowerCAmelCase = tokenizer.subword_tokenizer __lowerCAmelCase = subword_tokenizer.tokenize('国境 の 長い トンネル を 抜ける と 雪国 であった 。' ) self.assertListEqual(lowerCAmelCase_ , ['▁国境', '▁の', '▁長い', '▁トンネル', '▁を', '▁抜ける', '▁と', '▁雪', '国', '▁であった', '▁。'] ) __lowerCAmelCase = subword_tokenizer.tokenize('こんばんは こんばん にち は こんにちは' ) self.assertListEqual(lowerCAmelCase_ , ['▁こん', 'ばん', 'は', '▁こん', 'ばん', '▁に', 'ち', '▁は', '▁こんにちは'] ) def lowercase ( self : int ) -> str: __lowerCAmelCase = self.tokenizer_class.from_pretrained('cl-tohoku/bert-base-japanese' ) __lowerCAmelCase = tokenizer.encode('ありがとう。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.encode('どういたしまして。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ , lowerCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class _UpperCAmelCase ( _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = BertJapaneseTokenizer a_ = False def lowercase ( self : Optional[Any] ) -> Tuple: super().setUp() __lowerCAmelCase = ['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] __lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def lowercase ( self : str , **lowerCAmelCase_ : Tuple ) -> Union[str, Any]: return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='character' , **lowerCAmelCase_ ) def lowercase ( self : Tuple , lowerCAmelCase_ : Tuple ) -> Optional[int]: __lowerCAmelCase = 'こんにちは、世界。 \nこんばんは、世界。' __lowerCAmelCase = 'こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。' return input_text, output_text def lowercase ( self : Dict ) -> str: pass # TODO add if relevant def lowercase ( self : Any ) -> str: pass # TODO add if relevant def lowercase ( self : List[Any] ) -> int: pass # TODO add if relevant def lowercase ( self : str ) -> str: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , subword_tokenizer_type='character' ) __lowerCAmelCase = tokenizer.tokenize('こんにちは、世界。 \nこんばんは、世界。' ) self.assertListEqual( lowerCAmelCase_ , ['こ', 'ん', 'に', 'ち', 'は', '、', '世', '界', '。', 'こ', 'ん', 'ば', 'ん', 'は', '、', '世', '界', '。'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 4, 5, 6, 7, 1_1, 9, 1_0, 1_2, 3, 4, 8, 4, 7, 1_1, 9, 1_0, 1_2] ) def lowercase ( self : str ) -> Optional[int]: __lowerCAmelCase = ['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] __lowerCAmelCase = {} for i, token in enumerate(lowerCAmelCase_ ): __lowerCAmelCase = i __lowerCAmelCase = CharacterTokenizer(vocab=lowerCAmelCase_ , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('こんにちは' ) , ['こ', 'ん', 'に', 'ち', 'は'] ) self.assertListEqual(tokenizer.tokenize('こんにちほ' ) , ['こ', 'ん', 'に', 'ち', '[UNK]'] ) def lowercase ( self : int ) -> str: __lowerCAmelCase = self.tokenizer_class.from_pretrained('cl-tohoku/bert-base-japanese-char' ) __lowerCAmelCase = tokenizer.encode('ありがとう。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.encode('どういたしまして。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ , lowerCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : str ) -> Union[str, Any]: __lowerCAmelCase = 'cl-tohoku/bert-base-japanese' __lowerCAmelCase = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : List[str] ) -> Optional[int]: __lowerCAmelCase = 'cl-tohoku/bert-base-japanese' with self.assertLogs('transformers' , level='WARNING' ) as cm: BertTokenizer.from_pretrained(lowerCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( 'The tokenizer class you load from this checkpoint is not the same type as the class this function' ' is called from.' ) ) __lowerCAmelCase = 'bert-base-cased' with self.assertLogs('transformers' , level='WARNING' ) as cm: BertJapaneseTokenizer.from_pretrained(lowerCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( 'The tokenizer class you load from this checkpoint is not the same type as the class this function' ' is called from.' ) )
284
1
# DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class _UpperCAmelCase : """simple docstring""" a_ = 42 # setable values a_ = 42 a_ = 42 a_ = None @classmethod def lowercase ( cls : List[str] , lowerCAmelCase_ : CommonSchedulerState , lowerCAmelCase_ : jnp.ndarray , lowerCAmelCase_ : jnp.ndarray ) -> Dict: return cls(common=lowerCAmelCase_ , init_noise_sigma=lowerCAmelCase_ , timesteps=lowerCAmelCase_ ) @dataclass class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = 42 class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase ): """simple docstring""" a_ = [e.name for e in FlaxKarrasDiffusionSchedulers] a_ = 42 @property def lowercase ( self : str ) -> Optional[int]: return True @register_to_config def __init__( self : int , lowerCAmelCase_ : int = 1_0_0_0 , lowerCAmelCase_ : float = 0.00_01 , lowerCAmelCase_ : float = 0.02 , lowerCAmelCase_ : str = "linear" , lowerCAmelCase_ : Optional[jnp.ndarray] = None , lowerCAmelCase_ : str = "fixed_small" , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : str = "epsilon" , lowerCAmelCase_ : jnp.dtype = jnp.floataa , ) -> Optional[Any]: __lowerCAmelCase = dtype def lowercase ( self : Dict , lowerCAmelCase_ : Optional[CommonSchedulerState] = None ) -> DDPMSchedulerState: if common is None: __lowerCAmelCase = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution __lowerCAmelCase = jnp.array(1.0 , dtype=self.dtype ) __lowerCAmelCase = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=lowerCAmelCase_ , init_noise_sigma=lowerCAmelCase_ , timesteps=lowerCAmelCase_ , ) def lowercase ( self : str , lowerCAmelCase_ : DDPMSchedulerState , lowerCAmelCase_ : jnp.ndarray , lowerCAmelCase_ : Optional[int] = None ) -> jnp.ndarray: return sample def lowercase ( self : Tuple , lowerCAmelCase_ : DDPMSchedulerState , lowerCAmelCase_ : int , lowerCAmelCase_ : Tuple = () ) -> DDPMSchedulerState: __lowerCAmelCase = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 __lowerCAmelCase = (jnp.arange(0 , lowerCAmelCase_ ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=lowerCAmelCase_ , timesteps=lowerCAmelCase_ , ) def lowercase ( self : Tuple , lowerCAmelCase_ : DDPMSchedulerState , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : int=None , lowerCAmelCase_ : List[Any]=None ) -> Tuple: __lowerCAmelCase = state.common.alphas_cumprod[t] __lowerCAmelCase = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample __lowerCAmelCase = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: __lowerCAmelCase = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": __lowerCAmelCase = jnp.clip(lowerCAmelCase_ , a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": __lowerCAmelCase = jnp.log(jnp.clip(lowerCAmelCase_ , a_min=1e-20 ) ) elif variance_type == "fixed_large": __lowerCAmelCase = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log __lowerCAmelCase = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": __lowerCAmelCase = variance __lowerCAmelCase = state.common.betas[t] __lowerCAmelCase = (predicted_variance + 1) / 2 __lowerCAmelCase = frac * max_log + (1 - frac) * min_log return variance def lowercase ( self : Optional[int] , lowerCAmelCase_ : DDPMSchedulerState , lowerCAmelCase_ : jnp.ndarray , lowerCAmelCase_ : int , lowerCAmelCase_ : jnp.ndarray , lowerCAmelCase_ : Optional[jax.random.KeyArray] = None , lowerCAmelCase_ : bool = True , ) -> Union[FlaxDDPMSchedulerOutput, Tuple]: __lowerCAmelCase = timestep if key is None: __lowerCAmelCase = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: __lowerCAmelCase , __lowerCAmelCase = jnp.split(lowerCAmelCase_ , sample.shape[1] , axis=1 ) else: __lowerCAmelCase = None # 1. compute alphas, betas __lowerCAmelCase = state.common.alphas_cumprod[t] __lowerCAmelCase = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) __lowerCAmelCase = 1 - alpha_prod_t __lowerCAmelCase = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": __lowerCAmelCase = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": __lowerCAmelCase = model_output elif self.config.prediction_type == "v_prediction": __lowerCAmelCase = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( f"""prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` """ ' for the FlaxDDPMScheduler.' ) # 3. Clip "predicted x_0" if self.config.clip_sample: __lowerCAmelCase = jnp.clip(lowerCAmelCase_ , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf __lowerCAmelCase = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t __lowerCAmelCase = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf __lowerCAmelCase = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): __lowerCAmelCase = jax.random.split(lowerCAmelCase_ , num=1 ) __lowerCAmelCase = jax.random.normal(lowerCAmelCase_ , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(lowerCAmelCase_ , lowerCAmelCase_ , predicted_variance=lowerCAmelCase_ ) ** 0.5) * noise __lowerCAmelCase = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) __lowerCAmelCase = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=lowerCAmelCase_ , state=lowerCAmelCase_ ) def lowercase ( self : Optional[Any] , lowerCAmelCase_ : DDPMSchedulerState , lowerCAmelCase_ : jnp.ndarray , lowerCAmelCase_ : jnp.ndarray , lowerCAmelCase_ : jnp.ndarray , ) -> jnp.ndarray: return add_noise_common(state.common , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Optional[Any] , lowerCAmelCase_ : DDPMSchedulerState , lowerCAmelCase_ : jnp.ndarray , lowerCAmelCase_ : jnp.ndarray , lowerCAmelCase_ : jnp.ndarray , ) -> jnp.ndarray: return get_velocity_common(state.common , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def __len__( self : str ) -> Optional[int]: return self.config.num_train_timesteps
284
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : List[Any] = { 'microsoft/beit-base-patch16-224-pt22k': ( 'https://huggingface.co/microsoft/beit-base-patch16-224-pt22k/resolve/main/config.json' ), # See all BEiT models at https://huggingface.co/models?filter=beit } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """beit""" def __init__( self : List[Any] , lowerCAmelCase_ : Tuple=8_1_9_2 , lowerCAmelCase_ : Optional[int]=7_6_8 , lowerCAmelCase_ : int=1_2 , lowerCAmelCase_ : Optional[int]=1_2 , lowerCAmelCase_ : Any=3_0_7_2 , lowerCAmelCase_ : Optional[int]="gelu" , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Any=0.02 , lowerCAmelCase_ : int=1e-12 , lowerCAmelCase_ : int=2_2_4 , lowerCAmelCase_ : str=1_6 , lowerCAmelCase_ : int=3 , lowerCAmelCase_ : Dict=False , lowerCAmelCase_ : int=False , lowerCAmelCase_ : List[Any]=False , lowerCAmelCase_ : int=False , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : Union[str, Any]=0.1 , lowerCAmelCase_ : List[str]=True , lowerCAmelCase_ : List[Any]=[3, 5, 7, 1_1] , lowerCAmelCase_ : Optional[Any]=[1, 2, 3, 6] , lowerCAmelCase_ : Tuple=True , lowerCAmelCase_ : Dict=0.4 , lowerCAmelCase_ : Tuple=2_5_6 , lowerCAmelCase_ : Any=1 , lowerCAmelCase_ : Any=False , lowerCAmelCase_ : Optional[int]=2_5_5 , **lowerCAmelCase_ : Any , ) -> Dict: super().__init__(**lowerCAmelCase_ ) __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = initializer_range __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = image_size __lowerCAmelCase = patch_size __lowerCAmelCase = num_channels __lowerCAmelCase = use_mask_token __lowerCAmelCase = use_absolute_position_embeddings __lowerCAmelCase = use_relative_position_bias __lowerCAmelCase = use_shared_relative_position_bias __lowerCAmelCase = layer_scale_init_value __lowerCAmelCase = drop_path_rate __lowerCAmelCase = use_mean_pooling # decode head attributes (semantic segmentation) __lowerCAmelCase = out_indices __lowerCAmelCase = pool_scales # auxiliary head attributes (semantic segmentation) __lowerCAmelCase = use_auxiliary_head __lowerCAmelCase = auxiliary_loss_weight __lowerCAmelCase = auxiliary_channels __lowerCAmelCase = auxiliary_num_convs __lowerCAmelCase = auxiliary_concat_input __lowerCAmelCase = semantic_loss_ignore_index class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = version.parse("""1.11""" ) @property def lowercase ( self : Tuple ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def lowercase ( self : Optional[Any] ) -> float: return 1e-4
284
1
import argparse import torch from transformers import YosoConfig, YosoForMaskedLM def a_ ( lowerCAmelCase_ : int ): if "model" in orig_key: __lowerCAmelCase = orig_key.replace('model.', '' ) if "norm1" in orig_key: __lowerCAmelCase = orig_key.replace('norm1', 'attention.output.LayerNorm' ) if "norm2" in orig_key: __lowerCAmelCase = orig_key.replace('norm2', 'output.LayerNorm' ) if "norm" in orig_key: __lowerCAmelCase = orig_key.replace('norm', 'LayerNorm' ) if "transformer" in orig_key: __lowerCAmelCase = orig_key.split('.' )[0].split('_' )[-1] __lowerCAmelCase = orig_key.replace(F"""transformer_{layer_num}""", F"""encoder.layer.{layer_num}""" ) if "mha.attn" in orig_key: __lowerCAmelCase = orig_key.replace('mha.attn', 'attention.self' ) if "mha" in orig_key: __lowerCAmelCase = orig_key.replace('mha', 'attention' ) if "W_q" in orig_key: __lowerCAmelCase = orig_key.replace('W_q', 'self.query' ) if "W_k" in orig_key: __lowerCAmelCase = orig_key.replace('W_k', 'self.key' ) if "W_v" in orig_key: __lowerCAmelCase = orig_key.replace('W_v', 'self.value' ) if "ff1" in orig_key: __lowerCAmelCase = orig_key.replace('ff1', 'intermediate.dense' ) if "ff2" in orig_key: __lowerCAmelCase = orig_key.replace('ff2', 'output.dense' ) if "ff" in orig_key: __lowerCAmelCase = orig_key.replace('ff', 'output.dense' ) if "mlm_class" in orig_key: __lowerCAmelCase = orig_key.replace('mlm.mlm_class', 'cls.predictions.decoder' ) if "mlm" in orig_key: __lowerCAmelCase = orig_key.replace('mlm', 'cls.predictions.transform' ) if "cls" not in orig_key: __lowerCAmelCase = 'yoso.' + orig_key return orig_key def a_ ( lowerCAmelCase_ : Tuple, lowerCAmelCase_ : List[str] ): for key in orig_state_dict.copy().keys(): __lowerCAmelCase = orig_state_dict.pop(lowerCAmelCase_ ) if ("pooler" in key) or ("sen_class" in key): continue else: __lowerCAmelCase = val __lowerCAmelCase = orig_state_dict['cls.predictions.decoder.bias'] __lowerCAmelCase = torch.arange(lowerCAmelCase_ ).expand((1, -1) ) + 2 return orig_state_dict def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : Any, lowerCAmelCase_ : str ): __lowerCAmelCase = torch.load(lowerCAmelCase_, map_location='cpu' )['model_state_dict'] __lowerCAmelCase = YosoConfig.from_json_file(lowerCAmelCase_ ) __lowerCAmelCase = YosoForMaskedLM(lowerCAmelCase_ ) __lowerCAmelCase = convert_checkpoint_helper(config.max_position_embeddings, lowerCAmelCase_ ) print(model.load_state_dict(lowerCAmelCase_ ) ) model.eval() model.save_pretrained(lowerCAmelCase_ ) print(F"""Checkpoint successfuly converted. Model saved at {pytorch_dump_path}""" ) if __name__ == "__main__": _snake_case : str = argparse.ArgumentParser() # Required parameters parser.add_argument( '--pytorch_model_path', default=None, type=str, required=True, help='Path to YOSO pytorch checkpoint.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The json file for YOSO model config.', ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) _snake_case : Any = parser.parse_args() convert_yoso_checkpoint(args.pytorch_model_path, args.config_file, args.pytorch_dump_path)
284
def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : int ): return [sentence[i : i + ngram_size] for i in range(len(lowerCAmelCase_ ) - ngram_size + 1 )] if __name__ == "__main__": from doctest import testmod testmod()
284
1
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = StableDiffusionInstructPixaPixPipeline a_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width""", """cross_attention_kwargs"""} a_ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase ( self : Optional[int] ) -> Optional[int]: torch.manual_seed(0 ) __lowerCAmelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=8 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , ) __lowerCAmelCase = PNDMScheduler(skip_prk_steps=lowerCAmelCase_ ) torch.manual_seed(0 ) __lowerCAmelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) __lowerCAmelCase = CLIPTextModel(lowerCAmelCase_ ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __lowerCAmelCase = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowercase ( self : Optional[Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Tuple=0 ) -> Dict: __lowerCAmelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) __lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCAmelCase = Image.fromarray(np.uinta(lowerCAmelCase_ ) ).convert('RGB' ) if str(lowerCAmelCase_ ).startswith('mps' ): __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) else: __lowerCAmelCase = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'image_guidance_scale': 1, 'output_type': 'numpy', } return inputs def lowercase ( self : Tuple ) -> List[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_26, 0.37_50, 0.45_47, 0.61_17, 0.58_66, 0.50_16, 0.43_27, 0.56_42, 0.48_15] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Dict: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = 'french fries' __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ , negative_prompt=lowerCAmelCase_ ) __lowerCAmelCase = output.images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_11, 0.36_42, 0.45_53, 0.62_36, 0.57_97, 0.50_13, 0.43_43, 0.56_11, 0.48_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Any: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = [inputs['prompt']] * 2 __lowerCAmelCase = np.array(inputs['image'] ).astype(np.floataa ) / 2_55.0 __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ).unsqueeze(0 ).to(lowerCAmelCase_ ) __lowerCAmelCase = image / 2 + 0.5 __lowerCAmelCase = image.permute(0 , 3 , 1 , 2 ) __lowerCAmelCase = image.repeat(2 , 1 , 1 , 1 ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[-1, -3:, -3:, -1] assert image.shape == (2, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.58_12, 0.57_48, 0.52_22, 0.59_08, 0.56_95, 0.71_74, 0.68_04, 0.55_23, 0.55_79] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Dict ) -> Optional[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = EulerAncestralDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' ) __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = [round(lowerCAmelCase_ , 4 ) for x in image_slice.flatten().tolist()] print(','.join([str(lowerCAmelCase_ ) for x in slice] ) ) assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.74_17, 0.38_42, 0.47_32, 0.57_76, 0.58_91, 0.51_39, 0.40_52, 0.56_73, 0.49_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> Optional[int]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def lowercase ( self : Optional[Any] ) -> Optional[Any]: __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = VaeImageProcessor(do_resize=lowerCAmelCase_ , do_normalize=lowerCAmelCase_ ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = pipe(**self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) )[0] __lowerCAmelCase = components['vae'] __lowerCAmelCase = self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): __lowerCAmelCase = vae.encode(inputs[image_param] ).latent_dist.mode() __lowerCAmelCase = pipe(**lowerCAmelCase_ )[0] __lowerCAmelCase = np.abs(out - out_latents_inputs ).max() self.assertLess(lowerCAmelCase_ , 1e-4 , 'passing latents as image input generate different result from passing image' ) @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> Optional[int]: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase ( self : List[str] , lowerCAmelCase_ : List[Any]=0 ) -> Any: __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = load_image( 'https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg' ) __lowerCAmelCase = { 'prompt': 'turn him into a cyborg', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'image_guidance_scale': 1.0, 'output_type': 'numpy', } return inputs def lowercase ( self : List[Any] ) -> str: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.59_02, 0.60_15, 0.60_27, 0.59_83, 0.60_92, 0.60_61, 0.57_65, 0.57_85, 0.55_55] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Tuple ) -> List[str]: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.65_78, 0.68_17, 0.69_72, 0.67_61, 0.68_56, 0.69_16, 0.64_28, 0.65_16, 0.63_01] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[Any] ) -> Dict: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.38_28, 0.38_34, 0.38_18, 0.37_92, 0.38_65, 0.37_52, 0.37_92, 0.38_47, 0.37_53] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> int: __lowerCAmelCase = 0 def callback_fn(lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : torch.FloatTensor ) -> None: __lowerCAmelCase = True nonlocal number_of_steps number_of_steps += 1 if step == 1: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.24_63, -0.46_44, -0.97_56, 1.51_76, 1.44_14, 0.78_66, 0.98_97, 0.85_21, 0.79_83] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.26_44, -0.46_26, -0.96_53, 1.51_76, 1.45_51, 0.76_86, 0.98_05, 0.84_52, 0.81_15] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 __lowerCAmelCase = False __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() pipe(**lowerCAmelCase_ , callback=lowerCAmelCase_ , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def lowercase ( self : Optional[int] ) -> Any: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 1_0**9 def lowercase ( self : List[Any] ) -> Any: __lowerCAmelCase = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 __lowerCAmelCase = inputs['image'].resize((5_0_4, 5_0_4) ) __lowerCAmelCase = 'timbrooks/instruct-pix2pix' __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( lowerCAmelCase_ , safety_checker=lowerCAmelCase_ , ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = output.images[0] __lowerCAmelCase = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 5_0_4, 3) __lowerCAmelCase = np.array([0.27_26, 0.25_29, 0.26_64, 0.26_55, 0.26_41, 0.26_42, 0.25_91, 0.26_49, 0.25_90] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3
284
from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : Any = { 'google/pegasus-large': 'https://huggingface.co/google/pegasus-large/resolve/main/config.json', # See all PEGASUS models at https://huggingface.co/models?filter=pegasus } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """pegasus""" a_ = ["""past_key_values"""] a_ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[str] , lowerCAmelCase_ : Union[str, Any]=5_0_2_6_5 , lowerCAmelCase_ : Union[str, Any]=1_0_2_4 , lowerCAmelCase_ : Union[str, Any]=1_2 , lowerCAmelCase_ : Dict=4_0_9_6 , lowerCAmelCase_ : str=1_6 , lowerCAmelCase_ : List[Any]=1_2 , lowerCAmelCase_ : Union[str, Any]=4_0_9_6 , lowerCAmelCase_ : Union[str, Any]=1_6 , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Optional[int]=0.0 , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Any=True , lowerCAmelCase_ : Dict="gelu" , lowerCAmelCase_ : Optional[int]=1_0_2_4 , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : str=0.0 , lowerCAmelCase_ : Union[str, Any]=0.0 , lowerCAmelCase_ : Optional[int]=0.02 , lowerCAmelCase_ : Tuple=0 , lowerCAmelCase_ : Tuple=False , lowerCAmelCase_ : Union[str, Any]=0 , lowerCAmelCase_ : int=1 , lowerCAmelCase_ : Tuple=1 , **lowerCAmelCase_ : Tuple , ) -> List[str]: __lowerCAmelCase = vocab_size __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = d_model __lowerCAmelCase = encoder_ffn_dim __lowerCAmelCase = encoder_layers __lowerCAmelCase = encoder_attention_heads __lowerCAmelCase = decoder_ffn_dim __lowerCAmelCase = decoder_layers __lowerCAmelCase = decoder_attention_heads __lowerCAmelCase = dropout __lowerCAmelCase = attention_dropout __lowerCAmelCase = activation_dropout __lowerCAmelCase = activation_function __lowerCAmelCase = init_std __lowerCAmelCase = encoder_layerdrop __lowerCAmelCase = decoder_layerdrop __lowerCAmelCase = use_cache __lowerCAmelCase = encoder_layers __lowerCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , is_encoder_decoder=lowerCAmelCase_ , decoder_start_token_id=lowerCAmelCase_ , forced_eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ , ) @property def lowercase ( self : List[Any] ) -> int: return self.encoder_attention_heads @property def lowercase ( self : Optional[Any] ) -> int: return self.d_model
284
1
import importlib import os import sys # This is required to make the module import works (when the python process is running from the root of the repo) sys.path.append('.') def a_ ( lowerCAmelCase_ : Tuple ): __lowerCAmelCase = test_file.split(os.path.sep ) if components[0:2] != ["tests", "models"]: raise ValueError( '`test_file` should start with `tests/models/` (with `/` being the OS specific path separator). Got ' F"""{test_file} instead.""" ) __lowerCAmelCase = components[-1] if not test_fn.endswith('py' ): raise ValueError(F"""`test_file` should be a python file. Got {test_fn} instead.""" ) if not test_fn.startswith('test_modeling_' ): raise ValueError( F"""`test_file` should point to a file name of the form `test_modeling_*.py`. Got {test_fn} instead.""" ) __lowerCAmelCase = components[:-1] + [test_fn.replace('.py', '' )] __lowerCAmelCase = '.'.join(lowerCAmelCase_ ) return test_module_path def a_ ( lowerCAmelCase_ : Tuple ): __lowerCAmelCase = get_module_path(lowerCAmelCase_ ) __lowerCAmelCase = importlib.import_module(lowerCAmelCase_ ) return test_module def a_ ( lowerCAmelCase_ : Union[str, Any] ): __lowerCAmelCase = [] __lowerCAmelCase = get_test_module(lowerCAmelCase_ ) for attr in dir(lowerCAmelCase_ ): if attr.endswith('ModelTester' ): tester_classes.append(getattr(lowerCAmelCase_, lowerCAmelCase_ ) ) # sort with class names return sorted(lowerCAmelCase_, key=lambda lowerCAmelCase_ : x.__name__ ) def a_ ( lowerCAmelCase_ : int ): __lowerCAmelCase = [] __lowerCAmelCase = get_test_module(lowerCAmelCase_ ) for attr in dir(lowerCAmelCase_ ): __lowerCAmelCase = getattr(lowerCAmelCase_, lowerCAmelCase_ ) # (TF/Flax)ModelTesterMixin is also an attribute in specific model test module. Let's exclude them by checking # `all_model_classes` is not empty (which also excludes other special classes). __lowerCAmelCase = getattr(lowerCAmelCase_, 'all_model_classes', [] ) if len(lowerCAmelCase_ ) > 0: test_classes.append(lowerCAmelCase_ ) # sort with class names return sorted(lowerCAmelCase_, key=lambda lowerCAmelCase_ : x.__name__ ) def a_ ( lowerCAmelCase_ : int ): __lowerCAmelCase = get_test_classes(lowerCAmelCase_ ) __lowerCAmelCase = set() for test_class in test_classes: model_classes.update(test_class.all_model_classes ) # sort with class names return sorted(lowerCAmelCase_, key=lambda lowerCAmelCase_ : x.__name__ ) def a_ ( lowerCAmelCase_ : int ): __lowerCAmelCase = test_class() if hasattr(lowerCAmelCase_, 'setUp' ): test.setUp() __lowerCAmelCase = None if hasattr(lowerCAmelCase_, 'model_tester' ): # `(TF/Flax)ModelTesterMixin` has this attribute default to `None`. Let's skip this case. if test.model_tester is not None: __lowerCAmelCase = test.model_tester.__class__ return model_tester def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : Dict ): __lowerCAmelCase = get_test_classes(lowerCAmelCase_ ) __lowerCAmelCase = [] for test_class in test_classes: if model_class in test_class.all_model_classes: target_test_classes.append(lowerCAmelCase_ ) # sort with class names return sorted(lowerCAmelCase_, key=lambda lowerCAmelCase_ : x.__name__ ) def a_ ( lowerCAmelCase_ : List[str], lowerCAmelCase_ : Tuple ): __lowerCAmelCase = get_test_classes_for_model(lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = [] for test_class in test_classes: __lowerCAmelCase = get_model_tester_from_test_class(lowerCAmelCase_ ) if tester_class is not None: tester_classes.append(lowerCAmelCase_ ) # sort with class names return sorted(lowerCAmelCase_, key=lambda lowerCAmelCase_ : x.__name__ ) def a_ ( lowerCAmelCase_ : Optional[Any] ): __lowerCAmelCase = get_test_classes(lowerCAmelCase_ ) __lowerCAmelCase = {test_class: get_model_tester_from_test_class(lowerCAmelCase_ ) for test_class in test_classes} return test_tester_mapping def a_ ( lowerCAmelCase_ : Union[str, Any] ): __lowerCAmelCase = get_model_classes(lowerCAmelCase_ ) __lowerCAmelCase = { model_class: get_test_classes_for_model(lowerCAmelCase_, lowerCAmelCase_ ) for model_class in model_classes } return model_test_mapping def a_ ( lowerCAmelCase_ : int ): __lowerCAmelCase = get_model_classes(lowerCAmelCase_ ) __lowerCAmelCase = { model_class: get_tester_classes_for_model(lowerCAmelCase_, lowerCAmelCase_ ) for model_class in model_classes } return model_to_tester_mapping def a_ ( lowerCAmelCase_ : Tuple ): if isinstance(lowerCAmelCase_, lowerCAmelCase_ ): return o elif isinstance(lowerCAmelCase_, lowerCAmelCase_ ): return o.__name__ elif isinstance(lowerCAmelCase_, (list, tuple) ): return [to_json(lowerCAmelCase_ ) for x in o] elif isinstance(lowerCAmelCase_, lowerCAmelCase_ ): return {to_json(lowerCAmelCase_ ): to_json(lowerCAmelCase_ ) for k, v in o.items()} else: return o
284
def a_ ( lowerCAmelCase_ : int ): if p < 2: raise ValueError('p should not be less than 2!' ) elif p == 2: return True __lowerCAmelCase = 4 __lowerCAmelCase = (1 << p) - 1 for _ in range(p - 2 ): __lowerCAmelCase = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
284
1
import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def a_ ( lowerCAmelCase_ : Dict ): __lowerCAmelCase = args.pruning_method __lowerCAmelCase = args.threshold __lowerCAmelCase = args.model_name_or_path.rstrip('/' ) __lowerCAmelCase = args.target_model_path print(F"""Load fine-pruned model from {model_name_or_path}""" ) __lowerCAmelCase = torch.load(os.path.join(lowerCAmelCase_, 'pytorch_model.bin' ) ) __lowerCAmelCase = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: __lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: __lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) elif "bias" in name: __lowerCAmelCase = tensor print(F"""Copied layer {name}""" ) else: if pruning_method == "magnitude": __lowerCAmelCase = MagnitudeBinarizer.apply(inputs=lowerCAmelCase_, threshold=lowerCAmelCase_ ) __lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue __lowerCAmelCase = name[:-6] __lowerCAmelCase = model[F"""{prefix_}mask_scores"""] __lowerCAmelCase = TopKBinarizer.apply(lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue __lowerCAmelCase = name[:-6] __lowerCAmelCase = model[F"""{prefix_}mask_scores"""] __lowerCAmelCase = ThresholdBinarizer.apply(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue __lowerCAmelCase = name[:-6] __lowerCAmelCase = model[F"""{prefix_}mask_scores"""] __lowerCAmelCase , __lowerCAmelCase = -0.1, 1.1 __lowerCAmelCase = torch.sigmoid(lowerCAmelCase_ ) __lowerCAmelCase = s * (r - l) + l __lowerCAmelCase = s_bar.clamp(min=0.0, max=1.0 ) __lowerCAmelCase = tensor * mask print(F"""Pruned layer {name}""" ) else: raise ValueError('Unknown pruning method' ) if target_model_path is None: __lowerCAmelCase = os.path.join( os.path.dirname(lowerCAmelCase_ ), F"""bertarized_{os.path.basename(lowerCAmelCase_ )}""" ) if not os.path.isdir(lowerCAmelCase_ ): shutil.copytree(lowerCAmelCase_, lowerCAmelCase_ ) print(F"""\nCreated folder {target_model_path}""" ) torch.save(lowerCAmelCase_, os.path.join(lowerCAmelCase_, 'pytorch_model.bin' ) ) print('\nPruned model saved! See you later!' ) if __name__ == "__main__": _snake_case : Optional[Any] = argparse.ArgumentParser() parser.add_argument( '--pruning_method', choices=['l0', 'magnitude', 'topK', 'sigmoied_threshold'], type=str, required=True, help=( 'Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,' ' sigmoied_threshold = Soft movement pruning)' ), ) parser.add_argument( '--threshold', type=float, required=False, help=( 'For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.' 'For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.' 'Not needed for `l0`' ), ) parser.add_argument( '--model_name_or_path', type=str, required=True, help='Folder containing the model that was previously fine-pruned', ) parser.add_argument( '--target_model_path', default=None, type=str, required=False, help='Folder containing the model that was previously fine-pruned', ) _snake_case : Union[str, Any] = parser.parse_args() main(args)
284
from __future__ import annotations import math def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : int, lowerCAmelCase_ : bool, lowerCAmelCase_ : list[int], lowerCAmelCase_ : float ): if depth < 0: raise ValueError('Depth cannot be less than 0' ) if len(lowerCAmelCase_ ) == 0: raise ValueError('Scores cannot be empty' ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1, node_index * 2, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), minimax(depth + 1, node_index * 2 + 1, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), ) return min( minimax(depth + 1, node_index * 2, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), minimax(depth + 1, node_index * 2 + 1, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), ) def a_ ( ): __lowerCAmelCase = [90, 23, 6, 33, 21, 65, 123, 3_4423] __lowerCAmelCase = math.log(len(lowerCAmelCase_ ), 2 ) print('Optimal value : ', end='' ) print(minimax(0, 0, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
284
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : List[Any] = { 'microsoft/beit-base-patch16-224-pt22k': ( 'https://huggingface.co/microsoft/beit-base-patch16-224-pt22k/resolve/main/config.json' ), # See all BEiT models at https://huggingface.co/models?filter=beit } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """beit""" def __init__( self : List[Any] , lowerCAmelCase_ : Tuple=8_1_9_2 , lowerCAmelCase_ : Optional[int]=7_6_8 , lowerCAmelCase_ : int=1_2 , lowerCAmelCase_ : Optional[int]=1_2 , lowerCAmelCase_ : Any=3_0_7_2 , lowerCAmelCase_ : Optional[int]="gelu" , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Any=0.02 , lowerCAmelCase_ : int=1e-12 , lowerCAmelCase_ : int=2_2_4 , lowerCAmelCase_ : str=1_6 , lowerCAmelCase_ : int=3 , lowerCAmelCase_ : Dict=False , lowerCAmelCase_ : int=False , lowerCAmelCase_ : List[Any]=False , lowerCAmelCase_ : int=False , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : Union[str, Any]=0.1 , lowerCAmelCase_ : List[str]=True , lowerCAmelCase_ : List[Any]=[3, 5, 7, 1_1] , lowerCAmelCase_ : Optional[Any]=[1, 2, 3, 6] , lowerCAmelCase_ : Tuple=True , lowerCAmelCase_ : Dict=0.4 , lowerCAmelCase_ : Tuple=2_5_6 , lowerCAmelCase_ : Any=1 , lowerCAmelCase_ : Any=False , lowerCAmelCase_ : Optional[int]=2_5_5 , **lowerCAmelCase_ : Any , ) -> Dict: super().__init__(**lowerCAmelCase_ ) __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = initializer_range __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = image_size __lowerCAmelCase = patch_size __lowerCAmelCase = num_channels __lowerCAmelCase = use_mask_token __lowerCAmelCase = use_absolute_position_embeddings __lowerCAmelCase = use_relative_position_bias __lowerCAmelCase = use_shared_relative_position_bias __lowerCAmelCase = layer_scale_init_value __lowerCAmelCase = drop_path_rate __lowerCAmelCase = use_mean_pooling # decode head attributes (semantic segmentation) __lowerCAmelCase = out_indices __lowerCAmelCase = pool_scales # auxiliary head attributes (semantic segmentation) __lowerCAmelCase = use_auxiliary_head __lowerCAmelCase = auxiliary_loss_weight __lowerCAmelCase = auxiliary_channels __lowerCAmelCase = auxiliary_num_convs __lowerCAmelCase = auxiliary_concat_input __lowerCAmelCase = semantic_loss_ignore_index class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = version.parse("""1.11""" ) @property def lowercase ( self : Tuple ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def lowercase ( self : Optional[Any] ) -> float: return 1e-4
284
def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('number must not be negative' ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
284
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) _snake_case : Dict = { 'configuration_trocr': ['TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrOCRConfig'], 'processing_trocr': ['TrOCRProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[Any] = [ 'TROCR_PRETRAINED_MODEL_ARCHIVE_LIST', 'TrOCRForCausalLM', 'TrOCRPreTrainedModel', ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
284
1
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : Optional[Any] ) -> int: __lowerCAmelCase = 0 def lowercase ( self : Optional[Any] ) -> Dict: __lowerCAmelCase = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32' ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Any ) -> List[Any]: with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase = Path(lowerCAmelCase_ ) / 'preprocessor_config.json' __lowerCAmelCase = Path(lowerCAmelCase_ ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(lowerCAmelCase_ , 'w' ) , ) json.dump({'model_type': 'clip'} , open(lowerCAmelCase_ , 'w' ) ) __lowerCAmelCase = AutoImageProcessor.from_pretrained(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Optional[int] ) -> Dict: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase = Path(lowerCAmelCase_ ) / 'preprocessor_config.json' __lowerCAmelCase = Path(lowerCAmelCase_ ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(lowerCAmelCase_ , 'w' ) , ) json.dump({'model_type': 'clip'} , open(lowerCAmelCase_ , 'w' ) ) __lowerCAmelCase = AutoImageProcessor.from_pretrained(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Dict ) -> str: with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase = CLIPConfig() # Create a dummy config file with image_proceesor_type __lowerCAmelCase = Path(lowerCAmelCase_ ) / 'preprocessor_config.json' __lowerCAmelCase = Path(lowerCAmelCase_ ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(lowerCAmelCase_ , 'w' ) , ) json.dump({'model_type': 'clip'} , open(lowerCAmelCase_ , 'w' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally __lowerCAmelCase = AutoImageProcessor.from_pretrained(lowerCAmelCase_ ).to_dict() config_dict.pop('image_processor_type' ) __lowerCAmelCase = CLIPImageProcessor(**lowerCAmelCase_ ) # save in new folder model_config.save_pretrained(lowerCAmelCase_ ) config.save_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = AutoImageProcessor.from_pretrained(lowerCAmelCase_ ) # make sure private variable is not incorrectly saved __lowerCAmelCase = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : List[str] ) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase = Path(lowerCAmelCase_ ) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(lowerCAmelCase_ , 'w' ) , ) __lowerCAmelCase = AutoImageProcessor.from_pretrained(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : str ) -> Optional[int]: with self.assertRaisesRegex( lowerCAmelCase_ , 'clip-base is not a local folder and is not a valid model identifier' ): __lowerCAmelCase = AutoImageProcessor.from_pretrained('clip-base' ) def lowercase ( self : int ) -> Any: with self.assertRaisesRegex( lowerCAmelCase_ , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): __lowerCAmelCase = AutoImageProcessor.from_pretrained(lowerCAmelCase_ , revision='aaaaaa' ) def lowercase ( self : Tuple ) -> List[Any]: with self.assertRaisesRegex( lowerCAmelCase_ , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): __lowerCAmelCase = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model' ) def lowercase ( self : str ) -> List[str]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(lowerCAmelCase_ ): __lowerCAmelCase = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(lowerCAmelCase_ ): __lowerCAmelCase = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=lowerCAmelCase_ ) __lowerCAmelCase = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=lowerCAmelCase_ ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = AutoImageProcessor.from_pretrained(lowerCAmelCase_ , trust_remote_code=lowerCAmelCase_ ) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor' ) def lowercase ( self : str ) -> Dict: try: AutoConfig.register('custom' , lowerCAmelCase_ ) AutoImageProcessor.register(lowerCAmelCase_ , lowerCAmelCase_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowerCAmelCase_ ): AutoImageProcessor.register(lowerCAmelCase_ , lowerCAmelCase_ ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase = Path(lowerCAmelCase_ ) / 'preprocessor_config.json' __lowerCAmelCase = Path(lowerCAmelCase_ ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(lowerCAmelCase_ , 'w' ) , ) json.dump({'model_type': 'clip'} , open(lowerCAmelCase_ , 'w' ) ) __lowerCAmelCase = CustomImageProcessor.from_pretrained(lowerCAmelCase_ ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = AutoImageProcessor.from_pretrained(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def lowercase ( self : List[Any] ) -> Optional[Any]: class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = True try: AutoConfig.register('custom' , lowerCAmelCase_ ) AutoImageProcessor.register(lowerCAmelCase_ , lowerCAmelCase_ ) # If remote code is not set, the default is to use local __lowerCAmelCase = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. __lowerCAmelCase = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=lowerCAmelCase_ ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub __lowerCAmelCase = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=lowerCAmelCase_ ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(not hasattr(lowerCAmelCase_ , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
284
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def a_ ( lowerCAmelCase_ : List[str], lowerCAmelCase_ : Dict, lowerCAmelCase_ : Tuple=1024, lowerCAmelCase_ : Optional[Any]=1024, lowerCAmelCase_ : Tuple=False, **lowerCAmelCase_ : Union[str, Any] ): __lowerCAmelCase = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='train', **lowerCAmelCase_ ) __lowerCAmelCase = tok.pad_token_id def get_lens(lowerCAmelCase_ : Optional[Any] ): __lowerCAmelCase = tqdm( DataLoader(lowerCAmelCase_, batch_size=512, num_workers=8, shuffle=lowerCAmelCase_, collate_fn=ds.collate_fn ), desc=str(ds.len_file ), ) __lowerCAmelCase = [] for batch in dl: __lowerCAmelCase = batch['input_ids'].ne(lowerCAmelCase_ ).sum(1 ).tolist() __lowerCAmelCase = batch['labels'].ne(lowerCAmelCase_ ).sum(1 ).tolist() if consider_target: for src, tgt in zip(lowerCAmelCase_, lowerCAmelCase_ ): max_lens.append(max(lowerCAmelCase_, lowerCAmelCase_ ) ) else: max_lens.extend(lowerCAmelCase_ ) return max_lens __lowerCAmelCase = get_lens(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='val', **lowerCAmelCase_ ) __lowerCAmelCase = get_lens(lowerCAmelCase_ ) pickle_save(lowerCAmelCase_, train_ds.len_file ) pickle_save(lowerCAmelCase_, val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
284
1
from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging _snake_case : Tuple = logging.get_logger(__name__) _snake_case : Any = { 'Salesforce/codegen-350M-nl': 'https://huggingface.co/Salesforce/codegen-350M-nl/resolve/main/config.json', 'Salesforce/codegen-350M-multi': 'https://huggingface.co/Salesforce/codegen-350M-multi/resolve/main/config.json', 'Salesforce/codegen-350M-mono': 'https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/config.json', 'Salesforce/codegen-2B-nl': 'https://huggingface.co/Salesforce/codegen-2B-nl/resolve/main/config.json', 'Salesforce/codegen-2B-multi': 'https://huggingface.co/Salesforce/codegen-2B-multi/resolve/main/config.json', 'Salesforce/codegen-2B-mono': 'https://huggingface.co/Salesforce/codegen-2B-mono/resolve/main/config.json', 'Salesforce/codegen-6B-nl': 'https://huggingface.co/Salesforce/codegen-6B-nl/resolve/main/config.json', 'Salesforce/codegen-6B-multi': 'https://huggingface.co/Salesforce/codegen-6B-multi/resolve/main/config.json', 'Salesforce/codegen-6B-mono': 'https://huggingface.co/Salesforce/codegen-6B-mono/resolve/main/config.json', 'Salesforce/codegen-16B-nl': 'https://huggingface.co/Salesforce/codegen-16B-nl/resolve/main/config.json', 'Salesforce/codegen-16B-multi': 'https://huggingface.co/Salesforce/codegen-16B-multi/resolve/main/config.json', 'Salesforce/codegen-16B-mono': 'https://huggingface.co/Salesforce/codegen-16B-mono/resolve/main/config.json', } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """codegen""" a_ = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : List[Any] , lowerCAmelCase_ : Optional[int]=5_0_4_0_0 , lowerCAmelCase_ : List[Any]=2_0_4_8 , lowerCAmelCase_ : Dict=2_0_4_8 , lowerCAmelCase_ : List[Any]=4_0_9_6 , lowerCAmelCase_ : List[str]=2_8 , lowerCAmelCase_ : Tuple=1_6 , lowerCAmelCase_ : Dict=6_4 , lowerCAmelCase_ : int=None , lowerCAmelCase_ : List[Any]="gelu_new" , lowerCAmelCase_ : Dict=0.0 , lowerCAmelCase_ : Dict=0.0 , lowerCAmelCase_ : List[str]=0.0 , lowerCAmelCase_ : List[str]=1e-5 , lowerCAmelCase_ : List[Any]=0.02 , lowerCAmelCase_ : Union[str, Any]=True , lowerCAmelCase_ : List[str]=5_0_2_5_6 , lowerCAmelCase_ : Optional[int]=5_0_2_5_6 , lowerCAmelCase_ : int=False , **lowerCAmelCase_ : Union[str, Any] , ) -> str: __lowerCAmelCase = vocab_size __lowerCAmelCase = n_ctx __lowerCAmelCase = n_positions __lowerCAmelCase = n_embd __lowerCAmelCase = n_layer __lowerCAmelCase = n_head __lowerCAmelCase = n_inner __lowerCAmelCase = rotary_dim __lowerCAmelCase = activation_function __lowerCAmelCase = resid_pdrop __lowerCAmelCase = embd_pdrop __lowerCAmelCase = attn_pdrop __lowerCAmelCase = layer_norm_epsilon __lowerCAmelCase = initializer_range __lowerCAmelCase = use_cache __lowerCAmelCase = bos_token_id __lowerCAmelCase = eos_token_id super().__init__( bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , tie_word_embeddings=lowerCAmelCase_ , **lowerCAmelCase_ ) class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def __init__( self : Union[str, Any] , lowerCAmelCase_ : PretrainedConfig , lowerCAmelCase_ : str = "default" , lowerCAmelCase_ : List[PatchingSpec] = None , lowerCAmelCase_ : bool = False , ) -> str: super().__init__(lowerCAmelCase_ , task=lowerCAmelCase_ , patching_specs=lowerCAmelCase_ , use_past=lowerCAmelCase_ ) if not getattr(self._config , 'pad_token_id' , lowerCAmelCase_ ): # TODO: how to do that better? __lowerCAmelCase = 0 @property def lowercase ( self : int ) -> Mapping[str, Mapping[int, str]]: __lowerCAmelCase = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: self.fill_with_past_key_values_(lowerCAmelCase_ , direction='inputs' ) __lowerCAmelCase = {0: 'batch', 1: 'past_sequence + sequence'} else: __lowerCAmelCase = {0: 'batch', 1: 'sequence'} return common_inputs @property def lowercase ( self : Optional[Any] ) -> int: return self._config.n_layer @property def lowercase ( self : Any ) -> int: return self._config.n_head def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : PreTrainedTokenizer , lowerCAmelCase_ : int = -1 , lowerCAmelCase_ : int = -1 , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : Optional[TensorType] = None , ) -> Mapping[str, Any]: __lowerCAmelCase = super(lowerCAmelCase_ , self ).generate_dummy_inputs( lowerCAmelCase_ , batch_size=lowerCAmelCase_ , seq_length=lowerCAmelCase_ , is_pair=lowerCAmelCase_ , framework=lowerCAmelCase_ ) # We need to order the input in the way they appears in the forward() __lowerCAmelCase = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch __lowerCAmelCase , __lowerCAmelCase = common_inputs['input_ids'].shape # Not using the same length for past_key_values __lowerCAmelCase = seqlen + 2 __lowerCAmelCase = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) __lowerCAmelCase = [ (torch.zeros(lowerCAmelCase_ ), torch.zeros(lowerCAmelCase_ )) for _ in range(self.num_layers ) ] __lowerCAmelCase = common_inputs['attention_mask'] if self.use_past: __lowerCAmelCase = ordered_inputs['attention_mask'].dtype __lowerCAmelCase = torch.cat( [ordered_inputs['attention_mask'], torch.ones(lowerCAmelCase_ , lowerCAmelCase_ , dtype=lowerCAmelCase_ )] , dim=1 ) return ordered_inputs @property def lowercase ( self : Dict ) -> int: return 1_3
284
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def a_ ( lowerCAmelCase_ : List[Any], lowerCAmelCase_ : str, lowerCAmelCase_ : Optional[int]=None, lowerCAmelCase_ : List[Any]=None ): if attention_mask is None: __lowerCAmelCase = tf.cast(tf.math.not_equal(lowerCAmelCase_, config.pad_token_id ), tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class _UpperCAmelCase : """simple docstring""" a_ = OPTConfig a_ = {} a_ = """gelu""" def __init__( self : Dict , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : List[str]=1_3 , lowerCAmelCase_ : Tuple=7 , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : Union[str, Any]=False , lowerCAmelCase_ : Any=9_9 , lowerCAmelCase_ : Any=1_6 , lowerCAmelCase_ : List[str]=2 , lowerCAmelCase_ : Dict=4 , lowerCAmelCase_ : str=4 , lowerCAmelCase_ : Any="gelu" , lowerCAmelCase_ : Optional[Any]=0.1 , lowerCAmelCase_ : List[Any]=0.1 , lowerCAmelCase_ : Tuple=2_0 , lowerCAmelCase_ : Optional[Any]=2 , lowerCAmelCase_ : Any=1 , lowerCAmelCase_ : List[Any]=0 , lowerCAmelCase_ : Optional[int]=1_6 , lowerCAmelCase_ : Dict=1_6 , ) -> int: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = eos_token_id __lowerCAmelCase = pad_token_id __lowerCAmelCase = bos_token_id __lowerCAmelCase = embed_dim __lowerCAmelCase = word_embed_proj_dim __lowerCAmelCase = False def lowercase ( self : List[str] ) -> Optional[Any]: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowerCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowerCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowerCAmelCase = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=lowerCAmelCase_ , **self.config_updates , ) __lowerCAmelCase = prepare_opt_inputs_dict(lowerCAmelCase_ , lowerCAmelCase_ ) return config, inputs_dict def lowercase ( self : Any , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Dict ) -> List[str]: __lowerCAmelCase = TFOPTModel(config=lowerCAmelCase_ ) __lowerCAmelCase = inputs_dict['input_ids'] __lowerCAmelCase = input_ids[:1, :] __lowerCAmelCase = inputs_dict['attention_mask'][:1, :] __lowerCAmelCase = 1 # first forward pass __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , use_cache=lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowerCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowerCAmelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowerCAmelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowerCAmelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , past_key_values=lowerCAmelCase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowerCAmelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowerCAmelCase = output_from_no_past[:, -3:, random_slice_idx] __lowerCAmelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowerCAmelCase_ , lowerCAmelCase_ , rtol=1e-3 ) @require_tf class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () a_ = (TFOPTForCausalLM,) if is_tf_available() else () a_ = ( {"""feature-extraction""": TFOPTModel, """text-generation""": TFOPTForCausalLM} if is_tf_available() else {} ) a_ = False a_ = False a_ = False a_ = 10 def lowercase ( self : List[str] ) -> Optional[int]: __lowerCAmelCase = TFOPTModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=lowerCAmelCase_ ) def lowercase ( self : Tuple ) -> Tuple: self.config_tester.run_common_tests() def lowercase ( self : Tuple ) -> Optional[Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCAmelCase_ ) def lowercase ( self : Union[str, Any] ) -> Dict: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(lowerCAmelCase_ : str , lowerCAmelCase_ : Union[str, Any] ): if hasattr(lowerCAmelCase_ , 'weight' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(lowerCAmelCase_ , 'weight' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 1_0, config.vocab_size + 1_0]: # build the embeddings __lowerCAmelCase = model_class(config=lowerCAmelCase_ ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_input_embeddings() ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(lowerCAmelCase_ ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_input_embeddings() ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. __lowerCAmelCase = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , lowerCAmelCase_ ) # check that weights remain the same after resizing __lowerCAmelCase = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCAmelCase = False self.assertTrue(lowerCAmelCase_ ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , lowerCAmelCase_ ) __lowerCAmelCase = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCAmelCase = False self.assertTrue(lowerCAmelCase_ ) def a_ ( lowerCAmelCase_ : Union[str, Any] ): return tf.constant(lowerCAmelCase_, dtype=tf.intaa ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" a_ = 99 def lowercase ( self : Optional[int] ) -> Any: __lowerCAmelCase = tf.ones((4, 1) , dtype=tf.intaa ) * 2 __lowerCAmelCase = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) __lowerCAmelCase = input_ids.shape[0] __lowerCAmelCase = OPTConfig( vocab_size=self.vocab_size , hidden_size=2_4 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=3_2 , max_position_embeddings=4_8 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def lowercase ( self : str ) -> List[str]: __lowerCAmelCase = TFOPTModel.from_pretrained('facebook/opt-350m' ) __lowerCAmelCase = _long_tensor([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] ) __lowerCAmelCase = tf.not_equal(lowerCAmelCase_ , model.config.pad_token_id ) with tf.GradientTape(): __lowerCAmelCase = model(input_ids=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ ).last_hidden_state __lowerCAmelCase = (1, 1_1, 5_1_2) self.assertEqual(output.shape , lowerCAmelCase_ ) __lowerCAmelCase = tf.constant( [[-0.28_73, -1.92_18, -0.30_33], [-1.27_10, -0.13_38, -0.19_02], [0.40_95, 0.12_14, -1.31_21]] ) self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase_ , atol=4e-3 ) ) __lowerCAmelCase = tf.function(lowerCAmelCase_ , jit_compile=lowerCAmelCase_ ) __lowerCAmelCase = xla_generate(lowerCAmelCase_ , lowerCAmelCase_ )[0] self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase_ , atol=4e-2 ) ) @require_tf @slow class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> Dict: super().setUp() __lowerCAmelCase = 'facebook/opt-350m' def lowercase ( self : Dict ) -> Any: __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(self.path_model ) __lowerCAmelCase = GPTaTokenizer.from_pretrained(self.path_model ) __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of', 'Paris is the capital of France and', 'Computers and mobile phones have taken', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' , padding=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) __lowerCAmelCase = tf.constant( [ [1.38_51, -13.89_23, -10.52_29, -10.75_33, -0.23_09, -10.23_84, -0.53_65, -9.09_47, -5.16_70], [-4.70_73, -10.62_76, -3.94_15, -21.52_42, -0.28_22, -0.28_22, -0.28_22, -0.28_22, -0.28_22], [0.62_47, -3.42_29, -8.91_79, -1.42_97, -14.16_50, 1.41_46, -9.02_18, -0.27_03, -0.27_03], [6.47_83, -1.99_13, -10.79_26, -2.33_36, 1.50_92, -0.99_74, -6.82_13, 1.34_77, 1.34_77], ] ) self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1e-4 ) ) __lowerCAmelCase = tf.function(lowerCAmelCase_ , jit_compile=lowerCAmelCase_ ) __lowerCAmelCase = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1e-4 ) ) @require_tf @slow class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @property def lowercase ( self : Optional[int] ) -> int: return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def lowercase ( self : int ) -> str: __lowerCAmelCase = 'facebook/opt-125m' __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of New York, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] __lowerCAmelCase = [] __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) for prompt in self.prompts: __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(lowerCAmelCase_ , max_length=1_0 ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Optional[Any] ) -> str: __lowerCAmelCase = 'facebook/opt-350m' __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = 'left' # use different length sentences to test batching __lowerCAmelCase = [ 'Hello, my dog is a little', 'Today, I', ] __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' , padding=lowerCAmelCase_ ) __lowerCAmelCase = inputs['input_ids'] __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ , attention_mask=inputs['attention_mask'] ) __lowerCAmelCase = tokenizer(sentences[0] , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ ) __lowerCAmelCase = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['attention_mask'][-1] , tf.intaa ) ) __lowerCAmelCase = tokenizer(sentences[1] , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ , max_length=model.config.max_length - num_paddings ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = [ 'Hello, my dog is a little bit of a dork.\nI\'m a little bit', 'Today, I was in the middle of a conversation with a friend about the', ] self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , [non_padded_sentence, padded_sentence] ) def lowercase ( self : List[Any] ) -> List[Any]: __lowerCAmelCase = 'facebook/opt-350m' __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of San Francisco, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] __lowerCAmelCase = [] __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) for prompt in self.prompts: __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(lowerCAmelCase_ , max_length=1_0 ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ )
284
1
import datasets from .evaluate import evaluate _snake_case : List[Any] = '\\n@article{hendrycks2021cuad,\n title={CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review},\n author={Dan Hendrycks and Collin Burns and Anya Chen and Spencer Ball},\n journal={arXiv preprint arXiv:2103.06268},\n year={2021}\n}\n' _snake_case : Any = '\nThis metric wrap the official scoring script for version 1 of the Contract\nUnderstanding Atticus Dataset (CUAD).\nContract Understanding Atticus Dataset (CUAD) v1 is a corpus of more than 13,000 labels in 510\ncommercial legal contracts that have been manually labeled to identify 41 categories of important\nclauses that lawyers look for when reviewing contracts in connection with corporate transactions.\n' _snake_case : Optional[int] = '\nComputes CUAD scores (EM, F1, AUPR, Precision@80%Recall, and Precision@90%Recall).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair as given in the references (see below)\n - \'prediction_text\': list of possible texts for the answer, as a list of strings\n depending on a threshold on the confidence probability of each prediction.\n references: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair (see above),\n - \'answers\': a Dict in the CUAD dataset format\n {\n \'text\': list of possible texts for the answer, as a list of strings\n \'answer_start\': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n \'exact_match\': Exact match (the normalized answer exactly match the gold answer)\n \'f1\': The F-score of predicted tokens versus the gold answer\n \'aupr\': Area Under the Precision-Recall curve\n \'prec_at_80_recall\': Precision at 80% recall\n \'prec_at_90_recall\': Precision at 90% recall\nExamples:\n >>> predictions = [{\'prediction_text\': [\'The seller:\', \'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.\'], \'id\': \'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties\'}]\n >>> references = [{\'answers\': {\'answer_start\': [143, 49], \'text\': [\'The seller:\', \'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.\']}, \'id\': \'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties\'}]\n >>> cuad_metric = datasets.load_metric("cuad")\n >>> results = cuad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'exact_match\': 100.0, \'f1\': 100.0, \'aupr\': 0.0, \'prec_at_80_recall\': 1.0, \'prec_at_90_recall\': 1.0}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): """simple docstring""" def lowercase ( self : int ) -> int: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': { 'id': datasets.Value('string' ), 'prediction_text': datasets.features.Sequence(datasets.Value('string' ) ), }, 'references': { 'id': datasets.Value('string' ), 'answers': datasets.features.Sequence( { 'text': datasets.Value('string' ), 'answer_start': datasets.Value('int32' ), } ), }, } ) , codebase_urls=['https://www.atticusprojectai.org/cuad'] , reference_urls=['https://www.atticusprojectai.org/cuad'] , ) def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : str , lowerCAmelCase_ : List[Any] ) -> Union[str, Any]: __lowerCAmelCase = {prediction['id']: prediction['prediction_text'] for prediction in predictions} __lowerCAmelCase = [ { 'paragraphs': [ { 'qas': [ { 'answers': [{'text': answer_text} for answer_text in ref['answers']['text']], 'id': ref['id'], } for ref in references ] } ] } ] __lowerCAmelCase = evaluate(dataset=lowerCAmelCase_ , predictions=lowerCAmelCase_ ) return score
284
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { 'configuration_layoutlmv3': [ 'LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv3Config', 'LayoutLMv3OnnxConfig', ], 'processing_layoutlmv3': ['LayoutLMv3Processor'], 'tokenization_layoutlmv3': ['LayoutLMv3Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Union[str, Any] = ['LayoutLMv3TokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[str] = [ 'LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv3ForQuestionAnswering', 'LayoutLMv3ForSequenceClassification', 'LayoutLMv3ForTokenClassification', 'LayoutLMv3Model', 'LayoutLMv3PreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[Any] = [ 'TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLayoutLMv3ForQuestionAnswering', 'TFLayoutLMv3ForSequenceClassification', 'TFLayoutLMv3ForTokenClassification', 'TFLayoutLMv3Model', 'TFLayoutLMv3PreTrainedModel', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Tuple = ['LayoutLMv3FeatureExtractor'] _snake_case : str = ['LayoutLMv3ImageProcessor'] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys _snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
1
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = StableDiffusionXLImgaImgPipeline a_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} a_ = PipelineTesterMixin.required_optional_params - {"""latents"""} a_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase ( self : str ) -> Optional[int]: torch.manual_seed(0 ) __lowerCAmelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , attention_head_dim=(2, 4) , use_linear_projection=lowerCAmelCase_ , addition_embed_type='text_time' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=8_0 , cross_attention_dim=6_4 , ) __lowerCAmelCase = EulerDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , steps_offset=1 , beta_schedule='scaled_linear' , timestep_spacing='leading' , ) torch.manual_seed(0 ) __lowerCAmelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) __lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='gelu' , projection_dim=3_2 , ) __lowerCAmelCase = CLIPTextModel(lowerCAmelCase_ ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' , local_files_only=lowerCAmelCase_ ) __lowerCAmelCase = CLIPTextModelWithProjection(lowerCAmelCase_ ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' , local_files_only=lowerCAmelCase_ ) __lowerCAmelCase = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'text_encoder_2': text_encoder_a, 'tokenizer_2': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def lowercase ( self : str , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Union[str, Any]=0 ) -> Optional[Any]: __lowerCAmelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) __lowerCAmelCase = image / 2 + 0.5 if str(lowerCAmelCase_ ).startswith('mps' ): __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) else: __lowerCAmelCase = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 5.0, 'output_type': 'numpy', 'strength': 0.75, } return inputs def lowercase ( self : Union[str, Any] ) -> Tuple: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionXLImgaImgPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.46_56, 0.48_40, 0.44_39, 0.66_98, 0.55_74, 0.45_24, 0.57_99, 0.59_43, 0.51_65] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def lowercase ( self : Tuple ) -> Optional[int]: super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def lowercase ( self : Optional[int] ) -> Optional[Any]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def lowercase ( self : int ) -> Any: pass def lowercase ( self : Dict ) -> Tuple: __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionXLImgaImgPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) # forward without prompt embeds __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = 3 * ['this is a negative prompt'] __lowerCAmelCase = negative_prompt __lowerCAmelCase = 3 * [inputs['prompt']] __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ) __lowerCAmelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = 3 * ['this is a negative prompt'] __lowerCAmelCase = 3 * [inputs.pop('prompt' )] ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = sd_pipe.encode_prompt(lowerCAmelCase_ , negative_prompt=lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe( **lowerCAmelCase_ , prompt_embeds=lowerCAmelCase_ , negative_prompt_embeds=lowerCAmelCase_ , pooled_prompt_embeds=lowerCAmelCase_ , negative_pooled_prompt_embeds=lowerCAmelCase_ , ) __lowerCAmelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> str: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase ( self : int , lowerCAmelCase_ : str , lowerCAmelCase_ : List[str]="cpu" , lowerCAmelCase_ : Optional[Any]=torch.floataa , lowerCAmelCase_ : List[Any]=0 ) -> str: __lowerCAmelCase = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = np.random.RandomState(lowerCAmelCase_ ).standard_normal((1, 4, 6_4, 6_4) ) __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ).to(device=lowerCAmelCase_ , dtype=lowerCAmelCase_ ) __lowerCAmelCase = { 'prompt': 'a photograph of an astronaut riding a horse', 'latents': latents, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def lowercase ( self : Tuple ) -> int: __lowerCAmelCase = DiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-base' ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_inputs(lowerCAmelCase_ ) __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.4_94_93, 0.4_78_96, 0.4_07_98, 0.5_42_14, 0.5_32_12, 0.4_82_02, 0.4_76_56, 0.4_63_29, 0.4_85_06] ) assert np.abs(image_slice - expected_slice ).max() < 7e-3
284
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case : Dict = {'configuration_mra': ['MRA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MraConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'MRA_PRETRAINED_MODEL_ARCHIVE_LIST', 'MraForMaskedLM', 'MraForMultipleChoice', 'MraForQuestionAnswering', 'MraForSequenceClassification', 'MraForTokenClassification', 'MraLayer', 'MraModel', 'MraPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys _snake_case : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure)
284
1
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Sequence, Value from .base import TaskTemplate @dataclass(frozen=_UpperCamelCase ) class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = field(default="""question-answering-extractive""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) a_ = Features({"""question""": Value("""string""" ), """context""": Value("""string""" )} ) a_ = Features( { """answers""": Sequence( { """text""": Value("""string""" ), """answer_start""": Value("""int32""" ), } ) } ) a_ = "question" a_ = "context" a_ = "answers" @property def lowercase ( self : str ) -> Dict[str, str]: return {self.question_column: "question", self.context_column: "context", self.answers_column: "answers"}
284
from __future__ import annotations import string from itertools import cycle, product from pathlib import Path _snake_case : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) _snake_case : list[int] = [ord(letter) for letter in string.ascii_lowercase] _snake_case : set[int] = {ord(char) for char in VALID_CHARS} _snake_case : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def a_ ( lowerCAmelCase_ : list[int], lowerCAmelCase_ : tuple[int, ...] ): __lowerCAmelCase = "" __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 for keychar, cipherchar in zip(cycle(lowerCAmelCase_ ), lowerCAmelCase_ ): __lowerCAmelCase = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(lowerCAmelCase_ ) return decoded def a_ ( lowerCAmelCase_ : list[int] ): __lowerCAmelCase = [] for key in product(lowerCAmelCase_, repeat=3 ): __lowerCAmelCase = try_key(lowerCAmelCase_, lowerCAmelCase_ ) if encoded is not None: possibles.append(lowerCAmelCase_ ) return possibles def a_ ( lowerCAmelCase_ : list[str], lowerCAmelCase_ : str ): return [possible for possible in possibles if common_word in possible.lower()] def a_ ( lowerCAmelCase_ : str = "p059_cipher.txt" ): __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = Path(lowerCAmelCase_ ).parent.joinpath(lowerCAmelCase_ ).read_text(encoding='utf-8' ) __lowerCAmelCase = [int(lowerCAmelCase_ ) for number in data.strip().split(',' )] __lowerCAmelCase = filter_valid_chars(lowerCAmelCase_ ) for common_word in COMMON_WORDS: __lowerCAmelCase = filter_common_word(lowerCAmelCase_, lowerCAmelCase_ ) if len(lowerCAmelCase_ ) == 1: break __lowerCAmelCase = possibles[0] return sum(ord(lowerCAmelCase_ ) for char in decoded_text ) if __name__ == "__main__": print(F"""{solution() = }""")
284
1
def a_ ( lowerCAmelCase_ : Any ): __lowerCAmelCase = [] __lowerCAmelCase = [] __lowerCAmelCase = { '^': 3, '*': 2, '/': 2, '%': 2, '+': 1, '-': 1, } # Priority of each operator __lowerCAmelCase = len(lowerCAmelCase_ ) if (len(lowerCAmelCase_ ) > 7) else 7 # Print table header for output print( 'Symbol'.center(8 ), 'Stack'.center(lowerCAmelCase_ ), 'Postfix'.center(lowerCAmelCase_ ), sep=' | ', ) print('-' * (print_width * 3 + 7) ) for x in infix: if x.isalpha() or x.isdigit(): post_fix.append(lowerCAmelCase_ ) # if x is Alphabet / Digit, add it to Postfix elif x == "(": stack.append(lowerCAmelCase_ ) # if x is "(" push to Stack elif x == ")": # if x is ")" pop stack until "(" is encountered while stack[-1] != "(": post_fix.append(stack.pop() ) # Pop stack & add the content to Postfix stack.pop() else: if len(lowerCAmelCase_ ) == 0: stack.append(lowerCAmelCase_ ) # If stack is empty, push x to stack else: # while priority of x is not > priority of element in the stack while len(lowerCAmelCase_ ) > 0 and priority[x] <= priority[stack[-1]]: post_fix.append(stack.pop() ) # pop stack & add to Postfix stack.append(lowerCAmelCase_ ) # push x to stack print( x.center(8 ), (''.join(lowerCAmelCase_ )).ljust(lowerCAmelCase_ ), (''.join(lowerCAmelCase_ )).ljust(lowerCAmelCase_ ), sep=' | ', ) # Output in tabular format while len(lowerCAmelCase_ ) > 0: # while stack is not empty post_fix.append(stack.pop() ) # pop stack & add to Postfix print( ' '.center(8 ), (''.join(lowerCAmelCase_ )).ljust(lowerCAmelCase_ ), (''.join(lowerCAmelCase_ )).ljust(lowerCAmelCase_ ), sep=' | ', ) # Output in tabular format return "".join(lowerCAmelCase_ ) # return Postfix as str def a_ ( lowerCAmelCase_ : Optional[Any] ): __lowerCAmelCase = list(infix[::-1] ) # reverse the infix equation for i in range(len(lowerCAmelCase_ ) ): if infix[i] == "(": __lowerCAmelCase = ')' # change "(" to ")" elif infix[i] == ")": __lowerCAmelCase = '(' # change ")" to "(" return (infix_2_postfix(''.join(lowerCAmelCase_ ) ))[ ::-1 ] # call infix_2_postfix on Infix, return reverse of Postfix if __name__ == "__main__": _snake_case : List[str] = input('\nEnter an Infix Equation = ') # Input an Infix equation _snake_case : Union[str, Any] = ''.join(Infix.split()) # Remove spaces from the input print('\n\t', Infix, '(Infix) -> ', infix_2_prefix(Infix), '(Prefix)')
284
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem _snake_case : Tuple = importlib.util.find_spec('s3fs') is not None if _has_safs: from .safilesystem import SaFileSystem # noqa: F401 _snake_case : List[compression.BaseCompressedFileFileSystem] = [ compression.BzaFileSystem, compression.GzipFileSystem, compression.LzaFileSystem, compression.XzFileSystem, compression.ZstdFileSystem, ] # Register custom filesystems for fs_class in COMPRESSION_FILESYSTEMS + [HfFileSystem]: if fs_class.protocol in fsspec.registry and fsspec.registry[fs_class.protocol] is not fs_class: warnings.warn(F"""A filesystem protocol was already set for {fs_class.protocol} and will be overwritten.""") fsspec.register_implementation(fs_class.protocol, fs_class, clobber=True) def a_ ( lowerCAmelCase_ : str ): if "://" in dataset_path: __lowerCAmelCase = dataset_path.split('://' )[1] return dataset_path def a_ ( lowerCAmelCase_ : fsspec.AbstractFileSystem ): if fs is not None and fs.protocol != "file": return True else: return False def a_ ( lowerCAmelCase_ : fsspec.AbstractFileSystem, lowerCAmelCase_ : str, lowerCAmelCase_ : str ): __lowerCAmelCase = not is_remote_filesystem(lowerCAmelCase_ ) if is_local: # LocalFileSystem.mv does copy + rm, it is more efficient to simply move a local directory shutil.move(fs._strip_protocol(lowerCAmelCase_ ), fs._strip_protocol(lowerCAmelCase_ ) ) else: fs.mv(lowerCAmelCase_, lowerCAmelCase_, recursive=lowerCAmelCase_ ) def a_ ( ): if hasattr(fsspec.asyn, 'reset_lock' ): # for future fsspec>2022.05.0 fsspec.asyn.reset_lock() else: __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = threading.Lock()
284
1
def a_ ( lowerCAmelCase_ : int | float | str ): try: __lowerCAmelCase = float(lowerCAmelCase_ ) except ValueError: raise ValueError('Please enter a valid number' ) __lowerCAmelCase = decimal - int(lowerCAmelCase_ ) if fractional_part == 0: return int(lowerCAmelCase_ ), 1 else: __lowerCAmelCase = len(str(lowerCAmelCase_ ).split('.' )[1] ) __lowerCAmelCase = int(decimal * (10**number_of_frac_digits) ) __lowerCAmelCase = 10**number_of_frac_digits __lowerCAmelCase , __lowerCAmelCase = denominator, numerator while True: __lowerCAmelCase = dividend % divisor if remainder == 0: break __lowerCAmelCase , __lowerCAmelCase = divisor, remainder __lowerCAmelCase , __lowerCAmelCase = numerator / divisor, denominator / divisor return int(lowerCAmelCase_ ), int(lowerCAmelCase_ ) if __name__ == "__main__": print(F"""{decimal_to_fraction(2) = }""") print(F"""{decimal_to_fraction(89.0) = }""") print(F"""{decimal_to_fraction("67") = }""") print(F"""{decimal_to_fraction("45.0") = }""") print(F"""{decimal_to_fraction(1.5) = }""") print(F"""{decimal_to_fraction("6.25") = }""") print(F"""{decimal_to_fraction("78td") = }""")
284
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def a_ ( lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = filter(lambda lowerCAmelCase_ : p.requires_grad, model.parameters() ) __lowerCAmelCase = sum([np.prod(p.size() ) for p in model_parameters] ) return params _snake_case : Dict = logging.getLogger(__name__) def a_ ( lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Optional[int] ): if metric == "rouge2": __lowerCAmelCase = '{val_avg_rouge2:.4f}-{step_count}' elif metric == "bleu": __lowerCAmelCase = '{val_avg_bleu:.4f}-{step_count}' elif metric == "em": __lowerCAmelCase = '{val_avg_em:.4f}-{step_count}' else: raise NotImplementedError( F"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" ' function.' ) __lowerCAmelCase = ModelCheckpoint( dirpath=lowerCAmelCase_, filename=lowerCAmelCase_, monitor=F"""val_{metric}""", mode='max', save_top_k=3, every_n_epochs=1, ) return checkpoint_callback def a_ ( lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : Any ): return EarlyStopping( monitor=F"""val_{metric}""", mode='min' if 'loss' in metric else 'max', patience=lowerCAmelCase_, verbose=lowerCAmelCase_, ) class _UpperCAmelCase ( pl.Callback ): """simple docstring""" def lowercase ( self : Tuple , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : int ) -> Any: __lowerCAmelCase = {f"""lr_group_{i}""": param['lr'] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Optional[int] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule , lowerCAmelCase_ : str , lowerCAmelCase_ : List[Any]=True ) -> None: logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) __lowerCAmelCase = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['log', 'progress_bar', 'preds']} ) # Log results __lowerCAmelCase = Path(pl_module.hparams.output_dir ) if type_path == "test": __lowerCAmelCase = od / 'test_results.txt' __lowerCAmelCase = od / 'test_generations.txt' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __lowerCAmelCase = od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" __lowerCAmelCase = od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) generations_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'a+' ) as writer: for key in sorted(lowerCAmelCase_ ): if key in ["log", "progress_bar", "preds"]: continue __lowerCAmelCase = metrics[key] if isinstance(lowerCAmelCase_ , torch.Tensor ): __lowerCAmelCase = val.item() __lowerCAmelCase = f"""{key}: {val:.6f}\n""" writer.write(lowerCAmelCase_ ) if not save_generations: return if "preds" in metrics: __lowerCAmelCase = '\n'.join(metrics['preds'] ) generations_file.open('w+' ).write(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[str] ) -> Dict: try: __lowerCAmelCase = pl_module.model.model.num_parameters() except AttributeError: __lowerCAmelCase = pl_module.model.num_parameters() __lowerCAmelCase = count_trainable_parameters(lowerCAmelCase_ ) # mp stands for million parameters trainer.logger.log_metrics({'n_params': npars, 'mp': npars / 1e6, 'grad_mp': n_trainable_pars / 1e6} ) @rank_zero_only def lowercase ( self : int , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule ) -> Any: save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(lowerCAmelCase_ , lowerCAmelCase_ , 'test' ) @rank_zero_only def lowercase ( self : List[Any] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : Any ) -> int: save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
284
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers import ( TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, BertConfig, DPRConfig, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) class _UpperCAmelCase : """simple docstring""" def __init__( self : List[str] , lowerCAmelCase_ : Any , lowerCAmelCase_ : Any=1_3 , lowerCAmelCase_ : Optional[Any]=7 , lowerCAmelCase_ : str=True , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : Tuple=True , lowerCAmelCase_ : List[str]=9_9 , lowerCAmelCase_ : Any=3_2 , lowerCAmelCase_ : Dict=2 , lowerCAmelCase_ : Any=4 , lowerCAmelCase_ : Optional[Any]=3_7 , lowerCAmelCase_ : Any="gelu" , lowerCAmelCase_ : Optional[int]=0.1 , lowerCAmelCase_ : Tuple=0.1 , lowerCAmelCase_ : Union[str, Any]=5_1_2 , lowerCAmelCase_ : Dict=1_6 , lowerCAmelCase_ : Any=2 , lowerCAmelCase_ : Tuple=0.02 , lowerCAmelCase_ : List[str]=3 , lowerCAmelCase_ : Optional[Any]=4 , lowerCAmelCase_ : List[str]=None , lowerCAmelCase_ : List[str]=0 , ) -> str: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_input_mask __lowerCAmelCase = use_token_type_ids __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = type_vocab_size __lowerCAmelCase = type_sequence_label_size __lowerCAmelCase = initializer_range __lowerCAmelCase = num_labels __lowerCAmelCase = num_choices __lowerCAmelCase = scope __lowerCAmelCase = projection_dim def lowercase ( self : Dict ) -> List[Any]: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase = None if self.use_input_mask: # follow test_modeling_tf_ctrl.py __lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCAmelCase = None if self.use_token_type_ids: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = None if self.use_labels: __lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCAmelCase = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCAmelCase_ , initializer_range=self.initializer_range , ) __lowerCAmelCase = DPRConfig(projection_dim=self.projection_dim , **config.to_dict() ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowercase ( self : str , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : str , lowerCAmelCase_ : str , lowerCAmelCase_ : Dict , lowerCAmelCase_ : str , lowerCAmelCase_ : int , lowerCAmelCase_ : Any ) -> str: __lowerCAmelCase = TFDPRContextEncoder(config=lowerCAmelCase_ ) __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ ) __lowerCAmelCase = model(lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ ) __lowerCAmelCase = model(lowerCAmelCase_ ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def lowercase ( self : List[str] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Any ) -> Tuple: __lowerCAmelCase = TFDPRQuestionEncoder(config=lowerCAmelCase_ ) __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ ) __lowerCAmelCase = model(lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ ) __lowerCAmelCase = model(lowerCAmelCase_ ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def lowercase ( self : Dict , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Tuple ) -> Union[str, Any]: __lowerCAmelCase = TFDPRReader(config=lowerCAmelCase_ ) __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.relevance_logits.shape , (self.batch_size,) ) def lowercase ( self : int ) -> Optional[int]: __lowerCAmelCase = self.prepare_config_and_inputs() ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = config_and_inputs __lowerCAmelCase = {'input_ids': input_ids} return config, inputs_dict @require_tf class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = ( ( TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) if is_tf_available() else () ) a_ = {"""feature-extraction""": TFDPRQuestionEncoder} if is_tf_available() else {} a_ = False a_ = False a_ = False a_ = False a_ = False def lowercase ( self : List[Any] ) -> Optional[int]: __lowerCAmelCase = TFDPRModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=lowerCAmelCase_ , hidden_size=3_7 ) def lowercase ( self : Optional[int] ) -> List[str]: self.config_tester.run_common_tests() def lowercase ( self : Any ) -> Tuple: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_context_encoder(*lowerCAmelCase_ ) def lowercase ( self : Any ) -> List[Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_question_encoder(*lowerCAmelCase_ ) def lowercase ( self : str ) -> List[str]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_reader(*lowerCAmelCase_ ) @slow def lowercase ( self : List[Any] ) -> Tuple: for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = TFDPRContextEncoder.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = TFDPRContextEncoder.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) for model_name in TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = TFDPRQuestionEncoder.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) for model_name in TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = TFDPRReader.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def lowercase ( self : Tuple ) -> Optional[Any]: __lowerCAmelCase = TFDPRQuestionEncoder.from_pretrained('facebook/dpr-question_encoder-single-nq-base' ) __lowerCAmelCase = tf.constant( [[1_0_1, 7_5_9_2, 1_0_1_0, 2_0_0_3, 2_0_2_6, 3_8_9_9, 1_0_1_4_0, 1_0_2_9, 1_0_2]] ) # [CLS] hello, is my dog cute? [SEP] __lowerCAmelCase = model(lowerCAmelCase_ )[0] # embedding shape = (1, 768) # compare the actual values for a slice. __lowerCAmelCase = tf.constant( [ [ 0.03_23_62_53, 0.12_75_33_35, 0.16_81_85_09, 0.00_27_97_86, 0.3_89_69_33, 0.24_26_49_45, 0.2_17_89_71, -0.02_33_52_27, -0.08_48_19_59, -0.14_32_41_17, ] ] ) self.assertTrue(numpy.allclose(output[:, :1_0].numpy() , expected_slice.numpy() , atol=1e-4 ) )
284
import re from filelock import FileLock try: import nltk _snake_case : Any = True except (ImportError, ModuleNotFoundError): _snake_case : Union[str, Any] = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def a_ ( lowerCAmelCase_ : str ): re.sub('<n>', '', lowerCAmelCase_ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(lowerCAmelCase_ ) )
284
1
import datasets from .nmt_bleu import compute_bleu # From: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py _snake_case : Dict = '\\n@INPROCEEDINGS{Papineni02bleu:a,\n author = {Kishore Papineni and Salim Roukos and Todd Ward and Wei-jing Zhu},\n title = {BLEU: a Method for Automatic Evaluation of Machine Translation},\n booktitle = {},\n year = {2002},\n pages = {311--318}\n}\n@inproceedings{lin-och-2004-orange,\n title = "{ORANGE}: a Method for Evaluating Automatic Evaluation Metrics for Machine Translation",\n author = "Lin, Chin-Yew and\n Och, Franz Josef",\n booktitle = "{COLING} 2004: Proceedings of the 20th International Conference on Computational Linguistics",\n month = "aug 23{--}aug 27",\n year = "2004",\n address = "Geneva, Switzerland",\n publisher = "COLING",\n url = "https://www.aclweb.org/anthology/C04-1072",\n pages = "501--507",\n}\n' _snake_case : str = '\\nBLEU (bilingual evaluation understudy) is an algorithm for evaluating the quality of text which has been machine-translated from one natural language to another.\nQuality is considered to be the correspondence between a machine\'s output and that of a human: "the closer a machine translation is to a professional human translation,\nthe better it is" – this is the central idea behind BLEU. BLEU was one of the first metrics to claim a high correlation with human judgements of quality, and\nremains one of the most popular automated and inexpensive metrics.\n\nScores are calculated for individual translated segments—generally sentences—by comparing them with a set of good quality reference translations.\nThose scores are then averaged over the whole corpus to reach an estimate of the translation\'s overall quality. Intelligibility or grammatical correctness\nare not taken into account[citation needed].\n\nBLEU\'s output is always a number between 0 and 1. This value indicates how similar the candidate text is to the reference texts, with values closer to 1\nrepresenting more similar texts. Few human translations will attain a score of 1, since this would indicate that the candidate is identical to one of the\nreference translations. For this reason, it is not necessary to attain a score of 1. Because there are more opportunities to match, adding additional\nreference translations will increase the BLEU score.\n' _snake_case : Any = '\nComputes BLEU score of translated segments against one or more references.\nArgs:\n predictions: list of translations to score.\n Each translation should be tokenized into a list of tokens.\n references: list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\n max_order: Maximum n-gram order to use when computing BLEU score.\n smooth: Whether or not to apply Lin et al. 2004 smoothing.\nReturns:\n \'bleu\': bleu score,\n \'precisions\': geometric mean of n-gram precisions,\n \'brevity_penalty\': brevity penalty,\n \'length_ratio\': ratio of lengths,\n \'translation_length\': translation_length,\n \'reference_length\': reference_length\nExamples:\n\n >>> predictions = [\n ... ["hello", "there", "general", "kenobi"], # tokenized prediction of the first sample\n ... ["foo", "bar", "foobar"] # tokenized prediction of the second sample\n ... ]\n >>> references = [\n ... [["hello", "there", "general", "kenobi"], ["hello", "there", "!"]], # tokenized references for the first sample (2 references)\n ... [["foo", "bar", "foobar"]] # tokenized references for the second sample (1 reference)\n ... ]\n >>> bleu = datasets.load_metric("bleu")\n >>> results = bleu.compute(predictions=predictions, references=references)\n >>> print(results["bleu"])\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): """simple docstring""" def lowercase ( self : int ) -> Any: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('string' , id='token' ) , id='sequence' ), 'references': datasets.Sequence( datasets.Sequence(datasets.Value('string' , id='token' ) , id='sequence' ) , id='references' ), } ) , codebase_urls=['https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py'] , reference_urls=[ 'https://en.wikipedia.org/wiki/BLEU', 'https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213', ] , ) def lowercase ( self : List[Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : Any , lowerCAmelCase_ : Union[str, Any]=4 , lowerCAmelCase_ : str=False ) -> Any: __lowerCAmelCase = compute_bleu( reference_corpus=lowerCAmelCase_ , translation_corpus=lowerCAmelCase_ , max_order=lowerCAmelCase_ , smooth=lowerCAmelCase_ ) ((__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase) , (__lowerCAmelCase)) = score return { "bleu": bleu, "precisions": precisions, "brevity_penalty": bp, "length_ratio": ratio, "translation_length": translation_length, "reference_length": reference_length, }
284
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case : List[Any] = { 'configuration_biogpt': ['BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BioGptConfig'], 'tokenization_biogpt': ['BioGptTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BioGptForCausalLM', 'BioGptForTokenClassification', 'BioGptForSequenceClassification', 'BioGptModel', 'BioGptPreTrainedModel', ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys _snake_case : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
1
def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('number must not be negative' ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
284
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets _snake_case : Tuple = '\\n@article{hendrycksmath2021,\n title={Measuring Mathematical Problem Solving With the MATH Dataset},\n author={Dan Hendrycks\n and Collin Burns\n and Saurav Kadavath\n and Akul Arora\n and Steven Basart\n and Eric Tang\n and Dawn Song\n and Jacob Steinhardt},\n journal={arXiv preprint arXiv:2103.03874},\n year={2021}\n}\n' _snake_case : str = '\\nThis metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset.\nIt first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy.\n' _snake_case : List[str] = R'\nCalculates accuracy after canonicalizing inputs.\n\nArgs:\n predictions: list of predictions to score. Each prediction\n is a string that contains natural language and LaTex.\n references: list of reference for each prediction. Each\n reference is a string that contains natural language\n and LaTex.\nReturns:\n accuracy: accuracy after canonicalizing inputs\n (e.g., converting "1/2" to "\\frac{1}{2}")\n\nExamples:\n >>> metric = datasets.load_metric("competition_math")\n >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"])\n >>> print(results)\n {\'accuracy\': 1.0}\n' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): """simple docstring""" def lowercase ( self : str ) -> int: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' ), 'references': datasets.Value('string' ), } ) , homepage='https://github.com/hendrycks/math' , codebase_urls=['https://github.com/hendrycks/math'] , ) def lowercase ( self : str , lowerCAmelCase_ : Any , lowerCAmelCase_ : Optional[int] ) -> List[Any]: __lowerCAmelCase = 0.0 for i, j in zip(lowerCAmelCase_ , lowerCAmelCase_ ): n_correct += 1.0 if math_equivalence.is_equiv(lowerCAmelCase_ , lowerCAmelCase_ ) else 0.0 __lowerCAmelCase = n_correct / len(lowerCAmelCase_ ) return { "accuracy": accuracy, }
284
1
from manim import * class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : List[str] ) -> Union[str, Any]: __lowerCAmelCase = Rectangle(height=0.5 , width=0.5 ) __lowerCAmelCase = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) __lowerCAmelCase = Rectangle(height=0.25 , width=0.25 ) __lowerCAmelCase = [mem.copy() for i in range(6 )] __lowerCAmelCase = [mem.copy() for i in range(6 )] __lowerCAmelCase = VGroup(*lowerCAmelCase_ ).arrange(lowerCAmelCase_ , buff=0 ) __lowerCAmelCase = VGroup(*lowerCAmelCase_ ).arrange(lowerCAmelCase_ , buff=0 ) __lowerCAmelCase = VGroup(lowerCAmelCase_ , lowerCAmelCase_ ).arrange(lowerCAmelCase_ , buff=0 ) __lowerCAmelCase = Text('CPU' , font_size=2_4 ) __lowerCAmelCase = Group(lowerCAmelCase_ , lowerCAmelCase_ ).arrange(lowerCAmelCase_ , buff=0.5 , aligned_edge=lowerCAmelCase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowerCAmelCase_ ) __lowerCAmelCase = [mem.copy() for i in range(4 )] __lowerCAmelCase = VGroup(*lowerCAmelCase_ ).arrange(lowerCAmelCase_ , buff=0 ) __lowerCAmelCase = Text('GPU' , font_size=2_4 ) __lowerCAmelCase = Group(lowerCAmelCase_ , lowerCAmelCase_ ).arrange(lowerCAmelCase_ , buff=0.5 , aligned_edge=lowerCAmelCase_ ) gpu.move_to([-1, -1, 0] ) self.add(lowerCAmelCase_ ) __lowerCAmelCase = [mem.copy() for i in range(6 )] __lowerCAmelCase = VGroup(*lowerCAmelCase_ ).arrange(lowerCAmelCase_ , buff=0 ) __lowerCAmelCase = Text('Model' , font_size=2_4 ) __lowerCAmelCase = Group(lowerCAmelCase_ , lowerCAmelCase_ ).arrange(lowerCAmelCase_ , buff=0.5 , aligned_edge=lowerCAmelCase_ ) model.move_to([3, -1.0, 0] ) self.add(lowerCAmelCase_ ) __lowerCAmelCase = [] __lowerCAmelCase = [] for i, rect in enumerate(lowerCAmelCase_ ): __lowerCAmelCase = fill.copy().set_fill(lowerCAmelCase_ , opacity=0.8 ) target.move_to(lowerCAmelCase_ ) model_arr.append(lowerCAmelCase_ ) __lowerCAmelCase = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(lowerCAmelCase_ , opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(lowerCAmelCase_ ) self.add(*lowerCAmelCase_ , *lowerCAmelCase_ ) __lowerCAmelCase = [meta_mem.copy() for i in range(6 )] __lowerCAmelCase = [meta_mem.copy() for i in range(6 )] __lowerCAmelCase = VGroup(*lowerCAmelCase_ ).arrange(lowerCAmelCase_ , buff=0 ) __lowerCAmelCase = VGroup(*lowerCAmelCase_ ).arrange(lowerCAmelCase_ , buff=0 ) __lowerCAmelCase = VGroup(lowerCAmelCase_ , lowerCAmelCase_ ).arrange(lowerCAmelCase_ , buff=0 ) __lowerCAmelCase = Text('Disk' , font_size=2_4 ) __lowerCAmelCase = Group(lowerCAmelCase_ , lowerCAmelCase_ ).arrange(lowerCAmelCase_ , buff=0.5 , aligned_edge=lowerCAmelCase_ ) disk.move_to([-4, -1.25, 0] ) self.add(lowerCAmelCase_ , lowerCAmelCase_ ) __lowerCAmelCase = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) __lowerCAmelCase = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=1_8 , ) key_text.move_to([-5, 2.4, 0] ) self.add(lowerCAmelCase_ , lowerCAmelCase_ ) __lowerCAmelCase = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=1_8 , ) blue_text.next_to(lowerCAmelCase_ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(lowerCAmelCase_ ) __lowerCAmelCase = MarkupText( f"""Now watch as an input is passed through the model\nand how the memory is utilized and handled.""" , font_size=2_4 , ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCAmelCase_ ) ) __lowerCAmelCase = Square(0.3 ) input.set_fill(lowerCAmelCase_ , opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] , lowerCAmelCase_ , buff=0.5 ) self.play(Write(lowerCAmelCase_ ) ) input.generate_target() input.target.next_to(model_arr[0] , direction=lowerCAmelCase_ , buff=0.02 ) self.play(MoveToTarget(lowerCAmelCase_ ) ) self.play(FadeOut(lowerCAmelCase_ ) ) __lowerCAmelCase = Arrow(start=lowerCAmelCase_ , end=lowerCAmelCase_ , color=lowerCAmelCase_ , buff=0.5 ) a.next_to(model_arr[0].get_left() , lowerCAmelCase_ , buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) __lowerCAmelCase = MarkupText( f"""As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back.""" , font_size=2_4 , ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCAmelCase_ , run_time=3 ) ) __lowerCAmelCase = {'run_time': 1, 'fade_in': True, 'fade_out': True, 'buff': 0.02} self.play( Write(lowerCAmelCase_ ) , Circumscribe(model_arr[0] , color=lowerCAmelCase_ , **lowerCAmelCase_ ) , Circumscribe(model_cpu_arr[0] , color=lowerCAmelCase_ , **lowerCAmelCase_ ) , Circumscribe(gpu_rect[0] , color=lowerCAmelCase_ , **lowerCAmelCase_ ) , ) self.play(MoveToTarget(model_cpu_arr[0] ) ) __lowerCAmelCase = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 , lowerCAmelCase_ , buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) __lowerCAmelCase = AnimationGroup( FadeOut(lowerCAmelCase_ , run_time=0.5 ) , MoveToTarget(lowerCAmelCase_ , run_time=0.5 ) , FadeIn(lowerCAmelCase_ , run_time=0.5 ) , lag_ratio=0.2 ) self.play(lowerCAmelCase_ ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: __lowerCAmelCase = 0.7 self.play( Circumscribe(model_arr[i] , **lowerCAmelCase_ ) , Circumscribe(cpu_left_col_base[i] , **lowerCAmelCase_ ) , Circumscribe(cpu_left_col_base[i + 1] , color=lowerCAmelCase_ , **lowerCAmelCase_ ) , Circumscribe(gpu_rect[0] , color=lowerCAmelCase_ , **lowerCAmelCase_ ) , Circumscribe(model_arr[i + 1] , color=lowerCAmelCase_ , **lowerCAmelCase_ ) , ) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) , MoveToTarget(model_cpu_arr[i + 1] ) , ) else: self.play( MoveToTarget(model_cpu_arr[i] , run_time=0.7 ) , MoveToTarget(model_cpu_arr[i + 1] , run_time=0.7 ) , ) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() , RIGHT + 0.02 , buff=0.2 ) self.play( Circumscribe(model_arr[-1] , color=lowerCAmelCase_ , **lowerCAmelCase_ ) , Circumscribe(cpu_left_col_base[-1] , color=lowerCAmelCase_ , **lowerCAmelCase_ ) , Circumscribe(gpu_rect[0] , color=lowerCAmelCase_ , **lowerCAmelCase_ ) , ) self.play(MoveToTarget(model_cpu_arr[i] ) ) __lowerCAmelCase = a_c __lowerCAmelCase = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] , RIGHT + 0.02 , buff=0.5 ) self.play( FadeOut(lowerCAmelCase_ ) , FadeOut(lowerCAmelCase_ , run_time=0.5 ) , ) __lowerCAmelCase = MarkupText(f"""Inference on a model too large for GPU memory\nis successfully completed.""" , font_size=2_4 ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCAmelCase_ , run_time=3 ) , MoveToTarget(lowerCAmelCase_ ) ) self.wait()
284
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = StableDiffusionInstructPixaPixPipeline a_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width""", """cross_attention_kwargs"""} a_ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase ( self : Optional[int] ) -> Optional[int]: torch.manual_seed(0 ) __lowerCAmelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=8 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , ) __lowerCAmelCase = PNDMScheduler(skip_prk_steps=lowerCAmelCase_ ) torch.manual_seed(0 ) __lowerCAmelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) __lowerCAmelCase = CLIPTextModel(lowerCAmelCase_ ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __lowerCAmelCase = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowercase ( self : Optional[Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Tuple=0 ) -> Dict: __lowerCAmelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) __lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCAmelCase = Image.fromarray(np.uinta(lowerCAmelCase_ ) ).convert('RGB' ) if str(lowerCAmelCase_ ).startswith('mps' ): __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) else: __lowerCAmelCase = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'image_guidance_scale': 1, 'output_type': 'numpy', } return inputs def lowercase ( self : Tuple ) -> List[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_26, 0.37_50, 0.45_47, 0.61_17, 0.58_66, 0.50_16, 0.43_27, 0.56_42, 0.48_15] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Dict: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = 'french fries' __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ , negative_prompt=lowerCAmelCase_ ) __lowerCAmelCase = output.images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_11, 0.36_42, 0.45_53, 0.62_36, 0.57_97, 0.50_13, 0.43_43, 0.56_11, 0.48_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Any: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = [inputs['prompt']] * 2 __lowerCAmelCase = np.array(inputs['image'] ).astype(np.floataa ) / 2_55.0 __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ).unsqueeze(0 ).to(lowerCAmelCase_ ) __lowerCAmelCase = image / 2 + 0.5 __lowerCAmelCase = image.permute(0 , 3 , 1 , 2 ) __lowerCAmelCase = image.repeat(2 , 1 , 1 , 1 ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[-1, -3:, -3:, -1] assert image.shape == (2, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.58_12, 0.57_48, 0.52_22, 0.59_08, 0.56_95, 0.71_74, 0.68_04, 0.55_23, 0.55_79] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Dict ) -> Optional[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = EulerAncestralDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' ) __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = [round(lowerCAmelCase_ , 4 ) for x in image_slice.flatten().tolist()] print(','.join([str(lowerCAmelCase_ ) for x in slice] ) ) assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.74_17, 0.38_42, 0.47_32, 0.57_76, 0.58_91, 0.51_39, 0.40_52, 0.56_73, 0.49_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> Optional[int]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def lowercase ( self : Optional[Any] ) -> Optional[Any]: __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = VaeImageProcessor(do_resize=lowerCAmelCase_ , do_normalize=lowerCAmelCase_ ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = pipe(**self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) )[0] __lowerCAmelCase = components['vae'] __lowerCAmelCase = self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): __lowerCAmelCase = vae.encode(inputs[image_param] ).latent_dist.mode() __lowerCAmelCase = pipe(**lowerCAmelCase_ )[0] __lowerCAmelCase = np.abs(out - out_latents_inputs ).max() self.assertLess(lowerCAmelCase_ , 1e-4 , 'passing latents as image input generate different result from passing image' ) @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> Optional[int]: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase ( self : List[str] , lowerCAmelCase_ : List[Any]=0 ) -> Any: __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = load_image( 'https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg' ) __lowerCAmelCase = { 'prompt': 'turn him into a cyborg', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'image_guidance_scale': 1.0, 'output_type': 'numpy', } return inputs def lowercase ( self : List[Any] ) -> str: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.59_02, 0.60_15, 0.60_27, 0.59_83, 0.60_92, 0.60_61, 0.57_65, 0.57_85, 0.55_55] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Tuple ) -> List[str]: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.65_78, 0.68_17, 0.69_72, 0.67_61, 0.68_56, 0.69_16, 0.64_28, 0.65_16, 0.63_01] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[Any] ) -> Dict: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.38_28, 0.38_34, 0.38_18, 0.37_92, 0.38_65, 0.37_52, 0.37_92, 0.38_47, 0.37_53] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> int: __lowerCAmelCase = 0 def callback_fn(lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : torch.FloatTensor ) -> None: __lowerCAmelCase = True nonlocal number_of_steps number_of_steps += 1 if step == 1: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.24_63, -0.46_44, -0.97_56, 1.51_76, 1.44_14, 0.78_66, 0.98_97, 0.85_21, 0.79_83] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.26_44, -0.46_26, -0.96_53, 1.51_76, 1.45_51, 0.76_86, 0.98_05, 0.84_52, 0.81_15] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 __lowerCAmelCase = False __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() pipe(**lowerCAmelCase_ , callback=lowerCAmelCase_ , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def lowercase ( self : Optional[int] ) -> Any: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 1_0**9 def lowercase ( self : List[Any] ) -> Any: __lowerCAmelCase = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 __lowerCAmelCase = inputs['image'].resize((5_0_4, 5_0_4) ) __lowerCAmelCase = 'timbrooks/instruct-pix2pix' __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( lowerCAmelCase_ , safety_checker=lowerCAmelCase_ , ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = output.images[0] __lowerCAmelCase = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 5_0_4, 3) __lowerCAmelCase = np.array([0.27_26, 0.25_29, 0.26_64, 0.26_55, 0.26_41, 0.26_42, 0.25_91, 0.26_49, 0.25_90] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3
284
1
import os import re from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _snake_case : Dict = logging.get_logger(__name__) _snake_case : List[Any] = {'vocab_file': 'spiece.model'} _snake_case : int = { 'vocab_file': { 'google/bigbird-roberta-base': 'https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model', 'google/bigbird-roberta-large': ( 'https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model' ), 'google/bigbird-base-trivia-itc': ( 'https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model' ), } } _snake_case : Dict = { 'google/bigbird-roberta-base': 4096, 'google/bigbird-roberta-large': 4096, 'google/bigbird-base-trivia-itc': 4096, } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = VOCAB_FILES_NAMES a_ = PRETRAINED_VOCAB_FILES_MAP a_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ = ["""input_ids""", """attention_mask"""] a_ = [] def __init__( self : Dict , lowerCAmelCase_ : Any , lowerCAmelCase_ : List[Any]="<unk>" , lowerCAmelCase_ : Optional[Any]="<s>" , lowerCAmelCase_ : int="</s>" , lowerCAmelCase_ : Optional[Any]="<pad>" , lowerCAmelCase_ : Dict="[SEP]" , lowerCAmelCase_ : Optional[int]="[MASK]" , lowerCAmelCase_ : List[str]="[CLS]" , lowerCAmelCase_ : Optional[Dict[str, Any]] = None , **lowerCAmelCase_ : Optional[int] , ) -> None: __lowerCAmelCase = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else bos_token __lowerCAmelCase = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else eos_token __lowerCAmelCase = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else unk_token __lowerCAmelCase = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else pad_token __lowerCAmelCase = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else cls_token __lowerCAmelCase = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it __lowerCAmelCase = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else mask_token __lowerCAmelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowerCAmelCase_ , eos_token=lowerCAmelCase_ , unk_token=lowerCAmelCase_ , pad_token=lowerCAmelCase_ , sep_token=lowerCAmelCase_ , mask_token=lowerCAmelCase_ , cls_token=lowerCAmelCase_ , sp_model_kwargs=self.sp_model_kwargs , **lowerCAmelCase_ , ) __lowerCAmelCase = vocab_file __lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCAmelCase_ ) @property def lowercase ( self : List[Any] ) -> Optional[Any]: return self.sp_model.get_piece_size() def lowercase ( self : Optional[int] ) -> Optional[Any]: __lowerCAmelCase = {self.convert_ids_to_tokens(lowerCAmelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Dict ) -> List[str]: __lowerCAmelCase = self.__dict__.copy() __lowerCAmelCase = None return state def __setstate__( self : str , lowerCAmelCase_ : int ) -> Optional[int]: __lowerCAmelCase = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): __lowerCAmelCase = {} __lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowercase ( self : Any , lowerCAmelCase_ : str ) -> List[str]: return self.sp_model.encode(lowerCAmelCase_ , out_type=lowerCAmelCase_ ) def lowercase ( self : str , lowerCAmelCase_ : Optional[int] ) -> int: return self.sp_model.piece_to_id(lowerCAmelCase_ ) def lowercase ( self : Optional[Any] , lowerCAmelCase_ : Optional[int] ) -> Any: __lowerCAmelCase = self.sp_model.IdToPiece(lowerCAmelCase_ ) return token def lowercase ( self : Any , lowerCAmelCase_ : List[Any] ) -> Dict: __lowerCAmelCase = [] __lowerCAmelCase = '' __lowerCAmelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowerCAmelCase_ ) + token __lowerCAmelCase = True __lowerCAmelCase = [] else: current_sub_tokens.append(lowerCAmelCase_ ) __lowerCAmelCase = False out_string += self.sp_model.decode(lowerCAmelCase_ ) return out_string.strip() def lowercase ( self : Tuple , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = None , lowerCAmelCase_ : bool = True , **lowerCAmelCase_ : Tuple , ) -> str: __lowerCAmelCase = kwargs.pop('use_source_tokenizer' , lowerCAmelCase_ ) __lowerCAmelCase = self.convert_ids_to_tokens(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 __lowerCAmelCase = [] __lowerCAmelCase = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(lowerCAmelCase_ ) ) __lowerCAmelCase = [] sub_texts.append(lowerCAmelCase_ ) else: current_sub_text.append(lowerCAmelCase_ ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(lowerCAmelCase_ ) ) # Mimic the behavior of the Rust tokenizer: # No space before [MASK] and [SEP] if spaces_between_special_tokens: __lowerCAmelCase = re.sub(R' (\[(MASK|SEP)\])' , R'\1' , ' '.join(lowerCAmelCase_ ) ) else: __lowerCAmelCase = ''.join(lowerCAmelCase_ ) __lowerCAmelCase = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: __lowerCAmelCase = self.clean_up_tokenization(lowerCAmelCase_ ) return clean_text else: return text def lowercase ( self : List[str] , lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCAmelCase = os.path.join( lowerCAmelCase_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCAmelCase_ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCAmelCase_ , 'wb' ) as fi: __lowerCAmelCase = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase_ ) return (out_vocab_file,) def lowercase ( self : List[Any] , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __lowerCAmelCase = [self.cls_token_id] __lowerCAmelCase = [self.sep_token_id] return cls + token_ids_a + sep + token_ids_a + sep def lowercase ( self : Dict , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : Optional[List[int]] = None , lowerCAmelCase_ : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase_ , token_ids_a=lowerCAmelCase_ , already_has_special_tokens=lowerCAmelCase_ ) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase_ )) + [1] return [1] + ([0] * len(lowerCAmelCase_ )) + [1] + ([0] * len(lowerCAmelCase_ )) + [1] def lowercase ( self : Dict , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : Optional[List[int]] = None ) -> List[int]: __lowerCAmelCase = [self.sep_token_id] __lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1]
284
from timeit import timeit def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('the value of input must not be negative' ) __lowerCAmelCase = 0 while number: number &= number - 1 result += 1 return result def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('the value of input must not be negative' ) __lowerCAmelCase = 0 while number: if number % 2 == 1: result += 1 number >>= 1 return result def a_ ( ): def do_benchmark(lowerCAmelCase_ : int ) -> None: __lowerCAmelCase = 'import __main__ as z' print(F"""Benchmark when {number = }:""" ) print(F"""{get_set_bits_count_using_modulo_operator(lowerCAmelCase_ ) = }""" ) __lowerCAmelCase = timeit('z.get_set_bits_count_using_modulo_operator(25)', setup=lowerCAmelCase_ ) print(F"""timeit() runs in {timing} seconds""" ) print(F"""{get_set_bits_count_using_brian_kernighans_algorithm(lowerCAmelCase_ ) = }""" ) __lowerCAmelCase = timeit( 'z.get_set_bits_count_using_brian_kernighans_algorithm(25)', setup=lowerCAmelCase_, ) print(F"""timeit() runs in {timing} seconds""" ) for number in (25, 37, 58, 0): do_benchmark(lowerCAmelCase_ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
284
1
from dataclasses import dataclass, field from typing import Optional from transformers import AutoConfig, AutoImageProcessor, AutoTokenizer, FlaxVisionEncoderDecoderModel, HfArgumentParser @dataclass class _UpperCAmelCase : """simple docstring""" a_ = field( metadata={"""help""": """The output directory where the model will be written."""} , ) a_ = field( metadata={ """help""": ( """The encoder model checkpoint for weights initialization.""" """Don't set if you want to train an encoder model from scratch.""" ) } , ) a_ = field( metadata={ """help""": ( """The decoder model checkpoint for weights initialization.""" """Don't set if you want to train a decoder model from scratch.""" ) } , ) a_ = field( default=_UpperCamelCase , metadata={"""help""": """Pretrained encoder config name or path if not the same as encoder_model_name"""} ) a_ = field( default=_UpperCamelCase , metadata={"""help""": """Pretrained decoder config name or path if not the same as decoder_model_name"""} ) def a_ ( ): __lowerCAmelCase = HfArgumentParser((ModelArguments,) ) ((__lowerCAmelCase) , ) = parser.parse_args_into_dataclasses() # Load pretrained model and tokenizer # Use explicit specified encoder config if model_args.encoder_config_name: __lowerCAmelCase = AutoConfig.from_pretrained(model_args.encoder_config_name ) # Use pretrained encoder model's config else: __lowerCAmelCase = AutoConfig.from_pretrained(model_args.encoder_model_name_or_path ) # Use explicit specified decoder config if model_args.decoder_config_name: __lowerCAmelCase = AutoConfig.from_pretrained(model_args.decoder_config_name ) # Use pretrained decoder model's config else: __lowerCAmelCase = AutoConfig.from_pretrained(model_args.decoder_model_name_or_path ) # necessary for `from_encoder_decoder_pretrained` when `decoder_config` is passed __lowerCAmelCase = True __lowerCAmelCase = True __lowerCAmelCase = FlaxVisionEncoderDecoderModel.from_encoder_decoder_pretrained( encoder_pretrained_model_name_or_path=model_args.encoder_model_name_or_path, decoder_pretrained_model_name_or_path=model_args.decoder_model_name_or_path, encoder_config=lowerCAmelCase_, decoder_config=lowerCAmelCase_, ) # GPT2 only has bos/eos tokens but not decoder_start/pad tokens __lowerCAmelCase = decoder_config.decoder_start_token_id __lowerCAmelCase = decoder_config.pad_token_id if decoder_start_token_id is None: __lowerCAmelCase = decoder_config.bos_token_id if pad_token_id is None: __lowerCAmelCase = decoder_config.eos_token_id # This is necessary to make Flax's generate() work __lowerCAmelCase = decoder_config.eos_token_id __lowerCAmelCase = decoder_start_token_id __lowerCAmelCase = pad_token_id __lowerCAmelCase = AutoImageProcessor.from_pretrained(model_args.encoder_model_name_or_path ) __lowerCAmelCase = AutoTokenizer.from_pretrained(model_args.decoder_model_name_or_path ) __lowerCAmelCase = tokenizer.convert_ids_to_tokens(model.config.pad_token_id ) model.save_pretrained(model_args.output_dir ) image_processor.save_pretrained(model_args.output_dir ) tokenizer.save_pretrained(model_args.output_dir ) if __name__ == "__main__": main()
284
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss _snake_case : Dict = pytest.mark.integration @require_faiss class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : List[Any] ) -> Optional[Any]: __lowerCAmelCase = Dataset.from_dict({'filename': ['my_name-train' + '_' + str(lowerCAmelCase_ ) for x in np.arange(3_0 ).tolist()]} ) return dset def lowercase ( self : List[str] ) -> Tuple: import faiss __lowerCAmelCase = self._create_dummy_dataset() __lowerCAmelCase = dset.map( lambda lowerCAmelCase_ , lowerCAmelCase_ : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=lowerCAmelCase_ , keep_in_memory=lowerCAmelCase_ ) __lowerCAmelCase = dset.add_faiss_index('vecs' , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) dset.drop_index('vecs' ) def lowercase ( self : Optional[Any] ) -> str: import faiss __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT , ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def lowercase ( self : int ) -> Optional[Any]: import faiss __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCAmelCase_ ) as tmp_file: dset.save_faiss_index('vecs' , tmp_file.name ) dset.load_faiss_index('vecs2' , tmp_file.name ) os.unlink(tmp_file.name ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('vecs2' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) def lowercase ( self : Union[str, Any] ) -> List[Any]: __lowerCAmelCase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name='vecs' ) dset.drop_index('vecs' ) self.assertRaises(lowerCAmelCase_ , partial(dset.get_nearest_examples , 'vecs2' , np.ones(5 , dtype=np.floataa ) ) ) def lowercase ( self : Union[str, Any] ) -> Tuple: from elasticsearch import Elasticsearch __lowerCAmelCase = self._create_dummy_dataset() with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: __lowerCAmelCase = {'acknowledged': True} mocked_bulk.return_value([(True, None)] * 3_0 ) __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 2_9}]}} __lowerCAmelCase = Elasticsearch() dset.add_elasticsearch_index('filename' , es_client=lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase = dset.get_nearest_examples('filename' , 'my_name-train_29' ) self.assertEqual(examples['filename'][0] , 'my_name-train_29' ) @require_faiss class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : str ) -> int: import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 1_0 ) # single query __lowerCAmelCase = np.zeros(5 , dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertRaises(lowerCAmelCase_ , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries __lowerCAmelCase = np.eye(5 , dtype=np.floataa )[::-1] __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ ) self.assertRaises(lowerCAmelCase_ , index.search_batch , queries[0] ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , lowerCAmelCase_ ) def lowercase ( self : List[Any] ) -> List[str]: import faiss __lowerCAmelCase = FaissIndex(string_factory='Flat' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) __lowerCAmelCase = FaissIndex(string_factory='LSH' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(lowerCAmelCase_ ): __lowerCAmelCase = FaissIndex(string_factory='Flat' , custom_index=faiss.IndexFlat(5 ) ) def lowercase ( self : Union[str, Any] ) -> Dict: import faiss __lowerCAmelCase = faiss.IndexFlat(5 ) __lowerCAmelCase = FaissIndex(custom_index=lowerCAmelCase_ ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def lowercase ( self : str ) -> Any: import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowerCAmelCase_ ) as tmp_file: index.save(tmp_file.name ) __lowerCAmelCase = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) __lowerCAmelCase = np.zeros(5 , dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def a_ ( lowerCAmelCase_ : Union[str, Any] ): import faiss __lowerCAmelCase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5, dtype=np.floataa ) ) __lowerCAmelCase = 'index.faiss' __lowerCAmelCase = F"""mock://{index_name}""" index.save(lowerCAmelCase_, storage_options=mockfs.storage_options ) __lowerCAmelCase = FaissIndex.load(lowerCAmelCase_, storage_options=mockfs.storage_options ) __lowerCAmelCase = np.zeros(5, dtype=np.floataa ) __lowerCAmelCase = 1 __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : Any ) -> int: from elasticsearch import Elasticsearch with patch('elasticsearch.Elasticsearch.search' ) as mocked_search, patch( 'elasticsearch.client.IndicesClient.create' ) as mocked_index_create, patch('elasticsearch.helpers.streaming_bulk' ) as mocked_bulk: __lowerCAmelCase = Elasticsearch() __lowerCAmelCase = {'acknowledged': True} __lowerCAmelCase = ElasticSearchIndex(es_client=lowerCAmelCase_ ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(['foo', 'bar', 'foobar'] ) # single query __lowerCAmelCase = 'foo' __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout __lowerCAmelCase = 'foo' __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 0}]}} __lowerCAmelCase , __lowerCAmelCase = index.search(lowerCAmelCase_ , request_timeout=3_0 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries __lowerCAmelCase = ['foo', 'bar', 'foobar'] __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([1, 1, 1] , lowerCAmelCase_ ) # batched queries with timeout __lowerCAmelCase = ['foo', 'bar', 'foobar'] __lowerCAmelCase = {'hits': {'hits': [{'_score': 1, '_id': 1}]}} __lowerCAmelCase , __lowerCAmelCase = index.search_batch(lowerCAmelCase_ , request_timeout=3_0 ) __lowerCAmelCase = [scores[0] for scores in total_scores] __lowerCAmelCase = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowerCAmelCase_ ) , 0 ) self.assertListEqual([1, 1, 1] , lowerCAmelCase_ )
284
1
def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : str ): if len(lowerCAmelCase_ ) != len(lowerCAmelCase_ ): raise ValueError('String lengths must match!' ) __lowerCAmelCase = 0 for chara, chara in zip(lowerCAmelCase_, lowerCAmelCase_ ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
284
import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( 'kwargs, expected', [ ({'num_shards': 0, 'max_num_jobs': 1}, []), ({'num_shards': 10, 'max_num_jobs': 1}, [range(10 )]), ({'num_shards': 10, 'max_num_jobs': 10}, [range(lowerCAmelCase_, i + 1 ) for i in range(10 )]), ({'num_shards': 1, 'max_num_jobs': 10}, [range(1 )]), ({'num_shards': 10, 'max_num_jobs': 3}, [range(0, 4 ), range(4, 7 ), range(7, 10 )]), ({'num_shards': 3, 'max_num_jobs': 10}, [range(0, 1 ), range(1, 2 ), range(2, 3 )]), ], ) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : Any ): __lowerCAmelCase = _distribute_shards(**lowerCAmelCase_ ) assert out == expected @pytest.mark.parametrize( 'gen_kwargs, max_num_jobs, expected', [ ({'foo': 0}, 10, [{'foo': 0}]), ({'shards': [0, 1, 2, 3]}, 1, [{'shards': [0, 1, 2, 3]}]), ({'shards': [0, 1, 2, 3]}, 4, [{'shards': [0]}, {'shards': [1]}, {'shards': [2]}, {'shards': [3]}]), ({'shards': [0, 1]}, 4, [{'shards': [0]}, {'shards': [1]}]), ({'shards': [0, 1, 2, 3]}, 2, [{'shards': [0, 1]}, {'shards': [2, 3]}]), ], ) def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : List[str], lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = _split_gen_kwargs(lowerCAmelCase_, lowerCAmelCase_ ) assert out == expected @pytest.mark.parametrize( 'gen_kwargs, expected', [ ({'foo': 0}, 1), ({'shards': [0]}, 1), ({'shards': [0, 1, 2, 3]}, 4), ({'shards': [0, 1, 2, 3], 'foo': 0}, 4), ({'shards': [0, 1, 2, 3], 'other': (0, 1)}, 4), ({'shards': [0, 1, 2, 3], 'shards2': [0, 1]}, RuntimeError), ], ) def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : Any ): if expected is RuntimeError: with pytest.raises(lowerCAmelCase_ ): _number_of_shards_in_gen_kwargs(lowerCAmelCase_ ) else: __lowerCAmelCase = _number_of_shards_in_gen_kwargs(lowerCAmelCase_ ) assert out == expected
284
1
from statistics import mean, stdev def a_ ( lowerCAmelCase_ : list, lowerCAmelCase_ : int = 3 ): __lowerCAmelCase = min(lowerCAmelCase_ ) __lowerCAmelCase = max(lowerCAmelCase_ ) # normalize data return [round((x - x_min) / (x_max - x_min), lowerCAmelCase_ ) for x in data] def a_ ( lowerCAmelCase_ : list, lowerCAmelCase_ : int = 3 ): __lowerCAmelCase = mean(lowerCAmelCase_ ) __lowerCAmelCase = stdev(lowerCAmelCase_ ) # standardize data return [round((x - mu) / (sigma), lowerCAmelCase_ ) for x in data]
284
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class _UpperCAmelCase ( _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = BertJapaneseTokenizer a_ = False a_ = True def lowercase ( self : Optional[Any] ) -> List[str]: super().setUp() __lowerCAmelCase = [ '[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは', '世界', '##世界', '、', '##、', '。', '##。', ] __lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def lowercase ( self : List[Any] , lowerCAmelCase_ : Tuple ) -> str: __lowerCAmelCase = 'こんにちは、世界。 \nこんばんは、世界。' __lowerCAmelCase = 'こんにちは 、 世界 。 こんばんは 、 世界 。' return input_text, output_text def lowercase ( self : List[Any] , lowerCAmelCase_ : str ) -> Dict: __lowerCAmelCase , __lowerCAmelCase = self.get_input_output_texts(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(lowerCAmelCase_ , clean_up_tokenization_spaces=lowerCAmelCase_ ) return text, ids def lowercase ( self : List[str] ) -> Optional[int]: pass # TODO add if relevant def lowercase ( self : Optional[Any] ) -> Optional[Any]: pass # TODO add if relevant def lowercase ( self : Union[str, Any] ) -> Any: pass # TODO add if relevant def lowercase ( self : Dict ) -> Tuple: __lowerCAmelCase = self.tokenizer_class(self.vocab_file ) __lowerCAmelCase = tokenizer.tokenize('こんにちは、世界。\nこんばんは、世界。' ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) def lowercase ( self : List[str] ) -> List[str]: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='mecab' ) self.assertIsNotNone(lowerCAmelCase_ ) __lowerCAmelCase = 'こんにちは、世界。\nこんばんは、世界。' __lowerCAmelCase = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __lowerCAmelCase = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowerCAmelCase_ , 'wb' ) as handle: pickle.dump(lowerCAmelCase_ , lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'rb' ) as handle: __lowerCAmelCase = pickle.load(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer_new.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Dict ) -> Tuple: __lowerCAmelCase = MecabTokenizer(mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : List[Any] ) -> int: try: __lowerCAmelCase = MecabTokenizer(mecab_dic='unidic_lite' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : Tuple ) -> Optional[Any]: try: __lowerCAmelCase = MecabTokenizer(mecab_dic='unidic' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : Tuple ) -> Union[str, Any]: __lowerCAmelCase = MecabTokenizer(do_lower_case=lowerCAmelCase_ , mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iphone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def lowercase ( self : Union[str, Any] ) -> Optional[Any]: try: __lowerCAmelCase = MecabTokenizer( do_lower_case=lowerCAmelCase_ , normalize_text=lowerCAmelCase_ , mecab_option='-d /usr/local/lib/mecab/dic/jumandic' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れた', '\u3000', '。'] , ) def lowercase ( self : Any ) -> Union[str, Any]: __lowerCAmelCase = MecabTokenizer(normalize_text=lowerCAmelCase_ , mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', ' ', '。'] , ) @require_sudachi def lowercase ( self : List[str] ) -> List[str]: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='sudachi' ) self.assertIsNotNone(lowerCAmelCase_ ) __lowerCAmelCase = 'こんにちは、世界。\nこんばんは、世界。' __lowerCAmelCase = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __lowerCAmelCase = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowerCAmelCase_ , 'wb' ) as handle: pickle.dump(lowerCAmelCase_ , lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'rb' ) as handle: __lowerCAmelCase = pickle.load(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer_new.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) @require_sudachi def lowercase ( self : Union[str, Any] ) -> List[str]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iPhone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', ' ', '。', ' ', ' '] , ) @require_sudachi def lowercase ( self : Tuple ) -> Optional[Any]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='A' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国', '人', '参政', '権'] ) @require_sudachi def lowercase ( self : Tuple ) -> List[Any]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='B' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国人', '参政権'] ) @require_sudachi def lowercase ( self : List[str] ) -> Union[str, Any]: __lowerCAmelCase = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='C' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国人参政権'] ) @require_sudachi def lowercase ( self : Dict ) -> List[str]: __lowerCAmelCase = SudachiTokenizer(do_lower_case=lowerCAmelCase_ , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iphone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', ' ', '。', ' ', ' '] , ) @require_sudachi def lowercase ( self : Union[str, Any] ) -> List[Any]: __lowerCAmelCase = SudachiTokenizer(normalize_text=lowerCAmelCase_ , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iPhone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', '\u3000', '。', ' ', ' '] , ) @require_sudachi def lowercase ( self : int ) -> str: __lowerCAmelCase = SudachiTokenizer(trim_whitespace=lowerCAmelCase_ , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) @require_jumanpp def lowercase ( self : Union[str, Any] ) -> Any: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type='jumanpp' ) self.assertIsNotNone(lowerCAmelCase_ ) __lowerCAmelCase = 'こんにちは、世界。\nこんばんは、世界。' __lowerCAmelCase = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __lowerCAmelCase = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowerCAmelCase_ , 'wb' ) as handle: pickle.dump(lowerCAmelCase_ , lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'rb' ) as handle: __lowerCAmelCase = pickle.load(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer_new.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) @require_jumanpp def lowercase ( self : List[Any] ) -> Optional[Any]: __lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def lowercase ( self : Any ) -> Union[str, Any]: __lowerCAmelCase = JumanppTokenizer(do_lower_case=lowerCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iphone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def lowercase ( self : Dict ) -> Dict: __lowerCAmelCase = JumanppTokenizer(normalize_text=lowerCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['ア', 'ッ', 'フ', '゚', 'ル', 'ストア', 'で', 'iPhone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def lowercase ( self : List[str] ) -> List[str]: __lowerCAmelCase = JumanppTokenizer(trim_whitespace=lowerCAmelCase_ ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れた', '。'] , ) @require_jumanpp def lowercase ( self : Any ) -> Any: __lowerCAmelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('ありがとうございますm(_ _)m見つけるのが大変です。' ) , ['ありがとう', 'ございます', 'm(_ _)m', '見つける', 'の', 'が', '大変です', '。'] , ) def lowercase ( self : Any ) -> str: __lowerCAmelCase = ['[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは'] __lowerCAmelCase = {} for i, token in enumerate(lowerCAmelCase_ ): __lowerCAmelCase = i __lowerCAmelCase = WordpieceTokenizer(vocab=lowerCAmelCase_ , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('こんにちは' ) , ['こんにちは'] ) self.assertListEqual(tokenizer.tokenize('こんばんは' ) , ['こん', '##ばんは'] ) self.assertListEqual(tokenizer.tokenize('こんばんは こんばんにちは こんにちは' ) , ['こん', '##ばんは', '[UNK]', 'こんにちは'] ) def lowercase ( self : List[Any] ) -> Tuple: __lowerCAmelCase = BertJapaneseTokenizer.from_pretrained('nlp-waseda/roberta-base-japanese-with-auto-jumanpp' ) __lowerCAmelCase = tokenizer.subword_tokenizer __lowerCAmelCase = subword_tokenizer.tokenize('国境 の 長い トンネル を 抜ける と 雪国 であった 。' ) self.assertListEqual(lowerCAmelCase_ , ['▁国境', '▁の', '▁長い', '▁トンネル', '▁を', '▁抜ける', '▁と', '▁雪', '国', '▁であった', '▁。'] ) __lowerCAmelCase = subword_tokenizer.tokenize('こんばんは こんばん にち は こんにちは' ) self.assertListEqual(lowerCAmelCase_ , ['▁こん', 'ばん', 'は', '▁こん', 'ばん', '▁に', 'ち', '▁は', '▁こんにちは'] ) def lowercase ( self : int ) -> str: __lowerCAmelCase = self.tokenizer_class.from_pretrained('cl-tohoku/bert-base-japanese' ) __lowerCAmelCase = tokenizer.encode('ありがとう。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.encode('どういたしまして。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ , lowerCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class _UpperCAmelCase ( _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = BertJapaneseTokenizer a_ = False def lowercase ( self : Optional[Any] ) -> Tuple: super().setUp() __lowerCAmelCase = ['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] __lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def lowercase ( self : str , **lowerCAmelCase_ : Tuple ) -> Union[str, Any]: return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='character' , **lowerCAmelCase_ ) def lowercase ( self : Tuple , lowerCAmelCase_ : Tuple ) -> Optional[int]: __lowerCAmelCase = 'こんにちは、世界。 \nこんばんは、世界。' __lowerCAmelCase = 'こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。' return input_text, output_text def lowercase ( self : Dict ) -> str: pass # TODO add if relevant def lowercase ( self : Any ) -> str: pass # TODO add if relevant def lowercase ( self : List[Any] ) -> int: pass # TODO add if relevant def lowercase ( self : str ) -> str: __lowerCAmelCase = self.tokenizer_class(self.vocab_file , subword_tokenizer_type='character' ) __lowerCAmelCase = tokenizer.tokenize('こんにちは、世界。 \nこんばんは、世界。' ) self.assertListEqual( lowerCAmelCase_ , ['こ', 'ん', 'に', 'ち', 'は', '、', '世', '界', '。', 'こ', 'ん', 'ば', 'ん', 'は', '、', '世', '界', '。'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) , [3, 4, 5, 6, 7, 1_1, 9, 1_0, 1_2, 3, 4, 8, 4, 7, 1_1, 9, 1_0, 1_2] ) def lowercase ( self : str ) -> Optional[int]: __lowerCAmelCase = ['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] __lowerCAmelCase = {} for i, token in enumerate(lowerCAmelCase_ ): __lowerCAmelCase = i __lowerCAmelCase = CharacterTokenizer(vocab=lowerCAmelCase_ , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('こんにちは' ) , ['こ', 'ん', 'に', 'ち', 'は'] ) self.assertListEqual(tokenizer.tokenize('こんにちほ' ) , ['こ', 'ん', 'に', 'ち', '[UNK]'] ) def lowercase ( self : int ) -> str: __lowerCAmelCase = self.tokenizer_class.from_pretrained('cl-tohoku/bert-base-japanese-char' ) __lowerCAmelCase = tokenizer.encode('ありがとう。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.encode('どういたしまして。' , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase_ , lowerCAmelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : str ) -> Union[str, Any]: __lowerCAmelCase = 'cl-tohoku/bert-base-japanese' __lowerCAmelCase = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : List[str] ) -> Optional[int]: __lowerCAmelCase = 'cl-tohoku/bert-base-japanese' with self.assertLogs('transformers' , level='WARNING' ) as cm: BertTokenizer.from_pretrained(lowerCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( 'The tokenizer class you load from this checkpoint is not the same type as the class this function' ' is called from.' ) ) __lowerCAmelCase = 'bert-base-cased' with self.assertLogs('transformers' , level='WARNING' ) as cm: BertJapaneseTokenizer.from_pretrained(lowerCAmelCase_ ) self.assertTrue( cm.records[0].message.startswith( 'The tokenizer class you load from this checkpoint is not the same type as the class this function' ' is called from.' ) )
284
1
from __future__ import annotations import requests _snake_case : Dict = set( 'approved_at_utc approved_by author_flair_background_color\nauthor_flair_css_class author_flair_richtext author_flair_template_id author_fullname\nauthor_premium can_mod_post category clicked content_categories created_utc downs\nedited gilded gildings hidden hide_score is_created_from_ads_ui is_meta\nis_original_content is_reddit_media_domain is_video link_flair_css_class\nlink_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title\nname permalink pwls quarantine saved score secure_media secure_media_embed selftext\nsubreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type\ntotal_awards_received ups upvote_ratio url user_reports'.split() ) def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : int = 1, lowerCAmelCase_ : str = "new", lowerCAmelCase_ : list | None = None ): __lowerCAmelCase = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(lowerCAmelCase_ ) - valid_terms ) ): __lowerCAmelCase = F"""Invalid search term: {invalid_search_terms}""" raise ValueError(lowerCAmelCase_ ) __lowerCAmelCase = requests.get( F"""https://reddit.com/r/{subreddit}/{age}.json?limit={limit}""", headers={'User-agent': 'A random string'}, ) if response.status_code == 429: raise requests.HTTPError __lowerCAmelCase = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(lowerCAmelCase_ )} __lowerCAmelCase = {} for id_ in range(lowerCAmelCase_ ): __lowerCAmelCase = { item: data['data']['children'][id_]['data'][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data('learnpython', wanted_data=['title', 'url', 'selftext']))
284
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : List[Any] = { 'microsoft/beit-base-patch16-224-pt22k': ( 'https://huggingface.co/microsoft/beit-base-patch16-224-pt22k/resolve/main/config.json' ), # See all BEiT models at https://huggingface.co/models?filter=beit } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """beit""" def __init__( self : List[Any] , lowerCAmelCase_ : Tuple=8_1_9_2 , lowerCAmelCase_ : Optional[int]=7_6_8 , lowerCAmelCase_ : int=1_2 , lowerCAmelCase_ : Optional[int]=1_2 , lowerCAmelCase_ : Any=3_0_7_2 , lowerCAmelCase_ : Optional[int]="gelu" , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Any=0.02 , lowerCAmelCase_ : int=1e-12 , lowerCAmelCase_ : int=2_2_4 , lowerCAmelCase_ : str=1_6 , lowerCAmelCase_ : int=3 , lowerCAmelCase_ : Dict=False , lowerCAmelCase_ : int=False , lowerCAmelCase_ : List[Any]=False , lowerCAmelCase_ : int=False , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : Union[str, Any]=0.1 , lowerCAmelCase_ : List[str]=True , lowerCAmelCase_ : List[Any]=[3, 5, 7, 1_1] , lowerCAmelCase_ : Optional[Any]=[1, 2, 3, 6] , lowerCAmelCase_ : Tuple=True , lowerCAmelCase_ : Dict=0.4 , lowerCAmelCase_ : Tuple=2_5_6 , lowerCAmelCase_ : Any=1 , lowerCAmelCase_ : Any=False , lowerCAmelCase_ : Optional[int]=2_5_5 , **lowerCAmelCase_ : Any , ) -> Dict: super().__init__(**lowerCAmelCase_ ) __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = initializer_range __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = image_size __lowerCAmelCase = patch_size __lowerCAmelCase = num_channels __lowerCAmelCase = use_mask_token __lowerCAmelCase = use_absolute_position_embeddings __lowerCAmelCase = use_relative_position_bias __lowerCAmelCase = use_shared_relative_position_bias __lowerCAmelCase = layer_scale_init_value __lowerCAmelCase = drop_path_rate __lowerCAmelCase = use_mean_pooling # decode head attributes (semantic segmentation) __lowerCAmelCase = out_indices __lowerCAmelCase = pool_scales # auxiliary head attributes (semantic segmentation) __lowerCAmelCase = use_auxiliary_head __lowerCAmelCase = auxiliary_loss_weight __lowerCAmelCase = auxiliary_channels __lowerCAmelCase = auxiliary_num_convs __lowerCAmelCase = auxiliary_concat_input __lowerCAmelCase = semantic_loss_ignore_index class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = version.parse("""1.11""" ) @property def lowercase ( self : Tuple ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def lowercase ( self : Optional[Any] ) -> float: return 1e-4
284
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType _snake_case : Union[str, Any] = logging.get_logger(__name__) _snake_case : Any = { 'openai/whisper-base': 'https://huggingface.co/openai/whisper-base/resolve/main/config.json', } # fmt: off _snake_case : str = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 357, 366, 438, 532, 685, 705, 796, 930, 1058, 1220, 1267, 1279, 1303, 1343, 1377, 1391, 1635, 1782, 1875, 2162, 2361, 2488, 3467, 4008, 4211, 4600, 4808, 5299, 5855, 6329, 7203, 9609, 9959, 10563, 10786, 11420, 11709, 11907, 13163, 13697, 13700, 14808, 15306, 16410, 16791, 17992, 19203, 19510, 20724, 22305, 22935, 27007, 30109, 30420, 33409, 34949, 40283, 40493, 40549, 47282, 49146, 50257, 50359, 50360, 50361 ] _snake_case : Tuple = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1350, 1853, 1982, 2460, 2627, 3246, 3253, 3268, 3536, 3846, 3961, 4183, 4667, 6585, 6647, 7273, 9061, 9383, 10428, 10929, 11938, 12033, 12331, 12562, 13793, 14157, 14635, 15265, 15618, 16553, 16604, 18362, 18956, 20075, 21675, 22520, 26130, 26161, 26435, 28279, 29464, 31650, 32302, 32470, 36865, 42863, 47425, 49870, 50254, 50258, 50360, 50361, 50362 ] class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """whisper""" a_ = ["""past_key_values"""] a_ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : Optional[int] , lowerCAmelCase_ : Any=5_1_8_6_5 , lowerCAmelCase_ : int=8_0 , lowerCAmelCase_ : Dict=6 , lowerCAmelCase_ : Union[str, Any]=4 , lowerCAmelCase_ : Dict=6 , lowerCAmelCase_ : Union[str, Any]=4 , lowerCAmelCase_ : Dict=1_5_3_6 , lowerCAmelCase_ : Optional[int]=1_5_3_6 , lowerCAmelCase_ : Optional[Any]=0.0 , lowerCAmelCase_ : Dict=0.0 , lowerCAmelCase_ : str=5_0_2_5_7 , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : int=True , lowerCAmelCase_ : Tuple="gelu" , lowerCAmelCase_ : Tuple=2_5_6 , lowerCAmelCase_ : List[str]=0.0 , lowerCAmelCase_ : str=0.0 , lowerCAmelCase_ : List[str]=0.0 , lowerCAmelCase_ : Dict=0.02 , lowerCAmelCase_ : Optional[int]=False , lowerCAmelCase_ : Union[str, Any]=1_5_0_0 , lowerCAmelCase_ : Any=4_4_8 , lowerCAmelCase_ : Optional[Any]=5_0_2_5_6 , lowerCAmelCase_ : int=5_0_2_5_6 , lowerCAmelCase_ : Optional[Any]=5_0_2_5_6 , lowerCAmelCase_ : str=None , lowerCAmelCase_ : List[str]=[2_2_0, 5_0_2_5_6] , lowerCAmelCase_ : str=False , lowerCAmelCase_ : int=2_5_6 , lowerCAmelCase_ : str=False , lowerCAmelCase_ : Optional[Any]=0.05 , lowerCAmelCase_ : Tuple=1_0 , lowerCAmelCase_ : Dict=2 , lowerCAmelCase_ : int=0.0 , lowerCAmelCase_ : Optional[int]=1_0 , lowerCAmelCase_ : List[str]=0 , lowerCAmelCase_ : Dict=7 , **lowerCAmelCase_ : List[str] , ) -> Tuple: __lowerCAmelCase = vocab_size __lowerCAmelCase = num_mel_bins __lowerCAmelCase = d_model __lowerCAmelCase = encoder_layers __lowerCAmelCase = encoder_attention_heads __lowerCAmelCase = decoder_layers __lowerCAmelCase = decoder_attention_heads __lowerCAmelCase = decoder_ffn_dim __lowerCAmelCase = encoder_ffn_dim __lowerCAmelCase = dropout __lowerCAmelCase = attention_dropout __lowerCAmelCase = activation_dropout __lowerCAmelCase = activation_function __lowerCAmelCase = init_std __lowerCAmelCase = encoder_layerdrop __lowerCAmelCase = decoder_layerdrop __lowerCAmelCase = use_cache __lowerCAmelCase = encoder_layers __lowerCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True __lowerCAmelCase = max_source_positions __lowerCAmelCase = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. __lowerCAmelCase = classifier_proj_size __lowerCAmelCase = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowerCAmelCase = apply_spec_augment __lowerCAmelCase = mask_time_prob __lowerCAmelCase = mask_time_length __lowerCAmelCase = mask_time_min_masks __lowerCAmelCase = mask_feature_prob __lowerCAmelCase = mask_feature_length __lowerCAmelCase = mask_feature_min_masks __lowerCAmelCase = median_filter_width super().__init__( pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , is_encoder_decoder=lowerCAmelCase_ , decoder_start_token_id=lowerCAmelCase_ , suppress_tokens=lowerCAmelCase_ , begin_suppress_tokens=lowerCAmelCase_ , **lowerCAmelCase_ , ) class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" @property def lowercase ( self : Tuple ) -> Mapping[str, Mapping[int, str]]: __lowerCAmelCase = OrderedDict( [ ('input_features', {0: 'batch', 1: 'feature_size', 2: 'encoder_sequence'}), ] ) if self.use_past: __lowerCAmelCase = {0: 'batch'} else: __lowerCAmelCase = {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(lowerCAmelCase_ , direction='inputs' ) return common_inputs def lowercase ( self : int , lowerCAmelCase_ : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , lowerCAmelCase_ : int = -1 , lowerCAmelCase_ : int = -1 , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : Optional["TensorType"] = None , lowerCAmelCase_ : int = 2_2_0_5_0 , lowerCAmelCase_ : float = 5.0 , lowerCAmelCase_ : int = 2_2_0 , ) -> Mapping[str, Any]: __lowerCAmelCase = OrderedDict() __lowerCAmelCase = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=lowerCAmelCase_ , framework=lowerCAmelCase_ , sampling_rate=lowerCAmelCase_ , time_duration=lowerCAmelCase_ , frequency=lowerCAmelCase_ , ) __lowerCAmelCase = encoder_inputs['input_features'].shape[2] __lowerCAmelCase = encoder_sequence_length // 2 if self.use_past else seq_length __lowerCAmelCase = super().generate_dummy_inputs( preprocessor.tokenizer , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) __lowerCAmelCase = encoder_inputs.pop('input_features' ) __lowerCAmelCase = decoder_inputs.pop('decoder_input_ids' ) if "past_key_values" in decoder_inputs: __lowerCAmelCase = decoder_inputs.pop('past_key_values' ) return dummy_inputs @property def lowercase ( self : Optional[int] ) -> float: return 1e-3
284
def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : int ): return [sentence[i : i + ngram_size] for i in range(len(lowerCAmelCase_ ) - ngram_size + 1 )] if __name__ == "__main__": from doctest import testmod testmod()
284
1
import logging import os from typing import Dict, List, Optional, Union import torch import torch.nn as nn from accelerate.utils.imports import ( is_abit_bnb_available, is_abit_bnb_available, is_bnb_available, ) from ..big_modeling import dispatch_model, init_empty_weights from .dataclasses import BnbQuantizationConfig from .modeling import ( find_tied_parameters, get_balanced_memory, infer_auto_device_map, load_checkpoint_in_model, offload_weight, set_module_tensor_to_device, ) if is_bnb_available(): import bitsandbytes as bnb from copy import deepcopy _snake_case : Dict = logging.getLogger(__name__) def a_ ( lowerCAmelCase_ : torch.nn.Module, lowerCAmelCase_ : BnbQuantizationConfig, lowerCAmelCase_ : Union[str, os.PathLike] = None, lowerCAmelCase_ : Optional[Dict[str, Union[int, str, torch.device]]] = None, lowerCAmelCase_ : Optional[List[str]] = None, lowerCAmelCase_ : Optional[Dict[Union[int, str], Union[int, str]]] = None, lowerCAmelCase_ : Optional[Union[str, os.PathLike]] = None, lowerCAmelCase_ : bool = False, ): __lowerCAmelCase = bnb_quantization_config.load_in_abit __lowerCAmelCase = bnb_quantization_config.load_in_abit if load_in_abit and not is_abit_bnb_available(): raise ImportError( 'You have a version of `bitsandbytes` that is not compatible with 8bit quantization,' ' make sure you have the latest version of `bitsandbytes` installed.' ) if load_in_abit and not is_abit_bnb_available(): raise ValueError( 'You have a version of `bitsandbytes` that is not compatible with 4bit quantization,' 'make sure you have the latest version of `bitsandbytes` installed.' ) __lowerCAmelCase = [] # custom device map if isinstance(lowerCAmelCase_, lowerCAmelCase_ ) and len(device_map.keys() ) > 1: __lowerCAmelCase = [key for key, value in device_map.items() if value in ['disk', 'cpu']] # We keep some modules such as the lm_head in their original dtype for numerical stability reasons if bnb_quantization_config.skip_modules is None: __lowerCAmelCase = get_keys_to_not_convert(lowerCAmelCase_ ) # add cpu modules to skip modules only for 4-bit modules if load_in_abit: bnb_quantization_config.skip_modules.extend(lowerCAmelCase_ ) __lowerCAmelCase = bnb_quantization_config.skip_modules # We add the modules we want to keep in full precision if bnb_quantization_config.keep_in_fpaa_modules is None: __lowerCAmelCase = [] __lowerCAmelCase = bnb_quantization_config.keep_in_fpaa_modules modules_to_not_convert.extend(lowerCAmelCase_ ) # compatibility with peft __lowerCAmelCase = load_in_abit __lowerCAmelCase = load_in_abit __lowerCAmelCase = get_parameter_device(lowerCAmelCase_ ) if model_device.type != "meta": # quantization of an already loaded model logger.warning( 'It is not recommended to quantize a loaded model. ' 'The model should be instantiated under the `init_empty_weights` context manager.' ) __lowerCAmelCase = replace_with_bnb_layers(lowerCAmelCase_, lowerCAmelCase_, modules_to_not_convert=lowerCAmelCase_ ) # convert param to the right dtype __lowerCAmelCase = bnb_quantization_config.torch_dtype for name, param in model.state_dict().items(): if any(module_to_keep_in_fpaa in name for module_to_keep_in_fpaa in keep_in_fpaa_modules ): param.to(torch.floataa ) if param.dtype != torch.floataa: __lowerCAmelCase = name.replace('.weight', '' ).replace('.bias', '' ) __lowerCAmelCase = getattr(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) if param is not None: param.to(torch.floataa ) elif torch.is_floating_point(lowerCAmelCase_ ): param.to(lowerCAmelCase_ ) if model_device.type == "cuda": # move everything to cpu in the first place because we can't do quantization if the weights are already on cuda model.cuda(torch.cuda.current_device() ) torch.cuda.empty_cache() elif torch.cuda.is_available(): model.to(torch.cuda.current_device() ) else: raise RuntimeError('No GPU found. A GPU is needed for quantization.' ) logger.info( F"""The model device type is {model_device.type}. However, cuda is needed for quantization.""" 'We move the model to cuda.' ) return model elif weights_location is None: raise RuntimeError( F"""`weights_location` needs to be the folder path containing the weights of the model, but we found {weights_location} """ ) else: with init_empty_weights(): __lowerCAmelCase = replace_with_bnb_layers( lowerCAmelCase_, lowerCAmelCase_, modules_to_not_convert=lowerCAmelCase_ ) __lowerCAmelCase = get_quantized_model_device_map( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, max_memory=lowerCAmelCase_, no_split_module_classes=lowerCAmelCase_, ) if offload_state_dict is None and device_map is not None and "disk" in device_map.values(): __lowerCAmelCase = True __lowerCAmelCase = any(x in list(device_map.values() ) for x in ['cpu', 'disk'] ) load_checkpoint_in_model( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, dtype=bnb_quantization_config.torch_dtype, offload_folder=lowerCAmelCase_, offload_state_dict=lowerCAmelCase_, keep_in_fpaa_modules=bnb_quantization_config.keep_in_fpaa_modules, offload_abit_bnb=load_in_abit and offload, ) return dispatch_model(lowerCAmelCase_, device_map=lowerCAmelCase_, offload_dir=lowerCAmelCase_ ) def a_ ( lowerCAmelCase_ : List[Any], lowerCAmelCase_ : List[str], lowerCAmelCase_ : Any=None, lowerCAmelCase_ : Tuple=None, lowerCAmelCase_ : Any=None ): if device_map is None: if torch.cuda.is_available(): __lowerCAmelCase = {'': torch.cuda.current_device()} else: raise RuntimeError('No GPU found. A GPU is needed for quantization.' ) logger.info('The device_map was not initialized.' 'Setting device_map to `{\'\':torch.cuda.current_device()}`.' ) if isinstance(lowerCAmelCase_, lowerCAmelCase_ ): if device_map not in ["auto", "balanced", "balanced_low_0", "sequential"]: raise ValueError( 'If passing a string for `device_map`, please choose \'auto\', \'balanced\', \'balanced_low_0\' or ' '\'sequential\'.' ) __lowerCAmelCase = {} special_dtypes.update( { name: bnb_quantization_config.torch_dtype for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.skip_modules ) } ) special_dtypes.update( { name: torch.floataa for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.keep_in_fpaa_modules ) } ) __lowerCAmelCase = {} __lowerCAmelCase = special_dtypes __lowerCAmelCase = no_split_module_classes __lowerCAmelCase = bnb_quantization_config.target_dtype # get max_memory for each device. if device_map != "sequential": __lowerCAmelCase = get_balanced_memory( lowerCAmelCase_, low_zero=(device_map == 'balanced_low_0'), max_memory=lowerCAmelCase_, **lowerCAmelCase_, ) __lowerCAmelCase = max_memory __lowerCAmelCase = infer_auto_device_map(lowerCAmelCase_, **lowerCAmelCase_ ) if isinstance(lowerCAmelCase_, lowerCAmelCase_ ): # check if don't have any quantized module on the cpu __lowerCAmelCase = bnb_quantization_config.skip_modules + bnb_quantization_config.keep_in_fpaa_modules __lowerCAmelCase = { key: device_map[key] for key in device_map.keys() if key not in modules_not_to_convert } for device in ["cpu", "disk"]: if device in device_map_without_some_modules.values(): if bnb_quantization_config.load_in_abit: raise ValueError( '\n Some modules are dispatched on the CPU or the disk. Make sure you have enough GPU RAM to fit\n the quantized model. If you want to dispatch the model on the CPU or the disk while keeping\n these modules in `torch_dtype`, you need to pass a custom `device_map` to\n `load_and_quantize_model`. Check\n https://huggingface.co/docs/accelerate/main/en/usage_guides/quantization#offload-modules-to-cpu-and-disk\n for more details.\n ' ) else: logger.info( 'Some modules are are offloaded to the CPU or the disk. Note that these modules will be converted to 8-bit' ) del device_map_without_some_modules return device_map def a_ ( lowerCAmelCase_ : Any, lowerCAmelCase_ : List[Any], lowerCAmelCase_ : List[Any]=None, lowerCAmelCase_ : Tuple=None ): if modules_to_not_convert is None: __lowerCAmelCase = [] __lowerCAmelCase , __lowerCAmelCase = _replace_with_bnb_layers( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) if not has_been_replaced: logger.warning( 'You are loading your model in 8bit or 4bit but no linear modules were found in your model.' ' this can happen for some architectures such as gpt2 that uses Conv1D instead of Linear layers.' ' Please double check your model architecture, or submit an issue on github if you think this is' ' a bug.' ) return model def a_ ( lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : List[str], lowerCAmelCase_ : Union[str, Any]=None, lowerCAmelCase_ : List[str]=None, ): __lowerCAmelCase = False for name, module in model.named_children(): if current_key_name is None: __lowerCAmelCase = [] current_key_name.append(lowerCAmelCase_ ) if isinstance(lowerCAmelCase_, nn.Linear ) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` __lowerCAmelCase = '.'.join(lowerCAmelCase_ ) __lowerCAmelCase = True for key in modules_to_not_convert: if ( (key in current_key_name_str) and (key + "." in current_key_name_str) ) or key == current_key_name_str: __lowerCAmelCase = False break if proceed: # Load bnb module with empty weight and replace ``nn.Linear` module if bnb_quantization_config.load_in_abit: __lowerCAmelCase = bnb.nn.LinearabitLt( module.in_features, module.out_features, module.bias is not None, has_fpaa_weights=lowerCAmelCase_, threshold=bnb_quantization_config.llm_inta_threshold, ) elif bnb_quantization_config.load_in_abit: __lowerCAmelCase = bnb.nn.Linearabit( module.in_features, module.out_features, module.bias is not None, bnb_quantization_config.bnb_abit_compute_dtype, compress_statistics=bnb_quantization_config.bnb_abit_use_double_quant, quant_type=bnb_quantization_config.bnb_abit_quant_type, ) else: raise ValueError('load_in_8bit and load_in_4bit can\'t be both False' ) __lowerCAmelCase = module.weight.data if module.bias is not None: __lowerCAmelCase = module.bias.data bnb_module.requires_grad_(lowerCAmelCase_ ) setattr(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = True if len(list(module.children() ) ) > 0: __lowerCAmelCase , __lowerCAmelCase = _replace_with_bnb_layers( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = has_been_replaced | _has_been_replaced # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def a_ ( lowerCAmelCase_ : Union[str, Any] ): # Create a copy of the model with init_empty_weights(): __lowerCAmelCase = deepcopy(lowerCAmelCase_ ) # this has 0 cost since it is done inside `init_empty_weights` context manager` __lowerCAmelCase = find_tied_parameters(lowerCAmelCase_ ) # For compatibility with Accelerate < 0.18 if isinstance(lowerCAmelCase_, lowerCAmelCase_ ): __lowerCAmelCase = sum(list(tied_params.values() ), [] ) + list(tied_params.keys() ) else: __lowerCAmelCase = sum(lowerCAmelCase_, [] ) __lowerCAmelCase = len(lowerCAmelCase_ ) > 0 # Check if it is a base model __lowerCAmelCase = False if hasattr(lowerCAmelCase_, 'base_model_prefix' ): __lowerCAmelCase = not hasattr(lowerCAmelCase_, model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head __lowerCAmelCase = list(model.named_children() ) __lowerCAmelCase = [list_modules[-1][0]] # add last module together with tied weights __lowerCAmelCase = set(lowerCAmelCase_ ) - set(lowerCAmelCase_ ) __lowerCAmelCase = list(set(lowerCAmelCase_ ) ) + list(lowerCAmelCase_ ) # remove ".weight" from the keys __lowerCAmelCase = ['.weight', '.bias'] __lowerCAmelCase = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: __lowerCAmelCase = name.replace(lowerCAmelCase_, '' ) filtered_module_names.append(lowerCAmelCase_ ) return filtered_module_names def a_ ( lowerCAmelCase_ : Union[str, Any] ): for m in model.modules(): if isinstance(lowerCAmelCase_, bnb.nn.Linearabit ): return True return False def a_ ( lowerCAmelCase_ : nn.Module ): return next(parameter.parameters() ).device def a_ ( lowerCAmelCase_ : Dict, lowerCAmelCase_ : int, lowerCAmelCase_ : int, lowerCAmelCase_ : List[Any], lowerCAmelCase_ : Optional[Any], lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : int ): # if it is not quantized, we quantize and offload the quantized weights and the SCB stats if fpaa_statistics is None: set_module_tensor_to_device(lowerCAmelCase_, lowerCAmelCase_, 0, dtype=lowerCAmelCase_, value=lowerCAmelCase_ ) __lowerCAmelCase = param_name __lowerCAmelCase = model if "." in tensor_name: __lowerCAmelCase = tensor_name.split('.' ) for split in splits[:-1]: __lowerCAmelCase = getattr(lowerCAmelCase_, lowerCAmelCase_ ) if new_module is None: raise ValueError(F"""{module} has no attribute {split}.""" ) __lowerCAmelCase = new_module __lowerCAmelCase = splits[-1] # offload weights __lowerCAmelCase = False offload_weight(module._parameters[tensor_name], lowerCAmelCase_, lowerCAmelCase_, index=lowerCAmelCase_ ) if hasattr(module._parameters[tensor_name], 'SCB' ): offload_weight( module._parameters[tensor_name].SCB, param_name.replace('weight', 'SCB' ), lowerCAmelCase_, index=lowerCAmelCase_, ) else: offload_weight(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, index=lowerCAmelCase_ ) offload_weight(lowerCAmelCase_, param_name.replace('weight', 'SCB' ), lowerCAmelCase_, index=lowerCAmelCase_ ) set_module_tensor_to_device(lowerCAmelCase_, lowerCAmelCase_, 'meta', dtype=lowerCAmelCase_, value=torch.empty(*param.size() ) )
284
from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : Any = { 'google/pegasus-large': 'https://huggingface.co/google/pegasus-large/resolve/main/config.json', # See all PEGASUS models at https://huggingface.co/models?filter=pegasus } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """pegasus""" a_ = ["""past_key_values"""] a_ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[str] , lowerCAmelCase_ : Union[str, Any]=5_0_2_6_5 , lowerCAmelCase_ : Union[str, Any]=1_0_2_4 , lowerCAmelCase_ : Union[str, Any]=1_2 , lowerCAmelCase_ : Dict=4_0_9_6 , lowerCAmelCase_ : str=1_6 , lowerCAmelCase_ : List[Any]=1_2 , lowerCAmelCase_ : Union[str, Any]=4_0_9_6 , lowerCAmelCase_ : Union[str, Any]=1_6 , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Optional[int]=0.0 , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Any=True , lowerCAmelCase_ : Dict="gelu" , lowerCAmelCase_ : Optional[int]=1_0_2_4 , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : str=0.0 , lowerCAmelCase_ : Union[str, Any]=0.0 , lowerCAmelCase_ : Optional[int]=0.02 , lowerCAmelCase_ : Tuple=0 , lowerCAmelCase_ : Tuple=False , lowerCAmelCase_ : Union[str, Any]=0 , lowerCAmelCase_ : int=1 , lowerCAmelCase_ : Tuple=1 , **lowerCAmelCase_ : Tuple , ) -> List[str]: __lowerCAmelCase = vocab_size __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = d_model __lowerCAmelCase = encoder_ffn_dim __lowerCAmelCase = encoder_layers __lowerCAmelCase = encoder_attention_heads __lowerCAmelCase = decoder_ffn_dim __lowerCAmelCase = decoder_layers __lowerCAmelCase = decoder_attention_heads __lowerCAmelCase = dropout __lowerCAmelCase = attention_dropout __lowerCAmelCase = activation_dropout __lowerCAmelCase = activation_function __lowerCAmelCase = init_std __lowerCAmelCase = encoder_layerdrop __lowerCAmelCase = decoder_layerdrop __lowerCAmelCase = use_cache __lowerCAmelCase = encoder_layers __lowerCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , is_encoder_decoder=lowerCAmelCase_ , decoder_start_token_id=lowerCAmelCase_ , forced_eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ , ) @property def lowercase ( self : List[Any] ) -> int: return self.encoder_attention_heads @property def lowercase ( self : Optional[Any] ) -> int: return self.d_model
284
1
from __future__ import annotations from collections.abc import Sequence from typing import Literal def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : str ): __lowerCAmelCase = list(lowerCAmelCase_ ) __lowerCAmelCase = list(lowerCAmelCase_ ) __lowerCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count += 1 __lowerCAmelCase = '_' if count > 1: return False else: return "".join(lowerCAmelCase_ ) def a_ ( lowerCAmelCase_ : list[str] ): __lowerCAmelCase = [] while True: __lowerCAmelCase = ['$'] * len(lowerCAmelCase_ ) __lowerCAmelCase = [] for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1, len(lowerCAmelCase_ ) ): __lowerCAmelCase = compare_string(binary[i], binary[j] ) if k is False: __lowerCAmelCase = '*' __lowerCAmelCase = '*' temp.append('X' ) for i in range(len(lowerCAmelCase_ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(lowerCAmelCase_ ) == 0: return pi __lowerCAmelCase = list(set(lowerCAmelCase_ ) ) def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : Sequence[float] ): __lowerCAmelCase = [] for minterm in minterms: __lowerCAmelCase = '' for _ in range(lowerCAmelCase_ ): __lowerCAmelCase = str(minterm % 2 ) + string minterm //= 2 temp.append(lowerCAmelCase_ ) return temp def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : str, lowerCAmelCase_ : int ): __lowerCAmelCase = list(lowerCAmelCase_ ) __lowerCAmelCase = list(lowerCAmelCase_ ) __lowerCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def a_ ( lowerCAmelCase_ : list[list[int]], lowerCAmelCase_ : list[str] ): __lowerCAmelCase = [] __lowerCAmelCase = [0] * len(lowerCAmelCase_ ) for i in range(len(chart[0] ) ): __lowerCAmelCase = 0 __lowerCAmelCase = -1 for j in range(len(lowerCAmelCase_ ) ): if chart[j][i] == 1: count += 1 __lowerCAmelCase = j if count == 1: __lowerCAmelCase = 1 for i in range(len(lowerCAmelCase_ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(lowerCAmelCase_ ) ): __lowerCAmelCase = 0 temp.append(prime_implicants[i] ) while True: __lowerCAmelCase = 0 __lowerCAmelCase = -1 __lowerCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): __lowerCAmelCase = chart[i].count(1 ) if count_n > max_n: __lowerCAmelCase = count_n __lowerCAmelCase = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(lowerCAmelCase_ ) ): __lowerCAmelCase = 0 def a_ ( lowerCAmelCase_ : list[str], lowerCAmelCase_ : list[str] ): __lowerCAmelCase = [[0 for x in range(len(lowerCAmelCase_ ) )] for x in range(len(lowerCAmelCase_ ) )] for i in range(len(lowerCAmelCase_ ) ): __lowerCAmelCase = prime_implicants[i].count('_' ) for j in range(len(lowerCAmelCase_ ) ): if is_for_table(prime_implicants[i], binary[j], lowerCAmelCase_ ): __lowerCAmelCase = 1 return chart def a_ ( ): __lowerCAmelCase = int(input('Enter the no. of variables\n' ) ) __lowerCAmelCase = [ float(lowerCAmelCase_ ) for x in input( 'Enter the decimal representation of Minterms \'Spaces Separated\'\n' ).split() ] __lowerCAmelCase = decimal_to_binary(lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = check(lowerCAmelCase_ ) print('Prime Implicants are:' ) print(lowerCAmelCase_ ) __lowerCAmelCase = prime_implicant_chart(lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = selection(lowerCAmelCase_, lowerCAmelCase_ ) print('Essential Prime Implicants are:' ) print(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
284
def a_ ( lowerCAmelCase_ : int ): if p < 2: raise ValueError('p should not be less than 2!' ) elif p == 2: return True __lowerCAmelCase = 4 __lowerCAmelCase = (1 << p) - 1 for _ in range(p - 2 ): __lowerCAmelCase = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
284
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case : Dict = {'configuration_mra': ['MRA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MraConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'MRA_PRETRAINED_MODEL_ARCHIVE_LIST', 'MraForMaskedLM', 'MraForMultipleChoice', 'MraForQuestionAnswering', 'MraForSequenceClassification', 'MraForTokenClassification', 'MraLayer', 'MraModel', 'MraPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys _snake_case : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure)
284
from __future__ import annotations import math def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : int, lowerCAmelCase_ : bool, lowerCAmelCase_ : list[int], lowerCAmelCase_ : float ): if depth < 0: raise ValueError('Depth cannot be less than 0' ) if len(lowerCAmelCase_ ) == 0: raise ValueError('Scores cannot be empty' ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1, node_index * 2, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), minimax(depth + 1, node_index * 2 + 1, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), ) return min( minimax(depth + 1, node_index * 2, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), minimax(depth + 1, node_index * 2 + 1, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ), ) def a_ ( ): __lowerCAmelCase = [90, 23, 6, 33, 21, 65, 123, 3_4423] __lowerCAmelCase = math.log(len(lowerCAmelCase_ ), 2 ) print('Optimal value : ', end='' ) print(minimax(0, 0, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
284
1
def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : list[str] ): __lowerCAmelCase = '' for word_or_phrase in separated: if not isinstance(lowerCAmelCase_, lowerCAmelCase_ ): raise Exception('join() accepts only strings to be joined' ) joined += word_or_phrase + separator return joined.strip(lowerCAmelCase_ ) if __name__ == "__main__": from doctest import testmod testmod()
284
def a_ ( lowerCAmelCase_ : int ): if number < 0: raise ValueError('number must not be negative' ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
284
1
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : str ) -> Optional[int]: __lowerCAmelCase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCAmelCase_ , 'tf_padding' ) ) self.parent.assertTrue(hasattr(lowerCAmelCase_ , 'depth_multiplier' ) ) class _UpperCAmelCase : """simple docstring""" def __init__( self : Any , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Optional[Any]=1_3 , lowerCAmelCase_ : List[str]=3 , lowerCAmelCase_ : str=3_2 , lowerCAmelCase_ : Dict=0.25 , lowerCAmelCase_ : Optional[Any]=8 , lowerCAmelCase_ : Any=True , lowerCAmelCase_ : Dict=1_0_2_4 , lowerCAmelCase_ : Optional[Any]=3_2 , lowerCAmelCase_ : Optional[int]="relu6" , lowerCAmelCase_ : Any=0.1 , lowerCAmelCase_ : Any=0.02 , lowerCAmelCase_ : int=True , lowerCAmelCase_ : Optional[Any]=True , lowerCAmelCase_ : str=1_0 , lowerCAmelCase_ : Optional[Any]=None , ) -> Optional[Any]: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = num_channels __lowerCAmelCase = image_size __lowerCAmelCase = depth_multiplier __lowerCAmelCase = min_depth __lowerCAmelCase = tf_padding __lowerCAmelCase = int(last_hidden_size * depth_multiplier ) __lowerCAmelCase = output_stride __lowerCAmelCase = hidden_act __lowerCAmelCase = classifier_dropout_prob __lowerCAmelCase = use_labels __lowerCAmelCase = is_training __lowerCAmelCase = num_labels __lowerCAmelCase = initializer_range __lowerCAmelCase = scope def lowercase ( self : Union[str, Any] ) -> Tuple: __lowerCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCAmelCase = None __lowerCAmelCase = None if self.use_labels: __lowerCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) __lowerCAmelCase = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __lowerCAmelCase = self.get_config() return config, pixel_values, labels, pixel_labels def lowercase ( self : Optional[Any] ) -> Any: return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def lowercase ( self : Tuple , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Any , lowerCAmelCase_ : Any , lowerCAmelCase_ : Tuple ) -> str: __lowerCAmelCase = MobileNetVaModel(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() __lowerCAmelCase = model(lowerCAmelCase_ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : List[str] ) -> Optional[int]: __lowerCAmelCase = self.num_labels __lowerCAmelCase = MobileNetVaForImageClassification(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() __lowerCAmelCase = model(lowerCAmelCase_ , labels=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase ( self : Dict ) -> Any: __lowerCAmelCase = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = config_and_inputs __lowerCAmelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () a_ = ( {"""feature-extraction""": MobileNetVaModel, """image-classification""": MobileNetVaForImageClassification} if is_torch_available() else {} ) a_ = False a_ = False a_ = False a_ = False def lowercase ( self : str ) -> Tuple: __lowerCAmelCase = MobileNetVaModelTester(self ) __lowerCAmelCase = MobileNetVaConfigTester(self , config_class=lowerCAmelCase_ , has_text_modality=lowerCAmelCase_ ) def lowercase ( self : List[str] ) -> Union[str, Any]: self.config_tester.run_common_tests() @unittest.skip(reason='MobileNetV1 does not use inputs_embeds' ) def lowercase ( self : Dict ) -> Tuple: pass @unittest.skip(reason='MobileNetV1 does not support input and output embeddings' ) def lowercase ( self : Optional[int] ) -> List[str]: pass @unittest.skip(reason='MobileNetV1 does not output attentions' ) def lowercase ( self : List[Any] ) -> List[Any]: pass def lowercase ( self : int ) -> Optional[Any]: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase = model_class(lowerCAmelCase_ ) __lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCAmelCase = [*signature.parameters.keys()] __lowerCAmelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , lowerCAmelCase_ ) def lowercase ( self : Any ) -> Union[str, Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase_ ) def lowercase ( self : List[str] ) -> str: def check_hidden_states_output(lowerCAmelCase_ : Tuple , lowerCAmelCase_ : int , lowerCAmelCase_ : List[Any] ): __lowerCAmelCase = model_class(lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase = outputs.hidden_states __lowerCAmelCase = 2_6 self.assertEqual(len(lowerCAmelCase_ ) , lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase = True check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowerCAmelCase = True check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Union[str, Any] ) -> int: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase_ ) @slow def lowercase ( self : Any ) -> int: for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = MobileNetVaModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) def a_ ( ): __lowerCAmelCase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def lowercase ( self : List[str] ) -> List[Any]: return ( MobileNetVaImageProcessor.from_pretrained('google/mobilenet_v1_1.0_224' ) if is_vision_available() else None ) @slow def lowercase ( self : Union[str, Any] ) -> str: __lowerCAmelCase = MobileNetVaForImageClassification.from_pretrained('google/mobilenet_v1_1.0_224' ).to(lowerCAmelCase_ ) __lowerCAmelCase = self.default_image_processor __lowerCAmelCase = prepare_img() __lowerCAmelCase = image_processor(images=lowerCAmelCase_ , return_tensors='pt' ).to(lowerCAmelCase_ ) # forward pass with torch.no_grad(): __lowerCAmelCase = model(**lowerCAmelCase_ ) # verify the logits __lowerCAmelCase = torch.Size((1, 1_0_0_1) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase_ ) __lowerCAmelCase = torch.tensor([-4.17_39, -1.12_33, 3.12_05] ).to(lowerCAmelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase_ , atol=1e-4 ) )
284
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
284
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { 'configuration_layoutlmv3': [ 'LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv3Config', 'LayoutLMv3OnnxConfig', ], 'processing_layoutlmv3': ['LayoutLMv3Processor'], 'tokenization_layoutlmv3': ['LayoutLMv3Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Union[str, Any] = ['LayoutLMv3TokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[str] = [ 'LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv3ForQuestionAnswering', 'LayoutLMv3ForSequenceClassification', 'LayoutLMv3ForTokenClassification', 'LayoutLMv3Model', 'LayoutLMv3PreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[Any] = [ 'TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLayoutLMv3ForQuestionAnswering', 'TFLayoutLMv3ForSequenceClassification', 'TFLayoutLMv3ForTokenClassification', 'TFLayoutLMv3Model', 'TFLayoutLMv3PreTrainedModel', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Tuple = ['LayoutLMv3FeatureExtractor'] _snake_case : str = ['LayoutLMv3ImageProcessor'] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys _snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def a_ ( lowerCAmelCase_ : List[str], lowerCAmelCase_ : Dict, lowerCAmelCase_ : Tuple=1024, lowerCAmelCase_ : Optional[Any]=1024, lowerCAmelCase_ : Tuple=False, **lowerCAmelCase_ : Union[str, Any] ): __lowerCAmelCase = AutoTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='train', **lowerCAmelCase_ ) __lowerCAmelCase = tok.pad_token_id def get_lens(lowerCAmelCase_ : Optional[Any] ): __lowerCAmelCase = tqdm( DataLoader(lowerCAmelCase_, batch_size=512, num_workers=8, shuffle=lowerCAmelCase_, collate_fn=ds.collate_fn ), desc=str(ds.len_file ), ) __lowerCAmelCase = [] for batch in dl: __lowerCAmelCase = batch['input_ids'].ne(lowerCAmelCase_ ).sum(1 ).tolist() __lowerCAmelCase = batch['labels'].ne(lowerCAmelCase_ ).sum(1 ).tolist() if consider_target: for src, tgt in zip(lowerCAmelCase_, lowerCAmelCase_ ): max_lens.append(max(lowerCAmelCase_, lowerCAmelCase_ ) ) else: max_lens.extend(lowerCAmelCase_ ) return max_lens __lowerCAmelCase = get_lens(lowerCAmelCase_ ) __lowerCAmelCase = SeqaSeqDataset(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, type_path='val', **lowerCAmelCase_ ) __lowerCAmelCase = get_lens(lowerCAmelCase_ ) pickle_save(lowerCAmelCase_, train_ds.len_file ) pickle_save(lowerCAmelCase_, val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
284
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _snake_case : Union[str, Any] = logging.get_logger(__name__) _snake_case : List[str] = { 'xlm-roberta-base': 'https://huggingface.co/xlm-roberta-base/resolve/main/config.json', 'xlm-roberta-large': 'https://huggingface.co/xlm-roberta-large/resolve/main/config.json', 'xlm-roberta-large-finetuned-conll02-dutch': ( 'https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/config.json' ), 'xlm-roberta-large-finetuned-conll02-spanish': ( 'https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/config.json' ), 'xlm-roberta-large-finetuned-conll03-english': ( 'https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/config.json' ), 'xlm-roberta-large-finetuned-conll03-german': ( 'https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/config.json' ), } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """xlm-roberta""" def __init__( self : Optional[int] , lowerCAmelCase_ : Union[str, Any]=3_0_5_2_2 , lowerCAmelCase_ : List[str]=7_6_8 , lowerCAmelCase_ : Any=1_2 , lowerCAmelCase_ : Tuple=1_2 , lowerCAmelCase_ : Dict=3_0_7_2 , lowerCAmelCase_ : Union[str, Any]="gelu" , lowerCAmelCase_ : Optional[int]=0.1 , lowerCAmelCase_ : str=0.1 , lowerCAmelCase_ : Dict=5_1_2 , lowerCAmelCase_ : Optional[int]=2 , lowerCAmelCase_ : List[Any]=0.02 , lowerCAmelCase_ : List[Any]=1e-12 , lowerCAmelCase_ : Union[str, Any]=1 , lowerCAmelCase_ : Optional[int]=0 , lowerCAmelCase_ : Dict=2 , lowerCAmelCase_ : Union[str, Any]="absolute" , lowerCAmelCase_ : Optional[int]=True , lowerCAmelCase_ : str=None , **lowerCAmelCase_ : List[str] , ) -> Union[str, Any]: super().__init__(pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ ) __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = hidden_act __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = type_vocab_size __lowerCAmelCase = initializer_range __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = position_embedding_type __lowerCAmelCase = use_cache __lowerCAmelCase = classifier_dropout class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" @property def lowercase ( self : List[Any] ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": __lowerCAmelCase = {0: 'batch', 1: 'choice', 2: 'sequence'} else: __lowerCAmelCase = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
284
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def a_ ( lowerCAmelCase_ : List[Any], lowerCAmelCase_ : str, lowerCAmelCase_ : Optional[int]=None, lowerCAmelCase_ : List[Any]=None ): if attention_mask is None: __lowerCAmelCase = tf.cast(tf.math.not_equal(lowerCAmelCase_, config.pad_token_id ), tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class _UpperCAmelCase : """simple docstring""" a_ = OPTConfig a_ = {} a_ = """gelu""" def __init__( self : Dict , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : List[str]=1_3 , lowerCAmelCase_ : Tuple=7 , lowerCAmelCase_ : Dict=True , lowerCAmelCase_ : Union[str, Any]=False , lowerCAmelCase_ : Any=9_9 , lowerCAmelCase_ : Any=1_6 , lowerCAmelCase_ : List[str]=2 , lowerCAmelCase_ : Dict=4 , lowerCAmelCase_ : str=4 , lowerCAmelCase_ : Any="gelu" , lowerCAmelCase_ : Optional[Any]=0.1 , lowerCAmelCase_ : List[Any]=0.1 , lowerCAmelCase_ : Tuple=2_0 , lowerCAmelCase_ : Optional[Any]=2 , lowerCAmelCase_ : Any=1 , lowerCAmelCase_ : List[Any]=0 , lowerCAmelCase_ : Optional[int]=1_6 , lowerCAmelCase_ : Dict=1_6 , ) -> int: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = eos_token_id __lowerCAmelCase = pad_token_id __lowerCAmelCase = bos_token_id __lowerCAmelCase = embed_dim __lowerCAmelCase = word_embed_proj_dim __lowerCAmelCase = False def lowercase ( self : List[str] ) -> Optional[Any]: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowerCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowerCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowerCAmelCase = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=lowerCAmelCase_ , **self.config_updates , ) __lowerCAmelCase = prepare_opt_inputs_dict(lowerCAmelCase_ , lowerCAmelCase_ ) return config, inputs_dict def lowercase ( self : Any , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Dict ) -> List[str]: __lowerCAmelCase = TFOPTModel(config=lowerCAmelCase_ ) __lowerCAmelCase = inputs_dict['input_ids'] __lowerCAmelCase = input_ids[:1, :] __lowerCAmelCase = inputs_dict['attention_mask'][:1, :] __lowerCAmelCase = 1 # first forward pass __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , use_cache=lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowerCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowerCAmelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowerCAmelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowerCAmelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ )[0] __lowerCAmelCase = model(lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , past_key_values=lowerCAmelCase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowerCAmelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowerCAmelCase = output_from_no_past[:, -3:, random_slice_idx] __lowerCAmelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowerCAmelCase_ , lowerCAmelCase_ , rtol=1e-3 ) @require_tf class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () a_ = (TFOPTForCausalLM,) if is_tf_available() else () a_ = ( {"""feature-extraction""": TFOPTModel, """text-generation""": TFOPTForCausalLM} if is_tf_available() else {} ) a_ = False a_ = False a_ = False a_ = 10 def lowercase ( self : List[str] ) -> Optional[int]: __lowerCAmelCase = TFOPTModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=lowerCAmelCase_ ) def lowercase ( self : Tuple ) -> Tuple: self.config_tester.run_common_tests() def lowercase ( self : Tuple ) -> Optional[Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCAmelCase_ ) def lowercase ( self : Union[str, Any] ) -> Dict: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(lowerCAmelCase_ : str , lowerCAmelCase_ : Union[str, Any] ): if hasattr(lowerCAmelCase_ , 'weight' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(lowerCAmelCase_ , 'weight' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 1_0, config.vocab_size + 1_0]: # build the embeddings __lowerCAmelCase = model_class(config=lowerCAmelCase_ ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_input_embeddings() ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(lowerCAmelCase_ ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_input_embeddings() ) __lowerCAmelCase = _get_word_embedding_weight(lowerCAmelCase_ , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. __lowerCAmelCase = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , lowerCAmelCase_ ) # check that weights remain the same after resizing __lowerCAmelCase = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCAmelCase = False self.assertTrue(lowerCAmelCase_ ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , lowerCAmelCase_ ) __lowerCAmelCase = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCAmelCase = False self.assertTrue(lowerCAmelCase_ ) def a_ ( lowerCAmelCase_ : Union[str, Any] ): return tf.constant(lowerCAmelCase_, dtype=tf.intaa ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" a_ = 99 def lowercase ( self : Optional[int] ) -> Any: __lowerCAmelCase = tf.ones((4, 1) , dtype=tf.intaa ) * 2 __lowerCAmelCase = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) __lowerCAmelCase = input_ids.shape[0] __lowerCAmelCase = OPTConfig( vocab_size=self.vocab_size , hidden_size=2_4 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=3_2 , max_position_embeddings=4_8 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def lowercase ( self : str ) -> List[str]: __lowerCAmelCase = TFOPTModel.from_pretrained('facebook/opt-350m' ) __lowerCAmelCase = _long_tensor([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] ) __lowerCAmelCase = tf.not_equal(lowerCAmelCase_ , model.config.pad_token_id ) with tf.GradientTape(): __lowerCAmelCase = model(input_ids=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ ).last_hidden_state __lowerCAmelCase = (1, 1_1, 5_1_2) self.assertEqual(output.shape , lowerCAmelCase_ ) __lowerCAmelCase = tf.constant( [[-0.28_73, -1.92_18, -0.30_33], [-1.27_10, -0.13_38, -0.19_02], [0.40_95, 0.12_14, -1.31_21]] ) self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase_ , atol=4e-3 ) ) __lowerCAmelCase = tf.function(lowerCAmelCase_ , jit_compile=lowerCAmelCase_ ) __lowerCAmelCase = xla_generate(lowerCAmelCase_ , lowerCAmelCase_ )[0] self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase_ , atol=4e-2 ) ) @require_tf @slow class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> Dict: super().setUp() __lowerCAmelCase = 'facebook/opt-350m' def lowercase ( self : Dict ) -> Any: __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(self.path_model ) __lowerCAmelCase = GPTaTokenizer.from_pretrained(self.path_model ) __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of', 'Paris is the capital of France and', 'Computers and mobile phones have taken', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' , padding=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) __lowerCAmelCase = tf.constant( [ [1.38_51, -13.89_23, -10.52_29, -10.75_33, -0.23_09, -10.23_84, -0.53_65, -9.09_47, -5.16_70], [-4.70_73, -10.62_76, -3.94_15, -21.52_42, -0.28_22, -0.28_22, -0.28_22, -0.28_22, -0.28_22], [0.62_47, -3.42_29, -8.91_79, -1.42_97, -14.16_50, 1.41_46, -9.02_18, -0.27_03, -0.27_03], [6.47_83, -1.99_13, -10.79_26, -2.33_36, 1.50_92, -0.99_74, -6.82_13, 1.34_77, 1.34_77], ] ) self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1e-4 ) ) __lowerCAmelCase = tf.function(lowerCAmelCase_ , jit_compile=lowerCAmelCase_ ) __lowerCAmelCase = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(lowerCAmelCase_ , lowerCAmelCase_ , atol=1e-4 ) ) @require_tf @slow class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @property def lowercase ( self : Optional[int] ) -> int: return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def lowercase ( self : int ) -> str: __lowerCAmelCase = 'facebook/opt-125m' __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of New York, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] __lowerCAmelCase = [] __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) for prompt in self.prompts: __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(lowerCAmelCase_ , max_length=1_0 ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Optional[Any] ) -> str: __lowerCAmelCase = 'facebook/opt-350m' __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = 'left' # use different length sentences to test batching __lowerCAmelCase = [ 'Hello, my dog is a little', 'Today, I', ] __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' , padding=lowerCAmelCase_ ) __lowerCAmelCase = inputs['input_ids'] __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ , attention_mask=inputs['attention_mask'] ) __lowerCAmelCase = tokenizer(sentences[0] , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ ) __lowerCAmelCase = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['attention_mask'][-1] , tf.intaa ) ) __lowerCAmelCase = tokenizer(sentences[1] , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(input_ids=lowerCAmelCase_ , max_length=model.config.max_length - num_paddings ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=lowerCAmelCase_ ) __lowerCAmelCase = [ 'Hello, my dog is a little bit of a dork.\nI\'m a little bit', 'Today, I was in the middle of a conversation with a friend about the', ] self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , [non_padded_sentence, padded_sentence] ) def lowercase ( self : List[Any] ) -> List[Any]: __lowerCAmelCase = 'facebook/opt-350m' __lowerCAmelCase = [ 'Today is a beautiful day and I want to', 'In the city of San Francisco, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] __lowerCAmelCase = [] __lowerCAmelCase = GPTaTokenizer.from_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase_ ) for prompt in self.prompts: __lowerCAmelCase = tokenizer(lowerCAmelCase_ , return_tensors='tf' ).input_ids __lowerCAmelCase = model.generate(lowerCAmelCase_ , max_length=1_0 ) __lowerCAmelCase = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ )
284
1
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = (DDPMParallelScheduler,) def lowercase ( self : int , **lowerCAmelCase_ : Tuple ) -> List[Any]: __lowerCAmelCase = { 'num_train_timesteps': 1_0_0_0, 'beta_start': 0.00_01, 'beta_end': 0.02, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**lowerCAmelCase_ ) return config def lowercase ( self : Tuple ) -> str: for timesteps in [1, 5, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=lowerCAmelCase_ ) def lowercase ( self : Tuple ) -> Dict: for beta_start, beta_end in zip([0.00_01, 0.0_01, 0.01, 0.1] , [0.0_02, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=lowerCAmelCase_ , beta_end=lowerCAmelCase_ ) def lowercase ( self : List[Any] ) -> Optional[int]: for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowerCAmelCase_ ) def lowercase ( self : Dict ) -> Dict: for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=lowerCAmelCase_ ) def lowercase ( self : List[str] ) -> List[str]: for clip_sample in [True, False]: self.check_over_configs(clip_sample=lowerCAmelCase_ ) def lowercase ( self : int ) -> List[Any]: self.check_over_configs(thresholding=lowerCAmelCase_ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=lowerCAmelCase_ , prediction_type=lowerCAmelCase_ , sample_max_value=lowerCAmelCase_ , ) def lowercase ( self : Optional[int] ) -> Dict: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=lowerCAmelCase_ ) def lowercase ( self : str ) -> str: for t in [0, 5_0_0, 9_9_9]: self.check_over_forward(time_step=lowerCAmelCase_ ) def lowercase ( self : List[Any] ) -> Union[str, Any]: __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config() __lowerCAmelCase = scheduler_class(**lowerCAmelCase_ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(4_8_7 ) - 0.0_09_79 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(9_9_9 ) - 0.02 ) ) < 1e-5 def lowercase ( self : Any ) -> List[Any]: __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config() __lowerCAmelCase = scheduler_class(**lowerCAmelCase_ ) __lowerCAmelCase = len(lowerCAmelCase_ ) __lowerCAmelCase = self.dummy_model() __lowerCAmelCase = self.dummy_sample_deter __lowerCAmelCase = self.dummy_sample_deter + 0.1 __lowerCAmelCase = self.dummy_sample_deter - 0.1 __lowerCAmelCase = samplea.shape[0] __lowerCAmelCase = torch.stack([samplea, samplea, samplea] , dim=0 ) __lowerCAmelCase = torch.arange(lowerCAmelCase_ )[0:3, None].repeat(1 , lowerCAmelCase_ ) __lowerCAmelCase = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) __lowerCAmelCase = scheduler.batch_step_no_noise(lowerCAmelCase_ , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) __lowerCAmelCase = torch.sum(torch.abs(lowerCAmelCase_ ) ) __lowerCAmelCase = torch.mean(torch.abs(lowerCAmelCase_ ) ) assert abs(result_sum.item() - 11_53.18_33 ) < 1e-2 assert abs(result_mean.item() - 0.50_05 ) < 1e-3 def lowercase ( self : Union[str, Any] ) -> Union[str, Any]: __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config() __lowerCAmelCase = scheduler_class(**lowerCAmelCase_ ) __lowerCAmelCase = len(lowerCAmelCase_ ) __lowerCAmelCase = self.dummy_model() __lowerCAmelCase = self.dummy_sample_deter __lowerCAmelCase = torch.manual_seed(0 ) for t in reversed(range(lowerCAmelCase_ ) ): # 1. predict noise residual __lowerCAmelCase = model(lowerCAmelCase_ , lowerCAmelCase_ ) # 2. predict previous mean of sample x_t-1 __lowerCAmelCase = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , generator=lowerCAmelCase_ ).prev_sample __lowerCAmelCase = pred_prev_sample __lowerCAmelCase = torch.sum(torch.abs(lowerCAmelCase_ ) ) __lowerCAmelCase = torch.mean(torch.abs(lowerCAmelCase_ ) ) assert abs(result_sum.item() - 2_58.96_06 ) < 1e-2 assert abs(result_mean.item() - 0.33_72 ) < 1e-3 def lowercase ( self : Union[str, Any] ) -> Optional[Any]: __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config(prediction_type='v_prediction' ) __lowerCAmelCase = scheduler_class(**lowerCAmelCase_ ) __lowerCAmelCase = len(lowerCAmelCase_ ) __lowerCAmelCase = self.dummy_model() __lowerCAmelCase = self.dummy_sample_deter __lowerCAmelCase = torch.manual_seed(0 ) for t in reversed(range(lowerCAmelCase_ ) ): # 1. predict noise residual __lowerCAmelCase = model(lowerCAmelCase_ , lowerCAmelCase_ ) # 2. predict previous mean of sample x_t-1 __lowerCAmelCase = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , generator=lowerCAmelCase_ ).prev_sample __lowerCAmelCase = pred_prev_sample __lowerCAmelCase = torch.sum(torch.abs(lowerCAmelCase_ ) ) __lowerCAmelCase = torch.mean(torch.abs(lowerCAmelCase_ ) ) assert abs(result_sum.item() - 2_02.02_96 ) < 1e-2 assert abs(result_mean.item() - 0.26_31 ) < 1e-3 def lowercase ( self : Union[str, Any] ) -> Dict: __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config() __lowerCAmelCase = scheduler_class(**lowerCAmelCase_ ) __lowerCAmelCase = [1_0_0, 8_7, 5_0, 1, 0] scheduler.set_timesteps(timesteps=lowerCAmelCase_ ) __lowerCAmelCase = scheduler.timesteps for i, timestep in enumerate(lowerCAmelCase_ ): if i == len(lowerCAmelCase_ ) - 1: __lowerCAmelCase = -1 else: __lowerCAmelCase = timesteps[i + 1] __lowerCAmelCase = scheduler.previous_timestep(lowerCAmelCase_ ) __lowerCAmelCase = prev_t.item() self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Union[str, Any] ) -> List[str]: __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config() __lowerCAmelCase = scheduler_class(**lowerCAmelCase_ ) __lowerCAmelCase = [1_0_0, 8_7, 5_0, 5_1, 0] with self.assertRaises(lowerCAmelCase_ , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=lowerCAmelCase_ ) def lowercase ( self : int ) -> int: __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config() __lowerCAmelCase = scheduler_class(**lowerCAmelCase_ ) __lowerCAmelCase = [1_0_0, 8_7, 5_0, 1, 0] __lowerCAmelCase = len(lowerCAmelCase_ ) with self.assertRaises(lowerCAmelCase_ , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=lowerCAmelCase_ , timesteps=lowerCAmelCase_ ) def lowercase ( self : Optional[Any] ) -> int: __lowerCAmelCase = self.scheduler_classes[0] __lowerCAmelCase = self.get_scheduler_config() __lowerCAmelCase = scheduler_class(**lowerCAmelCase_ ) __lowerCAmelCase = [scheduler.config.num_train_timesteps] with self.assertRaises( lowerCAmelCase_ , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=lowerCAmelCase_ )
284
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { 'configuration_layoutlmv3': [ 'LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv3Config', 'LayoutLMv3OnnxConfig', ], 'processing_layoutlmv3': ['LayoutLMv3Processor'], 'tokenization_layoutlmv3': ['LayoutLMv3Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Union[str, Any] = ['LayoutLMv3TokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[str] = [ 'LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv3ForQuestionAnswering', 'LayoutLMv3ForSequenceClassification', 'LayoutLMv3ForTokenClassification', 'LayoutLMv3Model', 'LayoutLMv3PreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[Any] = [ 'TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLayoutLMv3ForQuestionAnswering', 'TFLayoutLMv3ForSequenceClassification', 'TFLayoutLMv3ForTokenClassification', 'TFLayoutLMv3Model', 'TFLayoutLMv3PreTrainedModel', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Tuple = ['LayoutLMv3FeatureExtractor'] _snake_case : str = ['LayoutLMv3ImageProcessor'] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys _snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
1
def a_ ( ): __lowerCAmelCase = [] __lowerCAmelCase = 1 while len(lowerCAmelCase_ ) < 1E6: constant.append(str(lowerCAmelCase_ ) ) i += 1 __lowerCAmelCase = ''.join(lowerCAmelCase_ ) return ( int(constant[0] ) * int(constant[9] ) * int(constant[99] ) * int(constant[999] ) * int(constant[9999] ) * int(constant[9_9999] ) * int(constant[99_9999] ) ) if __name__ == "__main__": print(solution())
284
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case : Dict = {'configuration_mra': ['MRA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MraConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'MRA_PRETRAINED_MODEL_ARCHIVE_LIST', 'MraForMaskedLM', 'MraForMultipleChoice', 'MraForQuestionAnswering', 'MraForSequenceClassification', 'MraForTokenClassification', 'MraLayer', 'MraModel', 'MraPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys _snake_case : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure)
284
1
import sacrebleu as scb from packaging import version from sacrebleu import TER import datasets _snake_case : Any = '\\n@inproceedings{snover-etal-2006-study,\n title = "A Study of Translation Edit Rate with Targeted Human Annotation",\n author = "Snover, Matthew and\n Dorr, Bonnie and\n Schwartz, Rich and\n Micciulla, Linnea and\n Makhoul, John",\n booktitle = "Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers",\n month = aug # " 8-12",\n year = "2006",\n address = "Cambridge, Massachusetts, USA",\n publisher = "Association for Machine Translation in the Americas",\n url = "https://aclanthology.org/2006.amta-papers.25",\n pages = "223--231",\n}\n@inproceedings{post-2018-call,\n title = "A Call for Clarity in Reporting {BLEU} Scores",\n author = "Post, Matt",\n booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers",\n month = oct,\n year = "2018",\n address = "Belgium, Brussels",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/W18-6319",\n pages = "186--191",\n}\n' _snake_case : Union[str, Any] = '\\nTER (Translation Edit Rate, also called Translation Error Rate) is a metric to quantify the edit operations that a\nhypothesis requires to match a reference translation. We use the implementation that is already present in sacrebleu\n(https://github.com/mjpost/sacreBLEU#ter), which in turn is inspired by the TERCOM implementation, which can be found\nhere: https://github.com/jhclark/tercom.\n\nThe implementation here is slightly different from sacrebleu in terms of the required input format. The length of\nthe references and hypotheses lists need to be the same, so you may need to transpose your references compared to\nsacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534\n\nSee the README.md file at https://github.com/mjpost/sacreBLEU#ter for more information.\n' _snake_case : List[str] = '\nProduces TER scores alongside the number of edits and reference length.\n\nArgs:\n predictions (list of str): The system stream (a sequence of segments).\n references (list of list of str): A list of one or more reference streams (each a sequence of segments).\n normalized (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`.\n ignore_punct (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`.\n support_zh_ja_chars (boolean): If `True`, tokenization/normalization supports processing of Chinese characters,\n as well as Japanese Kanji, Hiragana, Katakana, and Phonetic Extensions of Katakana.\n Only applies if `normalized = True`. Defaults to `False`.\n case_sensitive (boolean): If `False`, makes all predictions and references lowercase to ignore differences in case. Defaults to `False`.\n\nReturns:\n \'score\' (float): TER score (num_edits / sum_ref_lengths * 100)\n \'num_edits\' (int): The cumulative number of edits\n \'ref_length\' (float): The cumulative average reference length\n\nExamples:\n Example 1:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?",\n ... "What did the TER metric user say to the developer?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"],\n ... ["Your jokes are...", "...TERrible"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... case_sensitive=True)\n >>> print(results)\n {\'score\': 150.0, \'num_edits\': 15, \'ref_length\': 10.0}\n\n Example 2:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... case_sensitive=True)\n >>> print(results)\n {\'score\': 62.5, \'num_edits\': 5, \'ref_length\': 8.0}\n\n Example 3:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... normalized=True,\n ... case_sensitive=True)\n >>> print(results)\n {\'score\': 57.14285714285714, \'num_edits\': 6, \'ref_length\': 10.5}\n\n Example 4:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... ignore_punct=True,\n ... case_sensitive=False)\n >>> print(results)\n {\'score\': 0.0, \'num_edits\': 0, \'ref_length\': 8.0}\n\n Example 5:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?",\n ... "What did the TER metric user say to the developer?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"],\n ... ["Your jokes are...", "...TERrible"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... ignore_punct=True,\n ... case_sensitive=False)\n >>> print(results)\n {\'score\': 100.0, \'num_edits\': 10, \'ref_length\': 10.0}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): """simple docstring""" def lowercase ( self : Tuple ) -> Any: if version.parse(scb.__version__ ) < version.parse('1.4.12' ): raise ImportWarning( 'To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn\'t match this condition.\n' 'You can install it with `pip install "sacrebleu>=1.4.12"`.' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='http://www.cs.umd.edu/~snover/tercom/' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Sequence(datasets.Value('string' , id='sequence' ) , id='references' ), } ) , codebase_urls=['https://github.com/mjpost/sacreBLEU#ter'] , reference_urls=[ 'https://github.com/jhclark/tercom', ] , ) def lowercase ( self : Any , lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = False , ) -> Union[str, Any]: __lowerCAmelCase = len(references[0] ) if any(len(lowerCAmelCase_ ) != references_per_prediction for refs in references ): raise ValueError('Sacrebleu requires the same number of references for each prediction' ) __lowerCAmelCase = [[refs[i] for refs in references] for i in range(lowerCAmelCase_ )] __lowerCAmelCase = TER( normalized=lowerCAmelCase_ , no_punct=lowerCAmelCase_ , asian_support=lowerCAmelCase_ , case_sensitive=lowerCAmelCase_ , ) __lowerCAmelCase = sb_ter.corpus_score(lowerCAmelCase_ , lowerCAmelCase_ ) return {"score": output.score, "num_edits": output.num_edits, "ref_length": output.ref_length}
284
from __future__ import annotations import string from itertools import cycle, product from pathlib import Path _snake_case : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) _snake_case : list[int] = [ord(letter) for letter in string.ascii_lowercase] _snake_case : set[int] = {ord(char) for char in VALID_CHARS} _snake_case : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def a_ ( lowerCAmelCase_ : list[int], lowerCAmelCase_ : tuple[int, ...] ): __lowerCAmelCase = "" __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 for keychar, cipherchar in zip(cycle(lowerCAmelCase_ ), lowerCAmelCase_ ): __lowerCAmelCase = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(lowerCAmelCase_ ) return decoded def a_ ( lowerCAmelCase_ : list[int] ): __lowerCAmelCase = [] for key in product(lowerCAmelCase_, repeat=3 ): __lowerCAmelCase = try_key(lowerCAmelCase_, lowerCAmelCase_ ) if encoded is not None: possibles.append(lowerCAmelCase_ ) return possibles def a_ ( lowerCAmelCase_ : list[str], lowerCAmelCase_ : str ): return [possible for possible in possibles if common_word in possible.lower()] def a_ ( lowerCAmelCase_ : str = "p059_cipher.txt" ): __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = 42 __lowerCAmelCase = Path(lowerCAmelCase_ ).parent.joinpath(lowerCAmelCase_ ).read_text(encoding='utf-8' ) __lowerCAmelCase = [int(lowerCAmelCase_ ) for number in data.strip().split(',' )] __lowerCAmelCase = filter_valid_chars(lowerCAmelCase_ ) for common_word in COMMON_WORDS: __lowerCAmelCase = filter_common_word(lowerCAmelCase_, lowerCAmelCase_ ) if len(lowerCAmelCase_ ) == 1: break __lowerCAmelCase = possibles[0] return sum(ord(lowerCAmelCase_ ) for char in decoded_text ) if __name__ == "__main__": print(F"""{solution() = }""")
284
1
from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration _snake_case : int = HfArgumentParser(InitializationArguments) _snake_case : int = parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization _snake_case : str = AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks _snake_case : Any = { 'vocab_size': len(tokenizer), 'scale_attn_by_inverse_layer_idx': True, 'reorder_and_upcast_attn': True, } # Load model config (GPT-2 large in this case) _snake_case : Optional[int] = AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config _snake_case : List[str] = AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
284
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem _snake_case : Tuple = importlib.util.find_spec('s3fs') is not None if _has_safs: from .safilesystem import SaFileSystem # noqa: F401 _snake_case : List[compression.BaseCompressedFileFileSystem] = [ compression.BzaFileSystem, compression.GzipFileSystem, compression.LzaFileSystem, compression.XzFileSystem, compression.ZstdFileSystem, ] # Register custom filesystems for fs_class in COMPRESSION_FILESYSTEMS + [HfFileSystem]: if fs_class.protocol in fsspec.registry and fsspec.registry[fs_class.protocol] is not fs_class: warnings.warn(F"""A filesystem protocol was already set for {fs_class.protocol} and will be overwritten.""") fsspec.register_implementation(fs_class.protocol, fs_class, clobber=True) def a_ ( lowerCAmelCase_ : str ): if "://" in dataset_path: __lowerCAmelCase = dataset_path.split('://' )[1] return dataset_path def a_ ( lowerCAmelCase_ : fsspec.AbstractFileSystem ): if fs is not None and fs.protocol != "file": return True else: return False def a_ ( lowerCAmelCase_ : fsspec.AbstractFileSystem, lowerCAmelCase_ : str, lowerCAmelCase_ : str ): __lowerCAmelCase = not is_remote_filesystem(lowerCAmelCase_ ) if is_local: # LocalFileSystem.mv does copy + rm, it is more efficient to simply move a local directory shutil.move(fs._strip_protocol(lowerCAmelCase_ ), fs._strip_protocol(lowerCAmelCase_ ) ) else: fs.mv(lowerCAmelCase_, lowerCAmelCase_, recursive=lowerCAmelCase_ ) def a_ ( ): if hasattr(fsspec.asyn, 'reset_lock' ): # for future fsspec>2022.05.0 fsspec.asyn.reset_lock() else: __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = threading.Lock()
284
1
def a_ ( lowerCAmelCase_ : str ): return [ txt[:a] + txt[a].upper() + txt[a + 1 :] for a in range(len(lowerCAmelCase_ ) ) if txt[a].isalpha() ] if __name__ == "__main__": __import__('doctest').testmod()
284
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def a_ ( lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = filter(lambda lowerCAmelCase_ : p.requires_grad, model.parameters() ) __lowerCAmelCase = sum([np.prod(p.size() ) for p in model_parameters] ) return params _snake_case : Dict = logging.getLogger(__name__) def a_ ( lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Optional[int] ): if metric == "rouge2": __lowerCAmelCase = '{val_avg_rouge2:.4f}-{step_count}' elif metric == "bleu": __lowerCAmelCase = '{val_avg_bleu:.4f}-{step_count}' elif metric == "em": __lowerCAmelCase = '{val_avg_em:.4f}-{step_count}' else: raise NotImplementedError( F"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" ' function.' ) __lowerCAmelCase = ModelCheckpoint( dirpath=lowerCAmelCase_, filename=lowerCAmelCase_, monitor=F"""val_{metric}""", mode='max', save_top_k=3, every_n_epochs=1, ) return checkpoint_callback def a_ ( lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : Any ): return EarlyStopping( monitor=F"""val_{metric}""", mode='min' if 'loss' in metric else 'max', patience=lowerCAmelCase_, verbose=lowerCAmelCase_, ) class _UpperCAmelCase ( pl.Callback ): """simple docstring""" def lowercase ( self : Tuple , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : int ) -> Any: __lowerCAmelCase = {f"""lr_group_{i}""": param['lr'] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Optional[int] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule , lowerCAmelCase_ : str , lowerCAmelCase_ : List[Any]=True ) -> None: logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) __lowerCAmelCase = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['log', 'progress_bar', 'preds']} ) # Log results __lowerCAmelCase = Path(pl_module.hparams.output_dir ) if type_path == "test": __lowerCAmelCase = od / 'test_results.txt' __lowerCAmelCase = od / 'test_generations.txt' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __lowerCAmelCase = od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" __lowerCAmelCase = od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) generations_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'a+' ) as writer: for key in sorted(lowerCAmelCase_ ): if key in ["log", "progress_bar", "preds"]: continue __lowerCAmelCase = metrics[key] if isinstance(lowerCAmelCase_ , torch.Tensor ): __lowerCAmelCase = val.item() __lowerCAmelCase = f"""{key}: {val:.6f}\n""" writer.write(lowerCAmelCase_ ) if not save_generations: return if "preds" in metrics: __lowerCAmelCase = '\n'.join(metrics['preds'] ) generations_file.open('w+' ).write(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[str] ) -> Dict: try: __lowerCAmelCase = pl_module.model.model.num_parameters() except AttributeError: __lowerCAmelCase = pl_module.model.num_parameters() __lowerCAmelCase = count_trainable_parameters(lowerCAmelCase_ ) # mp stands for million parameters trainer.logger.log_metrics({'n_params': npars, 'mp': npars / 1e6, 'grad_mp': n_trainable_pars / 1e6} ) @rank_zero_only def lowercase ( self : int , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule ) -> Any: save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(lowerCAmelCase_ , lowerCAmelCase_ , 'test' ) @rank_zero_only def lowercase ( self : List[Any] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : Any ) -> int: save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
284
1
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem _snake_case : Tuple = importlib.util.find_spec('s3fs') is not None if _has_safs: from .safilesystem import SaFileSystem # noqa: F401 _snake_case : List[compression.BaseCompressedFileFileSystem] = [ compression.BzaFileSystem, compression.GzipFileSystem, compression.LzaFileSystem, compression.XzFileSystem, compression.ZstdFileSystem, ] # Register custom filesystems for fs_class in COMPRESSION_FILESYSTEMS + [HfFileSystem]: if fs_class.protocol in fsspec.registry and fsspec.registry[fs_class.protocol] is not fs_class: warnings.warn(F"""A filesystem protocol was already set for {fs_class.protocol} and will be overwritten.""") fsspec.register_implementation(fs_class.protocol, fs_class, clobber=True) def a_ ( lowerCAmelCase_ : str ): if "://" in dataset_path: __lowerCAmelCase = dataset_path.split('://' )[1] return dataset_path def a_ ( lowerCAmelCase_ : fsspec.AbstractFileSystem ): if fs is not None and fs.protocol != "file": return True else: return False def a_ ( lowerCAmelCase_ : fsspec.AbstractFileSystem, lowerCAmelCase_ : str, lowerCAmelCase_ : str ): __lowerCAmelCase = not is_remote_filesystem(lowerCAmelCase_ ) if is_local: # LocalFileSystem.mv does copy + rm, it is more efficient to simply move a local directory shutil.move(fs._strip_protocol(lowerCAmelCase_ ), fs._strip_protocol(lowerCAmelCase_ ) ) else: fs.mv(lowerCAmelCase_, lowerCAmelCase_, recursive=lowerCAmelCase_ ) def a_ ( ): if hasattr(fsspec.asyn, 'reset_lock' ): # for future fsspec>2022.05.0 fsspec.asyn.reset_lock() else: __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = threading.Lock()
284
import re from filelock import FileLock try: import nltk _snake_case : Any = True except (ImportError, ModuleNotFoundError): _snake_case : Union[str, Any] = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def a_ ( lowerCAmelCase_ : str ): re.sub('<n>', '', lowerCAmelCase_ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(lowerCAmelCase_ ) )
284
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _snake_case : Optional[Any] = { 'configuration_vivit': ['VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'VivitConfig'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : str = ['VivitImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[str] = [ 'VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'VivitModel', 'VivitPreTrainedModel', 'VivitForVideoClassification', ] if TYPE_CHECKING: from .configuration_vivit import VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, VivitConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_vivit import VivitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vivit import ( VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST, VivitForVideoClassification, VivitModel, VivitPreTrainedModel, ) else: import sys _snake_case : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case : List[Any] = { 'configuration_biogpt': ['BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BioGptConfig'], 'tokenization_biogpt': ['BioGptTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BioGptForCausalLM', 'BioGptForTokenClassification', 'BioGptForSequenceClassification', 'BioGptModel', 'BioGptPreTrainedModel', ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys _snake_case : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
284
1
import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : List[str] , lowerCAmelCase_ : List[Any] ) -> Optional[Any]: __lowerCAmelCase = 3 __lowerCAmelCase = 2_5_0 __lowerCAmelCase = ids_tensor((batch_size, length) , lowerCAmelCase_ ) __lowerCAmelCase = torch.ones((batch_size, length) , device=lowerCAmelCase_ , dtype=torch.float ) / length return input_ids, scores def lowercase ( self : List[Any] ) -> Any: __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(5 ) __lowerCAmelCase = StoppingCriteriaList( [ MaxLengthCriteria(max_length=1_0 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(1_0 ) self.assertTrue(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) def lowercase ( self : List[str] ) -> Union[str, Any]: __lowerCAmelCase = MaxLengthCriteria(max_length=1_0 ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(1_0 ) self.assertTrue(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) def lowercase ( self : List[Any] ) -> Optional[Any]: __lowerCAmelCase = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(1_0 ) self.assertTrue(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 1_0 ) def lowercase ( self : Dict ) -> int: __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(5 ) __lowerCAmelCase = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) def lowercase ( self : List[str] ) -> List[Any]: validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(1_0 )] ) , 1_0 ) with self.assertWarns(lowerCAmelCase_ ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(1_0 )] ) , 1_1 ) __lowerCAmelCase = validate_stopping_criteria(StoppingCriteriaList() , 1_1 ) self.assertEqual(len(lowerCAmelCase_ ) , 1 )
284
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets _snake_case : Tuple = '\\n@article{hendrycksmath2021,\n title={Measuring Mathematical Problem Solving With the MATH Dataset},\n author={Dan Hendrycks\n and Collin Burns\n and Saurav Kadavath\n and Akul Arora\n and Steven Basart\n and Eric Tang\n and Dawn Song\n and Jacob Steinhardt},\n journal={arXiv preprint arXiv:2103.03874},\n year={2021}\n}\n' _snake_case : str = '\\nThis metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset.\nIt first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy.\n' _snake_case : List[str] = R'\nCalculates accuracy after canonicalizing inputs.\n\nArgs:\n predictions: list of predictions to score. Each prediction\n is a string that contains natural language and LaTex.\n references: list of reference for each prediction. Each\n reference is a string that contains natural language\n and LaTex.\nReturns:\n accuracy: accuracy after canonicalizing inputs\n (e.g., converting "1/2" to "\\frac{1}{2}")\n\nExamples:\n >>> metric = datasets.load_metric("competition_math")\n >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"])\n >>> print(results)\n {\'accuracy\': 1.0}\n' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): """simple docstring""" def lowercase ( self : str ) -> int: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' ), 'references': datasets.Value('string' ), } ) , homepage='https://github.com/hendrycks/math' , codebase_urls=['https://github.com/hendrycks/math'] , ) def lowercase ( self : str , lowerCAmelCase_ : Any , lowerCAmelCase_ : Optional[int] ) -> List[Any]: __lowerCAmelCase = 0.0 for i, j in zip(lowerCAmelCase_ , lowerCAmelCase_ ): n_correct += 1.0 if math_equivalence.is_equiv(lowerCAmelCase_ , lowerCAmelCase_ ) else 0.0 __lowerCAmelCase = n_correct / len(lowerCAmelCase_ ) return { "accuracy": accuracy, }
284
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : Any = { 'google/pegasus-large': 'https://huggingface.co/google/pegasus-large/resolve/main/config.json', # See all PEGASUS models at https://huggingface.co/models?filter=pegasus } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = """pegasus""" a_ = ["""past_key_values"""] a_ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : List[str] , lowerCAmelCase_ : Union[str, Any]=5_0_2_6_5 , lowerCAmelCase_ : Union[str, Any]=1_0_2_4 , lowerCAmelCase_ : Union[str, Any]=1_2 , lowerCAmelCase_ : Dict=4_0_9_6 , lowerCAmelCase_ : str=1_6 , lowerCAmelCase_ : List[Any]=1_2 , lowerCAmelCase_ : Union[str, Any]=4_0_9_6 , lowerCAmelCase_ : Union[str, Any]=1_6 , lowerCAmelCase_ : Any=0.0 , lowerCAmelCase_ : Optional[int]=0.0 , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Any=True , lowerCAmelCase_ : Dict="gelu" , lowerCAmelCase_ : Optional[int]=1_0_2_4 , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : str=0.0 , lowerCAmelCase_ : Union[str, Any]=0.0 , lowerCAmelCase_ : Optional[int]=0.02 , lowerCAmelCase_ : Tuple=0 , lowerCAmelCase_ : Tuple=False , lowerCAmelCase_ : Union[str, Any]=0 , lowerCAmelCase_ : int=1 , lowerCAmelCase_ : Tuple=1 , **lowerCAmelCase_ : Tuple , ) -> List[str]: __lowerCAmelCase = vocab_size __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = d_model __lowerCAmelCase = encoder_ffn_dim __lowerCAmelCase = encoder_layers __lowerCAmelCase = encoder_attention_heads __lowerCAmelCase = decoder_ffn_dim __lowerCAmelCase = decoder_layers __lowerCAmelCase = decoder_attention_heads __lowerCAmelCase = dropout __lowerCAmelCase = attention_dropout __lowerCAmelCase = activation_dropout __lowerCAmelCase = activation_function __lowerCAmelCase = init_std __lowerCAmelCase = encoder_layerdrop __lowerCAmelCase = decoder_layerdrop __lowerCAmelCase = use_cache __lowerCAmelCase = encoder_layers __lowerCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , is_encoder_decoder=lowerCAmelCase_ , decoder_start_token_id=lowerCAmelCase_ , forced_eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ , ) @property def lowercase ( self : List[Any] ) -> int: return self.encoder_attention_heads @property def lowercase ( self : Optional[Any] ) -> int: return self.d_model
284
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = StableDiffusionInstructPixaPixPipeline a_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width""", """cross_attention_kwargs"""} a_ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase ( self : Optional[int] ) -> Optional[int]: torch.manual_seed(0 ) __lowerCAmelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=8 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , ) __lowerCAmelCase = PNDMScheduler(skip_prk_steps=lowerCAmelCase_ ) torch.manual_seed(0 ) __lowerCAmelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) __lowerCAmelCase = CLIPTextModel(lowerCAmelCase_ ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __lowerCAmelCase = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowercase ( self : Optional[Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Tuple=0 ) -> Dict: __lowerCAmelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) __lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCAmelCase = Image.fromarray(np.uinta(lowerCAmelCase_ ) ).convert('RGB' ) if str(lowerCAmelCase_ ).startswith('mps' ): __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) else: __lowerCAmelCase = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'image_guidance_scale': 1, 'output_type': 'numpy', } return inputs def lowercase ( self : Tuple ) -> List[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_26, 0.37_50, 0.45_47, 0.61_17, 0.58_66, 0.50_16, 0.43_27, 0.56_42, 0.48_15] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Dict: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = 'french fries' __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ , negative_prompt=lowerCAmelCase_ ) __lowerCAmelCase = output.images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_11, 0.36_42, 0.45_53, 0.62_36, 0.57_97, 0.50_13, 0.43_43, 0.56_11, 0.48_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Any: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = [inputs['prompt']] * 2 __lowerCAmelCase = np.array(inputs['image'] ).astype(np.floataa ) / 2_55.0 __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ).unsqueeze(0 ).to(lowerCAmelCase_ ) __lowerCAmelCase = image / 2 + 0.5 __lowerCAmelCase = image.permute(0 , 3 , 1 , 2 ) __lowerCAmelCase = image.repeat(2 , 1 , 1 , 1 ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[-1, -3:, -3:, -1] assert image.shape == (2, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.58_12, 0.57_48, 0.52_22, 0.59_08, 0.56_95, 0.71_74, 0.68_04, 0.55_23, 0.55_79] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Dict ) -> Optional[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = EulerAncestralDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' ) __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = [round(lowerCAmelCase_ , 4 ) for x in image_slice.flatten().tolist()] print(','.join([str(lowerCAmelCase_ ) for x in slice] ) ) assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.74_17, 0.38_42, 0.47_32, 0.57_76, 0.58_91, 0.51_39, 0.40_52, 0.56_73, 0.49_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> Optional[int]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def lowercase ( self : Optional[Any] ) -> Optional[Any]: __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = VaeImageProcessor(do_resize=lowerCAmelCase_ , do_normalize=lowerCAmelCase_ ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = pipe(**self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) )[0] __lowerCAmelCase = components['vae'] __lowerCAmelCase = self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): __lowerCAmelCase = vae.encode(inputs[image_param] ).latent_dist.mode() __lowerCAmelCase = pipe(**lowerCAmelCase_ )[0] __lowerCAmelCase = np.abs(out - out_latents_inputs ).max() self.assertLess(lowerCAmelCase_ , 1e-4 , 'passing latents as image input generate different result from passing image' ) @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> Optional[int]: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase ( self : List[str] , lowerCAmelCase_ : List[Any]=0 ) -> Any: __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = load_image( 'https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg' ) __lowerCAmelCase = { 'prompt': 'turn him into a cyborg', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'image_guidance_scale': 1.0, 'output_type': 'numpy', } return inputs def lowercase ( self : List[Any] ) -> str: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.59_02, 0.60_15, 0.60_27, 0.59_83, 0.60_92, 0.60_61, 0.57_65, 0.57_85, 0.55_55] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Tuple ) -> List[str]: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.65_78, 0.68_17, 0.69_72, 0.67_61, 0.68_56, 0.69_16, 0.64_28, 0.65_16, 0.63_01] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[Any] ) -> Dict: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.38_28, 0.38_34, 0.38_18, 0.37_92, 0.38_65, 0.37_52, 0.37_92, 0.38_47, 0.37_53] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> int: __lowerCAmelCase = 0 def callback_fn(lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : torch.FloatTensor ) -> None: __lowerCAmelCase = True nonlocal number_of_steps number_of_steps += 1 if step == 1: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.24_63, -0.46_44, -0.97_56, 1.51_76, 1.44_14, 0.78_66, 0.98_97, 0.85_21, 0.79_83] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.26_44, -0.46_26, -0.96_53, 1.51_76, 1.45_51, 0.76_86, 0.98_05, 0.84_52, 0.81_15] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 __lowerCAmelCase = False __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() pipe(**lowerCAmelCase_ , callback=lowerCAmelCase_ , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def lowercase ( self : Optional[int] ) -> Any: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 1_0**9 def lowercase ( self : List[Any] ) -> Any: __lowerCAmelCase = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 __lowerCAmelCase = inputs['image'].resize((5_0_4, 5_0_4) ) __lowerCAmelCase = 'timbrooks/instruct-pix2pix' __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( lowerCAmelCase_ , safety_checker=lowerCAmelCase_ , ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = output.images[0] __lowerCAmelCase = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 5_0_4, 3) __lowerCAmelCase = np.array([0.27_26, 0.25_29, 0.26_64, 0.26_55, 0.26_41, 0.26_42, 0.25_91, 0.26_49, 0.25_90] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3
284
1