code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def _lowercase( __a : Optional[Any] ): return EnvironmentCommand() class lowercase_ (lowercase__ ): @staticmethod def __UpperCamelCase ( lowercase_) -> Optional[int]: a__ =parser.add_parser('env') download_parser.set_defaults(func=lowercase_) def __UpperCamelCase ( self) -> Dict: a__ =huggingface_hub.__version__ a__ ='not installed' a__ ='NA' if is_torch_available(): import torch a__ =torch.__version__ a__ =torch.cuda.is_available() a__ ='not installed' if is_transformers_available(): import transformers a__ =transformers.__version__ a__ ='not installed' if is_accelerate_available(): import accelerate a__ =accelerate.__version__ a__ ='not installed' if is_xformers_available(): import xformers a__ =xformers.__version__ a__ ={ '`diffusers` version': version, 'Platform': platform.platform(), 'Python version': platform.python_version(), 'PyTorch version (GPU?)': F"""{pt_version} ({pt_cuda_available})""", 'Huggingface_hub version': hub_version, 'Transformers version': transformers_version, 'Accelerate version': accelerate_version, 'xFormers version': xformers_version, 'Using GPU in script?': '<fill in>', 'Using distributed or parallel set-up in script?': '<fill in>', } print('\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n') print(self.format_dict(lowercase_)) return info @staticmethod def __UpperCamelCase ( lowercase_) -> List[Any]: return "\n".join([F"""- {prop}: {val}""" for prop, val in d.items()]) + "\n"
20
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : Tuple = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : List[Any] = { """MIT/ast-finetuned-audioset-10-10-0.4593""": ( """https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json""" ), } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'audio-spectrogram-transformer' def __init__( self , _lowerCAmelCase=768 , _lowerCAmelCase=12 , _lowerCAmelCase=12 , _lowerCAmelCase=3072 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=1e-12 , _lowerCAmelCase=16 , _lowerCAmelCase=True , _lowerCAmelCase=10 , _lowerCAmelCase=10 , _lowerCAmelCase=1024 , _lowerCAmelCase=128 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : List[Any] = num_attention_heads UpperCAmelCase__ : Dict = intermediate_size UpperCAmelCase__ : Dict = hidden_act UpperCAmelCase__ : str = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : Tuple = initializer_range UpperCAmelCase__ : Dict = layer_norm_eps UpperCAmelCase__ : Optional[Any] = patch_size UpperCAmelCase__ : Tuple = qkv_bias UpperCAmelCase__ : Tuple = frequency_stride UpperCAmelCase__ : Union[str, Any] = time_stride UpperCAmelCase__ : Optional[Any] = max_length UpperCAmelCase__ : Optional[int] = num_mel_bins
79
0
def lowerCAmelCase_ ( lowerCamelCase = 1000 ): return sum(e for e in range(3 , lowerCamelCase ) if e % 3 == 0 or e % 5 == 0 ) if __name__ == "__main__": print(F"""{solution() = }""")
21
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) class UpperCAmelCase_ ( __lowerCamelCase ): def __init__( self , *_lowerCAmelCase , **_lowerCAmelCase ): warnings.warn( """The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use GLPNImageProcessor instead.""" , _lowerCAmelCase , ) super().__init__(*_lowerCAmelCase , **_lowerCAmelCase )
79
0
'''simple docstring''' def snake_case_ (UpperCamelCase : list ): '''simple docstring''' if len(UpperCamelCase ) <= 1: return lst _a = 1 while i < len(UpperCamelCase ): if lst[i - 1] <= lst[i]: i += 1 else: _a , _a = lst[i], lst[i - 1] i -= 1 if i == 0: _a = 1 return lst if __name__ == "__main__": _snake_case : Optional[int] = input('Enter numbers separated by a comma:\n').strip() _snake_case : Union[str, Any] = [int(item) for item in user_input.split(',')] print(gnome_sort(unsorted))
22
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : int = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__ : List[str] = { """vocab_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt""", }, """tokenizer_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json""" ), """google/realm-orqa-nq-openqa""": ( """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-nq-reader""": ( """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-openqa""": ( """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-reader""": ( """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json""" ), }, } SCREAMING_SNAKE_CASE__ : Optional[Any] = { """google/realm-cc-news-pretrained-embedder""": 5_12, """google/realm-cc-news-pretrained-encoder""": 5_12, """google/realm-cc-news-pretrained-scorer""": 5_12, """google/realm-cc-news-pretrained-openqa""": 5_12, """google/realm-orqa-nq-openqa""": 5_12, """google/realm-orqa-nq-reader""": 5_12, """google/realm-orqa-wq-openqa""": 5_12, """google/realm-orqa-wq-reader""": 5_12, } SCREAMING_SNAKE_CASE__ : Optional[Any] = { """google/realm-cc-news-pretrained-embedder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-encoder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-scorer""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-reader""": {"""do_lower_case""": True}, """google/realm-orqa-wq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-wq-reader""": {"""do_lower_case""": True}, } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = PRETRAINED_INIT_CONFIGURATION __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = RealmTokenizer def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=True , _lowerCAmelCase="[UNK]" , _lowerCAmelCase="[SEP]" , _lowerCAmelCase="[PAD]" , _lowerCAmelCase="[CLS]" , _lowerCAmelCase="[MASK]" , _lowerCAmelCase=True , _lowerCAmelCase=None , **_lowerCAmelCase , ): super().__init__( _lowerCAmelCase , tokenizer_file=_lowerCAmelCase , do_lower_case=_lowerCAmelCase , unk_token=_lowerCAmelCase , sep_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , cls_token=_lowerCAmelCase , mask_token=_lowerCAmelCase , tokenize_chinese_chars=_lowerCAmelCase , strip_accents=_lowerCAmelCase , **_lowerCAmelCase , ) UpperCAmelCase__ : Tuple = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , _lowerCAmelCase ) != do_lower_case or normalizer_state.get("""strip_accents""" , _lowerCAmelCase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , _lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase__ : Any = getattr(_lowerCAmelCase , normalizer_state.pop("""type""" ) ) UpperCAmelCase__ : str = do_lower_case UpperCAmelCase__ : Tuple = strip_accents UpperCAmelCase__ : Tuple = tokenize_chinese_chars UpperCAmelCase__ : Union[str, Any] = normalizer_class(**_lowerCAmelCase ) UpperCAmelCase__ : Dict = do_lower_case def __UpperCAmelCase ( self , _lowerCAmelCase , **_lowerCAmelCase ): UpperCAmelCase__ : List[Any] = PaddingStrategy.MAX_LENGTH UpperCAmelCase__ : Optional[int] = text UpperCAmelCase__ : Optional[int] = kwargs.pop("""text_pair""" , _lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = kwargs.pop("""return_tensors""" , _lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = { """input_ids""": [], """attention_mask""": [], """token_type_ids""": [], } for idx, candidate_text in enumerate(_lowerCAmelCase ): if batch_text_pair is not None: UpperCAmelCase__ : str = batch_text_pair[idx] else: UpperCAmelCase__ : Any = None UpperCAmelCase__ : str = super().__call__(_lowerCAmelCase , _lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = encoded_candidates.get("""input_ids""" ) UpperCAmelCase__ : str = encoded_candidates.get("""attention_mask""" ) UpperCAmelCase__ : Union[str, Any] = encoded_candidates.get("""token_type_ids""" ) if encoded_input_ids is not None: output_data["input_ids"].append(_lowerCAmelCase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(_lowerCAmelCase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = {key: item for key, item in output_data.items() if len(_lowerCAmelCase ) != 0} return BatchEncoding(_lowerCAmelCase , tensor_type=_lowerCAmelCase ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase = None ): UpperCAmelCase__ : Any = [self.sep_token_id] UpperCAmelCase__ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase = None ): UpperCAmelCase__ : List[str] = self._tokenizer.model.save(_lowerCAmelCase , name=_lowerCAmelCase ) return tuple(_lowerCAmelCase )
79
0
import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html snake_case__ : Union[str, Any] = """platform""" import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class _a : """simple docstring""" A_ = PegasusConfig A_ = {} A_ = """gelu""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase=13 , _UpperCAmelCase=7 , _UpperCAmelCase=True , _UpperCAmelCase=False , _UpperCAmelCase=99 , _UpperCAmelCase=32 , _UpperCAmelCase=5 , _UpperCAmelCase=4 , _UpperCAmelCase=37 , _UpperCAmelCase=0.1 , _UpperCAmelCase=0.1 , _UpperCAmelCase=20 , _UpperCAmelCase=2 , _UpperCAmelCase=1 , _UpperCAmelCase=0 , ) -> Optional[int]: UpperCamelCase_ = parent UpperCamelCase_ = batch_size UpperCamelCase_ = seq_length UpperCamelCase_ = is_training UpperCamelCase_ = use_labels UpperCamelCase_ = vocab_size UpperCamelCase_ = hidden_size UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = num_attention_heads UpperCamelCase_ = intermediate_size UpperCamelCase_ = hidden_dropout_prob UpperCamelCase_ = attention_probs_dropout_prob UpperCamelCase_ = max_position_embeddings UpperCamelCase_ = eos_token_id UpperCamelCase_ = pad_token_id UpperCamelCase_ = bos_token_id def _UpperCAmelCase ( self ) -> str: UpperCamelCase_ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ).clip(3 , self.vocab_size ) UpperCamelCase_ = np.expand_dims(np.array([self.eos_token_id] * self.batch_size ) , 1 ) UpperCamelCase_ = np.concatenate([input_ids, eos_tensor] , axis=1 ) UpperCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase_ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) UpperCamelCase_ = prepare_pegasus_inputs_dict(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return config, inputs_dict def _UpperCAmelCase ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Any: UpperCamelCase_ = 20 UpperCamelCase_ = model_class_name(_UpperCAmelCase ) UpperCamelCase_ = model.encode(inputs_dict['input_ids'] ) UpperCamelCase_ , UpperCamelCase_ = ( inputs_dict['decoder_input_ids'], inputs_dict['decoder_attention_mask'], ) UpperCamelCase_ = model.init_cache(decoder_input_ids.shape[0] , _UpperCAmelCase , _UpperCAmelCase ) UpperCamelCase_ = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype='i4' ) UpperCamelCase_ = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) UpperCamelCase_ = model.decode( decoder_input_ids[:, :-1] , _UpperCAmelCase , decoder_attention_mask=_UpperCAmelCase , past_key_values=_UpperCAmelCase , decoder_position_ids=_UpperCAmelCase , ) UpperCamelCase_ = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='i4' ) UpperCamelCase_ = model.decode( decoder_input_ids[:, -1:] , _UpperCAmelCase , decoder_attention_mask=_UpperCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_position_ids=_UpperCAmelCase , ) UpperCamelCase_ = model.decode(_UpperCAmelCase , _UpperCAmelCase ) UpperCamelCase_ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) def _UpperCAmelCase ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> List[Any]: UpperCamelCase_ = 20 UpperCamelCase_ = model_class_name(_UpperCAmelCase ) UpperCamelCase_ = model.encode(inputs_dict['input_ids'] ) UpperCamelCase_ , UpperCamelCase_ = ( inputs_dict['decoder_input_ids'], inputs_dict['decoder_attention_mask'], ) UpperCamelCase_ = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) UpperCamelCase_ = model.init_cache(decoder_input_ids.shape[0] , _UpperCAmelCase , _UpperCAmelCase ) UpperCamelCase_ = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) UpperCamelCase_ = model.decode( decoder_input_ids[:, :-1] , _UpperCAmelCase , decoder_attention_mask=_UpperCAmelCase , past_key_values=_UpperCAmelCase , decoder_position_ids=_UpperCAmelCase , ) UpperCamelCase_ = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='i4' ) UpperCamelCase_ = model.decode( decoder_input_ids[:, -1:] , _UpperCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=_UpperCAmelCase , decoder_position_ids=_UpperCAmelCase , ) UpperCamelCase_ = model.decode(_UpperCAmelCase , _UpperCAmelCase , decoder_attention_mask=_UpperCAmelCase ) UpperCamelCase_ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) def _snake_case (__lowercase , __lowercase , __lowercase , __lowercase=None , __lowercase=None , ): if attention_mask is None: UpperCamelCase_ = np.not_equal(__lowercase , config.pad_token_id).astype(np.inta) if decoder_attention_mask is None: UpperCamelCase_ = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape , dtype=np.inta), np.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id).astype(np.inta), ] , axis=-1 , ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class _a ( UpperCAmelCase__ , unittest.TestCase ): """simple docstring""" A_ = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) A_ = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () A_ = True A_ = False A_ = False A_ = False def _UpperCAmelCase ( self ) -> Dict: UpperCamelCase_ = FlaxPegasusModelTester(self ) UpperCamelCase_ = ConfigTester(self , config_class=_UpperCAmelCase ) def _UpperCAmelCase ( self ) -> str: self.config_tester.run_common_tests() def _UpperCAmelCase ( self ) -> Dict: UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) def _UpperCAmelCase ( self ) -> Tuple: UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) def _UpperCAmelCase ( self ) -> Dict: UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCamelCase_ = self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) UpperCamelCase_ = model_class(_UpperCAmelCase ) @jax.jit def encode_jitted(_UpperCAmelCase , _UpperCAmelCase=None , **_UpperCAmelCase ): return model.encode(input_ids=_UpperCAmelCase , attention_mask=_UpperCAmelCase ) with self.subTest('JIT Enabled' ): UpperCamelCase_ = encode_jitted(**_UpperCAmelCase ).to_tuple() with self.subTest('JIT Disabled' ): with jax.disable_jit(): UpperCamelCase_ = encode_jitted(**_UpperCAmelCase ).to_tuple() self.assertEqual(len(_UpperCAmelCase ) , len(_UpperCAmelCase ) ) for jitted_output, output in zip(_UpperCAmelCase , _UpperCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) def _UpperCAmelCase ( self ) -> List[str]: UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCamelCase_ = model_class(_UpperCAmelCase ) UpperCamelCase_ = model.encode(inputs_dict['input_ids'] , inputs_dict['attention_mask'] ) UpperCamelCase_ = { 'decoder_input_ids': inputs_dict['decoder_input_ids'], 'decoder_attention_mask': inputs_dict['decoder_attention_mask'], 'encoder_outputs': encoder_outputs, } @jax.jit def decode_jitted(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): return model.decode( decoder_input_ids=_UpperCAmelCase , decoder_attention_mask=_UpperCAmelCase , encoder_outputs=_UpperCAmelCase , ) with self.subTest('JIT Enabled' ): UpperCamelCase_ = decode_jitted(**_UpperCAmelCase ).to_tuple() with self.subTest('JIT Disabled' ): with jax.disable_jit(): UpperCamelCase_ = decode_jitted(**_UpperCAmelCase ).to_tuple() self.assertEqual(len(_UpperCAmelCase ) , len(_UpperCAmelCase ) ) for jitted_output, output in zip(_UpperCAmelCase , _UpperCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) @slow def _UpperCAmelCase ( self ) -> int: for model_class_name in self.all_model_classes: UpperCamelCase_ = model_class_name.from_pretrained('google/pegasus-large' , from_pt=_UpperCAmelCase ) UpperCamelCase_ = np.ones((1, 1) ) UpperCamelCase_ = model(_UpperCAmelCase ) self.assertIsNotNone(_UpperCAmelCase ) @slow def _UpperCAmelCase ( self ) -> List[Any]: UpperCamelCase_ = FlaxPegasusForConditionalGeneration.from_pretrained('google/pegasus-xsum' ) UpperCamelCase_ = PegasusTokenizer.from_pretrained('google/pegasus-xsum' ) UpperCamelCase_ = [ ' PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.', ' The London trio are up for best UK act and best album, as well as getting two nominations in the best song category."We got told like this morning \'Oh I think you\'re nominated\'", said Dappy."And I was like \'Oh yeah, which one?\' And now we\'ve got nominated for four awards. I mean, wow!"Bandmate Fazer added: "We thought it\'s best of us to come down and mingle with everyone and say hello to the cameras. And now we find we\'ve got four nominations."The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn\'t be too disappointed if they didn\'t win this time around."At the end of the day we\'re grateful to be where we are in our careers."If it don\'t happen then it don\'t happen - live to fight another day and keep on making albums and hits for the fans."Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers\' All These Things That I\'ve Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year\'s Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border."We just done Edinburgh the other day," said Dappy."We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!" ', ] UpperCamelCase_ = [ 'California\'s largest electricity provider has turned off power to hundreds of thousands of customers.', 'Pop group N-Dubz have revealed they were surprised to get four nominations for this year\'s Mobo Awards.', ] UpperCamelCase_ = tokenizer(_UpperCAmelCase , return_tensors='np' , truncation=_UpperCAmelCase , max_length=512 , padding=_UpperCAmelCase ) UpperCamelCase_ = model.generate(**_UpperCAmelCase , num_beams=2 ).sequences UpperCamelCase_ = tokenizer.batch_decode(_UpperCAmelCase , skip_special_tokens=_UpperCAmelCase ) assert tgt_text == decoded
23
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'facebook/bart-large-mnli' __lowerCamelCase = ( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) __lowerCamelCase = 'text_classifier' __lowerCamelCase = AutoTokenizer __lowerCamelCase = AutoModelForSequenceClassification __lowerCamelCase = ['text', ['text']] __lowerCamelCase = ['text'] def __UpperCAmelCase ( self ): super().setup() UpperCAmelCase__ : Optional[Any] = self.model.config UpperCAmelCase__ : Tuple = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("""entail""" ): UpperCAmelCase__ : Dict = int(_lowerCAmelCase ) if self.entailment_id == -1: raise ValueError("""Could not determine the entailment ID from the model config, please pass it at init.""" ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : List[Any] = labels return self.pre_processor( [text] * len(_lowerCAmelCase ) , [f"This example is {label}" for label in labels] , return_tensors="""pt""" , padding="""max_length""" , ) def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : str = outputs.logits UpperCAmelCase__ : List[Any] = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
79
0
'''simple docstring''' import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process UpperCAmelCase_ : Union[str, Any] = logging.getLogger(__name__) UpperCAmelCase_ : List[str] = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) UpperCAmelCase_ : Union[str, Any] = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowerCAmelCase : __lowercase : Optional[str] = field( default=__lowerCAmelCase , metadata={ '''help''': ( '''The model checkpoint for weights initialization. Leave None if you want to train a model from''' ''' scratch.''' ) } , ) __lowercase : Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''If training from scratch, pass a model type from the list: ''' + ''', '''.join(__lowerCAmelCase)} , ) __lowercase : Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''}) __lowercase : Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''}) __lowercase : Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) @dataclass class lowerCAmelCase : __lowercase : Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''The input training data file (a text file).'''}) __lowercase : Optional[str] = field( default=__lowerCAmelCase , metadata={ '''help''': ( '''The input training data files (multiple files in glob format). ''' '''Very often splitting large files to smaller files can prevent tokenizer going out of memory''' ) } , ) __lowercase : Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''An optional input evaluation data file to evaluate the perplexity on (a text file).'''} , ) __lowercase : Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''An optional input train ref data file for whole word mask in Chinese.'''} , ) __lowercase : Optional[str] = field( default=__lowerCAmelCase , metadata={'''help''': '''An optional input eval ref data file for whole word mask in Chinese.'''} , ) __lowercase : bool = field( default=__lowerCAmelCase , metadata={'''help''': '''Whether distinct lines of text in the dataset are to be handled as distinct sequences.'''} , ) __lowercase : bool = field( default=__lowerCAmelCase , metadata={'''help''': '''Train with masked-language modeling loss instead of language modeling.'''}) __lowercase : bool = field(default=__lowerCAmelCase , metadata={'''help''': '''Whether ot not to use whole word mask.'''}) __lowercase : float = field( default=0.15 , metadata={'''help''': '''Ratio of tokens to mask for masked language modeling loss'''}) __lowercase : float = field( default=1 / 6 , metadata={ '''help''': ( '''Ratio of length of a span of masked tokens to surrounding context length for permutation language''' ''' modeling.''' ) } , ) __lowercase : int = field( default=5 , metadata={'''help''': '''Maximum length of a span of masked tokens for permutation language modeling.'''}) __lowercase : int = field( default=-1 , metadata={ '''help''': ( '''Optional input sequence length after tokenization.''' '''The training dataset will be truncated in block of this size for training.''' '''Default to the model max input length for single sentence inputs (take into account special tokens).''' ) } , ) __lowercase : bool = field( default=__lowerCAmelCase , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''}) def _UpperCamelCase (_lowerCamelCase : DataTrainingArguments , _lowerCamelCase : PreTrainedTokenizer , _lowerCamelCase : bool = False , _lowerCamelCase : Optional[str] = None , )-> int: '''simple docstring''' def _dataset(_lowerCamelCase : Any , _lowerCamelCase : List[Any]=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError('''You need to set world whole masking and mlm to True for Chinese Whole Word Mask''' ) return LineByLineWithRefDataset( tokenizer=_lowerCamelCase , file_path=_lowerCamelCase , block_size=args.block_size , ref_path=_lowerCamelCase , ) return LineByLineTextDataset(tokenizer=_lowerCamelCase , file_path=_lowerCamelCase , block_size=args.block_size ) else: return TextDataset( tokenizer=_lowerCamelCase , file_path=_lowerCamelCase , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=_lowerCamelCase , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(_lowerCamelCase ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def _UpperCamelCase ()-> Any: '''simple docstring''' __snake_case = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) __snake_case , __snake_case , __snake_case = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( '''Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file ''' '''or remove the --do_eval argument.''' ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , _lowerCamelCase ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: __snake_case = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: __snake_case = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: __snake_case = CONFIG_MAPPING[model_args.model_type]() logger.warning('''You are instantiating a new config instance from scratch.''' ) if model_args.tokenizer_name: __snake_case = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: __snake_case = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( '''You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another''' ''' script, save it,and load it from here, using --tokenizer_name''' ) if model_args.model_name_or_path: __snake_case = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=_lowerCamelCase , cache_dir=model_args.cache_dir , ) else: logger.info('''Training new model from scratch''' ) __snake_case = AutoModelWithLMHead.from_config(_lowerCamelCase ) model.resize_token_embeddings(len(_lowerCamelCase ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( '''BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the''' '''--mlm flag (masked language modeling).''' ) if data_args.block_size <= 0: __snake_case = tokenizer.max_len # Our input block size will be the max possible for the model else: __snake_case = min(data_args.block_size , tokenizer.max_len ) # Get datasets __snake_case = ( get_dataset(_lowerCamelCase , tokenizer=_lowerCamelCase , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) __snake_case = ( get_dataset(_lowerCamelCase , tokenizer=_lowerCamelCase , evaluate=_lowerCamelCase , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": __snake_case = DataCollatorForPermutationLanguageModeling( tokenizer=_lowerCamelCase , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: __snake_case = DataCollatorForWholeWordMask( tokenizer=_lowerCamelCase , mlm_probability=data_args.mlm_probability ) else: __snake_case = DataCollatorForLanguageModeling( tokenizer=_lowerCamelCase , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer __snake_case = Trainer( model=_lowerCamelCase , args=_lowerCamelCase , data_collator=_lowerCamelCase , train_dataset=_lowerCamelCase , eval_dataset=_lowerCamelCase , prediction_loss_only=_lowerCamelCase , ) # Training if training_args.do_train: __snake_case = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=_lowerCamelCase ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation __snake_case = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) __snake_case = trainer.evaluate() __snake_case = math.exp(eval_output['''eval_loss'''] ) __snake_case = {'''perplexity''': perplexity} __snake_case = os.path.join(training_args.output_dir , '''eval_results_lm.txt''' ) if trainer.is_world_master(): with open(_lowerCamelCase , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key in sorted(result.keys() ): logger.info(''' %s = %s''' , _lowerCamelCase , str(result[key] ) ) writer.write('''%s = %s\n''' % (key, str(result[key] )) ) results.update(_lowerCamelCase ) return results def _UpperCamelCase (_lowerCamelCase : List[Any] )-> Optional[Any]: '''simple docstring''' main() if __name__ == "__main__": main()
24
from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=13 , _lowerCAmelCase=30 , _lowerCAmelCase=2 , _lowerCAmelCase=3 , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=32 , _lowerCAmelCase=2 , _lowerCAmelCase=4 , _lowerCAmelCase=37 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=10 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=3 , _lowerCAmelCase=None , ): UpperCAmelCase__ : Tuple = parent UpperCAmelCase__ : Optional[int] = batch_size UpperCAmelCase__ : Union[str, Any] = image_size UpperCAmelCase__ : int = patch_size UpperCAmelCase__ : str = num_channels UpperCAmelCase__ : int = is_training UpperCAmelCase__ : List[str] = use_labels UpperCAmelCase__ : List[Any] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : Tuple = num_attention_heads UpperCAmelCase__ : Optional[int] = intermediate_size UpperCAmelCase__ : Optional[Any] = hidden_act UpperCAmelCase__ : int = hidden_dropout_prob UpperCAmelCase__ : int = attention_probs_dropout_prob UpperCAmelCase__ : List[str] = type_sequence_label_size UpperCAmelCase__ : Optional[int] = initializer_range UpperCAmelCase__ : Any = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) UpperCAmelCase__ : Any = (image_size // patch_size) ** 2 UpperCAmelCase__ : Tuple = num_patches + 1 def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ : List[str] = None if self.use_labels: UpperCAmelCase__ : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ : Union[str, Any] = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self ): return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : str = TFViTModel(config=_lowerCAmelCase ) UpperCAmelCase__ : str = model(_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase__ : Optional[Any] = self.image_size // 2 UpperCAmelCase__ : List[str] = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase__ : List[Any] = model(_lowerCAmelCase , interpolate_pos_encoding=_lowerCAmelCase , training=_lowerCAmelCase ) UpperCAmelCase__ : str = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : Tuple = self.type_sequence_label_size UpperCAmelCase__ : List[Any] = TFViTForImageClassification(_lowerCAmelCase ) UpperCAmelCase__ : List[str] = model(_lowerCAmelCase , labels=_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase__ : Tuple = self.image_size // 2 UpperCAmelCase__ : Union[str, Any] = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase__ : List[str] = model(_lowerCAmelCase , interpolate_pos_encoding=_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase__ : Union[str, Any] = 1 UpperCAmelCase__ : Optional[Any] = TFViTForImageClassification(_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase__ : List[str] = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : int = config_and_inputs UpperCAmelCase__ : int = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class UpperCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () __lowerCamelCase = ( {'feature-extraction': TFViTModel, 'image-classification': TFViTForImageClassification} if is_tf_available() else {} ) __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = TFViTModelTester(self ) UpperCAmelCase__ : int = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def __UpperCAmelCase ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def __UpperCAmelCase ( self ): pass @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def __UpperCAmelCase ( self ): pass def __UpperCAmelCase ( self ): UpperCAmelCase__ , UpperCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : str = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , tf.keras.layers.Layer ) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : Optional[int] = model_class(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ : Tuple = [*signature.parameters.keys()] UpperCAmelCase__ : str = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = TFViTModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(_lowerCAmelCase ) def _lowerCamelCase ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ : List[str] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class UpperCAmelCase_ ( unittest.TestCase ): @cached_property def __UpperCAmelCase ( self ): return ViTImageProcessor.from_pretrained("""google/vit-base-patch16-224""" ) if is_vision_available() else None @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = TFViTForImageClassification.from_pretrained("""google/vit-base-patch16-224""" ) UpperCAmelCase__ : List[Any] = self.default_image_processor UpperCAmelCase__ : Union[str, Any] = prepare_img() UpperCAmelCase__ : Optional[Any] = image_processor(images=_lowerCAmelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase__ : int = model(**_lowerCAmelCase ) # verify the logits UpperCAmelCase__ : Tuple = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) UpperCAmelCase__ : int = tf.constant([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ) tf.debugging.assert_near(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 )
79
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) a_ = {'configuration_plbart': ['PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PLBartConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ['PLBartTokenizer'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ 'PLBART_PRETRAINED_MODEL_ARCHIVE_LIST', 'PLBartForCausalLM', 'PLBartForConditionalGeneration', 'PLBartForSequenceClassification', 'PLBartModel', 'PLBartPreTrainedModel', ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys a_ = _LazyModule(__name__, globals()['__file__'], _import_structure)
25
from functools import lru_cache @lru_cache def _lowerCamelCase ( __lowerCamelCase ) -> int: '''simple docstring''' if num < 0: raise ValueError("""Number should not be negative.""" ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
79
0
'''simple docstring''' from __future__ import annotations class _A : def __init__( self : int , __magic_name__ : str , __magic_name__ : str ) -> Any: """simple docstring""" __snake_case , __snake_case : str = text, pattern __snake_case , __snake_case : List[Any] = len(__magic_name__ ), len(__magic_name__ ) def lowercase__ ( self : Union[str, Any] , __magic_name__ : str ) -> int: """simple docstring""" for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def lowercase__ ( self : List[Any] , __magic_name__ : int ) -> int: """simple docstring""" for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def lowercase__ ( self : Union[str, Any] ) -> list[int]: """simple docstring""" __snake_case : Tuple = [] for i in range(self.textLen - self.patLen + 1 ): __snake_case : List[Any] = self.mismatch_in_text(__magic_name__ ) if mismatch_index == -1: positions.append(__magic_name__ ) else: __snake_case : List[Any] = self.match_in_pattern(self.text[mismatch_index] ) __snake_case : Tuple = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions __UpperCamelCase = "ABAABA" __UpperCamelCase = "AB" __UpperCamelCase = BoyerMooreSearch(text, pattern) __UpperCamelCase = bms.bad_character_heuristic() if len(positions) == 0: print("No match found") else: print("Pattern found in following positions: ") print(positions)
26
import argparse import hashlib # hashlib is only used inside the Test class import struct class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase ): UpperCAmelCase__ : Any = data UpperCAmelCase__ : List[Any] = [0X6745_2301, 0Xefcd_ab89, 0X98ba_dcfe, 0X1032_5476, 0Xc3d2_e1f0] @staticmethod def __UpperCAmelCase ( _lowerCAmelCase , _lowerCAmelCase ): return ((n << b) | (n >> (32 - b))) & 0Xffff_ffff def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) UpperCAmelCase__ : Optional[int] = self.data + padding + struct.pack(""">Q""" , 8 * len(self.data ) ) return padded_data def __UpperCAmelCase ( self ): return [ self.padded_data[i : i + 64] for i in range(0 , len(self.padded_data ) , 64 ) ] def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : Dict = list(struct.unpack(""">16L""" , _lowerCAmelCase ) ) + [0] * 64 for i in range(16 , 80 ): UpperCAmelCase__ : Optional[int] = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) , 1 ) return w def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[str] = self.padding() UpperCAmelCase__ : List[str] = self.split_blocks() for block in self.blocks: UpperCAmelCase__ : Tuple = self.expand_block(_lowerCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.h for i in range(0 , 80 ): if 0 <= i < 20: UpperCAmelCase__ : Optional[int] = (b & c) | ((~b) & d) UpperCAmelCase__ : int = 0X5a82_7999 elif 20 <= i < 40: UpperCAmelCase__ : Tuple = b ^ c ^ d UpperCAmelCase__ : int = 0X6ed9_eba1 elif 40 <= i < 60: UpperCAmelCase__ : List[str] = (b & c) | (b & d) | (c & d) UpperCAmelCase__ : Tuple = 0X8f1b_bcdc elif 60 <= i < 80: UpperCAmelCase__ : int = b ^ c ^ d UpperCAmelCase__ : str = 0Xca62_c1d6 UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = ( self.rotate(_lowerCAmelCase , 5 ) + f + e + k + expanded_block[i] & 0Xffff_ffff, a, self.rotate(_lowerCAmelCase , 30 ), c, d, ) UpperCAmelCase__ : int = ( self.h[0] + a & 0Xffff_ffff, self.h[1] + b & 0Xffff_ffff, self.h[2] + c & 0Xffff_ffff, self.h[3] + d & 0Xffff_ffff, self.h[4] + e & 0Xffff_ffff, ) return ("{:08x}" * 5).format(*self.h ) def _lowerCamelCase ( ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase__ : Optional[Any] = B"""Test String""" assert SHAaHash(__lowerCamelCase ).final_hash() == hashlib.shaa(__lowerCamelCase ).hexdigest() # noqa: S324 def _lowerCamelCase ( ) -> str: '''simple docstring''' UpperCAmelCase__ : Optional[int] = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument("""--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) UpperCAmelCase__ : str = parser.parse_args() UpperCAmelCase__ : Union[str, Any] = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: UpperCAmelCase__ : List[Any] = f.read() else: UpperCAmelCase__ : int = bytes(__lowerCamelCase , """utf-8""" ) print(SHAaHash(__lowerCamelCase ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
79
0
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableDiffusionUpscalePipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class lowerCamelCase( unittest.TestCase ): '''simple docstring''' def lowerCAmelCase__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def lowerCAmelCase__ ( self ): _A = 1 _A = 3 _A = (32, 32) _A = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(snake_case_ ) return image @property def lowerCAmelCase__ ( self ): torch.manual_seed(0 ) _A = UNetaDConditionModel( block_out_channels=(32, 32, 64) , layers_per_block=2 , sample_size=32 , in_channels=7 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , attention_head_dim=8 , use_linear_projection=snake_case_ , only_cross_attention=(True, True, False) , num_class_embeds=100 , ) return model @property def lowerCAmelCase__ ( self ): torch.manual_seed(0 ) _A = AutoencoderKL( block_out_channels=[32, 32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) return model @property def lowerCAmelCase__ ( self ): torch.manual_seed(0 ) _A = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=512 , ) return CLIPTextModel(snake_case_ ) def lowerCAmelCase__ ( self ): _A = 'cpu' # ensure determinism for the device-dependent torch.Generator _A = self.dummy_cond_unet_upscale _A = DDPMScheduler() _A = DDIMScheduler(prediction_type='v_prediction' ) _A = self.dummy_vae _A = self.dummy_text_encoder _A = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) _A = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] _A = Image.fromarray(np.uinta(snake_case_ ) ).convert('RGB' ).resize((64, 64) ) # make sure here that pndm scheduler skips prk _A = StableDiffusionUpscalePipeline( unet=snake_case_ , low_res_scheduler=snake_case_ , scheduler=snake_case_ , vae=snake_case_ , text_encoder=snake_case_ , tokenizer=snake_case_ , max_noise_level=350 , ) _A = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) _A = 'A painting of a squirrel eating a burger' _A = torch.Generator(device=snake_case_ ).manual_seed(0 ) _A = sd_pipe( [prompt] , image=snake_case_ , generator=snake_case_ , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='np' , ) _A = output.images _A = torch.Generator(device=snake_case_ ).manual_seed(0 ) _A = sd_pipe( [prompt] , image=snake_case_ , generator=snake_case_ , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='np' , return_dict=snake_case_ , )[0] _A = image[0, -3:, -3:, -1] _A = image_from_tuple[0, -3:, -3:, -1] _A = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) _A = np.array([0.3113, 0.3910, 0.4272, 0.4859, 0.5061, 0.4652, 0.5362, 0.5715, 0.5661] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCAmelCase__ ( self ): _A = 'cpu' # ensure determinism for the device-dependent torch.Generator _A = self.dummy_cond_unet_upscale _A = DDPMScheduler() _A = DDIMScheduler(prediction_type='v_prediction' ) _A = self.dummy_vae _A = self.dummy_text_encoder _A = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) _A = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] _A = Image.fromarray(np.uinta(snake_case_ ) ).convert('RGB' ).resize((64, 64) ) # make sure here that pndm scheduler skips prk _A = StableDiffusionUpscalePipeline( unet=snake_case_ , low_res_scheduler=snake_case_ , scheduler=snake_case_ , vae=snake_case_ , text_encoder=snake_case_ , tokenizer=snake_case_ , max_noise_level=350 , ) _A = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) _A = 'A painting of a squirrel eating a burger' _A = sd_pipe( 2 * [prompt] , image=2 * [low_res_image] , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='np' , ) _A = output.images assert image.shape[0] == 2 _A = torch.Generator(device=snake_case_ ).manual_seed(0 ) _A = sd_pipe( [prompt] , image=snake_case_ , generator=snake_case_ , num_images_per_prompt=2 , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type='np' , ) _A = output.images assert image.shape[0] == 2 @unittest.skipIf(torch_device != 'cuda' , 'This test requires a GPU' ) def lowerCAmelCase__ ( self ): _A = self.dummy_cond_unet_upscale _A = DDPMScheduler() _A = DDIMScheduler(prediction_type='v_prediction' ) _A = self.dummy_vae _A = self.dummy_text_encoder _A = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) _A = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] _A = Image.fromarray(np.uinta(snake_case_ ) ).convert('RGB' ).resize((64, 64) ) # put models in fp16, except vae as it overflows in fp16 _A = unet.half() _A = text_encoder.half() # make sure here that pndm scheduler skips prk _A = StableDiffusionUpscalePipeline( unet=snake_case_ , low_res_scheduler=snake_case_ , scheduler=snake_case_ , vae=snake_case_ , text_encoder=snake_case_ , tokenizer=snake_case_ , max_noise_level=350 , ) _A = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) _A = 'A painting of a squirrel eating a burger' _A = torch.manual_seed(0 ) _A = sd_pipe( [prompt] , image=snake_case_ , generator=snake_case_ , num_inference_steps=2 , output_type='np' , ).images _A = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) @slow @require_torch_gpu class lowerCamelCase( unittest.TestCase ): '''simple docstring''' def lowerCAmelCase__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase__ ( self ): _A = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-upscale/low_res_cat.png' ) _A = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale' '/upsampled_cat.npy' ) _A = 'stabilityai/stable-diffusion-x4-upscaler' _A = StableDiffusionUpscalePipeline.from_pretrained(snake_case_ ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing() _A = 'a cat sitting on a park bench' _A = torch.manual_seed(0 ) _A = pipe( prompt=snake_case_ , image=snake_case_ , generator=snake_case_ , output_type='np' , ) _A = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1E-3 def lowerCAmelCase__ ( self ): _A = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-upscale/low_res_cat.png' ) _A = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale' '/upsampled_cat_fp16.npy' ) _A = 'stabilityai/stable-diffusion-x4-upscaler' _A = StableDiffusionUpscalePipeline.from_pretrained( snake_case_ , torch_dtype=torch.floataa , ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing() _A = 'a cat sitting on a park bench' _A = torch.manual_seed(0 ) _A = pipe( prompt=snake_case_ , image=snake_case_ , generator=snake_case_ , output_type='np' , ) _A = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5E-1 def lowerCAmelCase__ ( self ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() _A = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/sd2-upscale/low_res_cat.png' ) _A = 'stabilityai/stable-diffusion-x4-upscaler' _A = StableDiffusionUpscalePipeline.from_pretrained( snake_case_ , torch_dtype=torch.floataa , ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() _A = 'a cat sitting on a park bench' _A = torch.manual_seed(0 ) _A = pipe( prompt=snake_case_ , image=snake_case_ , generator=snake_case_ , num_inference_steps=5 , output_type='np' , ) _A = torch.cuda.max_memory_allocated() # make sure that less than 2.9 GB is allocated assert mem_bytes < 2.9 * 10**9
27
from importlib import import_module from .logging import get_logger SCREAMING_SNAKE_CASE__ : Union[str, Any] = get_logger(__name__) class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : List[str] = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith("""__""" ): setattr(self , _lowerCAmelCase , getattr(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase__ : Tuple = module._original_module if isinstance(_lowerCAmelCase , _PatchedModuleObj ) else module class UpperCAmelCase_ : __lowerCamelCase = [] def __init__( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : str = obj UpperCAmelCase__ : List[str] = target UpperCAmelCase__ : List[str] = new UpperCAmelCase__ : Any = target.split(""".""" )[0] UpperCAmelCase__ : Union[str, Any] = {} UpperCAmelCase__ : str = attrs or [] def __enter__( self ): *UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.target.split(""".""" ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(_lowerCAmelCase ) ): try: UpperCAmelCase__ : Optional[int] = import_module(""".""".join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): UpperCAmelCase__ : Any = getattr(self.obj , _lowerCAmelCase ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(_lowerCAmelCase , _PatchedModuleObj ) and obj_attr._original_module is submodule) ): UpperCAmelCase__ : List[Any] = obj_attr # patch at top level setattr(self.obj , _lowerCAmelCase , _PatchedModuleObj(_lowerCAmelCase , attrs=self.attrs ) ) UpperCAmelCase__ : Optional[Any] = getattr(self.obj , _lowerCAmelCase ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(_lowerCAmelCase , _lowerCAmelCase , _PatchedModuleObj(getattr(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) , attrs=self.attrs ) ) UpperCAmelCase__ : Union[str, Any] = getattr(_lowerCAmelCase , _lowerCAmelCase ) # finally set the target attribute setattr(_lowerCAmelCase , _lowerCAmelCase , self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: UpperCAmelCase__ : Union[str, Any] = getattr(import_module(""".""".join(_lowerCAmelCase ) ) , _lowerCAmelCase ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , _lowerCAmelCase ) is attr_value: UpperCAmelCase__ : Optional[int] = getattr(self.obj , _lowerCAmelCase ) setattr(self.obj , _lowerCAmelCase , self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" UpperCAmelCase__ : Dict = globals()["""__builtins__"""][target_attr] setattr(self.obj , _lowerCAmelCase , self.new ) else: raise RuntimeError(f"Tried to patch attribute {target_attr} instead of a submodule." ) def __exit__( self , *_lowerCAmelCase ): for attr in list(self.original ): setattr(self.obj , _lowerCAmelCase , self.original.pop(_lowerCAmelCase ) ) def __UpperCAmelCase ( self ): self.__enter__() self._active_patches.append(self ) def __UpperCAmelCase ( self ): try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
79
0
'''simple docstring''' from __future__ import annotations import string from itertools import cycle, product from pathlib import Path UpperCamelCase_ = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) UpperCamelCase_ = [ord(letter) for letter in string.ascii_lowercase] UpperCamelCase_ = {ord(char) for char in VALID_CHARS} UpperCamelCase_ = ["the", "be", "to", "of", "and", "in", "that", "have"] def lowercase__( __UpperCamelCase: list[int] ,__UpperCamelCase: tuple[int, ...] ): """simple docstring""" SCREAMING_SNAKE_CASE : str = "" SCREAMING_SNAKE_CASE : int SCREAMING_SNAKE_CASE : int SCREAMING_SNAKE_CASE : int for keychar, cipherchar in zip(cycle(__UpperCamelCase ) ,__UpperCamelCase ): SCREAMING_SNAKE_CASE : Union[str, Any] = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(__UpperCamelCase ) return decoded def lowercase__( __UpperCamelCase: list[int] ): """simple docstring""" SCREAMING_SNAKE_CASE : list[str] = [] for key in product(__UpperCamelCase ,repeat=3 ): SCREAMING_SNAKE_CASE : List[str] = try_key(__UpperCamelCase ,__UpperCamelCase ) if encoded is not None: possibles.append(__UpperCamelCase ) return possibles def lowercase__( __UpperCamelCase: list[str] ,__UpperCamelCase: str ): """simple docstring""" return [possible for possible in possibles if common_word in possible.lower()] def lowercase__( __UpperCamelCase: str = "p059_cipher.txt" ): """simple docstring""" SCREAMING_SNAKE_CASE : list[int] SCREAMING_SNAKE_CASE : list[str] SCREAMING_SNAKE_CASE : str SCREAMING_SNAKE_CASE : str SCREAMING_SNAKE_CASE : str = Path(__UpperCamelCase ).parent.joinpath(__UpperCamelCase ).read_text(encoding='utf-8' ) SCREAMING_SNAKE_CASE : Tuple = [int(__UpperCamelCase ) for number in data.strip().split(',' )] SCREAMING_SNAKE_CASE : Dict = filter_valid_chars(__UpperCamelCase ) for common_word in COMMON_WORDS: SCREAMING_SNAKE_CASE : List[Any] = filter_common_word(__UpperCamelCase ,__UpperCamelCase ) if len(__UpperCamelCase ) == 1: break SCREAMING_SNAKE_CASE : Optional[int] = possibles[0] return sum(ord(__UpperCamelCase ) for char in decoded_text ) if __name__ == "__main__": print(F"""{solution() = }""")
28
from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : str = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Any = { """huggingface/informer-tourism-monthly""": ( """https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json""" ), # See all Informer models at https://huggingface.co/models?filter=informer } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'informer' __lowerCamelCase = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', 'num_hidden_layers': 'encoder_layers', } def __init__( self , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = "student_t" , _lowerCAmelCase = "nll" , _lowerCAmelCase = 1 , _lowerCAmelCase = None , _lowerCAmelCase = "mean" , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = 64 , _lowerCAmelCase = 32 , _lowerCAmelCase = 32 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = True , _lowerCAmelCase = "gelu" , _lowerCAmelCase = 0.0_5 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 100 , _lowerCAmelCase = 0.0_2 , _lowerCAmelCase=True , _lowerCAmelCase = "prob" , _lowerCAmelCase = 5 , _lowerCAmelCase = True , **_lowerCAmelCase , ): # time series specific configuration UpperCAmelCase__ : List[str] = prediction_length UpperCAmelCase__ : Optional[Any] = context_length or prediction_length UpperCAmelCase__ : str = distribution_output UpperCAmelCase__ : int = loss UpperCAmelCase__ : Optional[Any] = input_size UpperCAmelCase__ : Any = num_time_features UpperCAmelCase__ : int = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] UpperCAmelCase__ : Union[str, Any] = scaling UpperCAmelCase__ : Optional[Any] = num_dynamic_real_features UpperCAmelCase__ : List[str] = num_static_real_features UpperCAmelCase__ : str = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(_lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The cardinality should be a list of the same length as `num_static_categorical_features`""" ) UpperCAmelCase__ : List[str] = cardinality else: UpperCAmelCase__ : Optional[Any] = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(_lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The embedding dimension should be a list of the same length as `num_static_categorical_features`""" ) UpperCAmelCase__ : str = embedding_dimension else: UpperCAmelCase__ : List[str] = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] UpperCAmelCase__ : Union[str, Any] = num_parallel_samples # Transformer architecture configuration UpperCAmelCase__ : Dict = input_size * len(self.lags_sequence ) + self._number_of_features UpperCAmelCase__ : Any = d_model UpperCAmelCase__ : int = encoder_attention_heads UpperCAmelCase__ : Optional[Any] = decoder_attention_heads UpperCAmelCase__ : int = encoder_ffn_dim UpperCAmelCase__ : Tuple = decoder_ffn_dim UpperCAmelCase__ : List[Any] = encoder_layers UpperCAmelCase__ : Optional[Any] = decoder_layers UpperCAmelCase__ : Tuple = dropout UpperCAmelCase__ : int = attention_dropout UpperCAmelCase__ : List[str] = activation_dropout UpperCAmelCase__ : Any = encoder_layerdrop UpperCAmelCase__ : Union[str, Any] = decoder_layerdrop UpperCAmelCase__ : Tuple = activation_function UpperCAmelCase__ : Dict = init_std UpperCAmelCase__ : str = use_cache # Informer UpperCAmelCase__ : Union[str, Any] = attention_type UpperCAmelCase__ : int = sampling_factor UpperCAmelCase__ : Any = distil super().__init__(is_encoder_decoder=_lowerCAmelCase , **_lowerCAmelCase ) @property def __UpperCAmelCase ( self ): return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
79
0
"""simple docstring""" from math import log from scipy.constants import Boltzmann, physical_constants A_ = 300 # TEMPERATURE (unit = K) def lowercase ( lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,): if donor_conc <= 0: raise ValueError('''Donor concentration should be positive''' ) elif acceptor_conc <= 0: raise ValueError('''Acceptor concentration should be positive''' ) elif intrinsic_conc <= 0: raise ValueError('''Intrinsic concentration should be positive''' ) elif donor_conc <= intrinsic_conc: raise ValueError( '''Donor concentration should be greater than intrinsic concentration''' ) elif acceptor_conc <= intrinsic_conc: raise ValueError( '''Acceptor concentration should be greater than intrinsic concentration''' ) else: return ( Boltzmann * T * log((donor_conc * acceptor_conc) / intrinsic_conc**2 ) / physical_constants["electron volt"][0] ) if __name__ == "__main__": import doctest doctest.testmod()
29
def _lowerCamelCase ( __lowerCamelCase ) -> bool: '''simple docstring''' if p < 2: raise ValueError("""p should not be less than 2!""" ) elif p == 2: return True UpperCAmelCase__ : Tuple = 4 UpperCAmelCase__ : Tuple = (1 << p) - 1 for _ in range(p - 2 ): UpperCAmelCase__ : List[str] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
79
0
import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class __a: """simple docstring""" def __init__( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE=13 ,_SCREAMING_SNAKE_CASE=7 ,_SCREAMING_SNAKE_CASE=True ,_SCREAMING_SNAKE_CASE=True ,_SCREAMING_SNAKE_CASE=False ,_SCREAMING_SNAKE_CASE=True ,_SCREAMING_SNAKE_CASE=99 ,_SCREAMING_SNAKE_CASE=64 ,_SCREAMING_SNAKE_CASE=5 ,_SCREAMING_SNAKE_CASE=4 ,_SCREAMING_SNAKE_CASE=64 ,_SCREAMING_SNAKE_CASE="gelu" ,_SCREAMING_SNAKE_CASE=0.1 ,_SCREAMING_SNAKE_CASE=0.1 ,_SCREAMING_SNAKE_CASE=512 ,_SCREAMING_SNAKE_CASE=16 ,_SCREAMING_SNAKE_CASE=2 ,_SCREAMING_SNAKE_CASE=0.02 ,_SCREAMING_SNAKE_CASE=3 ,_SCREAMING_SNAKE_CASE=4 ,_SCREAMING_SNAKE_CASE=None ,) -> Any: UpperCAmelCase_ : str = parent UpperCAmelCase_ : str = batch_size UpperCAmelCase_ : Dict = seq_length UpperCAmelCase_ : Union[str, Any] = is_training UpperCAmelCase_ : Optional[int] = use_input_mask UpperCAmelCase_ : Tuple = use_token_type_ids UpperCAmelCase_ : Tuple = use_labels UpperCAmelCase_ : Dict = vocab_size UpperCAmelCase_ : Dict = hidden_size UpperCAmelCase_ : Dict = num_hidden_layers UpperCAmelCase_ : List[Any] = num_attention_heads UpperCAmelCase_ : Tuple = intermediate_size UpperCAmelCase_ : Union[str, Any] = hidden_act UpperCAmelCase_ : Any = hidden_dropout_prob UpperCAmelCase_ : Tuple = attention_probs_dropout_prob UpperCAmelCase_ : str = max_position_embeddings UpperCAmelCase_ : List[str] = type_vocab_size UpperCAmelCase_ : Optional[int] = type_sequence_label_size UpperCAmelCase_ : Dict = initializer_range UpperCAmelCase_ : Any = num_labels UpperCAmelCase_ : Optional[int] = num_choices UpperCAmelCase_ : Optional[int] = scope def a__ ( self ) -> Optional[int]: return MPNetConfig.from_pretrained('''microsoft/mpnet-base''' ) def a__ ( self ) -> Tuple: UpperCAmelCase_ : str = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase_ : int = None if self.use_input_mask: UpperCAmelCase_ : int = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_ : Tuple = None UpperCAmelCase_ : str = None UpperCAmelCase_ : List[str] = None if self.use_labels: UpperCAmelCase_ : Union[str, Any] = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase_ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) UpperCAmelCase_ : Dict = ids_tensor([self.batch_size] ,self.num_choices ) UpperCAmelCase_ : int = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def a__ ( self ) -> Any: return MPNetConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Optional[Any]: UpperCAmelCase_ : int = MPNetModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCAmelCase_ : Any = model(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[str] = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ : List[str] = MPNetForQuestionAnswering(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCAmelCase_ : Optional[int] = model( _SCREAMING_SNAKE_CASE ,attention_mask=_SCREAMING_SNAKE_CASE ,start_positions=_SCREAMING_SNAKE_CASE ,end_positions=_SCREAMING_SNAKE_CASE ,) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Optional[Any]: UpperCAmelCase_ : str = self.num_labels UpperCAmelCase_ : Optional[Any] = MPNetForSequenceClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCAmelCase_ : Any = model(_SCREAMING_SNAKE_CASE ,attention_mask=_SCREAMING_SNAKE_CASE ,labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Optional[int]: UpperCAmelCase_ : Any = self.num_choices UpperCAmelCase_ : List[str] = MPNetForMultipleChoice(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCAmelCase_ : Any = input_ids.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() UpperCAmelCase_ : Optional[int] = input_mask.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() UpperCAmelCase_ : Tuple = model( _SCREAMING_SNAKE_CASE ,attention_mask=_SCREAMING_SNAKE_CASE ,labels=_SCREAMING_SNAKE_CASE ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Any: UpperCAmelCase_ : Dict = self.num_labels UpperCAmelCase_ : int = MPNetForTokenClassification(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCAmelCase_ : Union[str, Any] = model(_SCREAMING_SNAKE_CASE ,attention_mask=_SCREAMING_SNAKE_CASE ,labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def a__ ( self ) -> Optional[int]: UpperCAmelCase_ : Optional[Any] = self.prepare_config_and_inputs() ((UpperCAmelCase_), (UpperCAmelCase_), (UpperCAmelCase_), (UpperCAmelCase_), (UpperCAmelCase_), (UpperCAmelCase_)) : Any = config_and_inputs UpperCAmelCase_ : Tuple = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __a( _a , _a , unittest.TestCase ): """simple docstring""" lowerCAmelCase = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) lowerCAmelCase = ( { '''feature-extraction''': MPNetModel, '''fill-mask''': MPNetForMaskedLM, '''question-answering''': MPNetForQuestionAnswering, '''text-classification''': MPNetForSequenceClassification, '''token-classification''': MPNetForTokenClassification, '''zero-shot''': MPNetForSequenceClassification, } if is_torch_available() else {} ) lowerCAmelCase = False lowerCAmelCase = True def a__ ( self ) -> str: UpperCAmelCase_ : str = MPNetModelTester(self ) UpperCAmelCase_ : str = ConfigTester(self ,config_class=_SCREAMING_SNAKE_CASE ,hidden_size=37 ) def a__ ( self ) -> Optional[int]: self.config_tester.run_common_tests() def a__ ( self ) -> Union[str, Any]: UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*_SCREAMING_SNAKE_CASE ) def a__ ( self ) -> str: UpperCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*_SCREAMING_SNAKE_CASE ) def a__ ( self ) -> List[Any]: UpperCAmelCase_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*_SCREAMING_SNAKE_CASE ) def a__ ( self ) -> Union[str, Any]: UpperCAmelCase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*_SCREAMING_SNAKE_CASE ) def a__ ( self ) -> int: UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*_SCREAMING_SNAKE_CASE ) @require_torch class __a( unittest.TestCase ): """simple docstring""" @slow def a__ ( self ) -> List[str]: UpperCAmelCase_ : Tuple = MPNetModel.from_pretrained('''microsoft/mpnet-base''' ) UpperCAmelCase_ : Dict = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) UpperCAmelCase_ : Any = model(_SCREAMING_SNAKE_CASE )[0] UpperCAmelCase_ : List[Any] = torch.Size((1, 11, 768) ) self.assertEqual(output.shape ,_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Dict = torch.tensor( [[[-0.05_50, 0.19_43, -0.07_40], [-0.05_62, 0.22_11, -0.05_79], [-0.04_37, 0.33_37, -0.06_41]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] ,_SCREAMING_SNAKE_CASE ,atol=1e-4 ) )
30
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE__ : Any = { """configuration_mobilevit""": ["""MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MobileViTConfig""", """MobileViTOnnxConfig"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : List[str] = ["""MobileViTFeatureExtractor"""] SCREAMING_SNAKE_CASE__ : Union[str, Any] = ["""MobileViTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Dict = [ """MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MobileViTForImageClassification""", """MobileViTForSemanticSegmentation""", """MobileViTModel""", """MobileViTPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Any = [ """TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFMobileViTForImageClassification""", """TFMobileViTForSemanticSegmentation""", """TFMobileViTModel""", """TFMobileViTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mobilevit import MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileViTConfig, MobileViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilevit import MobileViTFeatureExtractor from .image_processing_mobilevit import MobileViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilevit import ( MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel, MobileViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilevit import ( TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileViTForImageClassification, TFMobileViTForSemanticSegmentation, TFMobileViTModel, TFMobileViTPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
79
0
from math import isqrt def UpperCAmelCase_ ( __UpperCAmelCase : int ) -> bool: return all(number % divisor != 0 for divisor in range(2 , isqrt(__UpperCAmelCase ) + 1 ) ) def UpperCAmelCase_ ( __UpperCAmelCase : int = 10**6 ) -> int: SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = 1 SCREAMING_SNAKE_CASE_ = 7 while prime_candidate < max_prime: primes_count += is_prime(__UpperCAmelCase ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(f'''{solution() = }''')
31
from __future__ import annotations SCREAMING_SNAKE_CASE__ : List[str] = 8.988e9 # units = N * m^s * C^-2 def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> dict[str, float]: '''simple docstring''' UpperCAmelCase__ : int = abs(chargea * chargea ) if (force, chargea, chargea, distance).count(0 ) != 1: raise ValueError("""One and only one argument must be 0""" ) if distance < 0: raise ValueError("""Distance cannot be negative""" ) if force == 0: UpperCAmelCase__ : int = COULOMBS_CONSTANT * charge_product / (distance**2) return {"force": force} elif chargea == 0: UpperCAmelCase__ : str = abs(__lowerCamelCase ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge1": chargea} elif chargea == 0: UpperCAmelCase__ : Union[str, Any] = abs(__lowerCamelCase ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge2": chargea} elif distance == 0: UpperCAmelCase__ : Optional[Any] = (COULOMBS_CONSTANT * charge_product / abs(__lowerCamelCase )) ** 0.5 return {"distance": distance} raise ValueError("""Exactly one argument must be 0""" ) if __name__ == "__main__": import doctest doctest.testmod()
79
0
import copy import os import cva import numpy as np from matplotlib import pyplot as plt class __UpperCamelCase : def __init__( self ): _UpperCAmelCase = '''''' _UpperCAmelCase = '''''' _UpperCAmelCase = [] _UpperCAmelCase = 0 _UpperCAmelCase = 256 _UpperCAmelCase = 0 _UpperCAmelCase = 0 _UpperCAmelCase = 0 _UpperCAmelCase = 0 def UpperCamelCase( self , _UpperCamelCase ): _UpperCAmelCase = cva.imread(_UpperCamelCase , 0 ) _UpperCAmelCase = copy.deepcopy(self.img ) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = plt.hist(self.img.ravel() , 256 , [0, 256] , label='''x''' ) _UpperCAmelCase = np.sum(_UpperCamelCase ) for i in range(len(_UpperCamelCase ) ): _UpperCAmelCase = x[i] / self.k self.sk += prk _UpperCAmelCase = (self.L - 1) * self.sk if self.rem != 0: _UpperCAmelCase = int(last % last ) _UpperCAmelCase = int(last + 1 if self.rem >= 0.5 else last ) self.last_list.append(_UpperCamelCase ) _UpperCAmelCase = int(np.ma.count(self.img ) / self.img[1].size ) _UpperCAmelCase = self.img[1].size for i in range(self.number_of_cols ): for j in range(self.number_of_rows ): _UpperCAmelCase = self.img[j][i] if num != self.last_list[num]: _UpperCAmelCase = self.last_list[num] cva.imwrite('''output_data/output.jpg''' , self.img ) def UpperCamelCase( self ): plt.hist(self.img.ravel() , 256 , [0, 256] ) def UpperCamelCase( self ): cva.imshow('''Output-Image''' , self.img ) cva.imshow('''Input-Image''' , self.original_image ) cva.waitKey(5000 ) cva.destroyAllWindows() if __name__ == "__main__": UpperCAmelCase_ = os.path.join(os.path.basename(__file__), "image_data/input.jpg") UpperCAmelCase_ = ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
32
class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase ): # we need a list not a string, so do something to change the type UpperCAmelCase__ : Dict = arr.split(""",""" ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = [int(self.array[0] )] * len(self.array ) UpperCAmelCase__ : List[str] = [int(self.array[0] )] * len(self.array ) for i in range(1 , len(self.array ) ): UpperCAmelCase__ : Tuple = max( int(self.array[i] ) + sum_value[i - 1] , int(self.array[i] ) ) UpperCAmelCase__ : Union[str, Any] = max(sum_value[i] , rear[i - 1] ) return rear[len(self.array ) - 1] if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Tuple = input("""please input some numbers:""") SCREAMING_SNAKE_CASE__ : Dict = SubArray(whole_array) SCREAMING_SNAKE_CASE__ : Dict = array.solve_sub_array() print(("""the results is:""", re))
79
0
from __future__ import annotations def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> int: if len(__lowerCAmelCase ) < k or k < 0: raise ValueError('''Invalid Input''' ) snake_case__ = snake_case__ = sum(array[:k] ) for i in range(len(__lowerCAmelCase ) - k ): snake_case__ = current_sum - array[i] + array[i + k] snake_case__ = max(__lowerCAmelCase , __lowerCAmelCase ) return max_sum if __name__ == "__main__": from doctest import testmod from random import randint testmod() lowerCamelCase__ : List[Any] = [randint(-1_0_0_0, 1_0_0_0) for i in range(1_0_0)] lowerCamelCase__ : List[Any] = randint(0, 1_1_0) print(F"""The maximum sum of {k} consecutive elements is {max_sum_in_array(array,k)}""")
33
from ....configuration_utils import PretrainedConfig from ....utils import logging SCREAMING_SNAKE_CASE__ : List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Any = { """Visual-Attention-Network/van-base""": ( """https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json""" ), } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'van' def __init__( self , _lowerCAmelCase=224 , _lowerCAmelCase=3 , _lowerCAmelCase=[7, 3, 3, 3] , _lowerCAmelCase=[4, 2, 2, 2] , _lowerCAmelCase=[64, 128, 320, 512] , _lowerCAmelCase=[3, 3, 12, 3] , _lowerCAmelCase=[8, 8, 4, 4] , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=1e-6 , _lowerCAmelCase=1e-2 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) UpperCAmelCase__ : Tuple = image_size UpperCAmelCase__ : Optional[Any] = num_channels UpperCAmelCase__ : Optional[int] = patch_sizes UpperCAmelCase__ : int = strides UpperCAmelCase__ : Optional[int] = hidden_sizes UpperCAmelCase__ : str = depths UpperCAmelCase__ : Optional[Any] = mlp_ratios UpperCAmelCase__ : List[Any] = hidden_act UpperCAmelCase__ : Tuple = initializer_range UpperCAmelCase__ : Any = layer_norm_eps UpperCAmelCase__ : List[Any] = layer_scale_init_value UpperCAmelCase__ : int = drop_path_rate UpperCAmelCase__ : Dict = dropout_rate
79
0
"""simple docstring""" def __snake_case ( _lowercase ): """simple docstring""" if not isinstance(_lowercase ,_lowercase ): raise ValueError('''Input must be an integer''' ) if input_num <= 0: raise ValueError('''Input must be positive''' ) return sum( divisor for divisor in range(1 ,input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
34
import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Any: '''simple docstring''' UpperCAmelCase__ : List[str] = s.rsplit(__lowerCamelCase , __lowerCamelCase ) return new.join(__lowerCamelCase ) def _lowerCamelCase ( __lowerCamelCase ) -> str: '''simple docstring''' # encoder.embeddings are double copied in original FLAVA return sum(param.float().sum() if """encoder.embeddings""" not in key else 0 for key, param in state_dict.items() ) def _lowerCamelCase ( __lowerCamelCase ) -> int: '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = {} UpperCAmelCase__ : Union[str, Any] = ["""group_1""", """group_2""", """group_3""", """group_4"""] for key, value in state_dict.items(): for group_key in group_keys: if group_key in key: UpperCAmelCase__ : Optional[Any] = key.replace(F"{group_key}." , F"{group_key}.group." ) if "res_path" in key: UpperCAmelCase__ : Optional[int] = key.replace("""res_path.""" , """res_path.path.""" ) if key.endswith(""".w""" ): UpperCAmelCase__ : List[Any] = rreplace(__lowerCamelCase , """.w""" , """.weight""" , 1 ) if key.endswith(""".b""" ): UpperCAmelCase__ : Optional[int] = rreplace(__lowerCamelCase , """.b""" , """.bias""" , 1 ) UpperCAmelCase__ : Union[str, Any] = value.float() return upgrade @torch.no_grad() def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=None , __lowerCamelCase=True ) -> str: '''simple docstring''' from dall_e import Encoder UpperCAmelCase__ : Dict = Encoder() if os.path.exists(__lowerCamelCase ): UpperCAmelCase__ : Optional[Any] = torch.load(__lowerCamelCase ) else: UpperCAmelCase__ : Tuple = torch.hub.load_state_dict_from_url(__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase__ : Any = ckpt.state_dict() encoder.load_state_dict(__lowerCamelCase ) if config_path is not None: UpperCAmelCase__ : Dict = FlavaImageCodebookConfig.from_pretrained(__lowerCamelCase ) else: UpperCAmelCase__ : Optional[Any] = FlavaImageCodebookConfig() UpperCAmelCase__ : Optional[Any] = FlavaImageCodebook(__lowerCamelCase ).eval() UpperCAmelCase__ : str = encoder.state_dict() UpperCAmelCase__ : Optional[int] = upgrade_state_dict(__lowerCamelCase ) hf_model.load_state_dict(__lowerCamelCase ) UpperCAmelCase__ : List[str] = hf_model.state_dict() UpperCAmelCase__ : Tuple = count_parameters(__lowerCamelCase ) UpperCAmelCase__ : int = count_parameters(__lowerCamelCase ) assert torch.allclose(__lowerCamelCase , __lowerCamelCase , atol=1E-3 ) if save_checkpoint: hf_model.save_pretrained(__lowerCamelCase ) else: return hf_state_dict if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Any = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to flava checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") SCREAMING_SNAKE_CASE__ : int = parser.parse_args() convert_dalle_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
79
0
import importlib.metadata import warnings from copy import deepcopy from packaging import version from ..utils import logging from .import_utils import is_accelerate_available, is_bitsandbytes_available if is_bitsandbytes_available(): import bitsandbytes as bnb import torch import torch.nn as nn from ..pytorch_utils import ConvaD if is_accelerate_available(): from accelerate import init_empty_weights from accelerate.utils import find_tied_parameters a_ :Dict = logging.get_logger(__name__) def a ( A__ , A__ , A__ , A__=None , A__=None ) -> Dict: '''simple docstring''' if "." in tensor_name: SCREAMING_SNAKE_CASE__ : str = tensor_name.split('''.''' ) for split in splits[:-1]: SCREAMING_SNAKE_CASE__ : Union[str, Any] = getattr(A__ , A__ ) if new_module is None: raise ValueError(f"""{module} has no attribute {split}.""" ) SCREAMING_SNAKE_CASE__ : List[str] = new_module SCREAMING_SNAKE_CASE__ : Any = splits[-1] if tensor_name not in module._parameters and tensor_name not in module._buffers: raise ValueError(f"""{module} does not have a parameter or a buffer named {tensor_name}.""" ) SCREAMING_SNAKE_CASE__ : Dict = tensor_name in module._buffers SCREAMING_SNAKE_CASE__ : Dict = getattr(A__ , A__ ) if old_value.device == torch.device('''meta''' ) and device not in ["meta", torch.device('''meta''' )] and value is None: raise ValueError(f"""{tensor_name} is on the meta device, we need a `value` to put in on {device}.""" ) SCREAMING_SNAKE_CASE__ : Dict = False SCREAMING_SNAKE_CASE__ : List[str] = False if is_buffer or not is_bitsandbytes_available(): SCREAMING_SNAKE_CASE__ : Any = False SCREAMING_SNAKE_CASE__ : List[Any] = False else: SCREAMING_SNAKE_CASE__ : Union[str, Any] = hasattr(bnb.nn , '''Params4bit''' ) and isinstance(module._parameters[tensor_name] , bnb.nn.Paramsabit ) SCREAMING_SNAKE_CASE__ : Any = isinstance(module._parameters[tensor_name] , bnb.nn.IntaParams ) if is_abit or is_abit: SCREAMING_SNAKE_CASE__ : str = module._parameters[tensor_name] if param.device.type != "cuda": if value is None: SCREAMING_SNAKE_CASE__ : Dict = old_value.to(A__ ) elif isinstance(A__ , torch.Tensor ): SCREAMING_SNAKE_CASE__ : int = value.to('''cpu''' ) if value.dtype == torch.inta: SCREAMING_SNAKE_CASE__ : Optional[int] = version.parse(importlib.metadata.version('''bitsandbytes''' ) ) > version.parse( '''0.37.2''' ) if not is_abit_serializable: raise ValueError( '''Detected int8 weights but the version of bitsandbytes is not compatible with int8 serialization. ''' '''Make sure to download the latest `bitsandbytes` version. `pip install --upgrade bitsandbytes`.''' ) else: SCREAMING_SNAKE_CASE__ : List[Any] = torch.tensor(A__ , device='''cpu''' ) # Support models using `Conv1D` in place of `nn.Linear` (e.g. gpt2) by transposing the weight matrix prior to quantization. # Since weights are saved in the correct "orientation", we skip transposing when loading. if issubclass(module.source_cls , A__ ) and fpaa_statistics is None: SCREAMING_SNAKE_CASE__ : Optional[int] = new_value.T SCREAMING_SNAKE_CASE__ : int = old_value.__dict__ if is_abit: SCREAMING_SNAKE_CASE__ : int = bnb.nn.IntaParams(A__ , requires_grad=A__ , **A__ ).to(A__ ) elif is_abit: SCREAMING_SNAKE_CASE__ : Optional[int] = bnb.nn.Paramsabit(A__ , requires_grad=A__ , **A__ ).to(A__ ) SCREAMING_SNAKE_CASE__ : Dict = new_value if fpaa_statistics is not None: setattr(module.weight , '''SCB''' , fpaa_statistics.to(A__ ) ) else: if value is None: SCREAMING_SNAKE_CASE__ : Tuple = old_value.to(A__ ) elif isinstance(A__ , torch.Tensor ): SCREAMING_SNAKE_CASE__ : Tuple = value.to(A__ ) else: SCREAMING_SNAKE_CASE__ : Tuple = torch.tensor(A__ , device=A__ ) if is_buffer: SCREAMING_SNAKE_CASE__ : Dict = new_value else: SCREAMING_SNAKE_CASE__ : int = nn.Parameter(A__ , requires_grad=old_value.requires_grad ) SCREAMING_SNAKE_CASE__ : List[Any] = new_value def a ( A__ , A__=None , A__=None , A__=None , A__=False ) -> Tuple: '''simple docstring''' for name, module in model.named_children(): if current_key_name is None: SCREAMING_SNAKE_CASE__ : Any = [] current_key_name.append(A__ ) if (isinstance(A__ , nn.Linear ) or isinstance(A__ , A__ )) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` if not any(key in '''.'''.join(A__ ) for key in modules_to_not_convert ): with init_empty_weights(): if isinstance(A__ , A__ ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[Any] = module.weight.shape else: SCREAMING_SNAKE_CASE__ : Dict = module.in_features SCREAMING_SNAKE_CASE__ : Union[str, Any] = module.out_features if quantization_config.quantization_method() == "llm_int8": SCREAMING_SNAKE_CASE__ : str = bnb.nn.LinearabitLt( A__ , A__ , module.bias is not None , has_fpaa_weights=quantization_config.llm_inta_has_fpaa_weight , threshold=quantization_config.llm_inta_threshold , ) SCREAMING_SNAKE_CASE__ : Dict = True else: if ( quantization_config.llm_inta_skip_modules is not None and name in quantization_config.llm_inta_skip_modules ): pass else: SCREAMING_SNAKE_CASE__ : str = bnb.nn.Linearabit( A__ , A__ , module.bias is not None , quantization_config.bnb_abit_compute_dtype , compress_statistics=quantization_config.bnb_abit_use_double_quant , quant_type=quantization_config.bnb_abit_quant_type , ) SCREAMING_SNAKE_CASE__ : Optional[int] = True # Store the module class in case we need to transpose the weight later SCREAMING_SNAKE_CASE__ : Optional[Any] = type(A__ ) # Force requires grad to False to avoid unexpected errors model._modules[name].requires_grad_(A__ ) if len(list(module.children() ) ) > 0: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[Any] = _replace_with_bnb_linear( A__ , A__ , A__ , A__ , has_been_replaced=A__ , ) # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def a ( A__ , A__=None , A__=None , A__=None ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = ['''lm_head'''] if modules_to_not_convert is None else modules_to_not_convert SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Dict = _replace_with_bnb_linear( A__ , A__ , A__ , A__ ) if not has_been_replaced: logger.warning( '''You are loading your model in 8bit or 4bit but no linear modules were found in your model.''' ''' Please double check your model architecture, or submit an issue on github if you think this is''' ''' a bug.''' ) return model def a ( *A__ , **A__ ) -> Optional[int]: '''simple docstring''' warnings.warn( '''`replace_8bit_linear` will be deprecated in a future version, please use `replace_with_bnb_linear` instead''' , A__ , ) return replace_with_bnb_linear(*A__ , **A__ ) def a ( *A__ , **A__ ) -> Union[str, Any]: '''simple docstring''' warnings.warn( '''`set_module_8bit_tensor_to_device` will be deprecated in a future version, please use `set_module_quantized_tensor_to_device` instead''' , A__ , ) return set_module_quantized_tensor_to_device(*A__ , **A__ ) def a ( A__ ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = deepcopy(A__ ) # this has 0 cost since it is done inside `init_empty_weights` context manager` tied_model.tie_weights() SCREAMING_SNAKE_CASE__ : Union[str, Any] = find_tied_parameters(A__ ) # For compatibility with Accelerate < 0.18 if isinstance(A__ , A__ ): SCREAMING_SNAKE_CASE__ : Optional[Any] = sum(list(tied_params.values() ) , [] ) + list(tied_params.keys() ) else: SCREAMING_SNAKE_CASE__ : str = sum(A__ , [] ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = len(A__ ) > 0 # Check if it is a base model SCREAMING_SNAKE_CASE__ : Union[str, Any] = not hasattr(A__ , model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head SCREAMING_SNAKE_CASE__ : Tuple = list(model.named_children() ) SCREAMING_SNAKE_CASE__ : int = [list_modules[-1][0]] # add last module together with tied weights SCREAMING_SNAKE_CASE__ : str = set(A__ ) - set(A__ ) SCREAMING_SNAKE_CASE__ : Tuple = list(set(A__ ) ) + list(A__ ) # remove ".weight" from the keys SCREAMING_SNAKE_CASE__ : Tuple = ['''.weight''', '''.bias'''] SCREAMING_SNAKE_CASE__ : Union[str, Any] = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: SCREAMING_SNAKE_CASE__ : Optional[int] = name.replace(A__ , '''''' ) filtered_module_names.append(A__ ) return filtered_module_names
35
def _lowerCamelCase ( __lowerCamelCase ) -> int: '''simple docstring''' return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def _lowerCamelCase ( __lowerCamelCase ) -> bool: '''simple docstring''' UpperCAmelCase__ : Any = 0 UpperCAmelCase__ : Union[str, Any] = number while duplicate > 0: UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = divmod(__lowerCamelCase , 10 ) fact_sum += factorial(__lowerCamelCase ) return fact_sum == number if __name__ == "__main__": print("""Program to check whether a number is a Krisnamurthy Number or not.""") SCREAMING_SNAKE_CASE__ : Optional[Any] = int(input("""Enter number: """).strip()) print( f'''{number} is {"" if krishnamurthy(number) else "not "}a Krishnamurthy Number.''' )
79
0
__lowercase : List[str] = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' __lowercase : str = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] __lowercase : List[str] = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
36
def _lowerCamelCase ( __lowerCamelCase = 100_0000 ) -> int: '''simple docstring''' UpperCAmelCase__ : Tuple = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , __lowerCamelCase ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
79
0
def UpperCamelCase_ ( ) -> int: return [ a * b * (1_000 - a - b) for a in range(1 , 999 ) for b in range(__a , 999 ) if (a * a + b * b == (1_000 - a - b) ** 2) ][0] if __name__ == "__main__": print(f"""{solution() = }""")
37
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Tuple = { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json""", # See all REALM models at https://huggingface.co/models?filter=realm } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'realm' def __init__( self , _lowerCAmelCase=30522 , _lowerCAmelCase=768 , _lowerCAmelCase=128 , _lowerCAmelCase=12 , _lowerCAmelCase=12 , _lowerCAmelCase=8 , _lowerCAmelCase=3072 , _lowerCAmelCase="gelu_new" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=512 , _lowerCAmelCase=2 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=1e-12 , _lowerCAmelCase=256 , _lowerCAmelCase=10 , _lowerCAmelCase=1e-3 , _lowerCAmelCase=5 , _lowerCAmelCase=320 , _lowerCAmelCase=13353718 , _lowerCAmelCase=5000 , _lowerCAmelCase=1 , _lowerCAmelCase=0 , _lowerCAmelCase=2 , **_lowerCAmelCase , ): super().__init__(pad_token_id=_lowerCAmelCase , bos_token_id=_lowerCAmelCase , eos_token_id=_lowerCAmelCase , **_lowerCAmelCase ) # Common config UpperCAmelCase__ : List[Any] = vocab_size UpperCAmelCase__ : Dict = max_position_embeddings UpperCAmelCase__ : Any = hidden_size UpperCAmelCase__ : str = retriever_proj_size UpperCAmelCase__ : Tuple = num_hidden_layers UpperCAmelCase__ : List[str] = num_attention_heads UpperCAmelCase__ : List[Any] = num_candidates UpperCAmelCase__ : str = intermediate_size UpperCAmelCase__ : str = hidden_act UpperCAmelCase__ : Optional[Any] = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : Union[str, Any] = initializer_range UpperCAmelCase__ : Any = type_vocab_size UpperCAmelCase__ : Optional[Any] = layer_norm_eps # Reader config UpperCAmelCase__ : str = span_hidden_size UpperCAmelCase__ : Union[str, Any] = max_span_width UpperCAmelCase__ : List[str] = reader_layer_norm_eps UpperCAmelCase__ : Dict = reader_beam_size UpperCAmelCase__ : Union[str, Any] = reader_seq_len # Retrieval config UpperCAmelCase__ : List[Any] = num_block_records UpperCAmelCase__ : List[Any] = searcher_beam_size
79
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging A_ : Optional[Any] = logging.get_logger(__name__) A_ : int = { "uw-madison/mra-base-512-4": "https://huggingface.co/uw-madison/mra-base-512-4/resolve/main/config.json", } class __snake_case ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowerCamelCase__ = '''mra''' def __init__( self , __SCREAMING_SNAKE_CASE=5_0_2_6_5 , __SCREAMING_SNAKE_CASE=7_6_8 , __SCREAMING_SNAKE_CASE=1_2 , __SCREAMING_SNAKE_CASE=1_2 , __SCREAMING_SNAKE_CASE=3_0_7_2 , __SCREAMING_SNAKE_CASE="gelu" , __SCREAMING_SNAKE_CASE=0.1 , __SCREAMING_SNAKE_CASE=0.1 , __SCREAMING_SNAKE_CASE=5_1_2 , __SCREAMING_SNAKE_CASE=1 , __SCREAMING_SNAKE_CASE=0.02 , __SCREAMING_SNAKE_CASE=1e-5 , __SCREAMING_SNAKE_CASE="absolute" , __SCREAMING_SNAKE_CASE=4 , __SCREAMING_SNAKE_CASE="full" , __SCREAMING_SNAKE_CASE=0 , __SCREAMING_SNAKE_CASE=0 , __SCREAMING_SNAKE_CASE=1 , __SCREAMING_SNAKE_CASE=0 , __SCREAMING_SNAKE_CASE=2 , **__SCREAMING_SNAKE_CASE , ): super().__init__(pad_token_id=__SCREAMING_SNAKE_CASE , bos_token_id=__SCREAMING_SNAKE_CASE , eos_token_id=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) snake_case__ : List[Any] = vocab_size snake_case__ : Optional[int] = max_position_embeddings snake_case__ : List[Any] = hidden_size snake_case__ : Tuple = num_hidden_layers snake_case__ : List[Any] = num_attention_heads snake_case__ : Dict = intermediate_size snake_case__ : str = hidden_act snake_case__ : Optional[Any] = hidden_dropout_prob snake_case__ : Tuple = attention_probs_dropout_prob snake_case__ : Dict = initializer_range snake_case__ : List[str] = type_vocab_size snake_case__ : str = layer_norm_eps snake_case__ : Dict = position_embedding_type snake_case__ : Union[str, Any] = block_per_row snake_case__ : str = approx_mode snake_case__ : int = initial_prior_first_n_blocks snake_case__ : str = initial_prior_diagonal_n_blocks
38
import gc import unittest from parameterized import parameterized from diffusers import FlaxUNetaDConditionModel from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp @slow @require_flax class UpperCAmelCase_ ( unittest.TestCase ): def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase ): return f"gaussian_noise_s={seed}_shape={'_'.join([str(_lowerCAmelCase ) for s in shape] )}.npy" def __UpperCAmelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def __UpperCAmelCase ( self , _lowerCAmelCase=0 , _lowerCAmelCase=(4, 4, 64, 64) , _lowerCAmelCase=False ): UpperCAmelCase__ : Optional[Any] = jnp.bfloataa if fpaa else jnp.floataa UpperCAmelCase__ : Union[str, Any] = jnp.array(load_hf_numpy(self.get_file_format(_lowerCAmelCase , _lowerCAmelCase ) ) , dtype=_lowerCAmelCase ) return image def __UpperCAmelCase ( self , _lowerCAmelCase=False , _lowerCAmelCase="CompVis/stable-diffusion-v1-4" ): UpperCAmelCase__ : int = jnp.bfloataa if fpaa else jnp.floataa UpperCAmelCase__ : Optional[Any] = """bf16""" if fpaa else None UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = FlaxUNetaDConditionModel.from_pretrained( _lowerCAmelCase , subfolder="""unet""" , dtype=_lowerCAmelCase , revision=_lowerCAmelCase ) return model, params def __UpperCAmelCase ( self , _lowerCAmelCase=0 , _lowerCAmelCase=(4, 77, 768) , _lowerCAmelCase=False ): UpperCAmelCase__ : Optional[int] = jnp.bfloataa if fpaa else jnp.floataa UpperCAmelCase__ : Optional[int] = jnp.array(load_hf_numpy(self.get_file_format(_lowerCAmelCase , _lowerCAmelCase ) ) , dtype=_lowerCAmelCase ) return hidden_states @parameterized.expand( [ # fmt: off [83, 4, [-0.2_3_2_3, -0.1_3_0_4, 0.0_8_1_3, -0.3_0_9_3, -0.0_9_1_9, -0.1_5_7_1, -0.1_1_2_5, -0.5_8_0_6]], [17, 0.5_5, [-0.0_8_3_1, -0.2_4_4_3, 0.0_9_0_1, -0.0_9_1_9, 0.3_3_9_6, 0.0_1_0_3, -0.3_7_4_3, 0.0_7_0_1]], [8, 0.8_9, [-0.4_8_6_3, 0.0_8_5_9, 0.0_8_7_5, -0.1_6_5_8, 0.9_1_9_9, -0.0_1_1_4, 0.4_8_3_9, 0.4_6_3_9]], [3, 1000, [-0.5_6_4_9, 0.2_4_0_2, -0.5_5_1_8, 0.1_2_4_8, 1.1_3_2_8, -0.2_4_4_3, -0.0_3_2_5, -1.0_0_7_8]], # fmt: on ] ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.get_unet_model(model_id="""CompVis/stable-diffusion-v1-4""" , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = self.get_latents(_lowerCAmelCase , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Dict = self.get_encoder_hidden_states(_lowerCAmelCase , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = model.apply( {"""params""": params} , _lowerCAmelCase , jnp.array(_lowerCAmelCase , dtype=jnp.intaa ) , encoder_hidden_states=_lowerCAmelCase , ).sample assert sample.shape == latents.shape UpperCAmelCase__ : Dict = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) UpperCAmelCase__ : List[Any] = jnp.array(_lowerCAmelCase , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, in the same hardware assert jnp.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-2 ) @parameterized.expand( [ # fmt: off [83, 4, [0.1_5_1_4, 0.0_8_0_7, 0.1_6_2_4, 0.1_0_1_6, -0.1_8_9_6, 0.0_2_6_3, 0.0_6_7_7, 0.2_3_1_0]], [17, 0.5_5, [0.1_1_6_4, -0.0_2_1_6, 0.0_1_7_0, 0.1_5_8_9, -0.3_1_2_0, 0.1_0_0_5, -0.0_5_8_1, -0.1_4_5_8]], [8, 0.8_9, [-0.1_7_5_8, -0.0_1_6_9, 0.1_0_0_4, -0.1_4_1_1, 0.1_3_1_2, 0.1_1_0_3, -0.1_9_9_6, 0.2_1_3_9]], [3, 1000, [0.1_2_1_4, 0.0_3_5_2, -0.0_7_3_1, -0.1_5_6_2, -0.0_9_9_4, -0.0_9_0_6, -0.2_3_4_0, -0.0_5_3_9]], # fmt: on ] ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.get_unet_model(model_id="""stabilityai/stable-diffusion-2""" , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = self.get_latents(_lowerCAmelCase , shape=(4, 4, 96, 96) , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Any = self.get_encoder_hidden_states(_lowerCAmelCase , shape=(4, 77, 1024) , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Dict = model.apply( {"""params""": params} , _lowerCAmelCase , jnp.array(_lowerCAmelCase , dtype=jnp.intaa ) , encoder_hidden_states=_lowerCAmelCase , ).sample assert sample.shape == latents.shape UpperCAmelCase__ : Any = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) UpperCAmelCase__ : Any = jnp.array(_lowerCAmelCase , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, on the same hardware assert jnp.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-2 )
79
0
from math import factorial def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ = 20 ): snake_case_ = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... snake_case_ = n // 2 return int(factorial(SCREAMING_SNAKE_CASE__ ) / (factorial(SCREAMING_SNAKE_CASE__ ) * factorial(n - k )) ) if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution(20)) else: try: lowerCAmelCase_ = int(sys.argv[1]) print(solution(n)) except ValueError: print('''Invalid entry - please enter a number.''')
39
import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class UpperCAmelCase_ ( unittest.TestCase ): def __init__( self , _lowerCAmelCase , _lowerCAmelCase=7 , _lowerCAmelCase=3 , _lowerCAmelCase=18 , _lowerCAmelCase=30 , _lowerCAmelCase=400 , _lowerCAmelCase=True , _lowerCAmelCase=None , _lowerCAmelCase=True , ): UpperCAmelCase__ : List[str] = size if size is not None else {"""height""": 18, """width""": 18} UpperCAmelCase__ : Union[str, Any] = parent UpperCAmelCase__ : int = batch_size UpperCAmelCase__ : Tuple = num_channels UpperCAmelCase__ : Dict = image_size UpperCAmelCase__ : List[Any] = min_resolution UpperCAmelCase__ : str = max_resolution UpperCAmelCase__ : Union[str, Any] = do_resize UpperCAmelCase__ : Tuple = size UpperCAmelCase__ : int = do_normalize def __UpperCAmelCase ( self ): return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.8_8_6_6_4_4_3_6_3_4_0_3_3_2_0_3, 0.6_6_1_8_8_2_9_3_6_9_5_4_4_9_8_3, 0.3_8_9_1_7_4_6_4_0_1_7_8_6_8_0_4], [-0.6_0_4_2_5_5_9_1_4_6_8_8_1_1_0_4, -0.0_2_2_9_5_0_0_8_8_6_0_5_2_8_4_6_9, 0.5_4_2_3_7_9_7_3_6_9_0_0_3_2_9_6], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class UpperCAmelCase_ ( __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = ImageGPTImageProcessor if is_vision_available() else None def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = ImageGPTImageProcessingTester(self ) @property def __UpperCAmelCase ( self ): return self.image_processor_tester.prepare_image_processor_dict() def __UpperCAmelCase ( self ): UpperCAmelCase__ : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """clusters""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """do_normalize""" ) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 18} ) UpperCAmelCase__ : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) UpperCAmelCase__ : Optional[int] = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase , obj[key] ) ) else: self.assertEqual(obj[key] , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase__ : Union[str, Any] = os.path.join(_lowerCAmelCase , """image_processor.json""" ) image_processor_first.to_json_file(_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = self.image_processing_class.from_json_file(_lowerCAmelCase ).to_dict() UpperCAmelCase__ : Dict = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : str = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = self.image_processing_class.from_pretrained(_lowerCAmelCase ).to_dict() UpperCAmelCase__ : Tuple = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , _lowerCAmelCase ) @unittest.skip("""ImageGPT requires clusters at initialization""" ) def __UpperCAmelCase ( self ): pass def _lowerCamelCase ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ : Any = load_dataset("""hf-internal-testing/fixtures_image_utils""" , split="""test""" ) UpperCAmelCase__ : Dict = Image.open(dataset[4]["""file"""] ) UpperCAmelCase__ : Optional[Any] = Image.open(dataset[5]["""file"""] ) UpperCAmelCase__ : List[Any] = [imagea, imagea] return images @require_vision @require_torch class UpperCAmelCase_ ( unittest.TestCase ): @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = ImageGPTImageProcessor.from_pretrained("""openai/imagegpt-small""" ) UpperCAmelCase__ : int = prepare_images() # test non-batched UpperCAmelCase__ : List[str] = image_processing(images[0] , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 1024) ) UpperCAmelCase__ : List[Any] = [306, 191, 191] self.assertEqual(encoding.input_ids[0, :3].tolist() , _lowerCAmelCase ) # test batched UpperCAmelCase__ : List[str] = image_processing(_lowerCAmelCase , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 1024) ) UpperCAmelCase__ : Any = [303, 13, 13] self.assertEqual(encoding.input_ids[1, -3:].tolist() , _lowerCAmelCase )
79
0
from math import isqrt def UpperCamelCase ( snake_case__ : int ) -> list[int]: UpperCamelCase : Union[str, Any] = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , snake_case__ , snake_case__ ): UpperCamelCase : List[str] = False return [i for i in range(2 , snake_case__ ) if is_prime[i]] def UpperCamelCase ( snake_case__ : int = 10**8 ) -> int: UpperCamelCase : Union[str, Any] = calculate_prime_numbers(max_number // 2 ) UpperCamelCase : List[Any] = 0 UpperCamelCase : List[Any] = 0 UpperCamelCase : Tuple = len(snake_case__ ) - 1 while left <= right: while prime_numbers[left] * prime_numbers[right] >= max_number: right -= 1 semiprimes_count += right - left + 1 left += 1 return semiprimes_count if __name__ == "__main__": print(F"""{solution() = }""")
40
import os import unittest from transformers import MobileBertTokenizer, MobileBertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class UpperCAmelCase_ ( __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = MobileBertTokenizer __lowerCamelCase = MobileBertTokenizerFast __lowerCamelCase = True __lowerCamelCase = True __lowerCamelCase = filter_non_english __lowerCamelCase = 'google/mobilebert-uncased' def __UpperCAmelCase ( self ): super().setUp() UpperCAmelCase__ : Dict = [ """[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] UpperCAmelCase__ : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) UpperCAmelCase__ : List[str] = [ (tokenizer_def[0], self.pre_trained_model_path, tokenizer_def[2]) # else the 'google/' prefix is stripped for tokenizer_def in self.tokenizers_list ] def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : Tuple = """UNwant\u00E9d,running""" UpperCAmelCase__ : Union[str, Any] = """unwanted, running""" return input_text, output_text def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = self.tokenizer_class(self.vocab_file ) UpperCAmelCase__ : Tuple = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(_lowerCAmelCase , ["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [9, 6, 7, 12, 10, 11] ) def __UpperCAmelCase ( self ): if not self.test_rust_tokenizer: return UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : Dict = self.get_rust_tokenizer() UpperCAmelCase__ : List[str] = """UNwant\u00E9d,running""" UpperCAmelCase__ : Optional[int] = tokenizer.tokenize(_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = rust_tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = rust_tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.get_rust_tokenizer() UpperCAmelCase__ : Any = tokenizer.encode(_lowerCAmelCase ) UpperCAmelCase__ : str = rust_tokenizer.encode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) # With lower casing UpperCAmelCase__ : Tuple = self.get_tokenizer(do_lower_case=_lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.get_rust_tokenizer(do_lower_case=_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = """UNwant\u00E9d,running""" UpperCAmelCase__ : int = tokenizer.tokenize(_lowerCAmelCase ) UpperCAmelCase__ : Any = rust_tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = rust_tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = self.get_rust_tokenizer() UpperCAmelCase__ : List[str] = tokenizer.encode(_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = rust_tokenizer.encode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = BasicTokenizer() self.assertListEqual(tokenizer.tokenize("""ah\u535A\u63A8zz""" ) , ["""ah""", """\u535A""", """\u63A8""", """zz"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = BasicTokenizer(do_lower_case=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Union[str, Any] = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hällo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""h\u00E9llo"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[str] = BasicTokenizer(do_lower_case=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[str] = BasicTokenizer(do_lower_case=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : int = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HäLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HaLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = BasicTokenizer(do_lower_case=_lowerCAmelCase , never_split=["""[UNK]"""] ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? [UNK]""" ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?""", """[UNK]"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : int = ["""[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing"""] UpperCAmelCase__ : List[str] = {} for i, token in enumerate(_lowerCAmelCase ): UpperCAmelCase__ : Optional[Any] = i UpperCAmelCase__ : str = WordpieceTokenizer(vocab=_lowerCAmelCase , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""unwanted running""" ) , ["""un""", """##want""", """##ed""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.tokenize("""unwantedX running""" ) , ["""[UNK]""", """runn""", """##ing"""] ) def __UpperCAmelCase ( self ): self.assertTrue(_is_whitespace(""" """ ) ) self.assertTrue(_is_whitespace("""\t""" ) ) self.assertTrue(_is_whitespace("""\r""" ) ) self.assertTrue(_is_whitespace("""\n""" ) ) self.assertTrue(_is_whitespace("""\u00A0""" ) ) self.assertFalse(_is_whitespace("""A""" ) ) self.assertFalse(_is_whitespace("""-""" ) ) def __UpperCAmelCase ( self ): self.assertTrue(_is_control("""\u0005""" ) ) self.assertFalse(_is_control("""A""" ) ) self.assertFalse(_is_control(""" """ ) ) self.assertFalse(_is_control("""\t""" ) ) self.assertFalse(_is_control("""\r""" ) ) def __UpperCAmelCase ( self ): self.assertTrue(_is_punctuation("""-""" ) ) self.assertTrue(_is_punctuation("""$""" ) ) self.assertTrue(_is_punctuation("""`""" ) ) self.assertTrue(_is_punctuation(""".""" ) ) self.assertFalse(_is_punctuation("""A""" ) ) self.assertFalse(_is_punctuation(""" """ ) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = self.get_tokenizer() UpperCAmelCase__ : Union[str, Any] = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(_lowerCAmelCase ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) self.assertListEqual( [rust_tokenizer.tokenize(_lowerCAmelCase ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = self.tokenizer_class.from_pretrained("""google/mobilebert-uncased""" ) UpperCAmelCase__ : List[Any] = tokenizer.encode("""sequence builders""" , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer.encode("""multi-sequence build""" , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase ) UpperCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase , _lowerCAmelCase ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def __UpperCAmelCase ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): UpperCAmelCase__ : Any = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = f"A, naïve {tokenizer_r.mask_token} AllenNLP sentence." UpperCAmelCase__ : Optional[Any] = tokenizer_r.encode_plus( _lowerCAmelCase , return_attention_mask=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , ) UpperCAmelCase__ : Any = tokenizer_r.do_lower_case if hasattr(_lowerCAmelCase , """do_lower_case""" ) else False UpperCAmelCase__ : Optional[int] = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """A"""), ((1, 2), ""","""), ((3, 5), """na"""), ((5, 6), """##ï"""), ((6, 8), """##ve"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """Allen"""), ((21, 23), """##NL"""), ((23, 24), """##P"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """a"""), ((1, 2), ""","""), ((3, 8), """naive"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """allen"""), ((21, 23), """##nl"""), ((23, 24), """##p"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["""input_ids"""] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["""offset_mapping"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = ["""的""", """人""", """有"""] UpperCAmelCase__ : Tuple = """""".join(_lowerCAmelCase ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): UpperCAmelCase__ : Tuple = True UpperCAmelCase__ : Optional[Any] = self.tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = tokenizer_p.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer_r.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : Any = tokenizer_r.convert_ids_to_tokens(_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = tokenizer_p.convert_ids_to_tokens(_lowerCAmelCase ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : List[Any] = False UpperCAmelCase__ : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = tokenizer_r.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer_p.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = tokenizer_r.convert_ids_to_tokens(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = tokenizer_p.convert_ids_to_tokens(_lowerCAmelCase ) # it is expected that only the first Chinese character is not preceded by "##". UpperCAmelCase__ : List[str] = [ f"##{token}" if idx != 0 else token for idx, token in enumerate(_lowerCAmelCase ) ] self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase )
79
0
'''simple docstring''' from __future__ import annotations def _A ( A__ , A__ ): """simple docstring""" print(F"Vertex\tShortest Distance from vertex {src}" ) for i, d in enumerate(A__ ): print(F"{i}\t\t{d}" ) def _A ( A__ , A__ , A__ ): """simple docstring""" for j in range(A__ ): __lowercase , __lowercase , __lowercase = (graph[j][k] for k in ['''src''', '''dst''', '''weight''']) if distance[u] != float('''inf''' ) and distance[u] + w < distance[v]: return True return False def _A ( A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = [float('''inf''' )] * vertex_count __lowercase = 0.0 for _ in range(vertex_count - 1 ): for j in range(A__ ): __lowercase , __lowercase , __lowercase = (graph[j][k] for k in ['''src''', '''dst''', '''weight''']) if distance[u] != float('''inf''' ) and distance[u] + w < distance[v]: __lowercase = distance[u] + w __lowercase = check_negative_cycle(A__ , A__ , A__ ) if negative_cycle_exists: raise Exception('''Negative cycle found''' ) return distance if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = int(input('''Enter number of vertices: ''').strip()) lowerCAmelCase__ = int(input('''Enter number of edges: ''').strip()) lowerCAmelCase__ = [{} for _ in range(E)] for i in range(E): print('''Edge ''', i + 1) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = ( int(x) for x in input('''Enter source, destination, weight: ''').strip().split(''' ''') ) lowerCAmelCase__ = {'''src''': src, '''dst''': dest, '''weight''': weight} lowerCAmelCase__ = int(input('''\nEnter shortest path source:''').strip()) lowerCAmelCase__ = bellman_ford(graph, V, E, source) print_distance(shortest_distance, 0)
41
import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg""" UpperCAmelCase__ : int = Image.open(requests.get(__lowerCamelCase , stream=__lowerCamelCase ).raw ).convert("""RGB""" ) UpperCAmelCase__ : Any = transforms.Compose( [ transforms.Resize((image_size, image_size) , interpolation=InterpolationMode.BICUBIC ), transforms.ToTensor(), transforms.Normalize((0.48_145_466, 0.4_578_275, 0.40_821_073) , (0.26_862_954, 0.26_130_258, 0.27_577_711) ), ] ) UpperCAmelCase__ : Optional[int] = transform(__lowerCamelCase ).unsqueeze(0 ).to(__lowerCamelCase ) return image def _lowerCamelCase ( __lowerCamelCase ) -> str: '''simple docstring''' if "visual_encoder" in key: UpperCAmelCase__ : Dict = re.sub("""visual_encoder*""" , """vision_model.encoder""" , __lowerCamelCase ) if "blocks" in key: UpperCAmelCase__ : Optional[Any] = re.sub(r"""blocks""" , """layers""" , __lowerCamelCase ) if "attn" in key: UpperCAmelCase__ : List[str] = re.sub(r"""attn""" , """self_attn""" , __lowerCamelCase ) if "norm1" in key: UpperCAmelCase__ : Union[str, Any] = re.sub(r"""norm1""" , """layer_norm1""" , __lowerCamelCase ) if "norm2" in key: UpperCAmelCase__ : Any = re.sub(r"""norm2""" , """layer_norm2""" , __lowerCamelCase ) if "encoder.norm" in key: UpperCAmelCase__ : Dict = re.sub(r"""encoder.norm""" , """post_layernorm""" , __lowerCamelCase ) if "encoder.patch_embed.proj" in key: UpperCAmelCase__ : List[str] = re.sub(r"""encoder.patch_embed.proj""" , """embeddings.patch_embedding""" , __lowerCamelCase ) if "encoder.pos_embed" in key: UpperCAmelCase__ : List[str] = re.sub(r"""encoder.pos_embed""" , """embeddings.position_embedding""" , __lowerCamelCase ) if "encoder.cls_token" in key: UpperCAmelCase__ : List[Any] = re.sub(r"""encoder.cls_token""" , """embeddings.class_embedding""" , __lowerCamelCase ) if "self_attn" in key: UpperCAmelCase__ : List[Any] = re.sub(r"""self_attn.proj""" , """self_attn.projection""" , __lowerCamelCase ) return key @torch.no_grad() def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase=None ) -> Tuple: '''simple docstring''' if config_path is not None: UpperCAmelCase__ : Any = BlipConfig.from_pretrained(__lowerCamelCase ) else: UpperCAmelCase__ : str = BlipConfig(projection_dim=512 , text_config={} , vision_config={} ) UpperCAmelCase__ : int = BlipForConditionalGeneration(__lowerCamelCase ).eval() UpperCAmelCase__ : Any = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth""" UpperCAmelCase__ : List[str] = blip_decoder(pretrained=__lowerCamelCase , image_size=384 , vit="""base""" ) UpperCAmelCase__ : Union[str, Any] = pt_model.eval() UpperCAmelCase__ : Optional[int] = pt_model.state_dict() for key in modified_state_dict.copy(): UpperCAmelCase__ : Dict = modified_state_dict.pop(__lowerCamelCase ) UpperCAmelCase__ : Union[str, Any] = rename_key(__lowerCamelCase ) UpperCAmelCase__ : List[str] = value hf_model.load_state_dict(__lowerCamelCase ) UpperCAmelCase__ : Tuple = 384 UpperCAmelCase__ : str = load_demo_image(image_size=__lowerCamelCase , device="""cpu""" ) UpperCAmelCase__ : str = BertTokenizer.from_pretrained("""bert-base-uncased""" ) UpperCAmelCase__ : Dict = tokenizer(["""a picture of"""] ).input_ids UpperCAmelCase__ : int = hf_model.generate(__lowerCamelCase , __lowerCamelCase ) assert out[0].tolist() == [3_0522, 1037, 3861, 1997, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] UpperCAmelCase__ : Any = hf_model.generate(__lowerCamelCase ) assert out[0].tolist() == [3_0522, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(__lowerCamelCase ) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' UpperCAmelCase__ : Union[str, Any] = ( """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth""" ) UpperCAmelCase__ : List[Any] = blip_vqa(pretrained=__lowerCamelCase , image_size=__lowerCamelCase , vit="""base""" ) vqa_model.eval() UpperCAmelCase__ : str = vqa_model.state_dict() for key in modified_state_dict.copy(): UpperCAmelCase__ : Dict = modified_state_dict.pop(__lowerCamelCase ) UpperCAmelCase__ : Dict = rename_key(__lowerCamelCase ) UpperCAmelCase__ : int = value UpperCAmelCase__ : List[str] = BlipForQuestionAnswering(__lowerCamelCase ) hf_vqa_model.load_state_dict(__lowerCamelCase ) UpperCAmelCase__ : Tuple = ["""How many dogs are in this image?"""] UpperCAmelCase__ : Union[str, Any] = tokenizer(__lowerCamelCase , return_tensors="""pt""" ).input_ids UpperCAmelCase__ : Optional[Any] = hf_vqa_model.generate(__lowerCamelCase , __lowerCamelCase ) print(tokenizer.decode(answer[0] ) ) assert tokenizer.decode(answer[0] ) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + """_vqa""" ) UpperCAmelCase__ : int = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth""" UpperCAmelCase__ : Any = blip_itm(pretrained=__lowerCamelCase , image_size=__lowerCamelCase , vit="""base""" ) itm_model.eval() UpperCAmelCase__ : List[Any] = itm_model.state_dict() for key in modified_state_dict.copy(): UpperCAmelCase__ : Dict = modified_state_dict.pop(__lowerCamelCase ) UpperCAmelCase__ : int = rename_key(__lowerCamelCase ) UpperCAmelCase__ : Any = value UpperCAmelCase__ : Optional[int] = BlipForImageTextRetrieval(__lowerCamelCase ) UpperCAmelCase__ : Union[str, Any] = ["""A picture of a woman with a dog sitting in a beach"""] UpperCAmelCase__ : List[Any] = tokenizer( __lowerCamelCase , return_tensors="""pt""" , padding="""max_length""" , truncation=__lowerCamelCase , max_length=35 , ).input_ids hf_itm_model.load_state_dict(__lowerCamelCase ) hf_itm_model.eval() UpperCAmelCase__ : List[str] = hf_itm_model(__lowerCamelCase , __lowerCamelCase , use_itm_head=__lowerCamelCase ) UpperCAmelCase__ : List[str] = hf_itm_model(__lowerCamelCase , __lowerCamelCase , use_itm_head=__lowerCamelCase ) assert out[0].item() == 0.2_110_687_494_277_954 assert torch.nn.functional.softmax(out_itm[0] , dim=1 )[:, 1].item() == 0.45_698_845_386_505_127 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + """_itm""" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : str = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") SCREAMING_SNAKE_CASE__ : List[Any] = parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
79
0
'''simple docstring''' import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class UpperCAmelCase : '''simple docstring''' @property def UpperCamelCase( self ) -> int: '''simple docstring''' return self.get_dummy_input() @property def UpperCamelCase( self ) -> int: '''simple docstring''' if self.block_type == "down": return (4, 32, 16, 16) elif self.block_type == "mid": return (4, 32, 32, 32) elif self.block_type == "up": return (4, 32, 64, 64) raise ValueError(f'''\'{self.block_type}\' is not a supported block_type. Set it to \'up\', \'mid\', or \'down\'.''' ) def UpperCamelCase( self , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=False , ) -> int: '''simple docstring''' lowerCamelCase_ = 4 lowerCamelCase_ = 32 lowerCamelCase_ = (32, 32) lowerCamelCase_ = torch.manual_seed(0 ) lowerCamelCase_ = torch.device(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = (batch_size, num_channels) + sizes lowerCamelCase_ = randn_tensor(SCREAMING_SNAKE_CASE_ , generator=SCREAMING_SNAKE_CASE_ , device=SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = {'hidden_states': hidden_states} if include_temb: lowerCamelCase_ = 128 lowerCamelCase_ = randn_tensor((batch_size, temb_channels) , generator=SCREAMING_SNAKE_CASE_ , device=SCREAMING_SNAKE_CASE_ ) if include_res_hidden_states_tuple: lowerCamelCase_ = torch.manual_seed(1 ) lowerCamelCase_ = (randn_tensor(SCREAMING_SNAKE_CASE_ , generator=SCREAMING_SNAKE_CASE_ , device=SCREAMING_SNAKE_CASE_ ),) if include_encoder_hidden_states: lowerCamelCase_ = floats_tensor((batch_size, 32, 32) ).to(SCREAMING_SNAKE_CASE_ ) if include_skip_sample: lowerCamelCase_ = randn_tensor(((batch_size, 3) + sizes) , generator=SCREAMING_SNAKE_CASE_ , device=SCREAMING_SNAKE_CASE_ ) return dummy_input def UpperCamelCase( self ) -> Dict: '''simple docstring''' lowerCamelCase_ = { 'in_channels': 32, 'out_channels': 32, 'temb_channels': 128, } if self.block_type == "up": lowerCamelCase_ = 32 if self.block_type == "mid": init_dict.pop('out_channels' ) lowerCamelCase_ = self.dummy_input return init_dict, inputs_dict def UpperCamelCase( self , SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: '''simple docstring''' lowerCamelCase_ ,lowerCamelCase_ = self.prepare_init_args_and_inputs_for_common() lowerCamelCase_ = self.block_class(**SCREAMING_SNAKE_CASE_ ) unet_block.to(SCREAMING_SNAKE_CASE_ ) unet_block.eval() with torch.no_grad(): lowerCamelCase_ = unet_block(**SCREAMING_SNAKE_CASE_ ) if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowerCamelCase_ = output[0] self.assertEqual(output.shape , self.output_shape ) lowerCamelCase_ = output[0, -1, -3:, -3:] lowerCamelCase_ = torch.tensor(SCREAMING_SNAKE_CASE_ ).to(SCREAMING_SNAKE_CASE_ ) assert torch_all_close(output_slice.flatten() , SCREAMING_SNAKE_CASE_ , atol=5E-3 ) @unittest.skipIf(torch_device == 'mps' , 'Training is not supported in mps' ) def UpperCamelCase( self ) -> str: '''simple docstring''' lowerCamelCase_ ,lowerCamelCase_ = self.prepare_init_args_and_inputs_for_common() lowerCamelCase_ = self.block_class(**SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.train() lowerCamelCase_ = model(**SCREAMING_SNAKE_CASE_ ) if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowerCamelCase_ = output[0] lowerCamelCase_ = torch.device(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = randn_tensor(output.shape , device=SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = torch.nn.functional.mse_loss(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) loss.backward()
42
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : Tuple = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : List[Any] = { """MIT/ast-finetuned-audioset-10-10-0.4593""": ( """https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json""" ), } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'audio-spectrogram-transformer' def __init__( self , _lowerCAmelCase=768 , _lowerCAmelCase=12 , _lowerCAmelCase=12 , _lowerCAmelCase=3072 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=1e-12 , _lowerCAmelCase=16 , _lowerCAmelCase=True , _lowerCAmelCase=10 , _lowerCAmelCase=10 , _lowerCAmelCase=1024 , _lowerCAmelCase=128 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : List[Any] = num_attention_heads UpperCAmelCase__ : Dict = intermediate_size UpperCAmelCase__ : Dict = hidden_act UpperCAmelCase__ : str = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : Tuple = initializer_range UpperCAmelCase__ : Dict = layer_norm_eps UpperCAmelCase__ : Optional[Any] = patch_size UpperCAmelCase__ : Tuple = qkv_bias UpperCAmelCase__ : Tuple = frequency_stride UpperCAmelCase__ : Union[str, Any] = time_stride UpperCAmelCase__ : Optional[Any] = max_length UpperCAmelCase__ : Optional[int] = num_mel_bins
79
0
from __future__ import annotations from collections.abc import Generator def _a ( ): """simple docstring""" lowercase__ = {} lowercase__ = 2 while True: lowercase__ = factor_map.pop(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if factor: lowercase__ = factor + prime while x in factor_map: x += factor lowercase__ = factor else: lowercase__ = prime yield prime prime += 1 def _a ( SCREAMING_SNAKE_CASE = 1E10 ): """simple docstring""" lowercase__ = sieve() lowercase__ = 1 while True: lowercase__ = next(SCREAMING_SNAKE_CASE ) if (2 * prime * n) > limit: return n # Ignore the next prime as the reminder will be 2. next(SCREAMING_SNAKE_CASE ) n += 2 if __name__ == "__main__": print(solution())
43
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) class UpperCAmelCase_ ( __lowerCamelCase ): def __init__( self , *_lowerCAmelCase , **_lowerCAmelCase ): warnings.warn( """The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use GLPNImageProcessor instead.""" , _lowerCAmelCase , ) super().__init__(*_lowerCAmelCase , **_lowerCAmelCase )
79
0
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
44
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : int = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__ : List[str] = { """vocab_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt""", }, """tokenizer_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json""" ), """google/realm-orqa-nq-openqa""": ( """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-nq-reader""": ( """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-openqa""": ( """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-reader""": ( """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json""" ), }, } SCREAMING_SNAKE_CASE__ : Optional[Any] = { """google/realm-cc-news-pretrained-embedder""": 5_12, """google/realm-cc-news-pretrained-encoder""": 5_12, """google/realm-cc-news-pretrained-scorer""": 5_12, """google/realm-cc-news-pretrained-openqa""": 5_12, """google/realm-orqa-nq-openqa""": 5_12, """google/realm-orqa-nq-reader""": 5_12, """google/realm-orqa-wq-openqa""": 5_12, """google/realm-orqa-wq-reader""": 5_12, } SCREAMING_SNAKE_CASE__ : Optional[Any] = { """google/realm-cc-news-pretrained-embedder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-encoder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-scorer""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-reader""": {"""do_lower_case""": True}, """google/realm-orqa-wq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-wq-reader""": {"""do_lower_case""": True}, } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = PRETRAINED_INIT_CONFIGURATION __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = RealmTokenizer def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=True , _lowerCAmelCase="[UNK]" , _lowerCAmelCase="[SEP]" , _lowerCAmelCase="[PAD]" , _lowerCAmelCase="[CLS]" , _lowerCAmelCase="[MASK]" , _lowerCAmelCase=True , _lowerCAmelCase=None , **_lowerCAmelCase , ): super().__init__( _lowerCAmelCase , tokenizer_file=_lowerCAmelCase , do_lower_case=_lowerCAmelCase , unk_token=_lowerCAmelCase , sep_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , cls_token=_lowerCAmelCase , mask_token=_lowerCAmelCase , tokenize_chinese_chars=_lowerCAmelCase , strip_accents=_lowerCAmelCase , **_lowerCAmelCase , ) UpperCAmelCase__ : Tuple = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , _lowerCAmelCase ) != do_lower_case or normalizer_state.get("""strip_accents""" , _lowerCAmelCase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , _lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase__ : Any = getattr(_lowerCAmelCase , normalizer_state.pop("""type""" ) ) UpperCAmelCase__ : str = do_lower_case UpperCAmelCase__ : Tuple = strip_accents UpperCAmelCase__ : Tuple = tokenize_chinese_chars UpperCAmelCase__ : Union[str, Any] = normalizer_class(**_lowerCAmelCase ) UpperCAmelCase__ : Dict = do_lower_case def __UpperCAmelCase ( self , _lowerCAmelCase , **_lowerCAmelCase ): UpperCAmelCase__ : List[Any] = PaddingStrategy.MAX_LENGTH UpperCAmelCase__ : Optional[int] = text UpperCAmelCase__ : Optional[int] = kwargs.pop("""text_pair""" , _lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = kwargs.pop("""return_tensors""" , _lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = { """input_ids""": [], """attention_mask""": [], """token_type_ids""": [], } for idx, candidate_text in enumerate(_lowerCAmelCase ): if batch_text_pair is not None: UpperCAmelCase__ : str = batch_text_pair[idx] else: UpperCAmelCase__ : Any = None UpperCAmelCase__ : str = super().__call__(_lowerCAmelCase , _lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = encoded_candidates.get("""input_ids""" ) UpperCAmelCase__ : str = encoded_candidates.get("""attention_mask""" ) UpperCAmelCase__ : Union[str, Any] = encoded_candidates.get("""token_type_ids""" ) if encoded_input_ids is not None: output_data["input_ids"].append(_lowerCAmelCase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(_lowerCAmelCase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = {key: item for key, item in output_data.items() if len(_lowerCAmelCase ) != 0} return BatchEncoding(_lowerCAmelCase , tensor_type=_lowerCAmelCase ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase = None ): UpperCAmelCase__ : Any = [self.sep_token_id] UpperCAmelCase__ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase = None ): UpperCAmelCase__ : List[str] = self._tokenizer.model.save(_lowerCAmelCase , name=_lowerCAmelCase ) return tuple(_lowerCAmelCase )
79
0
from __future__ import annotations def A ( lowercase__ : list[int] , lowercase__ : int ) -> list[int]: UpperCamelCase__ :Tuple = 0 UpperCamelCase__ :Any = len(lowercase__ ) - 1 while i < j: if nums[i] + nums[j] == target: return [i, j] elif nums[i] + nums[j] < target: UpperCamelCase__ :Union[str, Any] = i + 1 else: UpperCamelCase__ :Optional[int] = j - 1 return [] if __name__ == "__main__": import doctest doctest.testmod() print(f'''{two_pointer([2, 7, 11, 15], 9) = }''')
45
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'facebook/bart-large-mnli' __lowerCamelCase = ( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) __lowerCamelCase = 'text_classifier' __lowerCamelCase = AutoTokenizer __lowerCamelCase = AutoModelForSequenceClassification __lowerCamelCase = ['text', ['text']] __lowerCamelCase = ['text'] def __UpperCAmelCase ( self ): super().setup() UpperCAmelCase__ : Optional[Any] = self.model.config UpperCAmelCase__ : Tuple = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("""entail""" ): UpperCAmelCase__ : Dict = int(_lowerCAmelCase ) if self.entailment_id == -1: raise ValueError("""Could not determine the entailment ID from the model config, please pass it at init.""" ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : List[Any] = labels return self.pre_processor( [text] * len(_lowerCAmelCase ) , [f"This example is {label}" for label in labels] , return_tensors="""pt""" , padding="""max_length""" , ) def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : str = outputs.logits UpperCAmelCase__ : List[Any] = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
79
0
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') _lowerCAmelCase : int = logging.getLogger(__name__) @dataclass class A_ : lowerCAmelCase__ = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) lowerCAmelCase__ = field( default=_a , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) lowerCAmelCase__ = field( default=_a , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) lowerCAmelCase__ = field( default=_a , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) lowerCAmelCase__ = field( default=_a , metadata={'help': 'Whether to use one of the fast tokenizer (backed by the tokenizers library) or not.'} , ) lowerCAmelCase__ = field( default='main' , metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'} , ) lowerCAmelCase__ = field( default=_a , metadata={ 'help': ( 'Will use the token generated when running `huggingface-cli login` (necessary to use this script ' 'with private models).' ) } , ) @dataclass class A_ : lowerCAmelCase__ = field(default=_a , metadata={'help': 'The input training data file (a text file).'} ) lowerCAmelCase__ = field( default=_a , metadata={'help': 'An optional input evaluation data file to evaluate the perplexity on (a text file).'} , ) lowerCAmelCase__ = field( default=_a , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) lowerCAmelCase__ = field( default=_a , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) lowerCAmelCase__ = field( default=_a , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. If passed, sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) lowerCAmelCase__ = field( default=_a , metadata={ 'help': ( 'Whether to pad all samples to the maximum sentence length. ' 'If False, will pad the samples dynamically when batching to the maximum length in the batch. More ' 'efficient on GPU but very bad for TPU.' ) } , ) lowerCAmelCase__ = field( default=_a , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) lowerCAmelCase__ = field( default=_a , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of evaluation examples to this ' 'value if set.' ) } , ) def _lowercase ( self: Tuple ): '''simple docstring''' if self.train_file is not None: _lowerCamelCase : List[str] = self.train_file.split("." )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: _lowerCamelCase : Tuple = self.validation_file.split("." )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class A_ : lowerCAmelCase__ = 42 lowerCAmelCase__ = True lowerCAmelCase__ = None lowerCAmelCase__ = None def __call__( self: Any ,__lowerCAmelCase: List[str] ): '''simple docstring''' _lowerCamelCase : Optional[Any] = "label" if "label" in features[0].keys() else "labels" _lowerCamelCase : List[Any] = [feature.pop(__lowerCAmelCase ) for feature in features] _lowerCamelCase : Optional[Any] = len(__lowerCAmelCase ) _lowerCamelCase : str = len(features[0]["input_ids"] ) _lowerCamelCase : Any = [ [{k: v[i] for k, v in feature.items()} for i in range(__lowerCAmelCase )] for feature in features ] _lowerCamelCase : List[str] = list(chain(*__lowerCAmelCase ) ) _lowerCamelCase : Optional[Any] = self.tokenizer.pad( __lowerCAmelCase ,padding=self.padding ,max_length=self.max_length ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors="pt" ,) # Un-flatten _lowerCamelCase : List[str] = {k: v.view(__lowerCAmelCase ,__lowerCAmelCase ,-1 ) for k, v in batch.items()} # Add back labels _lowerCamelCase : List[Any] = torch.tensor(__lowerCAmelCase ,dtype=torch.intaa ) return batch def lowerCamelCase_( ) -> Tuple: '''simple docstring''' _lowerCamelCase : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : str = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : int = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_swag" , _lowerCamelCase , _lowerCamelCase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _lowerCamelCase : Optional[int] = training_args.get_process_log_level() logger.setLevel(_lowerCamelCase ) datasets.utils.logging.set_verbosity(_lowerCamelCase ) transformers.utils.logging.set_verbosity(_lowerCamelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. _lowerCamelCase : str = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _lowerCamelCase : List[Any] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: _lowerCamelCase : str = {} if data_args.train_file is not None: _lowerCamelCase : List[Any] = data_args.train_file if data_args.validation_file is not None: _lowerCamelCase : Union[str, Any] = data_args.validation_file _lowerCamelCase : Optional[Any] = data_args.train_file.split("." )[-1] _lowerCamelCase : List[Any] = load_dataset( _lowerCamelCase , data_files=_lowerCamelCase , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. _lowerCamelCase : Optional[Any] = load_dataset( "swag" , "regular" , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _lowerCamelCase : Any = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _lowerCamelCase : Union[str, Any] = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=_lowerCamelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. _lowerCamelCase : str = [F"""ending{i}""" for i in range(4 )] _lowerCamelCase : Any = "sent1" _lowerCamelCase : int = "sent2" if data_args.max_seq_length is None: _lowerCamelCase : Union[str, Any] = tokenizer.model_max_length if max_seq_length > 1024: logger.warning( "The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value" " of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can" " override this default with `--block_size xxx`." ) _lowerCamelCase : Optional[Any] = 1024 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F"""The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the""" F"""model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.""" ) _lowerCamelCase : Tuple = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(_lowerCamelCase ): _lowerCamelCase : List[Any] = [[context] * 4 for context in examples[context_name]] _lowerCamelCase : str = examples[question_header_name] _lowerCamelCase : List[Any] = [ [F"""{header} {examples[end][i]}""" for end in ending_names] for i, header in enumerate(_lowerCamelCase ) ] # Flatten out _lowerCamelCase : Any = list(chain(*_lowerCamelCase ) ) _lowerCamelCase : List[str] = list(chain(*_lowerCamelCase ) ) # Tokenize _lowerCamelCase : Optional[Any] = tokenizer( _lowerCamelCase , _lowerCamelCase , truncation=_lowerCamelCase , max_length=_lowerCamelCase , padding="max_length" if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(_lowerCamelCase ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError("--do_train requires a train dataset" ) _lowerCamelCase : str = raw_datasets["train"] if data_args.max_train_samples is not None: _lowerCamelCase : str = min(len(_lowerCamelCase ) , data_args.max_train_samples ) _lowerCamelCase : List[str] = train_dataset.select(range(_lowerCamelCase ) ) with training_args.main_process_first(desc="train dataset map pre-processing" ): _lowerCamelCase : str = train_dataset.map( _lowerCamelCase , batched=_lowerCamelCase , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError("--do_eval requires a validation dataset" ) _lowerCamelCase : str = raw_datasets["validation"] if data_args.max_eval_samples is not None: _lowerCamelCase : int = min(len(_lowerCamelCase ) , data_args.max_eval_samples ) _lowerCamelCase : int = eval_dataset.select(range(_lowerCamelCase ) ) with training_args.main_process_first(desc="validation dataset map pre-processing" ): _lowerCamelCase : Optional[Any] = eval_dataset.map( _lowerCamelCase , batched=_lowerCamelCase , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator _lowerCamelCase : List[str] = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=_lowerCamelCase , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(_lowerCamelCase ): _lowerCamelCase, _lowerCamelCase : Optional[int] = eval_predictions _lowerCamelCase : List[Any] = np.argmax(_lowerCamelCase , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer _lowerCamelCase : Tuple = Trainer( model=_lowerCamelCase , args=_lowerCamelCase , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=_lowerCamelCase , data_collator=_lowerCamelCase , compute_metrics=_lowerCamelCase , ) # Training if training_args.do_train: _lowerCamelCase : Optional[Any] = None if training_args.resume_from_checkpoint is not None: _lowerCamelCase : List[Any] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _lowerCamelCase : Tuple = last_checkpoint _lowerCamelCase : Optional[int] = trainer.train(resume_from_checkpoint=_lowerCamelCase ) trainer.save_model() # Saves the tokenizer too for easy upload _lowerCamelCase : Optional[Any] = train_result.metrics _lowerCamelCase : List[str] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(_lowerCamelCase ) ) _lowerCamelCase : Optional[Any] = min(_lowerCamelCase , len(_lowerCamelCase ) ) trainer.log_metrics("train" , _lowerCamelCase ) trainer.save_metrics("train" , _lowerCamelCase ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info("*** Evaluate ***" ) _lowerCamelCase : int = trainer.evaluate() _lowerCamelCase : List[Any] = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(_lowerCamelCase ) _lowerCamelCase : List[Any] = min(_lowerCamelCase , len(_lowerCamelCase ) ) trainer.log_metrics("eval" , _lowerCamelCase ) trainer.save_metrics("eval" , _lowerCamelCase ) _lowerCamelCase : Union[str, Any] = { "finetuned_from": model_args.model_name_or_path, "tasks": "multiple-choice", "dataset_tags": "swag", "dataset_args": "regular", "dataset": "SWAG", "language": "en", } if training_args.push_to_hub: trainer.push_to_hub(**_lowerCamelCase ) else: trainer.create_model_card(**_lowerCamelCase ) def lowerCamelCase_( _lowerCamelCase ) -> Dict: '''simple docstring''' main() if __name__ == "__main__": main()
46
from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=13 , _lowerCAmelCase=30 , _lowerCAmelCase=2 , _lowerCAmelCase=3 , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=32 , _lowerCAmelCase=2 , _lowerCAmelCase=4 , _lowerCAmelCase=37 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=10 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=3 , _lowerCAmelCase=None , ): UpperCAmelCase__ : Tuple = parent UpperCAmelCase__ : Optional[int] = batch_size UpperCAmelCase__ : Union[str, Any] = image_size UpperCAmelCase__ : int = patch_size UpperCAmelCase__ : str = num_channels UpperCAmelCase__ : int = is_training UpperCAmelCase__ : List[str] = use_labels UpperCAmelCase__ : List[Any] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : Tuple = num_attention_heads UpperCAmelCase__ : Optional[int] = intermediate_size UpperCAmelCase__ : Optional[Any] = hidden_act UpperCAmelCase__ : int = hidden_dropout_prob UpperCAmelCase__ : int = attention_probs_dropout_prob UpperCAmelCase__ : List[str] = type_sequence_label_size UpperCAmelCase__ : Optional[int] = initializer_range UpperCAmelCase__ : Any = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) UpperCAmelCase__ : Any = (image_size // patch_size) ** 2 UpperCAmelCase__ : Tuple = num_patches + 1 def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ : List[str] = None if self.use_labels: UpperCAmelCase__ : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ : Union[str, Any] = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self ): return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : str = TFViTModel(config=_lowerCAmelCase ) UpperCAmelCase__ : str = model(_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase__ : Optional[Any] = self.image_size // 2 UpperCAmelCase__ : List[str] = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase__ : List[Any] = model(_lowerCAmelCase , interpolate_pos_encoding=_lowerCAmelCase , training=_lowerCAmelCase ) UpperCAmelCase__ : str = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : Tuple = self.type_sequence_label_size UpperCAmelCase__ : List[Any] = TFViTForImageClassification(_lowerCAmelCase ) UpperCAmelCase__ : List[str] = model(_lowerCAmelCase , labels=_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase__ : Tuple = self.image_size // 2 UpperCAmelCase__ : Union[str, Any] = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase__ : List[str] = model(_lowerCAmelCase , interpolate_pos_encoding=_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase__ : Union[str, Any] = 1 UpperCAmelCase__ : Optional[Any] = TFViTForImageClassification(_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase__ : List[str] = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : int = config_and_inputs UpperCAmelCase__ : int = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class UpperCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () __lowerCamelCase = ( {'feature-extraction': TFViTModel, 'image-classification': TFViTForImageClassification} if is_tf_available() else {} ) __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = TFViTModelTester(self ) UpperCAmelCase__ : int = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def __UpperCAmelCase ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def __UpperCAmelCase ( self ): pass @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def __UpperCAmelCase ( self ): pass def __UpperCAmelCase ( self ): UpperCAmelCase__ , UpperCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : str = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , tf.keras.layers.Layer ) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : Optional[int] = model_class(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ : Tuple = [*signature.parameters.keys()] UpperCAmelCase__ : str = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = TFViTModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(_lowerCAmelCase ) def _lowerCamelCase ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ : List[str] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class UpperCAmelCase_ ( unittest.TestCase ): @cached_property def __UpperCAmelCase ( self ): return ViTImageProcessor.from_pretrained("""google/vit-base-patch16-224""" ) if is_vision_available() else None @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = TFViTForImageClassification.from_pretrained("""google/vit-base-patch16-224""" ) UpperCAmelCase__ : List[Any] = self.default_image_processor UpperCAmelCase__ : Union[str, Any] = prepare_img() UpperCAmelCase__ : Optional[Any] = image_processor(images=_lowerCAmelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase__ : int = model(**_lowerCAmelCase ) # verify the logits UpperCAmelCase__ : Tuple = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) UpperCAmelCase__ : int = tf.constant([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ) tf.debugging.assert_near(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 )
79
0
import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def UpperCAmelCase__ ( ): with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(lowerCamelCase_ ): requests.request('GET' , 'https://huggingface.co' ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request('GET' , 'https://huggingface.co' , timeout=1.0 ) @pytest.mark.integration def UpperCAmelCase__ ( ): with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request('GET' , 'https://huggingface.co' ) def UpperCAmelCase__ ( ): with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(lowerCamelCase_ ): http_head('https://huggingface.co' )
47
from functools import lru_cache @lru_cache def _lowerCamelCase ( __lowerCamelCase ) -> int: '''simple docstring''' if num < 0: raise ValueError("""Number should not be negative.""" ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
79
0
'''simple docstring''' import copy import random from transformers import CLIPTokenizer class A ( SCREAMING_SNAKE_CASE__ ): def __init__( self : Any , *__magic_name__ : Optional[int] , **__magic_name__ : Union[str, Any] ): """simple docstring""" super().__init__(*__magic_name__ , **__magic_name__ ) lowerCAmelCase__ = {} def __SCREAMING_SNAKE_CASE ( self : int , __magic_name__ : Dict , *__magic_name__ : List[str] , **__magic_name__ : Optional[Any] ): """simple docstring""" lowerCAmelCase__ = super().add_tokens(__magic_name__ , *__magic_name__ , **__magic_name__ ) if num_added_tokens == 0: raise ValueError( f"""The tokenizer already contains the token {placeholder_token}. Please pass a different""" " `placeholder_token` that is not already in the tokenizer." ) def __SCREAMING_SNAKE_CASE ( self : List[Any] , __magic_name__ : Any , *__magic_name__ : Optional[Any] , __magic_name__ : Optional[int]=1 , **__magic_name__ : Optional[int] ): """simple docstring""" lowerCAmelCase__ = [] if num_vec_per_token == 1: self.try_adding_tokens(__magic_name__ , *__magic_name__ , **__magic_name__ ) output.append(__magic_name__ ) else: lowerCAmelCase__ = [] for i in range(__magic_name__ ): lowerCAmelCase__ = placeholder_token + f"""_{i}""" self.try_adding_tokens(__magic_name__ , *__magic_name__ , **__magic_name__ ) output.append(__magic_name__ ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( f"""The tokenizer already has placeholder token {token} that can get confused with""" f""" {placeholder_token}keep placeholder tokens independent""" ) lowerCAmelCase__ = output def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] , __magic_name__ : Union[str, Any] , __magic_name__ : Optional[int]=False , __magic_name__ : List[str]=1.0 ): """simple docstring""" if isinstance(__magic_name__ , __magic_name__ ): lowerCAmelCase__ = [] for i in range(len(__magic_name__ ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=__magic_name__ ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: lowerCAmelCase__ = self.token_map[placeholder_token] lowerCAmelCase__ = tokens[: 1 + int(len(__magic_name__ ) * prop_tokens_to_load )] if vector_shuffle: lowerCAmelCase__ = copy.copy(__magic_name__ ) random.shuffle(__magic_name__ ) lowerCAmelCase__ = text.replace(__magic_name__ , " ".join(__magic_name__ ) ) return text def __call__( self : List[Any] , __magic_name__ : Union[str, Any] , *__magic_name__ : int , __magic_name__ : Union[str, Any]=False , __magic_name__ : Union[str, Any]=1.0 , **__magic_name__ : List[str] ): """simple docstring""" return super().__call__( self.replace_placeholder_tokens_in_text( __magic_name__ , vector_shuffle=__magic_name__ , prop_tokens_to_load=__magic_name__ ) , *__magic_name__ , **__magic_name__ , ) def __SCREAMING_SNAKE_CASE ( self : int , __magic_name__ : Union[str, Any] , *__magic_name__ : Union[str, Any] , __magic_name__ : Union[str, Any]=False , __magic_name__ : Union[str, Any]=1.0 , **__magic_name__ : Any ): """simple docstring""" return super().encode( self.replace_placeholder_tokens_in_text( __magic_name__ , vector_shuffle=__magic_name__ , prop_tokens_to_load=__magic_name__ ) , *__magic_name__ , **__magic_name__ , )
48
import argparse import hashlib # hashlib is only used inside the Test class import struct class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase ): UpperCAmelCase__ : Any = data UpperCAmelCase__ : List[Any] = [0X6745_2301, 0Xefcd_ab89, 0X98ba_dcfe, 0X1032_5476, 0Xc3d2_e1f0] @staticmethod def __UpperCAmelCase ( _lowerCAmelCase , _lowerCAmelCase ): return ((n << b) | (n >> (32 - b))) & 0Xffff_ffff def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) UpperCAmelCase__ : Optional[int] = self.data + padding + struct.pack(""">Q""" , 8 * len(self.data ) ) return padded_data def __UpperCAmelCase ( self ): return [ self.padded_data[i : i + 64] for i in range(0 , len(self.padded_data ) , 64 ) ] def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : Dict = list(struct.unpack(""">16L""" , _lowerCAmelCase ) ) + [0] * 64 for i in range(16 , 80 ): UpperCAmelCase__ : Optional[int] = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) , 1 ) return w def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[str] = self.padding() UpperCAmelCase__ : List[str] = self.split_blocks() for block in self.blocks: UpperCAmelCase__ : Tuple = self.expand_block(_lowerCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.h for i in range(0 , 80 ): if 0 <= i < 20: UpperCAmelCase__ : Optional[int] = (b & c) | ((~b) & d) UpperCAmelCase__ : int = 0X5a82_7999 elif 20 <= i < 40: UpperCAmelCase__ : Tuple = b ^ c ^ d UpperCAmelCase__ : int = 0X6ed9_eba1 elif 40 <= i < 60: UpperCAmelCase__ : List[str] = (b & c) | (b & d) | (c & d) UpperCAmelCase__ : Tuple = 0X8f1b_bcdc elif 60 <= i < 80: UpperCAmelCase__ : int = b ^ c ^ d UpperCAmelCase__ : str = 0Xca62_c1d6 UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = ( self.rotate(_lowerCAmelCase , 5 ) + f + e + k + expanded_block[i] & 0Xffff_ffff, a, self.rotate(_lowerCAmelCase , 30 ), c, d, ) UpperCAmelCase__ : int = ( self.h[0] + a & 0Xffff_ffff, self.h[1] + b & 0Xffff_ffff, self.h[2] + c & 0Xffff_ffff, self.h[3] + d & 0Xffff_ffff, self.h[4] + e & 0Xffff_ffff, ) return ("{:08x}" * 5).format(*self.h ) def _lowerCamelCase ( ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase__ : Optional[Any] = B"""Test String""" assert SHAaHash(__lowerCamelCase ).final_hash() == hashlib.shaa(__lowerCamelCase ).hexdigest() # noqa: S324 def _lowerCamelCase ( ) -> str: '''simple docstring''' UpperCAmelCase__ : Optional[int] = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument("""--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) UpperCAmelCase__ : str = parser.parse_args() UpperCAmelCase__ : Union[str, Any] = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: UpperCAmelCase__ : List[Any] = f.read() else: UpperCAmelCase__ : int = bytes(__lowerCamelCase , """utf-8""" ) print(SHAaHash(__lowerCamelCase ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
79
0
"""simple docstring""" import argparse import random import joblib import numpy as np import torch from igf.igf import ( SecondaryLearner, collect_objective_set, compute_perplexity, generate_datasets, load_gpta, recopy_gpta, set_seed, train_secondary_learner, ) from torch.utils.data import DataLoader, RandomSampler from transformers import GPTaLMHeadModel def lowercase__ ( snake_case_ :Dict=32 , snake_case_ :Dict=10 , snake_case_ :Union[str, Any]=100 , snake_case_ :Union[str, Any]=1_026 , snake_case_ :str=True , snake_case_ :List[str]="data/tokenized_stories_train_wikitext103.jbl" , snake_case_ :List[str]="igf_context_pairs.jbl" , ): set_seed(3 ) # generate train_data and objective_set __UpperCAmelCase , __UpperCAmelCase = generate_datasets( snake_case_ , snake_case_ , number=snake_case_ , min_len=1_026 , trim=snake_case_ ) # keeps model same across runs set_seed(4 ) # model, lm_optimizer, lm_scheduler = recopy_gpt2(model, device, max_steps) # store original model weights # can we train on GPU? __UpperCAmelCase = torch.device('''cuda:0''' if torch.cuda.is_available() else '''cpu''' ) # load pretrained model __UpperCAmelCase = load_gpta('''gpt2''' ).to(snake_case_ ) print('''computing perplexity on objective set''' ) __UpperCAmelCase = compute_perplexity(snake_case_ , snake_case_ , snake_case_ ).item() print('''perplexity on objective set:''' , snake_case_ ) # collect igf pairs and save to file demo.jbl collect_objective_set(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # clean up, delete model and data we don't need anymore del model, train_data, objective_set torch.cuda.empty_cache() def lowercase__ ( snake_case_ :Dict , snake_case_ :List[Any]=15 , snake_case_ :Any=128 , snake_case_ :int=100 , snake_case_ :List[str]="igf_model.pt" , ): set_seed(42 ) # Load pre-trained model __UpperCAmelCase = GPTaLMHeadModel.from_pretrained('''gpt2''' ) # Initialize secondary learner to use embedding weights of model __UpperCAmelCase = SecondaryLearner(snake_case_ ) # Train secondary learner __UpperCAmelCase = train_secondary_learner( snake_case_ , snake_case_ , max_epochs=snake_case_ , batch_size=snake_case_ , eval_freq=100 , igf_model_path=snake_case_ , ) del model, secondary_learner_train_data torch.cuda.empty_cache() return secondary_learner def lowercase__ ( snake_case_ :str , snake_case_ :int , snake_case_ :str , snake_case_ :List[str]=32 , snake_case_ :Optional[Any]=1_000 , snake_case_ :Optional[Any]=16 , snake_case_ :List[str]=1.0 , snake_case_ :List[Any]=recopy_gpta , snake_case_ :Optional[int]=None , snake_case_ :Optional[int]=10 , snake_case_ :Tuple="gpt2_finetuned.pt" , ): __UpperCAmelCase = torch.device('''cuda:0''' if torch.cuda.is_available() else '''cpu''' ) __UpperCAmelCase = RandomSampler(snake_case_ ) __UpperCAmelCase = DataLoader(snake_case_ , sampler=snake_case_ ) __UpperCAmelCase = max_steps // (len(snake_case_ )) + 1 __UpperCAmelCase = 0 __UpperCAmelCase = torch.zeros((1, context_len) , dtype=torch.long , device=snake_case_ ) __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = recopy_model(snake_case_ , snake_case_ , snake_case_ ) model.train() if secondary_learner is not None: secondary_learner.to(snake_case_ ) secondary_learner.eval() __UpperCAmelCase = [] __UpperCAmelCase = 0 __UpperCAmelCase = [] __UpperCAmelCase = [] # Compute the performance of the transformer model at the beginning __UpperCAmelCase = compute_perplexity(snake_case_ , snake_case_ , snake_case_ ) test_perps.append(snake_case_ ) print('''Test perplexity, step''' , snake_case_ , ''':''' , snake_case_ ) for epoch in range(int(snake_case_ ) ): for step, example in enumerate(snake_case_ ): torch.cuda.empty_cache() __UpperCAmelCase = random.randint(0 , example.size(2 ) - context_len - 1 ) __UpperCAmelCase = example[0, 0, start : start + context_len] lm_optimizer.zero_grad() __UpperCAmelCase = model(snake_case_ , labels=snake_case_ ) __UpperCAmelCase = True if secondary_learner is not None: __UpperCAmelCase = secondary_learner.forward( torch.tensor(snake_case_ , dtype=torch.long , device=snake_case_ ).unsqueeze(0 ) )[0].item() observed_qs.append(float(snake_case_ ) ) # Here we implement the simple non-constant threshold for the predicted IG(X) value # We will decay the selectivity of our secondary learner filter from # 1 standard deviation above average to 1 below average after 10 batches. if global_step == 10: __UpperCAmelCase = -1 if predicted_q < threshold: __UpperCAmelCase = False # If we passed the filter, add the context to the batch! if do_backprop: contexts.append(np.array(context.cpu() ) ) __UpperCAmelCase = outputs[0] lm_loss.backward() examples += 1 del outputs # Once the batch is filled with enough contexts, backprop on the batch. if examples == batch_size: torch.cuda.empty_cache() __UpperCAmelCase = 0 # Do LM backprop torch.nn.utils.clip_grad_norm_(model.parameters() , 3.0 ) lm_optimizer.step() lm_scheduler.step() # Update learning rate schedule global_step += 1 # Compute the performance of the transformer model at this batch if global_step % eval_interval == 0: __UpperCAmelCase = compute_perplexity(snake_case_ , snake_case_ , snake_case_ ) test_perps.append(snake_case_ ) print('''Test perplexity, step''' , snake_case_ , ''':''' , snake_case_ ) # Break out of the loop after 60 batches if max_steps > 0 and global_step > 60: break if max_steps > 0 and global_step > 60: break # save finetuned transformer model torch.save(model.state_dict() , snake_case_ ) torch.cuda.empty_cache() # Do some cleaning up so we can reinitialize for the next run of this function del lm_optimizer del lm_scheduler return model def lowercase__ ( ): __UpperCAmelCase = argparse.ArgumentParser(description='''Fine-tune a transformer model with IGF on a language modeling task''' ) # Required parameters parser.add_argument( '''--data_dir''' , default=snake_case_ , type=snake_case_ , required=snake_case_ , help='''The input data dir. Should contain data files for WikiText.''' , ) parser.add_argument( '''--model_name_or_path''' , default=snake_case_ , type=snake_case_ , required=snake_case_ , help='''Path to pretrained model or model identifier from huggingface.co/models''' , ) parser.add_argument( '''--data_file''' , type=snake_case_ , default=snake_case_ , help=( '''A jbl file containing tokenized data which can be split as objective dataset, ''' '''train_dataset and test_dataset.''' ) , ) parser.add_argument( '''--igf_data_file''' , type=snake_case_ , default=snake_case_ , help='''A jbl file containing the context and information gain pairs to train secondary learner.''' , ) parser.add_argument( '''--output_dir''' , default=snake_case_ , type=snake_case_ , required=snake_case_ , help='''The output directory where the final fine-tuned model is stored.''' , ) parser.add_argument( '''--tokenizer_name''' , default=snake_case_ , type=snake_case_ , help='''Pretrained tokenizer name or path if not the same as model_name''' , ) parser.add_argument('''--seed''' , type=snake_case_ , default=snake_case_ , help='''A seed for reproducible training.''' ) parser.add_argument( '''--context_len''' , default=32 , type=snake_case_ , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--size_objective_set''' , default=100 , type=snake_case_ , help='''number of articles that are long enough to be used as our objective set''' , ) parser.add_argument( '''--eval_freq''' , default=100 , type=snake_case_ , help='''secondary model evaluation is triggered at eval_freq''' ) parser.add_argument('''--max_steps''' , default=1_000 , type=snake_case_ , help='''To calculate training epochs''' ) parser.add_argument( '''--secondary_learner_batch_size''' , default=128 , type=snake_case_ , help='''batch size of training data for secondary learner''' , ) parser.add_argument( '''--batch_size''' , default=16 , type=snake_case_ , help='''batch size of training data of language model(gpt2) ''' ) parser.add_argument( '''--eval_interval''' , default=10 , type=snake_case_ , help=( '''decay the selectivity of our secondary learner filter from''' '''1 standard deviation above average to 1 below average after 10 batches''' ) , ) parser.add_argument( '''--number''' , default=100 , type=snake_case_ , help='''The number of examples split to be used as objective_set/test_data''' ) parser.add_argument( '''--min_len''' , default=1_026 , type=snake_case_ , help='''The minimum length of the article to be used as objective set''' ) parser.add_argument( '''--secondary_learner_max_epochs''' , default=15 , type=snake_case_ , help='''number of epochs to train secondary learner''' ) parser.add_argument('''--trim''' , default=snake_case_ , type=snake_case_ , help='''truncate the example if it exceeds context length''' ) parser.add_argument( '''--threshold''' , default=1.0 , type=snake_case_ , help=( '''The threshold value used by secondary learner to filter the train_data and allow only''' ''' informative data as input to the model''' ) , ) parser.add_argument('''--finetuned_model_name''' , default='''gpt2_finetuned.pt''' , type=snake_case_ , help='''finetuned_model_name''' ) parser.add_argument( '''--recopy_model''' , default=snake_case_ , type=snake_case_ , help='''Reset the model to the original pretrained GPT-2 weights after each iteration''' , ) # function calls # Collecting *n* pairs of context and information gain(X, IG(X)) for training the secondary learner generate_n_pairs( context_len=32 , max_steps=10 , size_objective_set=100 , min_len=1_026 , trim=snake_case_ , data_file='''data/tokenized_stories_train_wikitext103.jbl''' , igf_data_file='''igf_context_pairs.jbl''' , ) # Load train data for secondary learner __UpperCAmelCase = joblib.load('''data/IGF_values.jbl''' ) # Train secondary learner __UpperCAmelCase = training_secondary_learner( snake_case_ , secondary_learner_max_epochs=15 , secondary_learner_batch_size=128 , eval_freq=100 , igf_model_path='''igf_model.pt''' , ) # load pretrained gpt2 model __UpperCAmelCase = GPTaLMHeadModel.from_pretrained('''gpt2''' ) set_seed(42 ) # Generate train and test data to train and evaluate gpt2 model __UpperCAmelCase , __UpperCAmelCase = generate_datasets( context_len=32 , file='''data/tokenized_stories_train_wikitext103.jbl''' , number=100 , min_len=1_026 , trim=snake_case_ ) # fine-tuning of the gpt2 model using igf (Information Gain Filtration) finetune( snake_case_ , snake_case_ , snake_case_ , context_len=32 , max_steps=1_000 , batch_size=16 , threshold=1.0 , recopy_model=snake_case_ , secondary_learner=snake_case_ , eval_interval=10 , finetuned_model_name='''gpt2_finetuned.pt''' , ) if __name__ == "__main__": main()
49
from importlib import import_module from .logging import get_logger SCREAMING_SNAKE_CASE__ : Union[str, Any] = get_logger(__name__) class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : List[str] = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith("""__""" ): setattr(self , _lowerCAmelCase , getattr(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase__ : Tuple = module._original_module if isinstance(_lowerCAmelCase , _PatchedModuleObj ) else module class UpperCAmelCase_ : __lowerCamelCase = [] def __init__( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : str = obj UpperCAmelCase__ : List[str] = target UpperCAmelCase__ : List[str] = new UpperCAmelCase__ : Any = target.split(""".""" )[0] UpperCAmelCase__ : Union[str, Any] = {} UpperCAmelCase__ : str = attrs or [] def __enter__( self ): *UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.target.split(""".""" ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(_lowerCAmelCase ) ): try: UpperCAmelCase__ : Optional[int] = import_module(""".""".join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): UpperCAmelCase__ : Any = getattr(self.obj , _lowerCAmelCase ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(_lowerCAmelCase , _PatchedModuleObj ) and obj_attr._original_module is submodule) ): UpperCAmelCase__ : List[Any] = obj_attr # patch at top level setattr(self.obj , _lowerCAmelCase , _PatchedModuleObj(_lowerCAmelCase , attrs=self.attrs ) ) UpperCAmelCase__ : Optional[Any] = getattr(self.obj , _lowerCAmelCase ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(_lowerCAmelCase , _lowerCAmelCase , _PatchedModuleObj(getattr(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) , attrs=self.attrs ) ) UpperCAmelCase__ : Union[str, Any] = getattr(_lowerCAmelCase , _lowerCAmelCase ) # finally set the target attribute setattr(_lowerCAmelCase , _lowerCAmelCase , self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: UpperCAmelCase__ : Union[str, Any] = getattr(import_module(""".""".join(_lowerCAmelCase ) ) , _lowerCAmelCase ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , _lowerCAmelCase ) is attr_value: UpperCAmelCase__ : Optional[int] = getattr(self.obj , _lowerCAmelCase ) setattr(self.obj , _lowerCAmelCase , self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" UpperCAmelCase__ : Dict = globals()["""__builtins__"""][target_attr] setattr(self.obj , _lowerCAmelCase , self.new ) else: raise RuntimeError(f"Tried to patch attribute {target_attr} instead of a submodule." ) def __exit__( self , *_lowerCAmelCase ): for attr in list(self.original ): setattr(self.obj , _lowerCAmelCase , self.original.pop(_lowerCAmelCase ) ) def __UpperCAmelCase ( self ): self.__enter__() self._active_patches.append(self ) def __UpperCAmelCase ( self ): try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
79
0
'''simple docstring''' from argparse import ArgumentParser, Namespace from ..utils import logging from . import BaseTransformersCLICommand def A__ ( __lowerCAmelCase : Namespace ): return ConvertCommand( args.model_type , args.tf_checkpoint , args.pytorch_dump_output , args.config , args.finetuning_task_name ) UpperCamelCase : Dict = '\ntransformers can only be used from the commandline to convert TensorFlow models in PyTorch, In that case, it requires\nTensorFlow to be installed. Please see https://www.tensorflow.org/install/ for installation instructions.\n' class UpperCamelCase__ (a ): '''simple docstring''' @staticmethod def UpperCamelCase_ ( _lowerCAmelCase ): lowerCamelCase__ = parser.add_parser( """convert""" ,help="""CLI tool to run convert model from original author checkpoints to Transformers PyTorch checkpoints.""" ,) train_parser.add_argument("""--model_type""" ,type=_lowerCAmelCase ,required=_lowerCAmelCase ,help="""Model's type.""" ) train_parser.add_argument( """--tf_checkpoint""" ,type=_lowerCAmelCase ,required=_lowerCAmelCase ,help="""TensorFlow checkpoint path or folder.""" ) train_parser.add_argument( """--pytorch_dump_output""" ,type=_lowerCAmelCase ,required=_lowerCAmelCase ,help="""Path to the PyTorch saved model output.""" ) train_parser.add_argument("""--config""" ,type=_lowerCAmelCase ,default="""""" ,help="""Configuration file path or folder.""" ) train_parser.add_argument( """--finetuning_task_name""" ,type=_lowerCAmelCase ,default=_lowerCAmelCase ,help="""Optional fine-tuning task name if the TF model was a finetuned model.""" ,) train_parser.set_defaults(func=_lowerCAmelCase ) def __init__( self ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,*_lowerCAmelCase ,): lowerCamelCase__ = logging.get_logger("""transformers-cli/converting""" ) self._logger.info(F'''Loading model {model_type}''' ) lowerCamelCase__ = model_type lowerCamelCase__ = tf_checkpoint lowerCamelCase__ = pytorch_dump_output lowerCamelCase__ = config lowerCamelCase__ = finetuning_task_name def UpperCamelCase_ ( self ): if self._model_type == "albert": try: from ..models.albert.convert_albert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_lowerCAmelCase ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) elif self._model_type == "bert": try: from ..models.bert.convert_bert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_lowerCAmelCase ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) elif self._model_type == "funnel": try: from ..models.funnel.convert_funnel_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_lowerCAmelCase ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) elif self._model_type == "t5": try: from ..models.ta.convert_ta_original_tf_checkpoint_to_pytorch import convert_tf_checkpoint_to_pytorch except ImportError: raise ImportError(_lowerCAmelCase ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) elif self._model_type == "gpt": from ..models.openai.convert_openai_original_tf_checkpoint_to_pytorch import ( convert_openai_checkpoint_to_pytorch, ) convert_openai_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) elif self._model_type == "transfo_xl": try: from ..models.transfo_xl.convert_transfo_xl_original_tf_checkpoint_to_pytorch import ( convert_transfo_xl_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_lowerCAmelCase ) if "ckpt" in self._tf_checkpoint.lower(): lowerCamelCase__ = self._tf_checkpoint lowerCamelCase__ = """""" else: lowerCamelCase__ = self._tf_checkpoint lowerCamelCase__ = """""" convert_transfo_xl_checkpoint_to_pytorch( _lowerCAmelCase ,self._config ,self._pytorch_dump_output ,_lowerCAmelCase ) elif self._model_type == "gpt2": try: from ..models.gpta.convert_gpta_original_tf_checkpoint_to_pytorch import ( convert_gpta_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_lowerCAmelCase ) convert_gpta_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) elif self._model_type == "xlnet": try: from ..models.xlnet.convert_xlnet_original_tf_checkpoint_to_pytorch import ( convert_xlnet_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_lowerCAmelCase ) convert_xlnet_checkpoint_to_pytorch( self._tf_checkpoint ,self._config ,self._pytorch_dump_output ,self._finetuning_task_name ) elif self._model_type == "xlm": from ..models.xlm.convert_xlm_original_pytorch_checkpoint_to_pytorch import ( convert_xlm_checkpoint_to_pytorch, ) convert_xlm_checkpoint_to_pytorch(self._tf_checkpoint ,self._pytorch_dump_output ) elif self._model_type == "lxmert": from ..models.lxmert.convert_lxmert_original_tf_checkpoint_to_pytorch import ( convert_lxmert_checkpoint_to_pytorch, ) convert_lxmert_checkpoint_to_pytorch(self._tf_checkpoint ,self._pytorch_dump_output ) elif self._model_type == "rembert": from ..models.rembert.convert_rembert_tf_checkpoint_to_pytorch import ( convert_rembert_tf_checkpoint_to_pytorch, ) convert_rembert_tf_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) else: raise ValueError( """--model_type should be selected in the list [bert, gpt, gpt2, t5, transfo_xl, xlnet, xlm, lxmert]""" )
50
from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : str = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Any = { """huggingface/informer-tourism-monthly""": ( """https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json""" ), # See all Informer models at https://huggingface.co/models?filter=informer } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'informer' __lowerCamelCase = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', 'num_hidden_layers': 'encoder_layers', } def __init__( self , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = "student_t" , _lowerCAmelCase = "nll" , _lowerCAmelCase = 1 , _lowerCAmelCase = None , _lowerCAmelCase = "mean" , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = 64 , _lowerCAmelCase = 32 , _lowerCAmelCase = 32 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = True , _lowerCAmelCase = "gelu" , _lowerCAmelCase = 0.0_5 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 100 , _lowerCAmelCase = 0.0_2 , _lowerCAmelCase=True , _lowerCAmelCase = "prob" , _lowerCAmelCase = 5 , _lowerCAmelCase = True , **_lowerCAmelCase , ): # time series specific configuration UpperCAmelCase__ : List[str] = prediction_length UpperCAmelCase__ : Optional[Any] = context_length or prediction_length UpperCAmelCase__ : str = distribution_output UpperCAmelCase__ : int = loss UpperCAmelCase__ : Optional[Any] = input_size UpperCAmelCase__ : Any = num_time_features UpperCAmelCase__ : int = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] UpperCAmelCase__ : Union[str, Any] = scaling UpperCAmelCase__ : Optional[Any] = num_dynamic_real_features UpperCAmelCase__ : List[str] = num_static_real_features UpperCAmelCase__ : str = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(_lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The cardinality should be a list of the same length as `num_static_categorical_features`""" ) UpperCAmelCase__ : List[str] = cardinality else: UpperCAmelCase__ : Optional[Any] = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(_lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The embedding dimension should be a list of the same length as `num_static_categorical_features`""" ) UpperCAmelCase__ : str = embedding_dimension else: UpperCAmelCase__ : List[str] = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] UpperCAmelCase__ : Union[str, Any] = num_parallel_samples # Transformer architecture configuration UpperCAmelCase__ : Dict = input_size * len(self.lags_sequence ) + self._number_of_features UpperCAmelCase__ : Any = d_model UpperCAmelCase__ : int = encoder_attention_heads UpperCAmelCase__ : Optional[Any] = decoder_attention_heads UpperCAmelCase__ : int = encoder_ffn_dim UpperCAmelCase__ : Tuple = decoder_ffn_dim UpperCAmelCase__ : List[Any] = encoder_layers UpperCAmelCase__ : Optional[Any] = decoder_layers UpperCAmelCase__ : Tuple = dropout UpperCAmelCase__ : int = attention_dropout UpperCAmelCase__ : List[str] = activation_dropout UpperCAmelCase__ : Any = encoder_layerdrop UpperCAmelCase__ : Union[str, Any] = decoder_layerdrop UpperCAmelCase__ : Tuple = activation_function UpperCAmelCase__ : Dict = init_std UpperCAmelCase__ : str = use_cache # Informer UpperCAmelCase__ : Union[str, Any] = attention_type UpperCAmelCase__ : int = sampling_factor UpperCAmelCase__ : Any = distil super().__init__(is_encoder_decoder=_lowerCAmelCase , **_lowerCAmelCase ) @property def __UpperCAmelCase ( self ): return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
79
0
'''simple docstring''' import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self : int ): UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = SamImageProcessor() UpperCAmelCase = SamProcessor(a__ ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : str , **a__ : Any ): return AutoProcessor.from_pretrained(self.tmpdirname , **a__ ).image_processor def __snake_case ( self : Tuple ): shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Any ): UpperCAmelCase = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] UpperCAmelCase = [Image.fromarray(np.moveaxis(a__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : Tuple ): UpperCAmelCase = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_image_processor(do_normalize=a__ , padding_value=1.0 ) UpperCAmelCase = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=a__ , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , a__ ) def __snake_case ( self : List[str] ): UpperCAmelCase = self.get_image_processor() UpperCAmelCase = SamProcessor(image_processor=a__ ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(a__ , return_tensors='''np''' ) UpperCAmelCase = processor(images=a__ , return_tensors='''np''' ) input_feat_extract.pop('''original_sizes''' ) # pop original_sizes as it is popped in the processor input_feat_extract.pop('''reshaped_input_sizes''' ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) @require_torch def __snake_case ( self : Any ): UpperCAmelCase = self.get_image_processor() UpperCAmelCase = SamProcessor(image_processor=a__ ) UpperCAmelCase = [torch.ones((1, 3, 5, 5) )] UpperCAmelCase = [[1764, 2646]] UpperCAmelCase = [[683, 1024]] UpperCAmelCase = processor.post_process_masks(a__ , a__ , a__ ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) UpperCAmelCase = processor.post_process_masks( a__ , torch.tensor(a__ ) , torch.tensor(a__ ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np UpperCAmelCase = [np.ones((1, 3, 5, 5) )] UpperCAmelCase = processor.post_process_masks(a__ , np.array(a__ ) , np.array(a__ ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) UpperCAmelCase = [[1, 0], [0, 1]] with self.assertRaises(a__ ): UpperCAmelCase = processor.post_process_masks(a__ , np.array(a__ ) , np.array(a__ ) ) @require_vision @require_tf class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self : List[str] ): UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = SamImageProcessor() UpperCAmelCase = SamProcessor(a__ ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[int] , **a__ : Any ): return AutoProcessor.from_pretrained(self.tmpdirname , **a__ ).image_processor def __snake_case ( self : int ): shutil.rmtree(self.tmpdirname ) def __snake_case ( self : int ): UpperCAmelCase = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] UpperCAmelCase = [Image.fromarray(np.moveaxis(a__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : str ): UpperCAmelCase = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_image_processor(do_normalize=a__ , padding_value=1.0 ) UpperCAmelCase = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=a__ , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , a__ ) def __snake_case ( self : Optional[int] ): UpperCAmelCase = self.get_image_processor() UpperCAmelCase = SamProcessor(image_processor=a__ ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(a__ , return_tensors='''np''' ) UpperCAmelCase = processor(images=a__ , return_tensors='''np''' ) input_feat_extract.pop('''original_sizes''' ) # pop original_sizes as it is popped in the processor input_feat_extract.pop('''reshaped_input_sizes''' ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) @require_tf def __snake_case ( self : Union[str, Any] ): UpperCAmelCase = self.get_image_processor() UpperCAmelCase = SamProcessor(image_processor=a__ ) UpperCAmelCase = [tf.ones((1, 3, 5, 5) )] UpperCAmelCase = [[1764, 2646]] UpperCAmelCase = [[683, 1024]] UpperCAmelCase = processor.post_process_masks(a__ , a__ , a__ , return_tensors='''tf''' ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) UpperCAmelCase = processor.post_process_masks( a__ , tf.convert_to_tensor(a__ ) , tf.convert_to_tensor(a__ ) , return_tensors='''tf''' , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np UpperCAmelCase = [np.ones((1, 3, 5, 5) )] UpperCAmelCase = processor.post_process_masks( a__ , np.array(a__ ) , np.array(a__ ) , return_tensors='''tf''' ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) UpperCAmelCase = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): UpperCAmelCase = processor.post_process_masks( a__ , np.array(a__ ) , np.array(a__ ) , return_tensors='''tf''' ) @require_vision @require_torchvision class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self : List[str] ): UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = SamImageProcessor() UpperCAmelCase = SamProcessor(a__ ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Union[str, Any] , **a__ : Tuple ): return AutoProcessor.from_pretrained(self.tmpdirname , **a__ ).image_processor def __snake_case ( self : List[Any] ): shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Union[str, Any] ): UpperCAmelCase = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] UpperCAmelCase = [Image.fromarray(np.moveaxis(a__ , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __snake_case ( self : List[Any] ): UpperCAmelCase = self.get_image_processor() UpperCAmelCase = SamProcessor(image_processor=a__ ) UpperCAmelCase = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) UpperCAmelCase = [tf.convert_to_tensor(a__ )] UpperCAmelCase = [torch.tensor(a__ )] UpperCAmelCase = [[1764, 2646]] UpperCAmelCase = [[683, 1024]] UpperCAmelCase = processor.post_process_masks( a__ , a__ , a__ , return_tensors='''tf''' ) UpperCAmelCase = processor.post_process_masks( a__ , a__ , a__ , return_tensors='''pt''' ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __snake_case ( self : List[Any] ): UpperCAmelCase = self.get_image_processor() UpperCAmelCase = SamProcessor(image_processor=a__ ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(a__ , return_tensors='''pt''' )['''pixel_values'''].numpy() UpperCAmelCase = processor(images=a__ , return_tensors='''pt''' )['''pixel_values'''].numpy() UpperCAmelCase = image_processor(a__ , return_tensors='''tf''' )['''pixel_values'''].numpy() UpperCAmelCase = processor(images=a__ , return_tensors='''tf''' )['''pixel_values'''].numpy() self.assertTrue(np.allclose(a__ , a__ ) ) self.assertTrue(np.allclose(a__ , a__ ) ) self.assertTrue(np.allclose(a__ , a__ ) )
51
def _lowerCamelCase ( __lowerCamelCase ) -> bool: '''simple docstring''' if p < 2: raise ValueError("""p should not be less than 2!""" ) elif p == 2: return True UpperCAmelCase__ : Tuple = 4 UpperCAmelCase__ : Tuple = (1 << p) - 1 for _ in range(p - 2 ): UpperCAmelCase__ : List[str] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
79
0
"""simple docstring""" def __A ( a_ :int) -> int: __a : str = abs(a_) __a : Any = 0 while n > 0: res += n % 10 n //= 10 return res def __A ( a_ :int) -> int: __a : int = abs(a_) return n if n < 10 else n % 10 + sum_of_digits(n // 10) def __A ( a_ :int) -> int: return sum(int(a_) for c in str(abs(a_))) def __A ( ) -> None: from collections.abc import Callable from timeit import timeit def benchmark_a_function(a_ :Callable , a_ :int) -> None: __a : Any = F"""{func.__name__}({value})""" __a : str = timeit(F"""__main__.{call}""" , setup='''import __main__''') print(F"""{call:56} = {func(a_)} -- {timing:.4f} seconds""") for value in (26_21_44, 11_25_89_99_06_84_26_24, 1_26_76_50_60_02_28_22_94_01_49_67_03_20_53_76): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(a_ , a_) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
52
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE__ : Any = { """configuration_mobilevit""": ["""MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MobileViTConfig""", """MobileViTOnnxConfig"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : List[str] = ["""MobileViTFeatureExtractor"""] SCREAMING_SNAKE_CASE__ : Union[str, Any] = ["""MobileViTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Dict = [ """MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MobileViTForImageClassification""", """MobileViTForSemanticSegmentation""", """MobileViTModel""", """MobileViTPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Any = [ """TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFMobileViTForImageClassification""", """TFMobileViTForSemanticSegmentation""", """TFMobileViTModel""", """TFMobileViTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mobilevit import MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileViTConfig, MobileViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilevit import MobileViTFeatureExtractor from .image_processing_mobilevit import MobileViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilevit import ( MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel, MobileViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilevit import ( TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileViTForImageClassification, TFMobileViTForSemanticSegmentation, TFMobileViTModel, TFMobileViTPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
79
0
from __future__ import annotations def a_ ( lowerCAmelCase_ : list[float] ): if len(lowerCAmelCase_ ) < 2: raise ValueError('Monogons and Digons are not polygons in the Euclidean space' ) if any(i <= 0 for i in nums ): raise ValueError('All values must be greater than 0' ) __lowerCAmelCase = nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
53
from __future__ import annotations SCREAMING_SNAKE_CASE__ : List[str] = 8.988e9 # units = N * m^s * C^-2 def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> dict[str, float]: '''simple docstring''' UpperCAmelCase__ : int = abs(chargea * chargea ) if (force, chargea, chargea, distance).count(0 ) != 1: raise ValueError("""One and only one argument must be 0""" ) if distance < 0: raise ValueError("""Distance cannot be negative""" ) if force == 0: UpperCAmelCase__ : int = COULOMBS_CONSTANT * charge_product / (distance**2) return {"force": force} elif chargea == 0: UpperCAmelCase__ : str = abs(__lowerCamelCase ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge1": chargea} elif chargea == 0: UpperCAmelCase__ : Union[str, Any] = abs(__lowerCamelCase ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge2": chargea} elif distance == 0: UpperCAmelCase__ : Optional[Any] = (COULOMBS_CONSTANT * charge_product / abs(__lowerCamelCase )) ** 0.5 return {"distance": distance} raise ValueError("""Exactly one argument must be 0""" ) if __name__ == "__main__": import doctest doctest.testmod()
79
0
import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def a__ ( lowercase__ , lowercase__=0.999 , lowercase__="cosine" , ): '''simple docstring''' if alpha_transform_type == "cosine": def alpha_bar_fn(lowercase__ ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(lowercase__ ): return math.exp(t * -12.0 ) else: raise ValueError(F'Unsupported alpha_tranform_type: {alpha_transform_type}' ) UpperCAmelCase_ =[] for i in range(lowercase__ ): UpperCAmelCase_ =i / num_diffusion_timesteps UpperCAmelCase_ =(i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(lowercase__ ) / alpha_bar_fn(lowercase__ ) , lowercase__ ) ) return torch.tensor(lowercase__ , dtype=torch.floataa ) class A ( __lowercase , __lowercase ): _snake_case =[e.name for e in KarrasDiffusionSchedulers] _snake_case =2 @register_to_config def __init__( self: Dict , _lowerCAmelCase: int = 1000 , _lowerCAmelCase: float = 0.0_00_85 , _lowerCAmelCase: float = 0.0_12 , _lowerCAmelCase: str = "linear" , _lowerCAmelCase: Optional[Union[np.ndarray, List[float]]] = None , _lowerCAmelCase: str = "epsilon" , _lowerCAmelCase: Optional[bool] = False , _lowerCAmelCase: Optional[bool] = False , _lowerCAmelCase: float = 1.0 , _lowerCAmelCase: str = "linspace" , _lowerCAmelCase: int = 0 , ) -> Tuple: '''simple docstring''' if trained_betas is not None: UpperCAmelCase_ =torch.tensor(_lowerCAmelCase , dtype=torch.floataa ) elif beta_schedule == "linear": UpperCAmelCase_ =torch.linspace(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. UpperCAmelCase_ =( torch.linspace(beta_start**0.5 , beta_end**0.5 , _lowerCAmelCase , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule UpperCAmelCase_ =betas_for_alpha_bar(_lowerCAmelCase , alpha_transform_type="cosine" ) elif beta_schedule == "exp": UpperCAmelCase_ =betas_for_alpha_bar(_lowerCAmelCase , alpha_transform_type="exp" ) else: raise NotImplementedError(F'{beta_schedule} does is not implemented for {self.__class__}' ) UpperCAmelCase_ =1.0 - self.betas UpperCAmelCase_ =torch.cumprod(self.alphas , dim=0 ) # set all values self.set_timesteps(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase_ =use_karras_sigmas def lowerCAmelCase__ ( self: List[Any] , _lowerCAmelCase: str , _lowerCAmelCase: Any=None ) -> List[Any]: '''simple docstring''' if schedule_timesteps is None: UpperCAmelCase_ =self.timesteps UpperCAmelCase_ =(schedule_timesteps == timestep).nonzero() # The sigma index that is taken for the **very** first `step` # is always the second index (or the last index if there is only 1) # This way we can ensure we don't accidentally skip a sigma in # case we start in the middle of the denoising schedule (e.g. for image-to-image) if len(self._index_counter ) == 0: UpperCAmelCase_ =1 if len(_lowerCAmelCase ) > 1 else 0 else: UpperCAmelCase_ =timestep.cpu().item() if torch.is_tensor(_lowerCAmelCase ) else timestep UpperCAmelCase_ =self._index_counter[timestep_int] return indices[pos].item() @property def lowerCAmelCase__ ( self: int ) -> List[str]: '''simple docstring''' if self.config.timestep_spacing in ["linspace", "trailing"]: return self.sigmas.max() return (self.sigmas.max() ** 2 + 1) ** 0.5 def lowerCAmelCase__ ( self: Optional[int] , _lowerCAmelCase: torch.FloatTensor , _lowerCAmelCase: Union[float, torch.FloatTensor] , ) -> torch.FloatTensor: '''simple docstring''' UpperCAmelCase_ =self.index_for_timestep(_lowerCAmelCase ) UpperCAmelCase_ =self.sigmas[step_index] UpperCAmelCase_ =sample / ((sigma**2 + 1) ** 0.5) return sample def lowerCAmelCase__ ( self: Union[str, Any] , _lowerCAmelCase: int , _lowerCAmelCase: Union[str, torch.device] = None , _lowerCAmelCase: Optional[int] = None , ) -> Dict: '''simple docstring''' UpperCAmelCase_ =num_inference_steps UpperCAmelCase_ =num_train_timesteps or self.config.num_train_timesteps # "linspace", "leading", "trailing" corresponds to annotation of Table 2. of https://arxiv.org/abs/2305.08891 if self.config.timestep_spacing == "linspace": UpperCAmelCase_ =np.linspace(0 , num_train_timesteps - 1 , _lowerCAmelCase , dtype=_lowerCAmelCase )[::-1].copy() elif self.config.timestep_spacing == "leading": UpperCAmelCase_ =num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 UpperCAmelCase_ =(np.arange(0 , _lowerCAmelCase ) * step_ratio).round()[::-1].copy().astype(_lowerCAmelCase ) timesteps += self.config.steps_offset elif self.config.timestep_spacing == "trailing": UpperCAmelCase_ =num_train_timesteps / self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 UpperCAmelCase_ =(np.arange(_lowerCAmelCase , 0 , -step_ratio )).round().copy().astype(_lowerCAmelCase ) timesteps -= 1 else: raise ValueError( F'{self.config.timestep_spacing} is not supported. Please make sure to choose one of \'linspace\', \'leading\' or \'trailing\'.' ) UpperCAmelCase_ =np.array(((1 - self.alphas_cumprod) / self.alphas_cumprod) ** 0.5 ) UpperCAmelCase_ =np.log(_lowerCAmelCase ) UpperCAmelCase_ =np.interp(_lowerCAmelCase , np.arange(0 , len(_lowerCAmelCase ) ) , _lowerCAmelCase ) if self.config.use_karras_sigmas: UpperCAmelCase_ =self._convert_to_karras(in_sigmas=_lowerCAmelCase , num_inference_steps=self.num_inference_steps ) UpperCAmelCase_ =np.array([self._sigma_to_t(_lowerCAmelCase , _lowerCAmelCase ) for sigma in sigmas] ) UpperCAmelCase_ =np.concatenate([sigmas, [0.0]] ).astype(np.floataa ) UpperCAmelCase_ =torch.from_numpy(_lowerCAmelCase ).to(device=_lowerCAmelCase ) UpperCAmelCase_ =torch.cat([sigmas[:1], sigmas[1:-1].repeat_interleave(2 ), sigmas[-1:]] ) UpperCAmelCase_ =torch.from_numpy(_lowerCAmelCase ) UpperCAmelCase_ =torch.cat([timesteps[:1], timesteps[1:].repeat_interleave(2 )] ) if str(_lowerCAmelCase ).startswith("mps" ): # mps does not support float64 UpperCAmelCase_ =timesteps.to(_lowerCAmelCase , dtype=torch.floataa ) else: UpperCAmelCase_ =timesteps.to(device=_lowerCAmelCase ) # empty dt and derivative UpperCAmelCase_ =None UpperCAmelCase_ =None # for exp beta schedules, such as the one for `pipeline_shap_e.py` # we need an index counter UpperCAmelCase_ =defaultdict(_lowerCAmelCase ) def lowerCAmelCase__ ( self: Union[str, Any] , _lowerCAmelCase: Optional[Any] , _lowerCAmelCase: Any ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ =np.log(_lowerCAmelCase ) # get distribution UpperCAmelCase_ =log_sigma - log_sigmas[:, np.newaxis] # get sigmas range UpperCAmelCase_ =np.cumsum((dists >= 0) , axis=0 ).argmax(axis=0 ).clip(max=log_sigmas.shape[0] - 2 ) UpperCAmelCase_ =low_idx + 1 UpperCAmelCase_ =log_sigmas[low_idx] UpperCAmelCase_ =log_sigmas[high_idx] # interpolate sigmas UpperCAmelCase_ =(low - log_sigma) / (low - high) UpperCAmelCase_ =np.clip(_lowerCAmelCase , 0 , 1 ) # transform interpolation to time range UpperCAmelCase_ =(1 - w) * low_idx + w * high_idx UpperCAmelCase_ =t.reshape(sigma.shape ) return t def lowerCAmelCase__ ( self: int , _lowerCAmelCase: torch.FloatTensor , _lowerCAmelCase: Tuple ) -> torch.FloatTensor: '''simple docstring''' UpperCAmelCase_ =in_sigmas[-1].item() UpperCAmelCase_ =in_sigmas[0].item() UpperCAmelCase_ =7.0 # 7.0 is the value used in the paper UpperCAmelCase_ =np.linspace(0 , 1 , _lowerCAmelCase ) UpperCAmelCase_ =sigma_min ** (1 / rho) UpperCAmelCase_ =sigma_max ** (1 / rho) UpperCAmelCase_ =(max_inv_rho + ramp * (min_inv_rho - max_inv_rho)) ** rho return sigmas @property def lowerCAmelCase__ ( self: Optional[Any] ) -> Optional[Any]: '''simple docstring''' return self.dt is None def lowerCAmelCase__ ( self: int , _lowerCAmelCase: Union[torch.FloatTensor, np.ndarray] , _lowerCAmelCase: Union[float, torch.FloatTensor] , _lowerCAmelCase: Union[torch.FloatTensor, np.ndarray] , _lowerCAmelCase: bool = True , ) -> Union[SchedulerOutput, Tuple]: '''simple docstring''' UpperCAmelCase_ =self.index_for_timestep(_lowerCAmelCase ) # advance index counter by 1 UpperCAmelCase_ =timestep.cpu().item() if torch.is_tensor(_lowerCAmelCase ) else timestep self._index_counter[timestep_int] += 1 if self.state_in_first_order: UpperCAmelCase_ =self.sigmas[step_index] UpperCAmelCase_ =self.sigmas[step_index + 1] else: # 2nd order / Heun's method UpperCAmelCase_ =self.sigmas[step_index - 1] UpperCAmelCase_ =self.sigmas[step_index] # currently only gamma=0 is supported. This usually works best anyways. # We can support gamma in the future but then need to scale the timestep before # passing it to the model which requires a change in API UpperCAmelCase_ =0 UpperCAmelCase_ =sigma * (gamma + 1) # Note: sigma_hat == sigma for now # 1. compute predicted original sample (x_0) from sigma-scaled predicted noise if self.config.prediction_type == "epsilon": UpperCAmelCase_ =sigma_hat if self.state_in_first_order else sigma_next UpperCAmelCase_ =sample - sigma_input * model_output elif self.config.prediction_type == "v_prediction": UpperCAmelCase_ =sigma_hat if self.state_in_first_order else sigma_next UpperCAmelCase_ =model_output * (-sigma_input / (sigma_input**2 + 1) ** 0.5) + ( sample / (sigma_input**2 + 1) ) elif self.config.prediction_type == "sample": UpperCAmelCase_ =model_output else: raise ValueError( F'prediction_type given as {self.config.prediction_type} must be one of `epsilon`, or `v_prediction`' ) if self.config.clip_sample: UpperCAmelCase_ =pred_original_sample.clamp( -self.config.clip_sample_range , self.config.clip_sample_range ) if self.state_in_first_order: # 2. Convert to an ODE derivative for 1st order UpperCAmelCase_ =(sample - pred_original_sample) / sigma_hat # 3. delta timestep UpperCAmelCase_ =sigma_next - sigma_hat # store for 2nd order step UpperCAmelCase_ =derivative UpperCAmelCase_ =dt UpperCAmelCase_ =sample else: # 2. 2nd order / Heun's method UpperCAmelCase_ =(sample - pred_original_sample) / sigma_next UpperCAmelCase_ =(self.prev_derivative + derivative) / 2 # 3. take prev timestep & sample UpperCAmelCase_ =self.dt UpperCAmelCase_ =self.sample # free dt and derivative # Note, this puts the scheduler in "first order mode" UpperCAmelCase_ =None UpperCAmelCase_ =None UpperCAmelCase_ =None UpperCAmelCase_ =sample + derivative * dt if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def lowerCAmelCase__ ( self: List[Any] , _lowerCAmelCase: torch.FloatTensor , _lowerCAmelCase: torch.FloatTensor , _lowerCAmelCase: torch.FloatTensor , ) -> torch.FloatTensor: '''simple docstring''' UpperCAmelCase_ =self.sigmas.to(device=original_samples.device , dtype=original_samples.dtype ) if original_samples.device.type == "mps" and torch.is_floating_point(_lowerCAmelCase ): # mps does not support float64 UpperCAmelCase_ =self.timesteps.to(original_samples.device , dtype=torch.floataa ) UpperCAmelCase_ =timesteps.to(original_samples.device , dtype=torch.floataa ) else: UpperCAmelCase_ =self.timesteps.to(original_samples.device ) UpperCAmelCase_ =timesteps.to(original_samples.device ) UpperCAmelCase_ =[self.index_for_timestep(_lowerCAmelCase , _lowerCAmelCase ) for t in timesteps] UpperCAmelCase_ =sigmas[step_indices].flatten() while len(sigma.shape ) < len(original_samples.shape ): UpperCAmelCase_ =sigma.unsqueeze(-1 ) UpperCAmelCase_ =original_samples + noise * sigma return noisy_samples def __len__( self: Dict ) -> Optional[Any]: '''simple docstring''' return self.config.num_train_timesteps
54
class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase ): # we need a list not a string, so do something to change the type UpperCAmelCase__ : Dict = arr.split(""",""" ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = [int(self.array[0] )] * len(self.array ) UpperCAmelCase__ : List[str] = [int(self.array[0] )] * len(self.array ) for i in range(1 , len(self.array ) ): UpperCAmelCase__ : Tuple = max( int(self.array[i] ) + sum_value[i - 1] , int(self.array[i] ) ) UpperCAmelCase__ : Union[str, Any] = max(sum_value[i] , rear[i - 1] ) return rear[len(self.array ) - 1] if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Tuple = input("""please input some numbers:""") SCREAMING_SNAKE_CASE__ : Dict = SubArray(whole_array) SCREAMING_SNAKE_CASE__ : Dict = array.solve_sub_array() print(("""the results is:""", re))
79
0
from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) SCREAMING_SNAKE_CASE :Union[str, Any] = logging.get_logger(__name__) # pylint: disable=invalid-name SCREAMING_SNAKE_CASE :str = '\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Img2ImgPipeline, KandinskyV22PriorPipeline\n >>> from diffusers.utils import load_image\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16\n ... )\n >>> pipe_prior.to("cuda")\n\n >>> prompt = "A red cartoon frog, 4k"\n >>> image_emb, zero_image_emb = pipe_prior(prompt, return_dict=False)\n\n >>> pipe = KandinskyV22Img2ImgPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-decoder", torch_dtype=torch.float16\n ... )\n >>> pipe.to("cuda")\n\n >>> init_image = load_image(\n ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main"\n ... "/kandinsky/frog.png"\n ... )\n\n >>> image = pipe(\n ... image=init_image,\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=100,\n ... strength=0.2,\n ... ).images\n\n >>> image[0].save("red_frog.png")\n ```\n' def UpperCAmelCase ( a_ , a_ , a_=8 ) -> List[str]: """simple docstring""" __A = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 __A = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor def UpperCAmelCase ( a_ , a_=5_1_2 , a_=5_1_2 ) -> List[Any]: """simple docstring""" __A = pil_image.resize((w, h) , resample=Image.BICUBIC , reducing_gap=1 ) __A = np.array(pil_image.convert("RGB" ) ) __A = arr.astype(np.floataa ) / 127.5 - 1 __A = np.transpose(a_ , [2, 0, 1] ) __A = torch.from_numpy(a_ ).unsqueeze(0 ) return image class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : Dict ,A : UNetaDConditionModel ,A : DDPMScheduler ,A : VQModel ,): super().__init__() self.register_modules( unet=A ,scheduler=A ,movq=A ,) __A = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCamelCase_ ( self : List[str] ,A : Optional[Any] ,A : Dict ,A : Union[str, Any] ): # get the original timestep using init_timestep __A = min(int(num_inference_steps * strength ) ,A ) __A = max(num_inference_steps - init_timestep ,0 ) __A = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def UpperCamelCase_ ( self : List[Any] ,A : Tuple ,A : Optional[int] ,A : List[str] ,A : Dict ,A : Union[str, Any] ,A : Tuple ,A : Tuple=None ): if not isinstance(A ,(torch.Tensor, PIL.Image.Image, list) ): raise ValueError( f'''`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(A )}''' ) __A = image.to(device=A ,dtype=A ) __A = batch_size * num_images_per_prompt if image.shape[1] == 4: __A = image else: if isinstance(A ,A ) and len(A ) != batch_size: raise ValueError( f'''You have passed a list of generators of length {len(A )}, but requested an effective batch''' f''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) elif isinstance(A ,A ): __A = [ self.movq.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(A ) ] __A = torch.cat(A ,dim=0 ) else: __A = self.movq.encode(A ).latent_dist.sample(A ) __A = self.movq.config.scaling_factor * init_latents __A = torch.cat([init_latents] ,dim=0 ) __A = init_latents.shape __A = randn_tensor(A ,generator=A ,device=A ,dtype=A ) # get latents __A = self.scheduler.add_noise(A ,A ,A ) __A = init_latents return latents def UpperCamelCase_ ( self : Optional[int] ,A : Union[str, Any]=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __A = torch.device(f'''cuda:{gpu_id}''' ) __A = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(A ,A ) def UpperCamelCase_ ( self : Dict ,A : List[Any]=0 ): if is_accelerate_available() and is_accelerate_version(">=" ,"0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) __A = torch.device(f'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to("cpu" ,silence_dtype_warnings=A ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __A = None for cpu_offloaded_model in [self.unet, self.movq]: __A , __A = cpu_offload_with_hook(A ,A ,prev_module_hook=A ) # We'll offload the last model manually. __A = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCamelCase_ ( self : Optional[Any] ): if not hasattr(self.unet ,"_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(A ,"_hf_hook" ) and hasattr(module._hf_hook ,"execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(A ) def __call__( self : Tuple ,A : Union[torch.FloatTensor, List[torch.FloatTensor]] ,A : Union[torch.FloatTensor, PIL.Image.Image, List[torch.FloatTensor], List[PIL.Image.Image]] ,A : Union[torch.FloatTensor, List[torch.FloatTensor]] ,A : int = 5_12 ,A : int = 5_12 ,A : int = 1_00 ,A : float = 4.0 ,A : float = 0.3 ,A : int = 1 ,A : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,A : Optional[str] = "pil" ,A : bool = True ,): __A = self._execution_device __A = guidance_scale > 1.0 if isinstance(A ,A ): __A = torch.cat(A ,dim=0 ) __A = image_embeds.shape[0] if isinstance(A ,A ): __A = torch.cat(A ,dim=0 ) if do_classifier_free_guidance: __A = image_embeds.repeat_interleave(A ,dim=0 ) __A = negative_image_embeds.repeat_interleave(A ,dim=0 ) __A = torch.cat([negative_image_embeds, image_embeds] ,dim=0 ).to(dtype=self.unet.dtype ,device=A ) if not isinstance(A ,A ): __A = [image] if not all(isinstance(A ,(PIL.Image.Image, torch.Tensor) ) for i in image ): raise ValueError( f'''Input is in incorrect format: {[type(A ) for i in image]}. Currently, we only support PIL image and pytorch tensor''' ) __A = torch.cat([prepare_image(A ,A ,A ) for i in image] ,dim=0 ) __A = image.to(dtype=image_embeds.dtype ,device=A ) __A = self.movq.encode(A )["latents"] __A = latents.repeat_interleave(A ,dim=0 ) self.scheduler.set_timesteps(A ,device=A ) __A , __A = self.get_timesteps(A ,A ,A ) __A = timesteps[:1].repeat(batch_size * num_images_per_prompt ) __A , __A = downscale_height_and_width(A ,A ,self.movq_scale_factor ) __A = self.prepare_latents( A ,A ,A ,A ,image_embeds.dtype ,A ,A ) for i, t in enumerate(self.progress_bar(A ) ): # expand the latents if we are doing classifier free guidance __A = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __A = {"image_embeds": image_embeds} __A = self.unet( sample=A ,timestep=A ,encoder_hidden_states=A ,added_cond_kwargs=A ,return_dict=A ,)[0] if do_classifier_free_guidance: __A , __A = noise_pred.split(latents.shape[1] ,dim=1 ) __A , __A = noise_pred.chunk(2 ) __A , __A = variance_pred.chunk(2 ) __A = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __A = torch.cat([noise_pred, variance_pred_text] ,dim=1 ) if not ( hasattr(self.scheduler.config ,"variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __A , __A = noise_pred.split(latents.shape[1] ,dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __A = self.scheduler.step( A ,A ,A ,generator=A ,)[0] # post-processing __A = self.movq.decode(A ,force_not_quantize=A )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(f'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: __A = image * 0.5 + 0.5 __A = image.clamp(0 ,1 ) __A = image.cpu().permute(0 ,2 ,3 ,1 ).float().numpy() if output_type == "pil": __A = self.numpy_to_pil(A ) if not return_dict: return (image,) return ImagePipelineOutput(images=A )
55
from ....configuration_utils import PretrainedConfig from ....utils import logging SCREAMING_SNAKE_CASE__ : List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Any = { """Visual-Attention-Network/van-base""": ( """https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json""" ), } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'van' def __init__( self , _lowerCAmelCase=224 , _lowerCAmelCase=3 , _lowerCAmelCase=[7, 3, 3, 3] , _lowerCAmelCase=[4, 2, 2, 2] , _lowerCAmelCase=[64, 128, 320, 512] , _lowerCAmelCase=[3, 3, 12, 3] , _lowerCAmelCase=[8, 8, 4, 4] , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=1e-6 , _lowerCAmelCase=1e-2 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) UpperCAmelCase__ : Tuple = image_size UpperCAmelCase__ : Optional[Any] = num_channels UpperCAmelCase__ : Optional[int] = patch_sizes UpperCAmelCase__ : int = strides UpperCAmelCase__ : Optional[int] = hidden_sizes UpperCAmelCase__ : str = depths UpperCAmelCase__ : Optional[Any] = mlp_ratios UpperCAmelCase__ : List[Any] = hidden_act UpperCAmelCase__ : Tuple = initializer_range UpperCAmelCase__ : Any = layer_norm_eps UpperCAmelCase__ : List[Any] = layer_scale_init_value UpperCAmelCase__ : int = drop_path_rate UpperCAmelCase__ : Dict = dropout_rate
79
0
'''simple docstring''' from PIL import Image def _a (lowercase__ : Image ) -> Image: """simple docstring""" __snake_case , __snake_case = image.size __snake_case = 0 __snake_case = image.load() for i in range(lowercase__ ): for j in range(lowercase__ ): __snake_case = pixels[j, i] mean += pixel mean //= width * height for j in range(lowercase__ ): for i in range(lowercase__ ): __snake_case = 2_5_5 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": _a : Dict = mean_threshold(Image.open("path_to_image").convert("L")) image.save("output_image_path")
56
import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Any: '''simple docstring''' UpperCAmelCase__ : List[str] = s.rsplit(__lowerCamelCase , __lowerCamelCase ) return new.join(__lowerCamelCase ) def _lowerCamelCase ( __lowerCamelCase ) -> str: '''simple docstring''' # encoder.embeddings are double copied in original FLAVA return sum(param.float().sum() if """encoder.embeddings""" not in key else 0 for key, param in state_dict.items() ) def _lowerCamelCase ( __lowerCamelCase ) -> int: '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = {} UpperCAmelCase__ : Union[str, Any] = ["""group_1""", """group_2""", """group_3""", """group_4"""] for key, value in state_dict.items(): for group_key in group_keys: if group_key in key: UpperCAmelCase__ : Optional[Any] = key.replace(F"{group_key}." , F"{group_key}.group." ) if "res_path" in key: UpperCAmelCase__ : Optional[int] = key.replace("""res_path.""" , """res_path.path.""" ) if key.endswith(""".w""" ): UpperCAmelCase__ : List[Any] = rreplace(__lowerCamelCase , """.w""" , """.weight""" , 1 ) if key.endswith(""".b""" ): UpperCAmelCase__ : Optional[int] = rreplace(__lowerCamelCase , """.b""" , """.bias""" , 1 ) UpperCAmelCase__ : Union[str, Any] = value.float() return upgrade @torch.no_grad() def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=None , __lowerCamelCase=True ) -> str: '''simple docstring''' from dall_e import Encoder UpperCAmelCase__ : Dict = Encoder() if os.path.exists(__lowerCamelCase ): UpperCAmelCase__ : Optional[Any] = torch.load(__lowerCamelCase ) else: UpperCAmelCase__ : Tuple = torch.hub.load_state_dict_from_url(__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase__ : Any = ckpt.state_dict() encoder.load_state_dict(__lowerCamelCase ) if config_path is not None: UpperCAmelCase__ : Dict = FlavaImageCodebookConfig.from_pretrained(__lowerCamelCase ) else: UpperCAmelCase__ : Optional[Any] = FlavaImageCodebookConfig() UpperCAmelCase__ : Optional[Any] = FlavaImageCodebook(__lowerCamelCase ).eval() UpperCAmelCase__ : str = encoder.state_dict() UpperCAmelCase__ : Optional[int] = upgrade_state_dict(__lowerCamelCase ) hf_model.load_state_dict(__lowerCamelCase ) UpperCAmelCase__ : List[str] = hf_model.state_dict() UpperCAmelCase__ : Tuple = count_parameters(__lowerCamelCase ) UpperCAmelCase__ : int = count_parameters(__lowerCamelCase ) assert torch.allclose(__lowerCamelCase , __lowerCamelCase , atol=1E-3 ) if save_checkpoint: hf_model.save_pretrained(__lowerCamelCase ) else: return hf_state_dict if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Any = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to flava checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") SCREAMING_SNAKE_CASE__ : int = parser.parse_args() convert_dalle_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
79
0
from __future__ import annotations from math import pi def snake_case (UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) -> dict[str, float]: if (inductance, frequency, reactance).count(0 ) != 1: raise ValueError('One and only one argument must be 0' ) if inductance < 0: raise ValueError('Inductance cannot be negative' ) if frequency < 0: raise ValueError('Frequency cannot be negative' ) if reactance < 0: raise ValueError('Inductive reactance cannot be negative' ) if inductance == 0: return {"inductance": reactance / (2 * pi * frequency)} elif frequency == 0: return {"frequency": reactance / (2 * pi * inductance)} elif reactance == 0: return {"reactance": 2 * pi * frequency * inductance} else: raise ValueError('Exactly one argument must be 0' ) if __name__ == "__main__": import doctest doctest.testmod()
57
def _lowerCamelCase ( __lowerCamelCase ) -> int: '''simple docstring''' return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def _lowerCamelCase ( __lowerCamelCase ) -> bool: '''simple docstring''' UpperCAmelCase__ : Any = 0 UpperCAmelCase__ : Union[str, Any] = number while duplicate > 0: UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = divmod(__lowerCamelCase , 10 ) fact_sum += factorial(__lowerCamelCase ) return fact_sum == number if __name__ == "__main__": print("""Program to check whether a number is a Krisnamurthy Number or not.""") SCREAMING_SNAKE_CASE__ : Optional[Any] = int(input("""Enter number: """).strip()) print( f'''{number} is {"" if krishnamurthy(number) else "not "}a Krishnamurthy Number.''' )
79
0
"""simple docstring""" from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging __lowerCAmelCase : Optional[int] = logging.get_logger(__name__) __lowerCAmelCase : Tuple = { '''EleutherAI/gpt-j-6B''': '''https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json''', # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = '''gptj''' _lowerCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self , _lowercase=5_0_4_0_0 , _lowercase=2_0_4_8 , _lowercase=4_0_9_6 , _lowercase=2_8 , _lowercase=1_6 , _lowercase=6_4 , _lowercase=None , _lowercase="gelu_new" , _lowercase=0.0 , _lowercase=0.0 , _lowercase=0.0 , _lowercase=1E-5 , _lowercase=0.02 , _lowercase=True , _lowercase=5_0_2_5_6 , _lowercase=5_0_2_5_6 , _lowercase=False , **_lowercase , ) -> Any: '''simple docstring''' snake_case_ : Dict = vocab_size snake_case_ : str = n_positions snake_case_ : List[str] = n_embd snake_case_ : List[str] = n_layer snake_case_ : str = n_head snake_case_ : List[Any] = n_inner snake_case_ : Union[str, Any] = rotary_dim snake_case_ : str = activation_function snake_case_ : Tuple = resid_pdrop snake_case_ : str = embd_pdrop snake_case_ : Optional[int] = attn_pdrop snake_case_ : Tuple = layer_norm_epsilon snake_case_ : Union[str, Any] = initializer_range snake_case_ : Dict = use_cache snake_case_ : Any = bos_token_id snake_case_ : Dict = eos_token_id super().__init__( bos_token_id=_lowercase , eos_token_id=_lowercase , tie_word_embeddings=_lowercase , **_lowercase ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , _lowercase , _lowercase = "default" , _lowercase = None , _lowercase = False , ) -> Optional[int]: '''simple docstring''' super().__init__(_lowercase , task=_lowercase , patching_specs=_lowercase , use_past=_lowercase ) if not getattr(self._config , """pad_token_id""" , _lowercase ): # TODO: how to do that better? snake_case_ : Dict = 0 @property def UpperCAmelCase__ ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' snake_case_ : Dict = OrderedDict({"""input_ids""": {0: """batch""", 1: """sequence"""}} ) if self.use_past: self.fill_with_past_key_values_(_lowercase , direction="""inputs""" ) snake_case_ : int = {0: """batch""", 1: """past_sequence + sequence"""} else: snake_case_ : Optional[int] = {0: """batch""", 1: """sequence"""} return common_inputs @property def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' return self._config.n_layer @property def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' return self._config.n_head def UpperCAmelCase__ ( self , _lowercase , _lowercase = -1 , _lowercase = -1 , _lowercase = False , _lowercase = None , ) -> Mapping[str, Any]: '''simple docstring''' snake_case_ : List[str] = super(_lowercase , self ).generate_dummy_inputs( _lowercase , batch_size=_lowercase , seq_length=_lowercase , is_pair=_lowercase , framework=_lowercase ) # We need to order the input in the way they appears in the forward() snake_case_ : Optional[Any] = OrderedDict({"""input_ids""": common_inputs["""input_ids"""]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch snake_case_ , snake_case_ : Dict = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values snake_case_ : List[Any] = seqlen + 2 snake_case_ : Dict = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) snake_case_ : str = [ (torch.zeros(_lowercase ), torch.zeros(_lowercase )) for _ in range(self.num_layers ) ] snake_case_ : Dict = common_inputs["""attention_mask"""] if self.use_past: snake_case_ : List[Any] = ordered_inputs["""attention_mask"""].dtype snake_case_ : Union[str, Any] = torch.cat( [ordered_inputs["""attention_mask"""], torch.ones(_lowercase , _lowercase , dtype=_lowercase )] , dim=1 ) return ordered_inputs @property def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' return 1_3
58
def _lowerCamelCase ( __lowerCamelCase = 100_0000 ) -> int: '''simple docstring''' UpperCAmelCase__ : Tuple = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , __lowerCamelCase ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
79
0
import argparse from collections import defaultdict def lowerCAmelCase_ ( __a , __a , __a , __a , __a ) -> Optional[int]: """simple docstring""" lowerCamelCase__: str =F"""{file}_{class_name}_{test_name}""" done_test[_id] += 1 with open(__a , "r" ) as f: lowerCamelCase__: Optional[int] =f.readlines() lowerCamelCase__: List[str] =F"""class {class_name}(""" lowerCamelCase__: Any =F"""{4 * " "}def {test_name}(""" lowerCamelCase__: Dict =F"""{8 * " "}{correct_line.split()[0]}""" lowerCamelCase__: Any =F"""{16 * " "}{correct_line.split()[0]}""" lowerCamelCase__: Tuple =False lowerCamelCase__: List[str] =False lowerCamelCase__: List[Any] =False lowerCamelCase__: int =False lowerCamelCase__: Optional[int] =0 lowerCamelCase__: Optional[Any] =0 lowerCamelCase__: List[str] =[] for line in lines: if line.startswith(__a ): lowerCamelCase__: List[str] =True elif in_class and line.startswith(__a ): lowerCamelCase__: Dict =True elif in_class and in_func and (line.startswith(__a ) or line.startswith(__a )): lowerCamelCase__: int =len(line.split(correct_line.split()[0] )[0] ) count += 1 if count == done_test[_id]: lowerCamelCase__: Dict =True if in_class and in_func and in_line: if ")" not in line: continue else: lowerCamelCase__: Union[str, Any] =True if in_class and in_func and in_line and insert_line: new_lines.append(F"""{spaces * " "}{correct_line}""" ) lowerCamelCase__: Any =False else: new_lines.append(__a ) with open(__a , "w" ) as f: for line in new_lines: f.write(__a ) def lowerCAmelCase_ ( __a , __a=None ) -> int: """simple docstring""" if fail is not None: with open(__a , "r" ) as f: lowerCamelCase__: str ={l.strip() for l in f.readlines()} else: lowerCamelCase__: List[str] =None with open(__a , "r" ) as f: lowerCamelCase__: Any =f.readlines() lowerCamelCase__: str =defaultdict(__a ) for line in correct_lines: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__: Optional[Any] =line.split(";" ) if test_failures is None or "::".join([file, class_name, test_name] ) in test_failures: overwrite_file(__a , __a , __a , __a , __a ) if __name__ == "__main__": __A = argparse.ArgumentParser() parser.add_argument("--correct_filename", help="filename of tests with expected result") parser.add_argument("--fail_filename", help="filename of test failures", type=str, default=None) __A = parser.parse_args() main(args.correct_filename, args.fail_filename)
59
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Tuple = { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json""", # See all REALM models at https://huggingface.co/models?filter=realm } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'realm' def __init__( self , _lowerCAmelCase=30522 , _lowerCAmelCase=768 , _lowerCAmelCase=128 , _lowerCAmelCase=12 , _lowerCAmelCase=12 , _lowerCAmelCase=8 , _lowerCAmelCase=3072 , _lowerCAmelCase="gelu_new" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=512 , _lowerCAmelCase=2 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=1e-12 , _lowerCAmelCase=256 , _lowerCAmelCase=10 , _lowerCAmelCase=1e-3 , _lowerCAmelCase=5 , _lowerCAmelCase=320 , _lowerCAmelCase=13353718 , _lowerCAmelCase=5000 , _lowerCAmelCase=1 , _lowerCAmelCase=0 , _lowerCAmelCase=2 , **_lowerCAmelCase , ): super().__init__(pad_token_id=_lowerCAmelCase , bos_token_id=_lowerCAmelCase , eos_token_id=_lowerCAmelCase , **_lowerCAmelCase ) # Common config UpperCAmelCase__ : List[Any] = vocab_size UpperCAmelCase__ : Dict = max_position_embeddings UpperCAmelCase__ : Any = hidden_size UpperCAmelCase__ : str = retriever_proj_size UpperCAmelCase__ : Tuple = num_hidden_layers UpperCAmelCase__ : List[str] = num_attention_heads UpperCAmelCase__ : List[Any] = num_candidates UpperCAmelCase__ : str = intermediate_size UpperCAmelCase__ : str = hidden_act UpperCAmelCase__ : Optional[Any] = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : Union[str, Any] = initializer_range UpperCAmelCase__ : Any = type_vocab_size UpperCAmelCase__ : Optional[Any] = layer_norm_eps # Reader config UpperCAmelCase__ : str = span_hidden_size UpperCAmelCase__ : Union[str, Any] = max_span_width UpperCAmelCase__ : List[str] = reader_layer_norm_eps UpperCAmelCase__ : Dict = reader_beam_size UpperCAmelCase__ : Union[str, Any] = reader_seq_len # Retrieval config UpperCAmelCase__ : List[Any] = num_block_records UpperCAmelCase__ : List[Any] = searcher_beam_size
79
0
lowerCAmelCase_ = range(2, 2_0 + 1) lowerCAmelCase_ = [1_0**k for k in range(ks[-1] + 1)] lowerCAmelCase_ = {} def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Any: """simple docstring""" snake_case_ : List[Any] = sum(a_i[j] for j in range(_UpperCamelCase , len(_UpperCamelCase ) ) ) snake_case_ : Optional[Any] = sum(a_i[j] * base[j] for j in range(min(len(_UpperCamelCase ) , _UpperCamelCase ) ) ) snake_case_ , snake_case_ : Any = 0, 0 snake_case_ : List[str] = n - i snake_case_ : Any = memo.get(_UpperCamelCase ) if sub_memo is not None: snake_case_ : Any = sub_memo.get(_UpperCamelCase ) if jumps is not None and len(_UpperCamelCase ) > 0: # find and make the largest jump without going over snake_case_ : int = -1 for _k in range(len(_UpperCamelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: snake_case_ : List[Any] = _k break if max_jump >= 0: snake_case_ , snake_case_ , snake_case_ : Union[str, Any] = jumps[max_jump] # since the difference between jumps is cached, add c snake_case_ : List[str] = diff + c for j in range(min(_UpperCamelCase , len(_UpperCamelCase ) ) ): snake_case_ , snake_case_ : Optional[Any] = divmod(_UpperCamelCase , 10 ) if new_c > 0: add(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) else: snake_case_ : List[Any] = [] else: snake_case_ : Optional[Any] = {c: []} snake_case_ : Union[str, Any] = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps snake_case_ , snake_case_ : List[str] = next_term(_UpperCamelCase , k - 1 , i + dn , _UpperCamelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead snake_case_ , snake_case_ : str = compute(_UpperCamelCase , _UpperCamelCase , i + dn , _UpperCamelCase ) diff += _diff dn += terms_jumped snake_case_ : str = sub_memo[c] # keep jumps sorted by # of terms skipped snake_case_ : List[Any] = 0 while j < len(_UpperCamelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(_UpperCamelCase , (diff, dn, k) ) return (diff, dn) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Optional[Any]: """simple docstring""" if i >= n: return 0, i if k > len(_UpperCamelCase ): a_i.extend([0 for _ in range(k - len(_UpperCamelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) snake_case_ : Union[str, Any] = i snake_case_ , snake_case_ , snake_case_ : List[Any] = 0, 0, 0 for j in range(len(_UpperCamelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 snake_case_ : Union[str, Any] = ds_c + ds_b diff += addend snake_case_ : int = 0 for j in range(_UpperCamelCase ): snake_case_ : List[str] = a_i[j] + addend snake_case_ , snake_case_ : Optional[Any] = divmod(_UpperCamelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) return diff, i - start_i def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Dict: """simple docstring""" for j in range(_UpperCamelCase , len(_UpperCamelCase ) ): snake_case_ : Any = digits[j] + addend if s >= 10: snake_case_ , snake_case_ : Tuple = divmod(_UpperCamelCase , 10 ) snake_case_ : Optional[int] = addend // 10 + quotient else: snake_case_ : Optional[int] = s snake_case_ : Optional[Any] = addend // 10 if addend == 0: break while addend > 0: snake_case_ , snake_case_ : int = divmod(_UpperCamelCase , 10 ) digits.append(_UpperCamelCase ) def lowerCamelCase_ ( _UpperCamelCase = 10**15 ) -> int: """simple docstring""" snake_case_ : Tuple = [1] snake_case_ : List[Any] = 1 snake_case_ : Union[str, Any] = 0 while True: snake_case_ , snake_case_ : List[Any] = next_term(_UpperCamelCase , 20 , i + dn , _UpperCamelCase ) dn += terms_jumped if dn == n - i: break snake_case_ : Any = 0 for j in range(len(_UpperCamelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'''{solution() = }''')
60
import gc import unittest from parameterized import parameterized from diffusers import FlaxUNetaDConditionModel from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp @slow @require_flax class UpperCAmelCase_ ( unittest.TestCase ): def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase ): return f"gaussian_noise_s={seed}_shape={'_'.join([str(_lowerCAmelCase ) for s in shape] )}.npy" def __UpperCAmelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def __UpperCAmelCase ( self , _lowerCAmelCase=0 , _lowerCAmelCase=(4, 4, 64, 64) , _lowerCAmelCase=False ): UpperCAmelCase__ : Optional[Any] = jnp.bfloataa if fpaa else jnp.floataa UpperCAmelCase__ : Union[str, Any] = jnp.array(load_hf_numpy(self.get_file_format(_lowerCAmelCase , _lowerCAmelCase ) ) , dtype=_lowerCAmelCase ) return image def __UpperCAmelCase ( self , _lowerCAmelCase=False , _lowerCAmelCase="CompVis/stable-diffusion-v1-4" ): UpperCAmelCase__ : int = jnp.bfloataa if fpaa else jnp.floataa UpperCAmelCase__ : Optional[Any] = """bf16""" if fpaa else None UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = FlaxUNetaDConditionModel.from_pretrained( _lowerCAmelCase , subfolder="""unet""" , dtype=_lowerCAmelCase , revision=_lowerCAmelCase ) return model, params def __UpperCAmelCase ( self , _lowerCAmelCase=0 , _lowerCAmelCase=(4, 77, 768) , _lowerCAmelCase=False ): UpperCAmelCase__ : Optional[int] = jnp.bfloataa if fpaa else jnp.floataa UpperCAmelCase__ : Optional[int] = jnp.array(load_hf_numpy(self.get_file_format(_lowerCAmelCase , _lowerCAmelCase ) ) , dtype=_lowerCAmelCase ) return hidden_states @parameterized.expand( [ # fmt: off [83, 4, [-0.2_3_2_3, -0.1_3_0_4, 0.0_8_1_3, -0.3_0_9_3, -0.0_9_1_9, -0.1_5_7_1, -0.1_1_2_5, -0.5_8_0_6]], [17, 0.5_5, [-0.0_8_3_1, -0.2_4_4_3, 0.0_9_0_1, -0.0_9_1_9, 0.3_3_9_6, 0.0_1_0_3, -0.3_7_4_3, 0.0_7_0_1]], [8, 0.8_9, [-0.4_8_6_3, 0.0_8_5_9, 0.0_8_7_5, -0.1_6_5_8, 0.9_1_9_9, -0.0_1_1_4, 0.4_8_3_9, 0.4_6_3_9]], [3, 1000, [-0.5_6_4_9, 0.2_4_0_2, -0.5_5_1_8, 0.1_2_4_8, 1.1_3_2_8, -0.2_4_4_3, -0.0_3_2_5, -1.0_0_7_8]], # fmt: on ] ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.get_unet_model(model_id="""CompVis/stable-diffusion-v1-4""" , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = self.get_latents(_lowerCAmelCase , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Dict = self.get_encoder_hidden_states(_lowerCAmelCase , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = model.apply( {"""params""": params} , _lowerCAmelCase , jnp.array(_lowerCAmelCase , dtype=jnp.intaa ) , encoder_hidden_states=_lowerCAmelCase , ).sample assert sample.shape == latents.shape UpperCAmelCase__ : Dict = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) UpperCAmelCase__ : List[Any] = jnp.array(_lowerCAmelCase , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, in the same hardware assert jnp.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-2 ) @parameterized.expand( [ # fmt: off [83, 4, [0.1_5_1_4, 0.0_8_0_7, 0.1_6_2_4, 0.1_0_1_6, -0.1_8_9_6, 0.0_2_6_3, 0.0_6_7_7, 0.2_3_1_0]], [17, 0.5_5, [0.1_1_6_4, -0.0_2_1_6, 0.0_1_7_0, 0.1_5_8_9, -0.3_1_2_0, 0.1_0_0_5, -0.0_5_8_1, -0.1_4_5_8]], [8, 0.8_9, [-0.1_7_5_8, -0.0_1_6_9, 0.1_0_0_4, -0.1_4_1_1, 0.1_3_1_2, 0.1_1_0_3, -0.1_9_9_6, 0.2_1_3_9]], [3, 1000, [0.1_2_1_4, 0.0_3_5_2, -0.0_7_3_1, -0.1_5_6_2, -0.0_9_9_4, -0.0_9_0_6, -0.2_3_4_0, -0.0_5_3_9]], # fmt: on ] ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.get_unet_model(model_id="""stabilityai/stable-diffusion-2""" , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = self.get_latents(_lowerCAmelCase , shape=(4, 4, 96, 96) , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Any = self.get_encoder_hidden_states(_lowerCAmelCase , shape=(4, 77, 1024) , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Dict = model.apply( {"""params""": params} , _lowerCAmelCase , jnp.array(_lowerCAmelCase , dtype=jnp.intaa ) , encoder_hidden_states=_lowerCAmelCase , ).sample assert sample.shape == latents.shape UpperCAmelCase__ : Any = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) UpperCAmelCase__ : Any = jnp.array(_lowerCAmelCase , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, on the same hardware assert jnp.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-2 )
79
0
from decimal import Decimal, getcontext from math import ceil, factorial def _A ( lowerCAmelCase_ : int ): """simple docstring""" if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("Undefined for non-integers" ) elif precision < 1: raise ValueError("Undefined for non-natural numbers" ) lowerCAmelCase__ = precision lowerCAmelCase__ = ceil(precision / 14 ) lowerCAmelCase__ = 42_6880 * Decimal(1_0005 ).sqrt() lowerCAmelCase__ = 1 lowerCAmelCase__ = 1359_1409 lowerCAmelCase__ = Decimal(lowerCAmelCase_ ) for k in range(1 , lowerCAmelCase_ ): lowerCAmelCase__ = factorial(6 * k ) // (factorial(3 * k ) * factorial(lowerCAmelCase_ ) ** 3) linear_term += 5_4514_0134 exponential_term *= -26_2537_4126_4076_8000 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": UpperCamelCase = 50 print(F"""The first {n} digits of pi is: {pi(n)}""")
61
import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class UpperCAmelCase_ ( unittest.TestCase ): def __init__( self , _lowerCAmelCase , _lowerCAmelCase=7 , _lowerCAmelCase=3 , _lowerCAmelCase=18 , _lowerCAmelCase=30 , _lowerCAmelCase=400 , _lowerCAmelCase=True , _lowerCAmelCase=None , _lowerCAmelCase=True , ): UpperCAmelCase__ : List[str] = size if size is not None else {"""height""": 18, """width""": 18} UpperCAmelCase__ : Union[str, Any] = parent UpperCAmelCase__ : int = batch_size UpperCAmelCase__ : Tuple = num_channels UpperCAmelCase__ : Dict = image_size UpperCAmelCase__ : List[Any] = min_resolution UpperCAmelCase__ : str = max_resolution UpperCAmelCase__ : Union[str, Any] = do_resize UpperCAmelCase__ : Tuple = size UpperCAmelCase__ : int = do_normalize def __UpperCAmelCase ( self ): return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.8_8_6_6_4_4_3_6_3_4_0_3_3_2_0_3, 0.6_6_1_8_8_2_9_3_6_9_5_4_4_9_8_3, 0.3_8_9_1_7_4_6_4_0_1_7_8_6_8_0_4], [-0.6_0_4_2_5_5_9_1_4_6_8_8_1_1_0_4, -0.0_2_2_9_5_0_0_8_8_6_0_5_2_8_4_6_9, 0.5_4_2_3_7_9_7_3_6_9_0_0_3_2_9_6], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class UpperCAmelCase_ ( __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = ImageGPTImageProcessor if is_vision_available() else None def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = ImageGPTImageProcessingTester(self ) @property def __UpperCAmelCase ( self ): return self.image_processor_tester.prepare_image_processor_dict() def __UpperCAmelCase ( self ): UpperCAmelCase__ : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """clusters""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """do_normalize""" ) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 18} ) UpperCAmelCase__ : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) UpperCAmelCase__ : Optional[int] = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase , obj[key] ) ) else: self.assertEqual(obj[key] , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase__ : Union[str, Any] = os.path.join(_lowerCAmelCase , """image_processor.json""" ) image_processor_first.to_json_file(_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = self.image_processing_class.from_json_file(_lowerCAmelCase ).to_dict() UpperCAmelCase__ : Dict = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : str = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = self.image_processing_class.from_pretrained(_lowerCAmelCase ).to_dict() UpperCAmelCase__ : Tuple = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , _lowerCAmelCase ) @unittest.skip("""ImageGPT requires clusters at initialization""" ) def __UpperCAmelCase ( self ): pass def _lowerCamelCase ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ : Any = load_dataset("""hf-internal-testing/fixtures_image_utils""" , split="""test""" ) UpperCAmelCase__ : Dict = Image.open(dataset[4]["""file"""] ) UpperCAmelCase__ : Optional[Any] = Image.open(dataset[5]["""file"""] ) UpperCAmelCase__ : List[Any] = [imagea, imagea] return images @require_vision @require_torch class UpperCAmelCase_ ( unittest.TestCase ): @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = ImageGPTImageProcessor.from_pretrained("""openai/imagegpt-small""" ) UpperCAmelCase__ : int = prepare_images() # test non-batched UpperCAmelCase__ : List[str] = image_processing(images[0] , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 1024) ) UpperCAmelCase__ : List[Any] = [306, 191, 191] self.assertEqual(encoding.input_ids[0, :3].tolist() , _lowerCAmelCase ) # test batched UpperCAmelCase__ : List[str] = image_processing(_lowerCAmelCase , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 1024) ) UpperCAmelCase__ : Any = [303, 13, 13] self.assertEqual(encoding.input_ids[1, -3:].tolist() , _lowerCAmelCase )
79
0
from __future__ import annotations from typing import Any class SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Optional[Any] , UpperCAmelCase_ : int , UpperCAmelCase_ : int , UpperCAmelCase_ : float = 0 ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : int = row, column SCREAMING_SNAKE_CASE : Optional[Any] = [[default_value for c in range(UpperCAmelCase_ )] for r in range(UpperCAmelCase_ )] def __str__( self : Optional[int] ): SCREAMING_SNAKE_CASE : Dict = f'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier SCREAMING_SNAKE_CASE : Dict = 0 for row_vector in self.array: for obj in row_vector: SCREAMING_SNAKE_CASE : Optional[Any] = max(UpperCAmelCase_ , len(str(UpperCAmelCase_ ) ) ) SCREAMING_SNAKE_CASE : Optional[int] = f'''%{max_element_length}s''' # Make string and return def single_line(UpperCAmelCase_ : list[float] ) -> str: nonlocal string_format_identifier SCREAMING_SNAKE_CASE : Optional[int] = "[" line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(UpperCAmelCase_ ) for row_vector in self.array ) return s def __repr__( self : Dict ): return str(self ) def _A ( self : Optional[int] , UpperCAmelCase_ : tuple[int, int] ): if not (isinstance(UpperCAmelCase_ , (list, tuple) ) and len(UpperCAmelCase_ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Any , UpperCAmelCase_ : tuple[int, int] ): assert self.validate_indicies(UpperCAmelCase_ ) return self.array[loc[0]][loc[1]] def __setitem__( self : int , UpperCAmelCase_ : tuple[int, int] , UpperCAmelCase_ : float ): assert self.validate_indicies(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Union[str, Any] = value def __add__( self : Optional[int] , UpperCAmelCase_ : Matrix ): assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) assert self.row == another.row and self.column == another.column # Add SCREAMING_SNAKE_CASE : str = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): SCREAMING_SNAKE_CASE : Any = self[r, c] + another[r, c] return result def __neg__( self : Union[str, Any] ): SCREAMING_SNAKE_CASE : Union[str, Any] = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): SCREAMING_SNAKE_CASE : str = -self[r, c] return result def __sub__( self : Optional[Any] , UpperCAmelCase_ : Matrix ): return self + (-another) def __mul__( self : Dict , UpperCAmelCase_ : int | float | Matrix ): if isinstance(UpperCAmelCase_ , (int, float) ): # Scalar multiplication SCREAMING_SNAKE_CASE : Any = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): SCREAMING_SNAKE_CASE : str = self[r, c] * another return result elif isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): # Matrix multiplication assert self.column == another.row SCREAMING_SNAKE_CASE : Any = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: SCREAMING_SNAKE_CASE : List[str] = f'''Unsupported type given for another ({type(UpperCAmelCase_ )})''' raise TypeError(UpperCAmelCase_ ) def _A ( self : int ): SCREAMING_SNAKE_CASE : List[str] = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): SCREAMING_SNAKE_CASE : List[str] = self[r, c] return result def _A ( self : Union[str, Any] , UpperCAmelCase_ : Matrix , UpperCAmelCase_ : Matrix ): assert isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate SCREAMING_SNAKE_CASE : Tuple = v.transpose() SCREAMING_SNAKE_CASE : int = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def lowerCamelCase__ ( ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = Matrix(3 , 3 , 0 ) for i in range(3 ): SCREAMING_SNAKE_CASE : str = 1 print(F'''a^(-1) is {ainv}''' ) # u, v SCREAMING_SNAKE_CASE : Optional[int] = Matrix(3 , 1 , 0 ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[int] = 1, 2, -3 SCREAMING_SNAKE_CASE : Tuple = Matrix(3 , 1 , 0 ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : int = 4, -2, 5 print(F'''u is {u}''' ) print(F'''v is {v}''' ) print(F'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(F'''(a + uv^T)^(-1) is {ainv.sherman_morrison(lowercase , lowercase )}''' ) def lowerCamelCase__ ( ): """simple docstring""" import doctest doctest.testmod() testa()
62
import os import unittest from transformers import MobileBertTokenizer, MobileBertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class UpperCAmelCase_ ( __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = MobileBertTokenizer __lowerCamelCase = MobileBertTokenizerFast __lowerCamelCase = True __lowerCamelCase = True __lowerCamelCase = filter_non_english __lowerCamelCase = 'google/mobilebert-uncased' def __UpperCAmelCase ( self ): super().setUp() UpperCAmelCase__ : Dict = [ """[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] UpperCAmelCase__ : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) UpperCAmelCase__ : List[str] = [ (tokenizer_def[0], self.pre_trained_model_path, tokenizer_def[2]) # else the 'google/' prefix is stripped for tokenizer_def in self.tokenizers_list ] def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : Tuple = """UNwant\u00E9d,running""" UpperCAmelCase__ : Union[str, Any] = """unwanted, running""" return input_text, output_text def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = self.tokenizer_class(self.vocab_file ) UpperCAmelCase__ : Tuple = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(_lowerCAmelCase , ["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [9, 6, 7, 12, 10, 11] ) def __UpperCAmelCase ( self ): if not self.test_rust_tokenizer: return UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : Dict = self.get_rust_tokenizer() UpperCAmelCase__ : List[str] = """UNwant\u00E9d,running""" UpperCAmelCase__ : Optional[int] = tokenizer.tokenize(_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = rust_tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = rust_tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.get_rust_tokenizer() UpperCAmelCase__ : Any = tokenizer.encode(_lowerCAmelCase ) UpperCAmelCase__ : str = rust_tokenizer.encode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) # With lower casing UpperCAmelCase__ : Tuple = self.get_tokenizer(do_lower_case=_lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.get_rust_tokenizer(do_lower_case=_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = """UNwant\u00E9d,running""" UpperCAmelCase__ : int = tokenizer.tokenize(_lowerCAmelCase ) UpperCAmelCase__ : Any = rust_tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = rust_tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = self.get_rust_tokenizer() UpperCAmelCase__ : List[str] = tokenizer.encode(_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = rust_tokenizer.encode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = BasicTokenizer() self.assertListEqual(tokenizer.tokenize("""ah\u535A\u63A8zz""" ) , ["""ah""", """\u535A""", """\u63A8""", """zz"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = BasicTokenizer(do_lower_case=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Union[str, Any] = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hällo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""h\u00E9llo"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[str] = BasicTokenizer(do_lower_case=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[str] = BasicTokenizer(do_lower_case=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : int = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HäLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HaLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = BasicTokenizer(do_lower_case=_lowerCAmelCase , never_split=["""[UNK]"""] ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? [UNK]""" ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?""", """[UNK]"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : int = ["""[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing"""] UpperCAmelCase__ : List[str] = {} for i, token in enumerate(_lowerCAmelCase ): UpperCAmelCase__ : Optional[Any] = i UpperCAmelCase__ : str = WordpieceTokenizer(vocab=_lowerCAmelCase , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""unwanted running""" ) , ["""un""", """##want""", """##ed""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.tokenize("""unwantedX running""" ) , ["""[UNK]""", """runn""", """##ing"""] ) def __UpperCAmelCase ( self ): self.assertTrue(_is_whitespace(""" """ ) ) self.assertTrue(_is_whitespace("""\t""" ) ) self.assertTrue(_is_whitespace("""\r""" ) ) self.assertTrue(_is_whitespace("""\n""" ) ) self.assertTrue(_is_whitespace("""\u00A0""" ) ) self.assertFalse(_is_whitespace("""A""" ) ) self.assertFalse(_is_whitespace("""-""" ) ) def __UpperCAmelCase ( self ): self.assertTrue(_is_control("""\u0005""" ) ) self.assertFalse(_is_control("""A""" ) ) self.assertFalse(_is_control(""" """ ) ) self.assertFalse(_is_control("""\t""" ) ) self.assertFalse(_is_control("""\r""" ) ) def __UpperCAmelCase ( self ): self.assertTrue(_is_punctuation("""-""" ) ) self.assertTrue(_is_punctuation("""$""" ) ) self.assertTrue(_is_punctuation("""`""" ) ) self.assertTrue(_is_punctuation(""".""" ) ) self.assertFalse(_is_punctuation("""A""" ) ) self.assertFalse(_is_punctuation(""" """ ) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = self.get_tokenizer() UpperCAmelCase__ : Union[str, Any] = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(_lowerCAmelCase ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) self.assertListEqual( [rust_tokenizer.tokenize(_lowerCAmelCase ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = self.tokenizer_class.from_pretrained("""google/mobilebert-uncased""" ) UpperCAmelCase__ : List[Any] = tokenizer.encode("""sequence builders""" , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer.encode("""multi-sequence build""" , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase ) UpperCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase , _lowerCAmelCase ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def __UpperCAmelCase ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): UpperCAmelCase__ : Any = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = f"A, naïve {tokenizer_r.mask_token} AllenNLP sentence." UpperCAmelCase__ : Optional[Any] = tokenizer_r.encode_plus( _lowerCAmelCase , return_attention_mask=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , ) UpperCAmelCase__ : Any = tokenizer_r.do_lower_case if hasattr(_lowerCAmelCase , """do_lower_case""" ) else False UpperCAmelCase__ : Optional[int] = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """A"""), ((1, 2), ""","""), ((3, 5), """na"""), ((5, 6), """##ï"""), ((6, 8), """##ve"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """Allen"""), ((21, 23), """##NL"""), ((23, 24), """##P"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """a"""), ((1, 2), ""","""), ((3, 8), """naive"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """allen"""), ((21, 23), """##nl"""), ((23, 24), """##p"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["""input_ids"""] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["""offset_mapping"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = ["""的""", """人""", """有"""] UpperCAmelCase__ : Tuple = """""".join(_lowerCAmelCase ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): UpperCAmelCase__ : Tuple = True UpperCAmelCase__ : Optional[Any] = self.tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = tokenizer_p.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer_r.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : Any = tokenizer_r.convert_ids_to_tokens(_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = tokenizer_p.convert_ids_to_tokens(_lowerCAmelCase ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : List[Any] = False UpperCAmelCase__ : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = tokenizer_r.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer_p.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = tokenizer_r.convert_ids_to_tokens(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = tokenizer_p.convert_ids_to_tokens(_lowerCAmelCase ) # it is expected that only the first Chinese character is not preceded by "##". UpperCAmelCase__ : List[str] = [ f"##{token}" if idx != 0 else token for idx, token in enumerate(_lowerCAmelCase ) ] self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase )
79
0
from math import asin, atan, cos, radians, sin, sqrt, tan a : Tuple = 6_37_81_37.0 a : Any = 6_35_67_52.31_42_45 a : Tuple = 6_378_137 def lowerCamelCase__ ( __lowerCamelCase : float , __lowerCamelCase : float , __lowerCamelCase : float , __lowerCamelCase : float ): __UpperCAmelCase : Union[str, Any] = (AXIS_A - AXIS_B) / AXIS_A __UpperCAmelCase : List[str] = atan((1 - flattening) * tan(radians(__lowerCamelCase ) ) ) __UpperCAmelCase : Optional[Any] = atan((1 - flattening) * tan(radians(__lowerCamelCase ) ) ) __UpperCAmelCase : Optional[Any] = radians(__lowerCamelCase ) __UpperCAmelCase : Optional[Any] = radians(__lowerCamelCase ) # Equation __UpperCAmelCase : Any = sin((phi_a - phi_a) / 2 ) __UpperCAmelCase : Any = sin((lambda_a - lambda_a) / 2 ) # Square both values sin_sq_phi *= sin_sq_phi sin_sq_lambda *= sin_sq_lambda __UpperCAmelCase : List[Any] = sqrt(sin_sq_phi + (cos(__lowerCamelCase ) * cos(__lowerCamelCase ) * sin_sq_lambda) ) return 2 * RADIUS * asin(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
63
import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg""" UpperCAmelCase__ : int = Image.open(requests.get(__lowerCamelCase , stream=__lowerCamelCase ).raw ).convert("""RGB""" ) UpperCAmelCase__ : Any = transforms.Compose( [ transforms.Resize((image_size, image_size) , interpolation=InterpolationMode.BICUBIC ), transforms.ToTensor(), transforms.Normalize((0.48_145_466, 0.4_578_275, 0.40_821_073) , (0.26_862_954, 0.26_130_258, 0.27_577_711) ), ] ) UpperCAmelCase__ : Optional[int] = transform(__lowerCamelCase ).unsqueeze(0 ).to(__lowerCamelCase ) return image def _lowerCamelCase ( __lowerCamelCase ) -> str: '''simple docstring''' if "visual_encoder" in key: UpperCAmelCase__ : Dict = re.sub("""visual_encoder*""" , """vision_model.encoder""" , __lowerCamelCase ) if "blocks" in key: UpperCAmelCase__ : Optional[Any] = re.sub(r"""blocks""" , """layers""" , __lowerCamelCase ) if "attn" in key: UpperCAmelCase__ : List[str] = re.sub(r"""attn""" , """self_attn""" , __lowerCamelCase ) if "norm1" in key: UpperCAmelCase__ : Union[str, Any] = re.sub(r"""norm1""" , """layer_norm1""" , __lowerCamelCase ) if "norm2" in key: UpperCAmelCase__ : Any = re.sub(r"""norm2""" , """layer_norm2""" , __lowerCamelCase ) if "encoder.norm" in key: UpperCAmelCase__ : Dict = re.sub(r"""encoder.norm""" , """post_layernorm""" , __lowerCamelCase ) if "encoder.patch_embed.proj" in key: UpperCAmelCase__ : List[str] = re.sub(r"""encoder.patch_embed.proj""" , """embeddings.patch_embedding""" , __lowerCamelCase ) if "encoder.pos_embed" in key: UpperCAmelCase__ : List[str] = re.sub(r"""encoder.pos_embed""" , """embeddings.position_embedding""" , __lowerCamelCase ) if "encoder.cls_token" in key: UpperCAmelCase__ : List[Any] = re.sub(r"""encoder.cls_token""" , """embeddings.class_embedding""" , __lowerCamelCase ) if "self_attn" in key: UpperCAmelCase__ : List[Any] = re.sub(r"""self_attn.proj""" , """self_attn.projection""" , __lowerCamelCase ) return key @torch.no_grad() def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase=None ) -> Tuple: '''simple docstring''' if config_path is not None: UpperCAmelCase__ : Any = BlipConfig.from_pretrained(__lowerCamelCase ) else: UpperCAmelCase__ : str = BlipConfig(projection_dim=512 , text_config={} , vision_config={} ) UpperCAmelCase__ : int = BlipForConditionalGeneration(__lowerCamelCase ).eval() UpperCAmelCase__ : Any = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth""" UpperCAmelCase__ : List[str] = blip_decoder(pretrained=__lowerCamelCase , image_size=384 , vit="""base""" ) UpperCAmelCase__ : Union[str, Any] = pt_model.eval() UpperCAmelCase__ : Optional[int] = pt_model.state_dict() for key in modified_state_dict.copy(): UpperCAmelCase__ : Dict = modified_state_dict.pop(__lowerCamelCase ) UpperCAmelCase__ : Union[str, Any] = rename_key(__lowerCamelCase ) UpperCAmelCase__ : List[str] = value hf_model.load_state_dict(__lowerCamelCase ) UpperCAmelCase__ : Tuple = 384 UpperCAmelCase__ : str = load_demo_image(image_size=__lowerCamelCase , device="""cpu""" ) UpperCAmelCase__ : str = BertTokenizer.from_pretrained("""bert-base-uncased""" ) UpperCAmelCase__ : Dict = tokenizer(["""a picture of"""] ).input_ids UpperCAmelCase__ : int = hf_model.generate(__lowerCamelCase , __lowerCamelCase ) assert out[0].tolist() == [3_0522, 1037, 3861, 1997, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] UpperCAmelCase__ : Any = hf_model.generate(__lowerCamelCase ) assert out[0].tolist() == [3_0522, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(__lowerCamelCase ) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' UpperCAmelCase__ : Union[str, Any] = ( """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth""" ) UpperCAmelCase__ : List[Any] = blip_vqa(pretrained=__lowerCamelCase , image_size=__lowerCamelCase , vit="""base""" ) vqa_model.eval() UpperCAmelCase__ : str = vqa_model.state_dict() for key in modified_state_dict.copy(): UpperCAmelCase__ : Dict = modified_state_dict.pop(__lowerCamelCase ) UpperCAmelCase__ : Dict = rename_key(__lowerCamelCase ) UpperCAmelCase__ : int = value UpperCAmelCase__ : List[str] = BlipForQuestionAnswering(__lowerCamelCase ) hf_vqa_model.load_state_dict(__lowerCamelCase ) UpperCAmelCase__ : Tuple = ["""How many dogs are in this image?"""] UpperCAmelCase__ : Union[str, Any] = tokenizer(__lowerCamelCase , return_tensors="""pt""" ).input_ids UpperCAmelCase__ : Optional[Any] = hf_vqa_model.generate(__lowerCamelCase , __lowerCamelCase ) print(tokenizer.decode(answer[0] ) ) assert tokenizer.decode(answer[0] ) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + """_vqa""" ) UpperCAmelCase__ : int = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth""" UpperCAmelCase__ : Any = blip_itm(pretrained=__lowerCamelCase , image_size=__lowerCamelCase , vit="""base""" ) itm_model.eval() UpperCAmelCase__ : List[Any] = itm_model.state_dict() for key in modified_state_dict.copy(): UpperCAmelCase__ : Dict = modified_state_dict.pop(__lowerCamelCase ) UpperCAmelCase__ : int = rename_key(__lowerCamelCase ) UpperCAmelCase__ : Any = value UpperCAmelCase__ : Optional[int] = BlipForImageTextRetrieval(__lowerCamelCase ) UpperCAmelCase__ : Union[str, Any] = ["""A picture of a woman with a dog sitting in a beach"""] UpperCAmelCase__ : List[Any] = tokenizer( __lowerCamelCase , return_tensors="""pt""" , padding="""max_length""" , truncation=__lowerCamelCase , max_length=35 , ).input_ids hf_itm_model.load_state_dict(__lowerCamelCase ) hf_itm_model.eval() UpperCAmelCase__ : List[str] = hf_itm_model(__lowerCamelCase , __lowerCamelCase , use_itm_head=__lowerCamelCase ) UpperCAmelCase__ : List[str] = hf_itm_model(__lowerCamelCase , __lowerCamelCase , use_itm_head=__lowerCamelCase ) assert out[0].item() == 0.2_110_687_494_277_954 assert torch.nn.functional.softmax(out_itm[0] , dim=1 )[:, 1].item() == 0.45_698_845_386_505_127 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + """_itm""" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : str = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") SCREAMING_SNAKE_CASE__ : List[Any] = parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
79
0
import os import pytest from attr import dataclass lowercase_ : Optional[Any] = 'us-east-1' # defaults region @dataclass class _lowerCamelCase : __a = 42 __a = "arn:aws:iam::558105141721:role/sagemaker_execution_role" __a = { "task_name": "mnli", "per_device_train_batch_size": 16, "per_device_eval_batch_size": 16, "do_train": True, "do_eval": True, "do_predict": True, "output_dir": "/opt/ml/model", "overwrite_output_dir": True, "max_steps": 500, "save_steps": 5500, } __a = {**hyperparameters, "max_steps": 1000} @property def UpperCamelCase_ ( self ) -> str: if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def UpperCamelCase_ ( self ) -> str: return f'{self.framework}-transfromers-test' @property def UpperCamelCase_ ( self ) -> str: return f'./tests/sagemaker/scripts/{self.framework}' @property def UpperCamelCase_ ( self ) -> str: if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='''class''' ) def A__ ( snake_case_ : Union[str, Any] ): SCREAMING_SNAKE_CASE__: Dict= SageMakerTestEnvironment(framework=request.cls.framework )
64
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : Tuple = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : List[Any] = { """MIT/ast-finetuned-audioset-10-10-0.4593""": ( """https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json""" ), } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'audio-spectrogram-transformer' def __init__( self , _lowerCAmelCase=768 , _lowerCAmelCase=12 , _lowerCAmelCase=12 , _lowerCAmelCase=3072 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=1e-12 , _lowerCAmelCase=16 , _lowerCAmelCase=True , _lowerCAmelCase=10 , _lowerCAmelCase=10 , _lowerCAmelCase=1024 , _lowerCAmelCase=128 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : List[Any] = num_attention_heads UpperCAmelCase__ : Dict = intermediate_size UpperCAmelCase__ : Dict = hidden_act UpperCAmelCase__ : str = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : Tuple = initializer_range UpperCAmelCase__ : Dict = layer_norm_eps UpperCAmelCase__ : Optional[Any] = patch_size UpperCAmelCase__ : Tuple = qkv_bias UpperCAmelCase__ : Tuple = frequency_stride UpperCAmelCase__ : Union[str, Any] = time_stride UpperCAmelCase__ : Optional[Any] = max_length UpperCAmelCase__ : Optional[int] = num_mel_bins
79
0
"""simple docstring""" __UpperCAmelCase = [ (1000, 'M'), (900, 'CM'), (500, 'D'), (400, 'CD'), (100, 'C'), (90, 'XC'), (50, 'L'), (40, 'XL'), (10, 'X'), (9, 'IX'), (5, 'V'), (4, 'IV'), (1, 'I'), ] def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = {"""I""": 1, """V""": 5, """X""": 10, """L""": 50, """C""": 100, """D""": 500, """M""": 1000} UpperCAmelCase__ : List[Any] = 0 UpperCAmelCase__ : Optional[Any] = 0 while place < len(__UpperCamelCase ): if (place + 1 < len(__UpperCamelCase )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def lowerCAmelCase ( __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = [] for arabic, roman in ROMAN: ((UpperCAmelCase__) , (UpperCAmelCase__)) : Optional[int] = divmod(__UpperCamelCase , __UpperCamelCase ) result.append(roman * factor ) if number == 0: break return "".join(__UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
65
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) class UpperCAmelCase_ ( __lowerCamelCase ): def __init__( self , *_lowerCAmelCase , **_lowerCAmelCase ): warnings.warn( """The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use GLPNImageProcessor instead.""" , _lowerCAmelCase , ) super().__init__(*_lowerCAmelCase , **_lowerCAmelCase )
79
0
import glob import os import random from string import ascii_lowercase, digits import cva UpperCamelCase = "" UpperCamelCase = "" UpperCamelCase = "" UpperCamelCase = 1 # (0 is vertical, 1 is horizontal) def __magic_name__ ( ) -> None: _lowercase , _lowercase : List[str] = get_dataset(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) print('Processing...' ) _lowercase , _lowercase , _lowercase : Optional[Any] = update_image_and_anno(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for index, image in enumerate(SCREAMING_SNAKE_CASE ): # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' _lowercase : Dict = random_chars(32 ) _lowercase : int = paths[index].split(os.sep )[-1].rsplit('.' , 1 )[0] _lowercase : Union[str, Any] = F"""{OUTPUT_DIR}/{file_name}_FLIP_{letter_code}""" cva.imwrite(F"""/{file_root}.jpg""" , SCREAMING_SNAKE_CASE , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"""Success {index+1}/{len(SCREAMING_SNAKE_CASE )} with {file_name}""" ) _lowercase : Tuple = [] for anno in new_annos[index]: _lowercase : Any = F"""{anno[0]} {anno[1]} {anno[2]} {anno[3]} {anno[4]}""" annos_list.append(SCREAMING_SNAKE_CASE ) with open(F"""/{file_root}.txt""" , 'w' ) as outfile: outfile.write('\n'.join(line for line in annos_list ) ) def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[list, list]: _lowercase : Tuple = [] _lowercase : Optional[int] = [] for label_file in glob.glob(os.path.join(SCREAMING_SNAKE_CASE , '*.txt' ) ): _lowercase : Any = label_file.split(os.sep )[-1].rsplit('.' , 1 )[0] with open(SCREAMING_SNAKE_CASE ) as in_file: _lowercase : Optional[int] = in_file.readlines() _lowercase : List[Any] = os.path.join(SCREAMING_SNAKE_CASE , F"""{label_name}.jpg""" ) _lowercase : List[Any] = [] for obj_list in obj_lists: _lowercase : Optional[int] = obj_list.rstrip('\n' ).split(' ' ) boxes.append( [ int(obj[0] ), float(obj[1] ), float(obj[2] ), float(obj[3] ), float(obj[4] ), ] ) if not boxes: continue img_paths.append(SCREAMING_SNAKE_CASE ) labels.append(SCREAMING_SNAKE_CASE ) return img_paths, labels def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 1 ) -> tuple[list, list, list]: _lowercase : Optional[int] = [] _lowercase : Tuple = [] _lowercase : str = [] for idx in range(len(SCREAMING_SNAKE_CASE ) ): _lowercase : Dict = [] _lowercase : Tuple = img_list[idx] path_list.append(SCREAMING_SNAKE_CASE ) _lowercase : int = anno_list[idx] _lowercase : int = cva.imread(SCREAMING_SNAKE_CASE ) if flip_type == 1: _lowercase : int = cva.flip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for bbox in img_annos: _lowercase : Union[str, Any] = 1 - bbox[1] new_annos.append([bbox[0], x_center_new, bbox[2], bbox[3], bbox[4]] ) elif flip_type == 0: _lowercase : str = cva.flip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for bbox in img_annos: _lowercase : List[Any] = 1 - bbox[2] new_annos.append([bbox[0], bbox[1], y_center_new, bbox[3], bbox[4]] ) new_annos_lists.append(SCREAMING_SNAKE_CASE ) new_imgs_list.append(SCREAMING_SNAKE_CASE ) return new_imgs_list, new_annos_lists, path_list def __magic_name__ ( SCREAMING_SNAKE_CASE = 32 ) -> str: assert number_char > 1, "The number of character should greater than 1" _lowercase : Optional[int] = ascii_lowercase + digits return "".join(random.choice(SCREAMING_SNAKE_CASE ) for _ in range(SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": main() print("DONE ✅")
66
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : int = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__ : List[str] = { """vocab_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt""", }, """tokenizer_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json""" ), """google/realm-orqa-nq-openqa""": ( """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-nq-reader""": ( """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-openqa""": ( """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-reader""": ( """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json""" ), }, } SCREAMING_SNAKE_CASE__ : Optional[Any] = { """google/realm-cc-news-pretrained-embedder""": 5_12, """google/realm-cc-news-pretrained-encoder""": 5_12, """google/realm-cc-news-pretrained-scorer""": 5_12, """google/realm-cc-news-pretrained-openqa""": 5_12, """google/realm-orqa-nq-openqa""": 5_12, """google/realm-orqa-nq-reader""": 5_12, """google/realm-orqa-wq-openqa""": 5_12, """google/realm-orqa-wq-reader""": 5_12, } SCREAMING_SNAKE_CASE__ : Optional[Any] = { """google/realm-cc-news-pretrained-embedder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-encoder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-scorer""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-reader""": {"""do_lower_case""": True}, """google/realm-orqa-wq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-wq-reader""": {"""do_lower_case""": True}, } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = PRETRAINED_INIT_CONFIGURATION __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = RealmTokenizer def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=True , _lowerCAmelCase="[UNK]" , _lowerCAmelCase="[SEP]" , _lowerCAmelCase="[PAD]" , _lowerCAmelCase="[CLS]" , _lowerCAmelCase="[MASK]" , _lowerCAmelCase=True , _lowerCAmelCase=None , **_lowerCAmelCase , ): super().__init__( _lowerCAmelCase , tokenizer_file=_lowerCAmelCase , do_lower_case=_lowerCAmelCase , unk_token=_lowerCAmelCase , sep_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , cls_token=_lowerCAmelCase , mask_token=_lowerCAmelCase , tokenize_chinese_chars=_lowerCAmelCase , strip_accents=_lowerCAmelCase , **_lowerCAmelCase , ) UpperCAmelCase__ : Tuple = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , _lowerCAmelCase ) != do_lower_case or normalizer_state.get("""strip_accents""" , _lowerCAmelCase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , _lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase__ : Any = getattr(_lowerCAmelCase , normalizer_state.pop("""type""" ) ) UpperCAmelCase__ : str = do_lower_case UpperCAmelCase__ : Tuple = strip_accents UpperCAmelCase__ : Tuple = tokenize_chinese_chars UpperCAmelCase__ : Union[str, Any] = normalizer_class(**_lowerCAmelCase ) UpperCAmelCase__ : Dict = do_lower_case def __UpperCAmelCase ( self , _lowerCAmelCase , **_lowerCAmelCase ): UpperCAmelCase__ : List[Any] = PaddingStrategy.MAX_LENGTH UpperCAmelCase__ : Optional[int] = text UpperCAmelCase__ : Optional[int] = kwargs.pop("""text_pair""" , _lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = kwargs.pop("""return_tensors""" , _lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = { """input_ids""": [], """attention_mask""": [], """token_type_ids""": [], } for idx, candidate_text in enumerate(_lowerCAmelCase ): if batch_text_pair is not None: UpperCAmelCase__ : str = batch_text_pair[idx] else: UpperCAmelCase__ : Any = None UpperCAmelCase__ : str = super().__call__(_lowerCAmelCase , _lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = encoded_candidates.get("""input_ids""" ) UpperCAmelCase__ : str = encoded_candidates.get("""attention_mask""" ) UpperCAmelCase__ : Union[str, Any] = encoded_candidates.get("""token_type_ids""" ) if encoded_input_ids is not None: output_data["input_ids"].append(_lowerCAmelCase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(_lowerCAmelCase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = {key: item for key, item in output_data.items() if len(_lowerCAmelCase ) != 0} return BatchEncoding(_lowerCAmelCase , tensor_type=_lowerCAmelCase ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase = None ): UpperCAmelCase__ : Any = [self.sep_token_id] UpperCAmelCase__ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase = None ): UpperCAmelCase__ : List[str] = self._tokenizer.model.save(_lowerCAmelCase , name=_lowerCAmelCase ) return tuple(_lowerCAmelCase )
79
0
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case = { """configuration_xmod""": [ """XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XmodConfig""", """XmodOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case = [ """XMOD_PRETRAINED_MODEL_ARCHIVE_LIST""", """XmodForCausalLM""", """XmodForMaskedLM""", """XmodForMultipleChoice""", """XmodForQuestionAnswering""", """XmodForSequenceClassification""", """XmodForTokenClassification""", """XmodModel""", """XmodPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
67
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'facebook/bart-large-mnli' __lowerCamelCase = ( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) __lowerCamelCase = 'text_classifier' __lowerCamelCase = AutoTokenizer __lowerCamelCase = AutoModelForSequenceClassification __lowerCamelCase = ['text', ['text']] __lowerCamelCase = ['text'] def __UpperCAmelCase ( self ): super().setup() UpperCAmelCase__ : Optional[Any] = self.model.config UpperCAmelCase__ : Tuple = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("""entail""" ): UpperCAmelCase__ : Dict = int(_lowerCAmelCase ) if self.entailment_id == -1: raise ValueError("""Could not determine the entailment ID from the model config, please pass it at init.""" ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : List[Any] = labels return self.pre_processor( [text] * len(_lowerCAmelCase ) , [f"This example is {label}" for label in labels] , return_tensors="""pt""" , padding="""max_length""" , ) def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : str = outputs.logits UpperCAmelCase__ : List[Any] = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
79
0
def lowercase__ ( A_: int , A_: int ) -> int: """simple docstring""" return 1 if input_a == input_a else 0 def lowercase__ ( ) -> None: """simple docstring""" assert xnor_gate(0 , 0 ) == 1 assert xnor_gate(0 , 1 ) == 0 assert xnor_gate(1 , 0 ) == 0 assert xnor_gate(1 , 1 ) == 1 if __name__ == "__main__": print(xnor_gate(0, 0)) print(xnor_gate(0, 1)) print(xnor_gate(1, 0)) print(xnor_gate(1, 1))
68
from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=13 , _lowerCAmelCase=30 , _lowerCAmelCase=2 , _lowerCAmelCase=3 , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=32 , _lowerCAmelCase=2 , _lowerCAmelCase=4 , _lowerCAmelCase=37 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=10 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=3 , _lowerCAmelCase=None , ): UpperCAmelCase__ : Tuple = parent UpperCAmelCase__ : Optional[int] = batch_size UpperCAmelCase__ : Union[str, Any] = image_size UpperCAmelCase__ : int = patch_size UpperCAmelCase__ : str = num_channels UpperCAmelCase__ : int = is_training UpperCAmelCase__ : List[str] = use_labels UpperCAmelCase__ : List[Any] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : Tuple = num_attention_heads UpperCAmelCase__ : Optional[int] = intermediate_size UpperCAmelCase__ : Optional[Any] = hidden_act UpperCAmelCase__ : int = hidden_dropout_prob UpperCAmelCase__ : int = attention_probs_dropout_prob UpperCAmelCase__ : List[str] = type_sequence_label_size UpperCAmelCase__ : Optional[int] = initializer_range UpperCAmelCase__ : Any = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) UpperCAmelCase__ : Any = (image_size // patch_size) ** 2 UpperCAmelCase__ : Tuple = num_patches + 1 def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ : List[str] = None if self.use_labels: UpperCAmelCase__ : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ : Union[str, Any] = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self ): return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : str = TFViTModel(config=_lowerCAmelCase ) UpperCAmelCase__ : str = model(_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase__ : Optional[Any] = self.image_size // 2 UpperCAmelCase__ : List[str] = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase__ : List[Any] = model(_lowerCAmelCase , interpolate_pos_encoding=_lowerCAmelCase , training=_lowerCAmelCase ) UpperCAmelCase__ : str = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : Tuple = self.type_sequence_label_size UpperCAmelCase__ : List[Any] = TFViTForImageClassification(_lowerCAmelCase ) UpperCAmelCase__ : List[str] = model(_lowerCAmelCase , labels=_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase__ : Tuple = self.image_size // 2 UpperCAmelCase__ : Union[str, Any] = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase__ : List[str] = model(_lowerCAmelCase , interpolate_pos_encoding=_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase__ : Union[str, Any] = 1 UpperCAmelCase__ : Optional[Any] = TFViTForImageClassification(_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase__ : List[str] = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : int = config_and_inputs UpperCAmelCase__ : int = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class UpperCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () __lowerCamelCase = ( {'feature-extraction': TFViTModel, 'image-classification': TFViTForImageClassification} if is_tf_available() else {} ) __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = TFViTModelTester(self ) UpperCAmelCase__ : int = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def __UpperCAmelCase ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def __UpperCAmelCase ( self ): pass @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def __UpperCAmelCase ( self ): pass def __UpperCAmelCase ( self ): UpperCAmelCase__ , UpperCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : str = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , tf.keras.layers.Layer ) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : Optional[int] = model_class(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ : Tuple = [*signature.parameters.keys()] UpperCAmelCase__ : str = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = TFViTModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(_lowerCAmelCase ) def _lowerCamelCase ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ : List[str] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class UpperCAmelCase_ ( unittest.TestCase ): @cached_property def __UpperCAmelCase ( self ): return ViTImageProcessor.from_pretrained("""google/vit-base-patch16-224""" ) if is_vision_available() else None @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = TFViTForImageClassification.from_pretrained("""google/vit-base-patch16-224""" ) UpperCAmelCase__ : List[Any] = self.default_image_processor UpperCAmelCase__ : Union[str, Any] = prepare_img() UpperCAmelCase__ : Optional[Any] = image_processor(images=_lowerCAmelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase__ : int = model(**_lowerCAmelCase ) # verify the logits UpperCAmelCase__ : Tuple = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) UpperCAmelCase__ : int = tf.constant([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ) tf.debugging.assert_near(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 )
79
0
'''simple docstring''' import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging a : int = logging.get_logger(__name__) a : Any = r''' Args: input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`): Indices of input sequence tokens in the vocabulary. Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and [`PreTrainedTokenizer.__call__`] for details. [What are input IDs?](../glossary#input-ids) scores (`torch.FloatTensor` of shape `(batch_size, config.vocab_size)`): Prediction scores of a language modeling head. These can be scores for each vocabulary token before SoftMax or scores for each vocabulary token after SoftMax. kwargs (`Dict[str, Any]`, *optional*): Additional stopping criteria specific kwargs. Return: `bool`. `False` indicates we should continue, `True` indicates we should stop. ''' class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): @add_start_docstrings(a_ ) def __call__( self : int , a_ : torch.LongTensor , a_ : torch.FloatTensor , **a_ : Tuple ): """simple docstring""" raise NotImplementedError("StoppingCriteria needs to be subclassed" ) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def __init__( self : Dict , a_ : int , a_ : Optional[int] = None ): """simple docstring""" __snake_case = max_length __snake_case = max_position_embeddings @add_start_docstrings(a_ ) def __call__( self : str , a_ : torch.LongTensor , a_ : torch.FloatTensor , **a_ : Union[str, Any] ): """simple docstring""" __snake_case = input_ids.shape[-1] __snake_case = cur_len >= self.max_length if self.max_position_embeddings is not None and not is_done and cur_len >= self.max_position_embeddings: logger.warning_once( "This is a friendly reminder - the current text generation call will exceed the model's predefined " f'''maximum length ({self.max_position_embeddings}). Depending on the model, you may observe ''' "exceptions, performance degradation, or nothing at all." ) return is_done class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def __init__( self : List[str] , a_ : int , a_ : int ): """simple docstring""" warnings.warn( "The class `MaxNewTokensCriteria` is deprecated. " f'''Please use `MaxLengthCriteria(max_length={start_length + max_new_tokens})` ''' "with `max_length = start_length + max_new_tokens` instead." , a_ , ) __snake_case = start_length __snake_case = max_new_tokens __snake_case = start_length + max_new_tokens @add_start_docstrings(a_ ) def __call__( self : Union[str, Any] , a_ : torch.LongTensor , a_ : torch.FloatTensor , **a_ : Tuple ): """simple docstring""" return input_ids.shape[-1] >= self.max_length class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def __init__( self : Tuple , a_ : float , a_ : Optional[float] = None ): """simple docstring""" __snake_case = max_time __snake_case = time.time() if initial_timestamp is None else initial_timestamp @add_start_docstrings(a_ ) def __call__( self : str , a_ : torch.LongTensor , a_ : torch.FloatTensor , **a_ : str ): """simple docstring""" return time.time() - self.initial_timestamp > self.max_time class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): @add_start_docstrings(a_ ) def __call__( self : Tuple , a_ : torch.LongTensor , a_ : torch.FloatTensor , **a_ : Tuple ): """simple docstring""" return any(criteria(a_ , a_ ) for criteria in self ) @property def A ( self : Any ): """simple docstring""" for stopping_criterium in self: if isinstance(a_ , a_ ): return stopping_criterium.max_length elif isinstance(a_ , a_ ): return stopping_criterium.max_length return None def __UpperCAmelCase ( _UpperCAmelCase : StoppingCriteriaList , _UpperCAmelCase : int ) -> StoppingCriteriaList: __snake_case = stopping_criteria.max_length __snake_case = deepcopy(_UpperCAmelCase ) if stopping_max_length is not None and stopping_max_length != max_length: warnings.warn("You set different `max_length` for stopping criteria and `max_length` parameter" , _UpperCAmelCase ) elif stopping_max_length is None: new_stopping_criteria.append(MaxLengthCriteria(max_length=_UpperCAmelCase ) ) return new_stopping_criteria
69
from functools import lru_cache @lru_cache def _lowerCamelCase ( __lowerCamelCase ) -> int: '''simple docstring''' if num < 0: raise ValueError("""Number should not be negative.""" ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
79
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase : Dict = {"configuration_plbart": ["PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP", "PLBartConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : List[str] = ["PLBartTokenizer"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : List[str] = [ "PLBART_PRETRAINED_MODEL_ARCHIVE_LIST", "PLBartForCausalLM", "PLBartForConditionalGeneration", "PLBartForSequenceClassification", "PLBartModel", "PLBartPreTrainedModel", ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys lowerCamelCase : Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure)
70
import argparse import hashlib # hashlib is only used inside the Test class import struct class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase ): UpperCAmelCase__ : Any = data UpperCAmelCase__ : List[Any] = [0X6745_2301, 0Xefcd_ab89, 0X98ba_dcfe, 0X1032_5476, 0Xc3d2_e1f0] @staticmethod def __UpperCAmelCase ( _lowerCAmelCase , _lowerCAmelCase ): return ((n << b) | (n >> (32 - b))) & 0Xffff_ffff def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) UpperCAmelCase__ : Optional[int] = self.data + padding + struct.pack(""">Q""" , 8 * len(self.data ) ) return padded_data def __UpperCAmelCase ( self ): return [ self.padded_data[i : i + 64] for i in range(0 , len(self.padded_data ) , 64 ) ] def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : Dict = list(struct.unpack(""">16L""" , _lowerCAmelCase ) ) + [0] * 64 for i in range(16 , 80 ): UpperCAmelCase__ : Optional[int] = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) , 1 ) return w def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[str] = self.padding() UpperCAmelCase__ : List[str] = self.split_blocks() for block in self.blocks: UpperCAmelCase__ : Tuple = self.expand_block(_lowerCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.h for i in range(0 , 80 ): if 0 <= i < 20: UpperCAmelCase__ : Optional[int] = (b & c) | ((~b) & d) UpperCAmelCase__ : int = 0X5a82_7999 elif 20 <= i < 40: UpperCAmelCase__ : Tuple = b ^ c ^ d UpperCAmelCase__ : int = 0X6ed9_eba1 elif 40 <= i < 60: UpperCAmelCase__ : List[str] = (b & c) | (b & d) | (c & d) UpperCAmelCase__ : Tuple = 0X8f1b_bcdc elif 60 <= i < 80: UpperCAmelCase__ : int = b ^ c ^ d UpperCAmelCase__ : str = 0Xca62_c1d6 UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = ( self.rotate(_lowerCAmelCase , 5 ) + f + e + k + expanded_block[i] & 0Xffff_ffff, a, self.rotate(_lowerCAmelCase , 30 ), c, d, ) UpperCAmelCase__ : int = ( self.h[0] + a & 0Xffff_ffff, self.h[1] + b & 0Xffff_ffff, self.h[2] + c & 0Xffff_ffff, self.h[3] + d & 0Xffff_ffff, self.h[4] + e & 0Xffff_ffff, ) return ("{:08x}" * 5).format(*self.h ) def _lowerCamelCase ( ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase__ : Optional[Any] = B"""Test String""" assert SHAaHash(__lowerCamelCase ).final_hash() == hashlib.shaa(__lowerCamelCase ).hexdigest() # noqa: S324 def _lowerCamelCase ( ) -> str: '''simple docstring''' UpperCAmelCase__ : Optional[int] = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument("""--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) UpperCAmelCase__ : str = parser.parse_args() UpperCAmelCase__ : Union[str, Any] = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: UpperCAmelCase__ : List[Any] = f.read() else: UpperCAmelCase__ : int = bytes(__lowerCamelCase , """utf-8""" ) print(SHAaHash(__lowerCamelCase ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
79
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _lowerCamelCase = {"""configuration_plbart""": ["""PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PLBartConfig"""]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ["""PLBartTokenizer"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ """PLBART_PRETRAINED_MODEL_ARCHIVE_LIST""", """PLBartForCausalLM""", """PLBartForConditionalGeneration""", """PLBartForSequenceClassification""", """PLBartModel""", """PLBartPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
71
from importlib import import_module from .logging import get_logger SCREAMING_SNAKE_CASE__ : Union[str, Any] = get_logger(__name__) class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : List[str] = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith("""__""" ): setattr(self , _lowerCAmelCase , getattr(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase__ : Tuple = module._original_module if isinstance(_lowerCAmelCase , _PatchedModuleObj ) else module class UpperCAmelCase_ : __lowerCamelCase = [] def __init__( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : str = obj UpperCAmelCase__ : List[str] = target UpperCAmelCase__ : List[str] = new UpperCAmelCase__ : Any = target.split(""".""" )[0] UpperCAmelCase__ : Union[str, Any] = {} UpperCAmelCase__ : str = attrs or [] def __enter__( self ): *UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.target.split(""".""" ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(_lowerCAmelCase ) ): try: UpperCAmelCase__ : Optional[int] = import_module(""".""".join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): UpperCAmelCase__ : Any = getattr(self.obj , _lowerCAmelCase ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(_lowerCAmelCase , _PatchedModuleObj ) and obj_attr._original_module is submodule) ): UpperCAmelCase__ : List[Any] = obj_attr # patch at top level setattr(self.obj , _lowerCAmelCase , _PatchedModuleObj(_lowerCAmelCase , attrs=self.attrs ) ) UpperCAmelCase__ : Optional[Any] = getattr(self.obj , _lowerCAmelCase ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(_lowerCAmelCase , _lowerCAmelCase , _PatchedModuleObj(getattr(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) , attrs=self.attrs ) ) UpperCAmelCase__ : Union[str, Any] = getattr(_lowerCAmelCase , _lowerCAmelCase ) # finally set the target attribute setattr(_lowerCAmelCase , _lowerCAmelCase , self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: UpperCAmelCase__ : Union[str, Any] = getattr(import_module(""".""".join(_lowerCAmelCase ) ) , _lowerCAmelCase ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , _lowerCAmelCase ) is attr_value: UpperCAmelCase__ : Optional[int] = getattr(self.obj , _lowerCAmelCase ) setattr(self.obj , _lowerCAmelCase , self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" UpperCAmelCase__ : Dict = globals()["""__builtins__"""][target_attr] setattr(self.obj , _lowerCAmelCase , self.new ) else: raise RuntimeError(f"Tried to patch attribute {target_attr} instead of a submodule." ) def __exit__( self , *_lowerCAmelCase ): for attr in list(self.original ): setattr(self.obj , _lowerCAmelCase , self.original.pop(_lowerCAmelCase ) ) def __UpperCAmelCase ( self ): self.__enter__() self._active_patches.append(self ) def __UpperCAmelCase ( self ): try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
79
0
'''simple docstring''' from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo _UpperCAmelCase : Any = '''\ @misc{wu2016googles, title={Google\'s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation}, author={Yonghui Wu and Mike Schuster and Zhifeng Chen and Quoc V. Le and Mohammad Norouzi and Wolfgang Macherey and Maxim Krikun and Yuan Cao and Qin Gao and Klaus Macherey and Jeff Klingner and Apurva Shah and Melvin Johnson and Xiaobing Liu and Łukasz Kaiser and Stephan Gouws and Yoshikiyo Kato and Taku Kudo and Hideto Kazawa and Keith Stevens and George Kurian and Nishant Patil and Wei Wang and Cliff Young and Jason Smith and Jason Riesa and Alex Rudnick and Oriol Vinyals and Greg Corrado and Macduff Hughes and Jeffrey Dean}, year={2016}, eprint={1609.08144}, archivePrefix={arXiv}, primaryClass={cs.CL} } ''' _UpperCAmelCase : str = '''\ The BLEU score has some undesirable properties when used for single sentences, as it was designed to be a corpus measure. We therefore use a slightly different score for our RL experiments which we call the \'GLEU score\'. For the GLEU score, we record all sub-sequences of 1, 2, 3 or 4 tokens in output and target sequence (n-grams). We then compute a recall, which is the ratio of the number of matching n-grams to the number of total n-grams in the target (ground truth) sequence, and a precision, which is the ratio of the number of matching n-grams to the number of total n-grams in the generated output sequence. Then GLEU score is simply the minimum of recall and precision. This GLEU score\'s range is always between 0 (no matches) and 1 (all match) and it is symmetrical when switching output and target. According to our experiments, GLEU score correlates quite well with the BLEU metric on a corpus level but does not have its drawbacks for our per sentence reward objective. ''' _UpperCAmelCase : List[str] = '''\ Computes corpus-level Google BLEU (GLEU) score of translated segments against one or more references. Instead of averaging the sentence level GLEU scores (i.e. macro-average precision), Wu et al. (2016) sum up the matching tokens and the max of hypothesis and reference tokens for each sentence, then compute using the aggregate values. Args: predictions (list of str): list of translations to score. Each translation should be tokenized into a list of tokens. references (list of list of str): list of lists of references for each translation. Each reference should be tokenized into a list of tokens. min_len (int): The minimum order of n-gram this function should extract. Defaults to 1. max_len (int): The maximum order of n-gram this function should extract. Defaults to 4. Returns: \'google_bleu\': google_bleu score Examples: Example 1: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results["google_bleu"], 2)) 0.44 Example 2: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\', ... \'heed\', \'the\', \'cat\', \'commands\'] >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\', ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\', ... \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results["google_bleu"], 2)) 0.61 Example 3: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\', ... \'heed\', \'the\', \'cat\', \'commands\'] >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\', ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\', ... \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references, min_len=2) >>> print(round(results["google_bleu"], 2)) 0.53 Example 4: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\', ... \'heed\', \'the\', \'cat\', \'commands\'] >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\', ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\', ... \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses,references=list_of_references, min_len=2, max_len=6) >>> print(round(results["google_bleu"], 2)) 0.4 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' , id='''token''' ) , id='''sequence''' ), '''references''': datasets.Sequence( datasets.Sequence(datasets.Value('''string''' , id='''token''' ) , id='''sequence''' ) , id='''references''' ), } ) , ) def _A( self , snake_case_ , snake_case_ , snake_case_ = 1 , snake_case_ = 4 , ): return { "google_bleu": gleu_score.corpus_gleu( list_of_references=snake_case_ , hypotheses=snake_case_ , min_len=snake_case_ , max_len=snake_case_ ) }
72
from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : str = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Any = { """huggingface/informer-tourism-monthly""": ( """https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json""" ), # See all Informer models at https://huggingface.co/models?filter=informer } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'informer' __lowerCamelCase = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', 'num_hidden_layers': 'encoder_layers', } def __init__( self , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = "student_t" , _lowerCAmelCase = "nll" , _lowerCAmelCase = 1 , _lowerCAmelCase = None , _lowerCAmelCase = "mean" , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = 64 , _lowerCAmelCase = 32 , _lowerCAmelCase = 32 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = True , _lowerCAmelCase = "gelu" , _lowerCAmelCase = 0.0_5 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 100 , _lowerCAmelCase = 0.0_2 , _lowerCAmelCase=True , _lowerCAmelCase = "prob" , _lowerCAmelCase = 5 , _lowerCAmelCase = True , **_lowerCAmelCase , ): # time series specific configuration UpperCAmelCase__ : List[str] = prediction_length UpperCAmelCase__ : Optional[Any] = context_length or prediction_length UpperCAmelCase__ : str = distribution_output UpperCAmelCase__ : int = loss UpperCAmelCase__ : Optional[Any] = input_size UpperCAmelCase__ : Any = num_time_features UpperCAmelCase__ : int = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] UpperCAmelCase__ : Union[str, Any] = scaling UpperCAmelCase__ : Optional[Any] = num_dynamic_real_features UpperCAmelCase__ : List[str] = num_static_real_features UpperCAmelCase__ : str = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(_lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The cardinality should be a list of the same length as `num_static_categorical_features`""" ) UpperCAmelCase__ : List[str] = cardinality else: UpperCAmelCase__ : Optional[Any] = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(_lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The embedding dimension should be a list of the same length as `num_static_categorical_features`""" ) UpperCAmelCase__ : str = embedding_dimension else: UpperCAmelCase__ : List[str] = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] UpperCAmelCase__ : Union[str, Any] = num_parallel_samples # Transformer architecture configuration UpperCAmelCase__ : Dict = input_size * len(self.lags_sequence ) + self._number_of_features UpperCAmelCase__ : Any = d_model UpperCAmelCase__ : int = encoder_attention_heads UpperCAmelCase__ : Optional[Any] = decoder_attention_heads UpperCAmelCase__ : int = encoder_ffn_dim UpperCAmelCase__ : Tuple = decoder_ffn_dim UpperCAmelCase__ : List[Any] = encoder_layers UpperCAmelCase__ : Optional[Any] = decoder_layers UpperCAmelCase__ : Tuple = dropout UpperCAmelCase__ : int = attention_dropout UpperCAmelCase__ : List[str] = activation_dropout UpperCAmelCase__ : Any = encoder_layerdrop UpperCAmelCase__ : Union[str, Any] = decoder_layerdrop UpperCAmelCase__ : Tuple = activation_function UpperCAmelCase__ : Dict = init_std UpperCAmelCase__ : str = use_cache # Informer UpperCAmelCase__ : Union[str, Any] = attention_type UpperCAmelCase__ : int = sampling_factor UpperCAmelCase__ : Any = distil super().__init__(is_encoder_decoder=_lowerCAmelCase , **_lowerCAmelCase ) @property def __UpperCAmelCase ( self ): return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
79
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a_ : int = { 'configuration_rag': ['RagConfig'], 'retrieval_rag': ['RagRetriever'], 'tokenization_rag': ['RagTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : List[Any] = [ 'RagModel', 'RagPreTrainedModel', 'RagSequenceForGeneration', 'RagTokenForGeneration', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : Tuple = [ 'TFRagModel', 'TFRagPreTrainedModel', 'TFRagSequenceForGeneration', 'TFRagTokenForGeneration', ] if TYPE_CHECKING: from .configuration_rag import RagConfig from .retrieval_rag import RagRetriever from .tokenization_rag import RagTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rag import RagModel, RagPreTrainedModel, RagSequenceForGeneration, RagTokenForGeneration try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rag import ( TFRagModel, TFRagPreTrainedModel, TFRagSequenceForGeneration, TFRagTokenForGeneration, ) else: import sys a_ : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
73
def _lowerCamelCase ( __lowerCamelCase ) -> bool: '''simple docstring''' if p < 2: raise ValueError("""p should not be less than 2!""" ) elif p == 2: return True UpperCAmelCase__ : Tuple = 4 UpperCAmelCase__ : Tuple = (1 << p) - 1 for _ in range(p - 2 ): UpperCAmelCase__ : List[str] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
79
0
import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class __UpperCamelCase : """simple docstring""" def __init__( self : str , _A : List[Any] , _A : Union[str, Any]=13 , _A : List[Any]=7 , _A : Any=True , _A : Dict=True , _A : Union[str, Any]=False , _A : Optional[Any]=True , _A : str=99 , _A : Any=32 , _A : Optional[int]=5 , _A : Dict=4 , _A : Dict=37 , _A : Any="gelu" , _A : Union[str, Any]=0.1 , _A : Tuple=0.1 , _A : List[Any]=512 , _A : Dict=16 , _A : Any=2 , _A : List[Any]=0.02 , _A : Tuple=3 , _A : Dict=4 , _A : Tuple=None , ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = parent __SCREAMING_SNAKE_CASE : Optional[Any] = batch_size __SCREAMING_SNAKE_CASE : Optional[Any] = seq_length __SCREAMING_SNAKE_CASE : str = is_training __SCREAMING_SNAKE_CASE : Tuple = use_input_mask __SCREAMING_SNAKE_CASE : Optional[int] = use_token_type_ids __SCREAMING_SNAKE_CASE : Optional[Any] = use_labels __SCREAMING_SNAKE_CASE : str = vocab_size __SCREAMING_SNAKE_CASE : List[Any] = hidden_size __SCREAMING_SNAKE_CASE : str = num_hidden_layers __SCREAMING_SNAKE_CASE : Union[str, Any] = num_attention_heads __SCREAMING_SNAKE_CASE : Optional[Any] = intermediate_size __SCREAMING_SNAKE_CASE : str = hidden_act __SCREAMING_SNAKE_CASE : Union[str, Any] = hidden_dropout_prob __SCREAMING_SNAKE_CASE : str = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE : Optional[int] = max_position_embeddings __SCREAMING_SNAKE_CASE : Optional[int] = type_vocab_size __SCREAMING_SNAKE_CASE : List[str] = type_sequence_label_size __SCREAMING_SNAKE_CASE : List[str] = initializer_range __SCREAMING_SNAKE_CASE : str = num_labels __SCREAMING_SNAKE_CASE : str = num_choices __SCREAMING_SNAKE_CASE : List[Any] = scope def UpperCAmelCase__ ( self : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __SCREAMING_SNAKE_CASE : Any = None if self.use_input_mask: __SCREAMING_SNAKE_CASE : str = random_attention_mask([self.batch_size, self.seq_length] ) __SCREAMING_SNAKE_CASE : str = None if self.use_token_type_ids: __SCREAMING_SNAKE_CASE : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __SCREAMING_SNAKE_CASE : List[Any] = None __SCREAMING_SNAKE_CASE : List[str] = None __SCREAMING_SNAKE_CASE : List[str] = None if self.use_labels: __SCREAMING_SNAKE_CASE : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __SCREAMING_SNAKE_CASE : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __SCREAMING_SNAKE_CASE : Tuple = ids_tensor([self.batch_size] , self.num_choices ) __SCREAMING_SNAKE_CASE : List[str] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase__ ( self : Union[str, Any] ): """simple docstring""" return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_A , initializer_range=self.initializer_range , ) def UpperCAmelCase__ ( self : str , _A : Optional[Any] , _A : Optional[int] , _A : Tuple , _A : Tuple , _A : Optional[Any] , _A : List[Any] , _A : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = BioGptModel(config=_A ) model.to(_A ) model.eval() __SCREAMING_SNAKE_CASE : str = model(_A , attention_mask=_A ) __SCREAMING_SNAKE_CASE : List[Any] = model(_A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase__ ( self : Union[str, Any] , _A : List[str] , _A : Optional[Any] , _A : Any , _A : List[str] , _A : List[str] , _A : Optional[Any] , _A : Optional[Any] , _A : int , _A : Tuple , ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = BioGptForCausalLM(config=_A ) model.to(_A ) model.eval() __SCREAMING_SNAKE_CASE : List[str] = model(_A , attention_mask=_A , token_type_ids=_A , labels=_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase__ ( self : Optional[Any] , _A : Any , _A : Union[str, Any] , _A : Dict , _A : List[str] , _A : Union[str, Any] , *_A : Union[str, Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = BioGptModel(config=_A ) model.to(_A ) model.eval() # create attention mask __SCREAMING_SNAKE_CASE : Any = torch.ones(input_ids.shape , dtype=torch.long , device=_A ) __SCREAMING_SNAKE_CASE : Any = self.seq_length // 2 __SCREAMING_SNAKE_CASE : List[Any] = 0 # first forward pass __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : Optional[int] = model(_A , attention_mask=_A ).to_tuple() # create hypothetical next token and extent to next_input_ids __SCREAMING_SNAKE_CASE : Union[str, Any] = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids __SCREAMING_SNAKE_CASE : Optional[int] = ids_tensor((1,) , _A ).item() + 1 __SCREAMING_SNAKE_CASE : int = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) __SCREAMING_SNAKE_CASE : Any = random_other_next_tokens # append to next input_ids and attn_mask __SCREAMING_SNAKE_CASE : int = torch.cat([input_ids, next_tokens] , dim=-1 ) __SCREAMING_SNAKE_CASE : Optional[Any] = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=_A )] , dim=1 , ) # get two different outputs __SCREAMING_SNAKE_CASE : Union[str, Any] = model(_A , attention_mask=_A )['''last_hidden_state'''] __SCREAMING_SNAKE_CASE : List[str] = model(_A , past_key_values=_A , attention_mask=_A )['''last_hidden_state'''] # select random slice __SCREAMING_SNAKE_CASE : Union[str, Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() __SCREAMING_SNAKE_CASE : List[Any] = output_from_no_past[:, -1, random_slice_idx].detach() __SCREAMING_SNAKE_CASE : Tuple = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(_A , _A , atol=1e-3 ) ) def UpperCAmelCase__ ( self : Dict , _A : List[str] , _A : List[Any] , _A : List[str] , _A : Dict , _A : int , *_A : Dict ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = BioGptModel(config=_A ).to(_A ).eval() __SCREAMING_SNAKE_CASE : str = torch.ones(input_ids.shape , dtype=torch.long , device=_A ) # first forward pass __SCREAMING_SNAKE_CASE : List[Any] = model(_A , attention_mask=_A , use_cache=_A ) __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : Optional[Any] = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids __SCREAMING_SNAKE_CASE : Optional[int] = ids_tensor((self.batch_size, 3) , config.vocab_size ) __SCREAMING_SNAKE_CASE : Tuple = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and __SCREAMING_SNAKE_CASE : List[Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) __SCREAMING_SNAKE_CASE : Any = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) __SCREAMING_SNAKE_CASE : Optional[Any] = model(_A , attention_mask=_A )['''last_hidden_state'''] __SCREAMING_SNAKE_CASE : str = model(_A , attention_mask=_A , past_key_values=_A )[ '''last_hidden_state''' ] # select random slice __SCREAMING_SNAKE_CASE : str = ids_tensor((1,) , output_from_past.shape[-1] ).item() __SCREAMING_SNAKE_CASE : Dict = output_from_no_past[:, -3:, random_slice_idx].detach() __SCREAMING_SNAKE_CASE : Dict = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(_A , _A , atol=1e-3 ) ) def UpperCAmelCase__ ( self : Any , _A : List[str] , _A : str , _A : Dict , _A : List[str] , _A : int , *_A : Optional[Any] , _A : str=False ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = BioGptForCausalLM(_A ) model.to(_A ) if gradient_checkpointing: model.gradient_checkpointing_enable() __SCREAMING_SNAKE_CASE : Dict = model(_A , labels=_A ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def UpperCAmelCase__ ( self : List[Any] , _A : str , *_A : Dict ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = BioGptModel(_A ) __SCREAMING_SNAKE_CASE : int = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.0_01 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.01 ) def UpperCAmelCase__ ( self : str , _A : Union[str, Any] , _A : str , _A : Dict , _A : List[str] , _A : Dict , *_A : List[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = self.num_labels __SCREAMING_SNAKE_CASE : Any = BioGptForTokenClassification(_A ) model.to(_A ) model.eval() __SCREAMING_SNAKE_CASE : Any = model(_A , attention_mask=_A , token_type_ids=_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase__ ( self : Tuple ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = self.prepare_config_and_inputs() ( ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ( __SCREAMING_SNAKE_CASE ), ) : List[str] = config_and_inputs __SCREAMING_SNAKE_CASE : Any = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): """simple docstring""" lowerCAmelCase_ = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) lowerCAmelCase_ = (BioGptForCausalLM,) if is_torch_available() else () lowerCAmelCase_ = ( { '''feature-extraction''': BioGptModel, '''text-classification''': BioGptForSequenceClassification, '''text-generation''': BioGptForCausalLM, '''token-classification''': BioGptForTokenClassification, '''zero-shot''': BioGptForSequenceClassification, } if is_torch_available() else {} ) lowerCAmelCase_ = False def UpperCAmelCase__ ( self : Any ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = BioGptModelTester(self ) __SCREAMING_SNAKE_CASE : int = ConfigTester(self , config_class=_A , hidden_size=37 ) def UpperCAmelCase__ ( self : int ): """simple docstring""" self.config_tester.run_common_tests() def UpperCAmelCase__ ( self : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_A ) def UpperCAmelCase__ ( self : List[str] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __SCREAMING_SNAKE_CASE : Any = type self.model_tester.create_and_check_model(*_A ) def UpperCAmelCase__ ( self : List[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*_A ) def UpperCAmelCase__ ( self : Dict ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*_A , gradient_checkpointing=_A ) def UpperCAmelCase__ ( self : List[str] ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*_A ) def UpperCAmelCase__ ( self : Union[str, Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*_A ) def UpperCAmelCase__ ( self : str ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*_A ) @slow def UpperCAmelCase__ ( self : Dict ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = BioGptForCausalLM.from_pretrained('''microsoft/biogpt''' ) model.to(_A ) __SCREAMING_SNAKE_CASE : List[Any] = BioGptTokenizer.from_pretrained('''microsoft/biogpt''' ) __SCREAMING_SNAKE_CASE : List[str] = '''left''' # Define PAD Token = EOS Token = 50256 __SCREAMING_SNAKE_CASE : Dict = tokenizer.eos_token __SCREAMING_SNAKE_CASE : Optional[Any] = model.config.eos_token_id # use different length sentences to test batching __SCREAMING_SNAKE_CASE : Optional[int] = [ '''Hello, my dog is a little''', '''Today, I''', ] __SCREAMING_SNAKE_CASE : Union[str, Any] = tokenizer(_A , return_tensors='''pt''' , padding=_A ) __SCREAMING_SNAKE_CASE : List[Any] = inputs['''input_ids'''].to(_A ) __SCREAMING_SNAKE_CASE : Union[str, Any] = model.generate( input_ids=_A , attention_mask=inputs['''attention_mask'''].to(_A ) , ) __SCREAMING_SNAKE_CASE : Optional[int] = tokenizer(sentences[0] , return_tensors='''pt''' ).input_ids.to(_A ) __SCREAMING_SNAKE_CASE : int = model.generate(input_ids=_A ) __SCREAMING_SNAKE_CASE : Tuple = inputs_non_padded.shape[-1] - inputs['''attention_mask'''][-1].long().sum().cpu().item() __SCREAMING_SNAKE_CASE : Any = tokenizer(sentences[1] , return_tensors='''pt''' ).input_ids.to(_A ) __SCREAMING_SNAKE_CASE : Dict = model.generate(input_ids=_A , max_length=model.config.max_length - num_paddings ) __SCREAMING_SNAKE_CASE : List[Any] = tokenizer.batch_decode(_A , skip_special_tokens=_A ) __SCREAMING_SNAKE_CASE : Optional[int] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=_A ) __SCREAMING_SNAKE_CASE : Union[str, Any] = tokenizer.decode(output_padded[0] , skip_special_tokens=_A ) __SCREAMING_SNAKE_CASE : Any = [ '''Hello, my dog is a little bit bigger than a little bit.''', '''Today, I have a good idea of how to use the information''', ] self.assertListEqual(_A , _A ) self.assertListEqual(_A , [non_padded_sentence, padded_sentence] ) @slow def UpperCAmelCase__ ( self : List[str] ): """simple docstring""" for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE : List[Any] = BioGptModel.from_pretrained(_A ) self.assertIsNotNone(_A ) def UpperCAmelCase__ ( self : int ): """simple docstring""" __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : int = self.model_tester.prepare_config_and_inputs_for_common() __SCREAMING_SNAKE_CASE : Optional[Any] = 3 __SCREAMING_SNAKE_CASE : str = input_dict['''input_ids'''] __SCREAMING_SNAKE_CASE : List[Any] = input_ids.ne(1 ).to(_A ) __SCREAMING_SNAKE_CASE : Tuple = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __SCREAMING_SNAKE_CASE : Union[str, Any] = BioGptForSequenceClassification(_A ) model.to(_A ) model.eval() __SCREAMING_SNAKE_CASE : Dict = model(_A , attention_mask=_A , labels=_A ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def UpperCAmelCase__ ( self : Any ): """simple docstring""" __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : Tuple = self.model_tester.prepare_config_and_inputs_for_common() __SCREAMING_SNAKE_CASE : List[Any] = 3 __SCREAMING_SNAKE_CASE : Optional[int] = '''multi_label_classification''' __SCREAMING_SNAKE_CASE : List[Any] = input_dict['''input_ids'''] __SCREAMING_SNAKE_CASE : Optional[int] = input_ids.ne(1 ).to(_A ) __SCREAMING_SNAKE_CASE : Union[str, Any] = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __SCREAMING_SNAKE_CASE : Optional[int] = BioGptForSequenceClassification(_A ) model.to(_A ) model.eval() __SCREAMING_SNAKE_CASE : List[str] = model(_A , attention_mask=_A , labels=_A ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class __UpperCamelCase ( unittest.TestCase ): """simple docstring""" @slow def UpperCAmelCase__ ( self : Tuple ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = BioGptForCausalLM.from_pretrained('''microsoft/biogpt''' ) __SCREAMING_SNAKE_CASE : Optional[Any] = torch.tensor([[2, 4805, 9, 656, 21]] ) __SCREAMING_SNAKE_CASE : Tuple = model(_A )[0] __SCREAMING_SNAKE_CASE : int = 4_2384 __SCREAMING_SNAKE_CASE : List[Any] = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , _A ) __SCREAMING_SNAKE_CASE : List[str] = torch.tensor( [[[-9.52_36, -9.89_18, 10.45_57], [-11.04_69, -9.64_23, 8.10_22], [-8.86_64, -7.88_26, 5.53_25]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , _A , atol=1e-4 ) ) @slow def UpperCAmelCase__ ( self : Any ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = BioGptTokenizer.from_pretrained('''microsoft/biogpt''' ) __SCREAMING_SNAKE_CASE : str = BioGptForCausalLM.from_pretrained('''microsoft/biogpt''' ) model.to(_A ) torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : List[str] = tokenizer('''COVID-19 is''' , return_tensors='''pt''' ).to(_A ) __SCREAMING_SNAKE_CASE : Union[str, Any] = model.generate( **_A , min_length=100 , max_length=1024 , num_beams=5 , early_stopping=_A , ) __SCREAMING_SNAKE_CASE : int = tokenizer.decode(output_ids[0] , skip_special_tokens=_A ) __SCREAMING_SNAKE_CASE : Union[str, Any] = ( '''COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the''' ''' causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and''' ''' territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),''' ''' and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and''' ''' more than 800,000 deaths.''' ) self.assertEqual(_A , _A )
74
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE__ : Any = { """configuration_mobilevit""": ["""MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MobileViTConfig""", """MobileViTOnnxConfig"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : List[str] = ["""MobileViTFeatureExtractor"""] SCREAMING_SNAKE_CASE__ : Union[str, Any] = ["""MobileViTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Dict = [ """MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MobileViTForImageClassification""", """MobileViTForSemanticSegmentation""", """MobileViTModel""", """MobileViTPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Any = [ """TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFMobileViTForImageClassification""", """TFMobileViTForSemanticSegmentation""", """TFMobileViTModel""", """TFMobileViTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mobilevit import MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileViTConfig, MobileViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilevit import MobileViTFeatureExtractor from .image_processing_mobilevit import MobileViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilevit import ( MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel, MobileViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilevit import ( TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileViTForImageClassification, TFMobileViTForSemanticSegmentation, TFMobileViTModel, TFMobileViTPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
79
0
'''simple docstring''' import json import sys def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: with open(lowerCAmelCase__ , encoding='''utf-8''' ) as f: UpperCAmelCase__ : Union[str, Any] = json.load(lowerCAmelCase__ ) UpperCAmelCase__ : str = ['''<details>''', '''<summary>Show updated benchmarks!</summary>''', ''' '''] for benchmark_name in sorted(lowerCAmelCase__ ): UpperCAmelCase__ : int = results[benchmark_name] UpperCAmelCase__ : Optional[int] = benchmark_name.split('''/''' )[-1] output_md.append(F"""### Benchmark: {benchmark_file_name}""" ) UpperCAmelCase__ : Optional[Any] = '''| metric |''' UpperCAmelCase__ : str = '''|--------|''' UpperCAmelCase__ : List[Any] = '''| new / old (diff) |''' for metric_name in sorted(lowerCAmelCase__ ): UpperCAmelCase__ : Union[str, Any] = benchmark_res[metric_name] UpperCAmelCase__ : List[Any] = metric_vals['''new'''] UpperCAmelCase__ : int = metric_vals.get('''old''' , lowerCAmelCase__ ) UpperCAmelCase__ : Union[str, Any] = metric_vals.get('''diff''' , lowerCAmelCase__ ) UpperCAmelCase__ : Any = F""" {new_val:f}""" if isinstance(lowerCAmelCase__ , (int, float) ) else '''None''' if old_val is not None: val_str += F""" / {old_val:f}""" if isinstance(lowerCAmelCase__ , (int, float) ) else "None" if dif_val is not None: val_str += F""" ({dif_val:f})""" if isinstance(lowerCAmelCase__ , (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append('''</details>''' ) with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.writelines('''\n'''.join(lowerCAmelCase__ ) ) if __name__ == "__main__": UpperCamelCase__ = sys.argv[1] UpperCamelCase__ = sys.argv[2] format_json_to_md(input_json_file, output_md_file)
75
from __future__ import annotations SCREAMING_SNAKE_CASE__ : List[str] = 8.988e9 # units = N * m^s * C^-2 def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> dict[str, float]: '''simple docstring''' UpperCAmelCase__ : int = abs(chargea * chargea ) if (force, chargea, chargea, distance).count(0 ) != 1: raise ValueError("""One and only one argument must be 0""" ) if distance < 0: raise ValueError("""Distance cannot be negative""" ) if force == 0: UpperCAmelCase__ : int = COULOMBS_CONSTANT * charge_product / (distance**2) return {"force": force} elif chargea == 0: UpperCAmelCase__ : str = abs(__lowerCamelCase ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge1": chargea} elif chargea == 0: UpperCAmelCase__ : Union[str, Any] = abs(__lowerCamelCase ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge2": chargea} elif distance == 0: UpperCAmelCase__ : Optional[Any] = (COULOMBS_CONSTANT * charge_product / abs(__lowerCamelCase )) ** 0.5 return {"distance": distance} raise ValueError("""Exactly one argument must be 0""" ) if __name__ == "__main__": import doctest doctest.testmod()
79
0
"""simple docstring""" import os import time import warnings from dataclasses import dataclass, field from enum import Enum from typing import List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import logging from ..processors.glue import glue_convert_examples_to_features, glue_output_modes, glue_processors from ..processors.utils import InputFeatures a_ = logging.get_logger(__name__) @dataclass class UpperCAmelCase_ : UpperCamelCase =field(metadata={"help": "The name of the task to train on: " + ", ".join(glue_processors.keys() )} ) UpperCamelCase =field( metadata={"help": "The input data dir. Should contain the .tsv files (or other data files) for the task."} ) UpperCamelCase =field( default=1_28 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) UpperCamelCase =field( default=snake_case , metadata={"help": "Overwrite the cached training and evaluation sets"} ) def _lowerCamelCase ( self ) -> Optional[Any]: __lowercase : List[str] = self.task_name.lower() class UpperCAmelCase_ ( snake_case ): UpperCamelCase ="train" UpperCamelCase ="dev" UpperCamelCase ="test" class UpperCAmelCase_ ( snake_case ): UpperCamelCase =42 UpperCamelCase =42 UpperCamelCase =42 def __init__( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = None , UpperCamelCase_ = Split.train , UpperCamelCase_ = None , ) -> List[str]: warnings.warn( '''This dataset will be removed from the library soon, preprocessing should be handled with the 🤗 Datasets ''' '''library. You can have a look at this example script for pointers: ''' '''https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py''' , UpperCamelCase_ , ) __lowercase : Optional[int] = args __lowercase : Optional[int] = glue_processors[args.task_name]() __lowercase : List[str] = glue_output_modes[args.task_name] if isinstance(UpperCamelCase_ , UpperCamelCase_ ): try: __lowercase : Union[str, Any] = Split[mode] except KeyError: raise KeyError('''mode is not a valid split name''' ) # Load data features from cache or dataset file __lowercase : Union[str, Any] = os.path.join( cache_dir if cache_dir is not None else args.data_dir , F"""cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{args.task_name}""" , ) __lowercase : Optional[int] = self.processor.get_labels() if args.task_name in ["mnli", "mnli-mm"] and tokenizer.__class__.__name__ in ( "RobertaTokenizer", "RobertaTokenizerFast", "XLMRobertaTokenizer", "BartTokenizer", "BartTokenizerFast", ): # HACK(label indices are swapped in RoBERTa pretrained model) __lowercase ,__lowercase : str = label_list[2], label_list[1] __lowercase : Optional[int] = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. __lowercase : List[Any] = cached_features_file + '''.lock''' with FileLock(UpperCamelCase_ ): if os.path.exists(UpperCamelCase_ ) and not args.overwrite_cache: __lowercase : List[Any] = time.time() __lowercase : Dict = torch.load(UpperCamelCase_ ) logger.info( F"""Loading features from cached file {cached_features_file} [took %.3f s]""" , time.time() - start ) else: logger.info(F"""Creating features from dataset file at {args.data_dir}""" ) if mode == Split.dev: __lowercase : Dict = self.processor.get_dev_examples(args.data_dir ) elif mode == Split.test: __lowercase : Tuple = self.processor.get_test_examples(args.data_dir ) else: __lowercase : int = self.processor.get_train_examples(args.data_dir ) if limit_length is not None: __lowercase : str = examples[:limit_length] __lowercase : Optional[int] = glue_convert_examples_to_features( UpperCamelCase_ , UpperCamelCase_ , max_length=args.max_seq_length , label_list=UpperCamelCase_ , output_mode=self.output_mode , ) __lowercase : List[Any] = time.time() torch.save(self.features , UpperCamelCase_ ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( F"""Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]""" ) def __len__( self ) -> str: return len(self.features ) def __getitem__( self , UpperCamelCase_ ) -> InputFeatures: return self.features[i] def _lowerCamelCase ( self ) -> Optional[int]: return self.label_list
76
class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase ): # we need a list not a string, so do something to change the type UpperCAmelCase__ : Dict = arr.split(""",""" ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = [int(self.array[0] )] * len(self.array ) UpperCAmelCase__ : List[str] = [int(self.array[0] )] * len(self.array ) for i in range(1 , len(self.array ) ): UpperCAmelCase__ : Tuple = max( int(self.array[i] ) + sum_value[i - 1] , int(self.array[i] ) ) UpperCAmelCase__ : Union[str, Any] = max(sum_value[i] , rear[i - 1] ) return rear[len(self.array ) - 1] if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Tuple = input("""please input some numbers:""") SCREAMING_SNAKE_CASE__ : Dict = SubArray(whole_array) SCREAMING_SNAKE_CASE__ : Dict = array.solve_sub_array() print(("""the results is:""", re))
79
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) A = { """configuration_deberta""": ["""DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """DebertaConfig""", """DebertaOnnxConfig"""], """tokenization_deberta""": ["""DebertaTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ["""DebertaTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """DebertaForMaskedLM""", """DebertaForQuestionAnswering""", """DebertaForSequenceClassification""", """DebertaForTokenClassification""", """DebertaModel""", """DebertaPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFDebertaForMaskedLM""", """TFDebertaForQuestionAnswering""", """TFDebertaForSequenceClassification""", """TFDebertaForTokenClassification""", """TFDebertaModel""", """TFDebertaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_deberta import DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, DebertaConfig, DebertaOnnxConfig from .tokenization_deberta import DebertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_deberta_fast import DebertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deberta import ( DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, DebertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deberta import ( TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFDebertaForMaskedLM, TFDebertaForQuestionAnswering, TFDebertaForSequenceClassification, TFDebertaForTokenClassification, TFDebertaModel, TFDebertaPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
from ....configuration_utils import PretrainedConfig from ....utils import logging SCREAMING_SNAKE_CASE__ : List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Any = { """Visual-Attention-Network/van-base""": ( """https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json""" ), } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'van' def __init__( self , _lowerCAmelCase=224 , _lowerCAmelCase=3 , _lowerCAmelCase=[7, 3, 3, 3] , _lowerCAmelCase=[4, 2, 2, 2] , _lowerCAmelCase=[64, 128, 320, 512] , _lowerCAmelCase=[3, 3, 12, 3] , _lowerCAmelCase=[8, 8, 4, 4] , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=1e-6 , _lowerCAmelCase=1e-2 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) UpperCAmelCase__ : Tuple = image_size UpperCAmelCase__ : Optional[Any] = num_channels UpperCAmelCase__ : Optional[int] = patch_sizes UpperCAmelCase__ : int = strides UpperCAmelCase__ : Optional[int] = hidden_sizes UpperCAmelCase__ : str = depths UpperCAmelCase__ : Optional[Any] = mlp_ratios UpperCAmelCase__ : List[Any] = hidden_act UpperCAmelCase__ : Tuple = initializer_range UpperCAmelCase__ : Any = layer_norm_eps UpperCAmelCase__ : List[Any] = layer_scale_init_value UpperCAmelCase__ : int = drop_path_rate UpperCAmelCase__ : Dict = dropout_rate
79
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE_: Optional[int] ={'configuration_beit': ['BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BeitConfig', 'BeitOnnxConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_: Optional[int] =['BeitFeatureExtractor'] SCREAMING_SNAKE_CASE_: int =['BeitImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_: Optional[int] =[ 'BEIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BeitForImageClassification', 'BeitForMaskedImageModeling', 'BeitForSemanticSegmentation', 'BeitModel', 'BeitPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_: int =[ 'FlaxBeitForImageClassification', 'FlaxBeitForMaskedImageModeling', 'FlaxBeitModel', 'FlaxBeitPreTrainedModel', ] if TYPE_CHECKING: from .configuration_beit import BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, BeitConfig, BeitOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_beit import BeitFeatureExtractor from .image_processing_beit import BeitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_beit import ( BEIT_PRETRAINED_MODEL_ARCHIVE_LIST, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, BeitPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_beit import ( FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel, FlaxBeitPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE_: Dict =_LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
78
import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Any: '''simple docstring''' UpperCAmelCase__ : List[str] = s.rsplit(__lowerCamelCase , __lowerCamelCase ) return new.join(__lowerCamelCase ) def _lowerCamelCase ( __lowerCamelCase ) -> str: '''simple docstring''' # encoder.embeddings are double copied in original FLAVA return sum(param.float().sum() if """encoder.embeddings""" not in key else 0 for key, param in state_dict.items() ) def _lowerCamelCase ( __lowerCamelCase ) -> int: '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = {} UpperCAmelCase__ : Union[str, Any] = ["""group_1""", """group_2""", """group_3""", """group_4"""] for key, value in state_dict.items(): for group_key in group_keys: if group_key in key: UpperCAmelCase__ : Optional[Any] = key.replace(F"{group_key}." , F"{group_key}.group." ) if "res_path" in key: UpperCAmelCase__ : Optional[int] = key.replace("""res_path.""" , """res_path.path.""" ) if key.endswith(""".w""" ): UpperCAmelCase__ : List[Any] = rreplace(__lowerCamelCase , """.w""" , """.weight""" , 1 ) if key.endswith(""".b""" ): UpperCAmelCase__ : Optional[int] = rreplace(__lowerCamelCase , """.b""" , """.bias""" , 1 ) UpperCAmelCase__ : Union[str, Any] = value.float() return upgrade @torch.no_grad() def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=None , __lowerCamelCase=True ) -> str: '''simple docstring''' from dall_e import Encoder UpperCAmelCase__ : Dict = Encoder() if os.path.exists(__lowerCamelCase ): UpperCAmelCase__ : Optional[Any] = torch.load(__lowerCamelCase ) else: UpperCAmelCase__ : Tuple = torch.hub.load_state_dict_from_url(__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase__ : Any = ckpt.state_dict() encoder.load_state_dict(__lowerCamelCase ) if config_path is not None: UpperCAmelCase__ : Dict = FlavaImageCodebookConfig.from_pretrained(__lowerCamelCase ) else: UpperCAmelCase__ : Optional[Any] = FlavaImageCodebookConfig() UpperCAmelCase__ : Optional[Any] = FlavaImageCodebook(__lowerCamelCase ).eval() UpperCAmelCase__ : str = encoder.state_dict() UpperCAmelCase__ : Optional[int] = upgrade_state_dict(__lowerCamelCase ) hf_model.load_state_dict(__lowerCamelCase ) UpperCAmelCase__ : List[str] = hf_model.state_dict() UpperCAmelCase__ : Tuple = count_parameters(__lowerCamelCase ) UpperCAmelCase__ : int = count_parameters(__lowerCamelCase ) assert torch.allclose(__lowerCamelCase , __lowerCamelCase , atol=1E-3 ) if save_checkpoint: hf_model.save_pretrained(__lowerCamelCase ) else: return hf_state_dict if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Any = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to flava checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") SCREAMING_SNAKE_CASE__ : int = parser.parse_args() convert_dalle_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
79
0
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class __UpperCamelCase ( _lowerCAmelCase ): def __init__( self : List[str] , _lowerCAmelCase : pyspark.sql.DataFrame , _lowerCAmelCase : Optional[NamedSplit] = None , _lowerCAmelCase : Optional[Features] = None , _lowerCAmelCase : bool = True , _lowerCAmelCase : str = None , _lowerCAmelCase : bool = False , _lowerCAmelCase : str = None , _lowerCAmelCase : bool = True , _lowerCAmelCase : str = "arrow" , **_lowerCAmelCase : List[str] , ) -> Optional[Any]: """simple docstring""" super().__init__( split=_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase , streaming=_lowerCAmelCase , **_lowerCAmelCase , ) __lowercase = load_from_cache_file __lowercase = file_format __lowercase = Spark( df=_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase , working_dir=_lowerCAmelCase , **_lowerCAmelCase , ) def _a ( self : Dict ) -> str: """simple docstring""" if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) __lowercase = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=_lowerCAmelCase , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
80
def _lowerCamelCase ( __lowerCamelCase ) -> int: '''simple docstring''' return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def _lowerCamelCase ( __lowerCamelCase ) -> bool: '''simple docstring''' UpperCAmelCase__ : Any = 0 UpperCAmelCase__ : Union[str, Any] = number while duplicate > 0: UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = divmod(__lowerCamelCase , 10 ) fact_sum += factorial(__lowerCamelCase ) return fact_sum == number if __name__ == "__main__": print("""Program to check whether a number is a Krisnamurthy Number or not.""") SCREAMING_SNAKE_CASE__ : Optional[Any] = int(input("""Enter number: """).strip()) print( f'''{number} is {"" if krishnamurthy(number) else "not "}a Krishnamurthy Number.''' )
79
0
import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, BartConfig, BertConfig, CamembertConfig, CTRLConfig, DistilBertConfig, DPRConfig, ElectraConfig, FlaubertConfig, GPTaConfig, LayoutLMConfig, LxmertConfig, OpenAIGPTConfig, RobertaConfig, TaConfig, TFAlbertForPreTraining, TFBartForConditionalGeneration, TFBartForSequenceClassification, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFCamembertForMaskedLM, TFCTRLLMHeadModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, TFElectraForPreTraining, TFFlaubertWithLMHeadModel, TFGPTaLMHeadModel, TFLayoutLMForMaskedLM, TFLxmertForPreTraining, TFLxmertVisualFeatureEncoder, TFOpenAIGPTLMHeadModel, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForSequenceClassification, TFTaForConditionalGeneration, TFTransfoXLLMHeadModel, TFWavaVecaModel, TFXLMRobertaForMaskedLM, TFXLMWithLMHeadModel, TFXLNetLMHeadModel, TransfoXLConfig, WavaVecaConfig, WavaVecaModel, XLMConfig, XLMRobertaConfig, XLNetConfig, is_torch_available, load_pytorch_checkpoint_in_tfa_model, ) from .utils import CONFIG_NAME, WEIGHTS_NAME, cached_file, logging if is_torch_available(): import numpy as np import torch from . import ( AlbertForPreTraining, BartForConditionalGeneration, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, CamembertForMaskedLM, CTRLLMHeadModel, DistilBertForMaskedLM, DistilBertForQuestionAnswering, DPRContextEncoder, DPRQuestionEncoder, DPRReader, ElectraForPreTraining, FlaubertWithLMHeadModel, GPTaLMHeadModel, LayoutLMForMaskedLM, LxmertForPreTraining, LxmertVisualFeatureEncoder, OpenAIGPTLMHeadModel, RobertaForMaskedLM, RobertaForSequenceClassification, TaForConditionalGeneration, TransfoXLLMHeadModel, XLMRobertaForMaskedLM, XLMWithLMHeadModel, XLNetLMHeadModel, ) logging.set_verbosity_info() _snake_case : Dict = { "bart": ( BartConfig, TFBartForConditionalGeneration, TFBartForSequenceClassification, BartForConditionalGeneration, BART_PRETRAINED_MODEL_ARCHIVE_LIST, ), "bert": ( BertConfig, TFBertForPreTraining, BertForPreTraining, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "bert-large-uncased-whole-word-masking-finetuned-squad": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "bert-large-cased-whole-word-masking-finetuned-squad": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "bert-base-cased-finetuned-mrpc": ( BertConfig, TFBertForSequenceClassification, BertForSequenceClassification, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "dpr": ( DPRConfig, TFDPRQuestionEncoder, TFDPRContextEncoder, TFDPRReader, DPRQuestionEncoder, DPRContextEncoder, DPRReader, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ), "gpt2": ( GPTaConfig, TFGPTaLMHeadModel, GPTaLMHeadModel, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "xlnet": ( XLNetConfig, TFXLNetLMHeadModel, XLNetLMHeadModel, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "xlm": ( XLMConfig, TFXLMWithLMHeadModel, XLMWithLMHeadModel, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "xlm-roberta": ( XLMRobertaConfig, TFXLMRobertaForMaskedLM, XLMRobertaForMaskedLM, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "transfo-xl": ( TransfoXLConfig, TFTransfoXLLMHeadModel, TransfoXLLMHeadModel, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "openai-gpt": ( OpenAIGPTConfig, TFOpenAIGPTLMHeadModel, OpenAIGPTLMHeadModel, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "roberta": ( RobertaConfig, TFRobertaForCausalLM, TFRobertaForMaskedLM, RobertaForMaskedLM, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "layoutlm": ( LayoutLMConfig, TFLayoutLMForMaskedLM, LayoutLMForMaskedLM, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, ), "roberta-large-mnli": ( RobertaConfig, TFRobertaForSequenceClassification, RobertaForSequenceClassification, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "camembert": ( CamembertConfig, TFCamembertForMaskedLM, CamembertForMaskedLM, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "flaubert": ( FlaubertConfig, TFFlaubertWithLMHeadModel, FlaubertWithLMHeadModel, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "distilbert": ( DistilBertConfig, TFDistilBertForMaskedLM, DistilBertForMaskedLM, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "distilbert-base-distilled-squad": ( DistilBertConfig, TFDistilBertForQuestionAnswering, DistilBertForQuestionAnswering, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "lxmert": ( LxmertConfig, TFLxmertForPreTraining, LxmertForPreTraining, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "lxmert-visual-feature-encoder": ( LxmertConfig, TFLxmertVisualFeatureEncoder, LxmertVisualFeatureEncoder, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "ctrl": ( CTRLConfig, TFCTRLLMHeadModel, CTRLLMHeadModel, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "albert": ( AlbertConfig, TFAlbertForPreTraining, AlbertForPreTraining, ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "t5": ( TaConfig, TFTaForConditionalGeneration, TaForConditionalGeneration, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "electra": ( ElectraConfig, TFElectraForPreTraining, ElectraForPreTraining, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "wav2vec2": ( WavaVecaConfig, TFWavaVecaModel, WavaVecaModel, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), } def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=False , __lowerCamelCase=True ): if model_type not in MODEL_CLASSES: raise ValueError(F'Unrecognized model type, should be one of {list(MODEL_CLASSES.keys() )}.' ) __snake_case , __snake_case , __snake_case , __snake_case : Optional[Any] = MODEL_CLASSES[model_type] # Initialise TF model if config_file in aws_config_map: __snake_case : List[Any] = cached_file(__lowerCamelCase , __lowerCamelCase , force_download=not use_cached_models ) __snake_case : Tuple = config_class.from_json_file(__lowerCamelCase ) __snake_case : Optional[int] = True __snake_case : Optional[int] = True print(F'Building TensorFlow model from configuration: {config}' ) __snake_case : str = model_class(__lowerCamelCase ) # Load weights from tf checkpoint if pytorch_checkpoint_path in aws_config_map.keys(): __snake_case : str = cached_file( __lowerCamelCase , __lowerCamelCase , force_download=not use_cached_models ) # Load PyTorch checkpoint in tf2 model: __snake_case : int = load_pytorch_checkpoint_in_tfa_model(__lowerCamelCase , __lowerCamelCase ) if compare_with_pt_model: __snake_case : Any = tf_model(tf_model.dummy_inputs , training=__lowerCamelCase ) # build the network __snake_case : Tuple = torch.load(__lowerCamelCase , map_location="cpu" ) __snake_case : Tuple = pt_model_class.from_pretrained( pretrained_model_name_or_path=__lowerCamelCase , config=__lowerCamelCase , state_dict=__lowerCamelCase ) with torch.no_grad(): __snake_case : List[Any] = pt_model(**pt_model.dummy_inputs ) __snake_case : Optional[int] = pto[0].numpy() __snake_case : Union[str, Any] = tfo[0].numpy() __snake_case : Dict = np.amax(np.abs(np_pt - np_tf ) ) print(F'Max absolute difference between models outputs {diff}' ) assert diff <= 2e-2, F'Error, model absolute difference is >2e-2: {diff}' # Save pytorch-model print(F'Save TensorFlow model to {tf_dump_path}' ) tf_model.save_weights(__lowerCamelCase , save_format="h5" ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=False , __lowerCamelCase=False , __lowerCamelCase=False , __lowerCamelCase=False , ): if args_model_type is None: __snake_case : Any = list(MODEL_CLASSES.keys() ) else: __snake_case : Optional[int] = [args_model_type] for j, model_type in enumerate(__lowerCamelCase , start=1 ): print("=" * 1_0_0 ) print(F' Converting model type {j}/{len(__lowerCamelCase )}: {model_type}' ) print("=" * 1_0_0 ) if model_type not in MODEL_CLASSES: raise ValueError(F'Unrecognized model type {model_type}, should be one of {list(MODEL_CLASSES.keys() )}.' ) __snake_case , __snake_case , __snake_case , __snake_case , __snake_case : Any = MODEL_CLASSES[model_type] if model_shortcut_names_or_path is None: __snake_case : Optional[int] = list(aws_model_maps.keys() ) if config_shortcut_names_or_path is None: __snake_case : List[str] = model_shortcut_names_or_path for i, (model_shortcut_name, config_shortcut_name) in enumerate( zip(__lowerCamelCase , __lowerCamelCase ) , start=1 ): print("-" * 1_0_0 ) if "-squad" in model_shortcut_name or "-mrpc" in model_shortcut_name or "-mnli" in model_shortcut_name: if not only_convert_finetuned_models: print(F' Skipping finetuned checkpoint {model_shortcut_name}' ) continue __snake_case : Dict = model_shortcut_name elif only_convert_finetuned_models: print(F' Skipping not finetuned checkpoint {model_shortcut_name}' ) continue print( F' Converting checkpoint {i}/{len(__lowerCamelCase )}: {model_shortcut_name} - model_type {model_type}' ) print("-" * 1_0_0 ) if config_shortcut_name in aws_config_map: __snake_case : List[str] = cached_file(__lowerCamelCase , __lowerCamelCase , force_download=not use_cached_models ) else: __snake_case : List[Any] = config_shortcut_name if model_shortcut_name in aws_model_maps: __snake_case : str = cached_file(__lowerCamelCase , __lowerCamelCase , force_download=not use_cached_models ) else: __snake_case : str = model_shortcut_name if os.path.isfile(__lowerCamelCase ): __snake_case : List[Any] = "converted_model" convert_pt_checkpoint_to_tf( model_type=__lowerCamelCase , pytorch_checkpoint_path=__lowerCamelCase , config_file=__lowerCamelCase , tf_dump_path=os.path.join(__lowerCamelCase , model_shortcut_name + "-tf_model.h5" ) , compare_with_pt_model=__lowerCamelCase , ) if remove_cached_files: os.remove(__lowerCamelCase ) os.remove(__lowerCamelCase ) if __name__ == "__main__": _snake_case : int = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_dump_path", default=None, type=str, required=True, help="Path to the output Tensorflow dump file." ) parser.add_argument( "--model_type", default=None, type=str, help=( f'''Model type selected in the list of {list(MODEL_CLASSES.keys())}. If not given, will download and ''' "convert all the models from AWS." ), ) parser.add_argument( "--pytorch_checkpoint_path", default=None, type=str, help=( "Path to the PyTorch checkpoint path or shortcut name to download from AWS. " "If not given, will download and convert all the checkpoints from AWS." ), ) parser.add_argument( "--config_file", default=None, type=str, help=( "The config json file corresponding to the pre-trained model. \n" "This specifies the model architecture. If not given and " "--pytorch_checkpoint_path is not given or is a shortcut name " "use the configuration associated to the shortcut name on the AWS" ), ) parser.add_argument( "--compare_with_pt_model", action="store_true", help="Compare Tensorflow and PyTorch model predictions." ) parser.add_argument( "--use_cached_models", action="store_true", help="Use cached models if possible instead of updating to latest checkpoint versions.", ) parser.add_argument( "--remove_cached_files", action="store_true", help="Remove pytorch models after conversion (save memory when converting in batches).", ) parser.add_argument("--only_convert_finetuned_models", action="store_true", help="Only convert finetuned models.") _snake_case : str = parser.parse_args() # if args.pytorch_checkpoint_path is not None: # convert_pt_checkpoint_to_tf(args.model_type.lower(), # args.pytorch_checkpoint_path, # args.config_file if args.config_file is not None else args.pytorch_checkpoint_path, # args.tf_dump_path, # compare_with_pt_model=args.compare_with_pt_model, # use_cached_models=args.use_cached_models) # else: convert_all_pt_checkpoints_to_tf( args.model_type.lower() if args.model_type is not None else None, args.tf_dump_path, model_shortcut_names_or_path=[args.pytorch_checkpoint_path] if args.pytorch_checkpoint_path is not None else None, config_shortcut_names_or_path=[args.config_file] if args.config_file is not None else None, compare_with_pt_model=args.compare_with_pt_model, use_cached_models=args.use_cached_models, remove_cached_files=args.remove_cached_files, only_convert_finetuned_models=args.only_convert_finetuned_models, )
81
def _lowerCamelCase ( __lowerCamelCase = 100_0000 ) -> int: '''simple docstring''' UpperCAmelCase__ : Tuple = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , __lowerCamelCase ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
79
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase = { """configuration_longformer""": [ """LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LongformerConfig""", """LongformerOnnxConfig""", ], """tokenization_longformer""": ["""LongformerTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = ["""LongformerTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ """LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """LongformerForMaskedLM""", """LongformerForMultipleChoice""", """LongformerForQuestionAnswering""", """LongformerForSequenceClassification""", """LongformerForTokenClassification""", """LongformerModel""", """LongformerPreTrainedModel""", """LongformerSelfAttention""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ """TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFLongformerForMaskedLM""", """TFLongformerForMultipleChoice""", """TFLongformerForQuestionAnswering""", """TFLongformerForSequenceClassification""", """TFLongformerForTokenClassification""", """TFLongformerModel""", """TFLongformerPreTrainedModel""", """TFLongformerSelfAttention""", ] if TYPE_CHECKING: from .configuration_longformer import ( LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, LongformerConfig, LongformerOnnxConfig, ) from .tokenization_longformer import LongformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_longformer_fast import LongformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longformer import ( LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, LongformerForMaskedLM, LongformerForMultipleChoice, LongformerForQuestionAnswering, LongformerForSequenceClassification, LongformerForTokenClassification, LongformerModel, LongformerPreTrainedModel, LongformerSelfAttention, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_longformer import ( TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFLongformerForMaskedLM, TFLongformerForMultipleChoice, TFLongformerForQuestionAnswering, TFLongformerForSequenceClassification, TFLongformerForTokenClassification, TFLongformerModel, TFLongformerPreTrainedModel, TFLongformerSelfAttention, ) else: import sys lowerCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
82
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Tuple = { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json""", # See all REALM models at https://huggingface.co/models?filter=realm } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'realm' def __init__( self , _lowerCAmelCase=30522 , _lowerCAmelCase=768 , _lowerCAmelCase=128 , _lowerCAmelCase=12 , _lowerCAmelCase=12 , _lowerCAmelCase=8 , _lowerCAmelCase=3072 , _lowerCAmelCase="gelu_new" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=512 , _lowerCAmelCase=2 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=1e-12 , _lowerCAmelCase=256 , _lowerCAmelCase=10 , _lowerCAmelCase=1e-3 , _lowerCAmelCase=5 , _lowerCAmelCase=320 , _lowerCAmelCase=13353718 , _lowerCAmelCase=5000 , _lowerCAmelCase=1 , _lowerCAmelCase=0 , _lowerCAmelCase=2 , **_lowerCAmelCase , ): super().__init__(pad_token_id=_lowerCAmelCase , bos_token_id=_lowerCAmelCase , eos_token_id=_lowerCAmelCase , **_lowerCAmelCase ) # Common config UpperCAmelCase__ : List[Any] = vocab_size UpperCAmelCase__ : Dict = max_position_embeddings UpperCAmelCase__ : Any = hidden_size UpperCAmelCase__ : str = retriever_proj_size UpperCAmelCase__ : Tuple = num_hidden_layers UpperCAmelCase__ : List[str] = num_attention_heads UpperCAmelCase__ : List[Any] = num_candidates UpperCAmelCase__ : str = intermediate_size UpperCAmelCase__ : str = hidden_act UpperCAmelCase__ : Optional[Any] = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : Union[str, Any] = initializer_range UpperCAmelCase__ : Any = type_vocab_size UpperCAmelCase__ : Optional[Any] = layer_norm_eps # Reader config UpperCAmelCase__ : str = span_hidden_size UpperCAmelCase__ : Union[str, Any] = max_span_width UpperCAmelCase__ : List[str] = reader_layer_norm_eps UpperCAmelCase__ : Dict = reader_beam_size UpperCAmelCase__ : Union[str, Any] = reader_seq_len # Retrieval config UpperCAmelCase__ : List[Any] = num_block_records UpperCAmelCase__ : List[Any] = searcher_beam_size
79
0
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union lowerCAmelCase__ = TypeVar('''T''') lowerCAmelCase__ = Union[List[T], Tuple[T, ...]] lowerCAmelCase__ = Union[T, List[T], Dict[str, T]] lowerCAmelCase__ = Union[str, bytes, os.PathLike]
83
import gc import unittest from parameterized import parameterized from diffusers import FlaxUNetaDConditionModel from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp @slow @require_flax class UpperCAmelCase_ ( unittest.TestCase ): def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase ): return f"gaussian_noise_s={seed}_shape={'_'.join([str(_lowerCAmelCase ) for s in shape] )}.npy" def __UpperCAmelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def __UpperCAmelCase ( self , _lowerCAmelCase=0 , _lowerCAmelCase=(4, 4, 64, 64) , _lowerCAmelCase=False ): UpperCAmelCase__ : Optional[Any] = jnp.bfloataa if fpaa else jnp.floataa UpperCAmelCase__ : Union[str, Any] = jnp.array(load_hf_numpy(self.get_file_format(_lowerCAmelCase , _lowerCAmelCase ) ) , dtype=_lowerCAmelCase ) return image def __UpperCAmelCase ( self , _lowerCAmelCase=False , _lowerCAmelCase="CompVis/stable-diffusion-v1-4" ): UpperCAmelCase__ : int = jnp.bfloataa if fpaa else jnp.floataa UpperCAmelCase__ : Optional[Any] = """bf16""" if fpaa else None UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = FlaxUNetaDConditionModel.from_pretrained( _lowerCAmelCase , subfolder="""unet""" , dtype=_lowerCAmelCase , revision=_lowerCAmelCase ) return model, params def __UpperCAmelCase ( self , _lowerCAmelCase=0 , _lowerCAmelCase=(4, 77, 768) , _lowerCAmelCase=False ): UpperCAmelCase__ : Optional[int] = jnp.bfloataa if fpaa else jnp.floataa UpperCAmelCase__ : Optional[int] = jnp.array(load_hf_numpy(self.get_file_format(_lowerCAmelCase , _lowerCAmelCase ) ) , dtype=_lowerCAmelCase ) return hidden_states @parameterized.expand( [ # fmt: off [83, 4, [-0.2_3_2_3, -0.1_3_0_4, 0.0_8_1_3, -0.3_0_9_3, -0.0_9_1_9, -0.1_5_7_1, -0.1_1_2_5, -0.5_8_0_6]], [17, 0.5_5, [-0.0_8_3_1, -0.2_4_4_3, 0.0_9_0_1, -0.0_9_1_9, 0.3_3_9_6, 0.0_1_0_3, -0.3_7_4_3, 0.0_7_0_1]], [8, 0.8_9, [-0.4_8_6_3, 0.0_8_5_9, 0.0_8_7_5, -0.1_6_5_8, 0.9_1_9_9, -0.0_1_1_4, 0.4_8_3_9, 0.4_6_3_9]], [3, 1000, [-0.5_6_4_9, 0.2_4_0_2, -0.5_5_1_8, 0.1_2_4_8, 1.1_3_2_8, -0.2_4_4_3, -0.0_3_2_5, -1.0_0_7_8]], # fmt: on ] ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.get_unet_model(model_id="""CompVis/stable-diffusion-v1-4""" , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = self.get_latents(_lowerCAmelCase , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Dict = self.get_encoder_hidden_states(_lowerCAmelCase , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = model.apply( {"""params""": params} , _lowerCAmelCase , jnp.array(_lowerCAmelCase , dtype=jnp.intaa ) , encoder_hidden_states=_lowerCAmelCase , ).sample assert sample.shape == latents.shape UpperCAmelCase__ : Dict = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) UpperCAmelCase__ : List[Any] = jnp.array(_lowerCAmelCase , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, in the same hardware assert jnp.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-2 ) @parameterized.expand( [ # fmt: off [83, 4, [0.1_5_1_4, 0.0_8_0_7, 0.1_6_2_4, 0.1_0_1_6, -0.1_8_9_6, 0.0_2_6_3, 0.0_6_7_7, 0.2_3_1_0]], [17, 0.5_5, [0.1_1_6_4, -0.0_2_1_6, 0.0_1_7_0, 0.1_5_8_9, -0.3_1_2_0, 0.1_0_0_5, -0.0_5_8_1, -0.1_4_5_8]], [8, 0.8_9, [-0.1_7_5_8, -0.0_1_6_9, 0.1_0_0_4, -0.1_4_1_1, 0.1_3_1_2, 0.1_1_0_3, -0.1_9_9_6, 0.2_1_3_9]], [3, 1000, [0.1_2_1_4, 0.0_3_5_2, -0.0_7_3_1, -0.1_5_6_2, -0.0_9_9_4, -0.0_9_0_6, -0.2_3_4_0, -0.0_5_3_9]], # fmt: on ] ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.get_unet_model(model_id="""stabilityai/stable-diffusion-2""" , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = self.get_latents(_lowerCAmelCase , shape=(4, 4, 96, 96) , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Any = self.get_encoder_hidden_states(_lowerCAmelCase , shape=(4, 77, 1024) , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Dict = model.apply( {"""params""": params} , _lowerCAmelCase , jnp.array(_lowerCAmelCase , dtype=jnp.intaa ) , encoder_hidden_states=_lowerCAmelCase , ).sample assert sample.shape == latents.shape UpperCAmelCase__ : Any = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) UpperCAmelCase__ : Any = jnp.array(_lowerCAmelCase , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, on the same hardware assert jnp.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-2 )
79
0
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A_ ( __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : List[str] = LongformerTokenizer _UpperCamelCase : Any = True _UpperCamelCase : Optional[Any] = LongformerTokenizerFast _UpperCamelCase : Union[str, Any] = True def SCREAMING_SNAKE_CASE__ ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowercase = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] lowercase = dict(zip(snake_case , range(len(snake_case ) ) ) ) lowercase = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] lowercase = {'unk_token': '<unk>'} lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(snake_case ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(snake_case ) ) def SCREAMING_SNAKE_CASE__ ( self , **snake_case ): kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self , **snake_case ): kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case ): lowercase = 'lower newer' lowercase = 'lower newer' return input_text, output_text def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) lowercase = 'lower newer' lowercase = ['l', 'o', 'w', 'er', '\u0120', 'n', 'e', 'w', 'er'] lowercase = tokenizer.tokenize(snake_case ) # , add_prefix_space=True) self.assertListEqual(snake_case , snake_case ) lowercase = tokens + [tokenizer.unk_token] lowercase = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case ) , snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.get_tokenizer() self.assertListEqual(tokenizer.encode('Hello world!' , add_special_tokens=snake_case ) , [0, 3_1414, 232, 328, 2] ) self.assertListEqual( tokenizer.encode('Hello world! cécé herlolip 418' , add_special_tokens=snake_case ) , [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2] , ) @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.tokenizer_class.from_pretrained('allenai/longformer-base-4096' ) lowercase = tokenizer.encode('sequence builders' , add_special_tokens=snake_case ) lowercase = tokenizer.encode('multi-sequence build' , add_special_tokens=snake_case ) lowercase = tokenizer.encode( 'sequence builders' , add_special_tokens=snake_case , add_prefix_space=snake_case ) lowercase = tokenizer.encode( 'sequence builders' , 'multi-sequence build' , add_special_tokens=snake_case , add_prefix_space=snake_case ) lowercase = tokenizer.build_inputs_with_special_tokens(snake_case ) lowercase = tokenizer.build_inputs_with_special_tokens(snake_case , snake_case ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.get_tokenizer() lowercase = 'Encode this sequence.' lowercase = tokenizer.byte_encoder[' '.encode('utf-8' )[0]] # Testing encoder arguments lowercase = tokenizer.encode(snake_case , add_special_tokens=snake_case , add_prefix_space=snake_case ) lowercase = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(snake_case , snake_case ) lowercase = tokenizer.encode(snake_case , add_special_tokens=snake_case , add_prefix_space=snake_case ) lowercase = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(snake_case , snake_case ) tokenizer.add_special_tokens({'bos_token': '<s>'} ) lowercase = tokenizer.encode(snake_case , add_special_tokens=snake_case ) lowercase = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(snake_case , snake_case ) # Testing spaces after special tokens lowercase = '<mask>' tokenizer.add_special_tokens( {'mask_token': AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case )} ) # mask token has a left space lowercase = tokenizer.convert_tokens_to_ids(snake_case ) lowercase = 'Encode <mask> sequence' lowercase = 'Encode <mask>sequence' lowercase = tokenizer.encode(snake_case ) lowercase = encoded.index(snake_case ) lowercase = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(snake_case , snake_case ) lowercase = tokenizer.encode(snake_case ) lowercase = encoded.index(snake_case ) lowercase = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): pass def SCREAMING_SNAKE_CASE__ ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowercase = self.rust_tokenizer_class.from_pretrained(snake_case , **snake_case ) lowercase = self.tokenizer_class.from_pretrained(snake_case , **snake_case ) lowercase = 'A, <mask> AllenNLP sentence.' lowercase = tokenizer_r.encode_plus(snake_case , add_special_tokens=snake_case , return_token_type_ids=snake_case ) lowercase = tokenizer_p.encode_plus(snake_case , add_special_tokens=snake_case , return_token_type_ids=snake_case ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r['token_type_ids'] ) , sum(tokens_p['token_type_ids'] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r['attention_mask'] ) / len(tokens_r['attention_mask'] ) , sum(tokens_p['attention_mask'] ) / len(tokens_p['attention_mask'] ) , ) lowercase = tokenizer_r.convert_ids_to_tokens(tokens_r['input_ids'] ) lowercase = tokenizer_p.convert_ids_to_tokens(tokens_p['input_ids'] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p['input_ids'] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r['input_ids'] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual( snake_case , ['<s>', 'A', ',', '<mask>', 'ĠAllen', 'N', 'LP', 'Ġsentence', '.', '</s>'] ) self.assertSequenceEqual( snake_case , ['<s>', 'A', ',', '<mask>', 'ĠAllen', 'N', 'LP', 'Ġsentence', '.', '</s>'] ) def SCREAMING_SNAKE_CASE__ ( self ): for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): lowercase = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=snake_case , add_prefix_space=snake_case , trim_offsets=snake_case ) lowercase = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) lowercase = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state['add_prefix_space'] , snake_case ) self.assertEqual(post_processor_state['add_prefix_space'] , snake_case ) self.assertEqual(post_processor_state['trim_offsets'] , snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): # Test which aims to verify that the offsets are well adapted to the argument `add_prefix_space` and # `trim_offsets` for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowercase = 'hello' # `hello` is a token in the vocabulary of `pretrained_name` lowercase = F'''{text_of_1_token} {text_of_1_token}''' lowercase = self.rust_tokenizer_class.from_pretrained( snake_case , use_fast=snake_case , add_prefix_space=snake_case , trim_offsets=snake_case ) lowercase = tokenizer_r(snake_case , return_offsets_mapping=snake_case , add_special_tokens=snake_case ) self.assertEqual(encoding.offset_mapping[0] , (0, len(snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (len(snake_case ) + 1, len(snake_case ) + 1 + len(snake_case )) , ) lowercase = self.rust_tokenizer_class.from_pretrained( snake_case , use_fast=snake_case , add_prefix_space=snake_case , trim_offsets=snake_case ) lowercase = tokenizer_r(snake_case , return_offsets_mapping=snake_case , add_special_tokens=snake_case ) self.assertEqual(encoding.offset_mapping[0] , (0, len(snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (len(snake_case ) + 1, len(snake_case ) + 1 + len(snake_case )) , ) lowercase = self.rust_tokenizer_class.from_pretrained( snake_case , use_fast=snake_case , add_prefix_space=snake_case , trim_offsets=snake_case ) lowercase = tokenizer_r(snake_case , return_offsets_mapping=snake_case , add_special_tokens=snake_case ) self.assertEqual(encoding.offset_mapping[0] , (0, len(snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (len(snake_case ), len(snake_case ) + 1 + len(snake_case )) , ) lowercase = self.rust_tokenizer_class.from_pretrained( snake_case , use_fast=snake_case , add_prefix_space=snake_case , trim_offsets=snake_case ) lowercase = tokenizer_r(snake_case , return_offsets_mapping=snake_case , add_special_tokens=snake_case ) self.assertEqual(encoding.offset_mapping[0] , (0, len(snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (len(snake_case ), len(snake_case ) + 1 + len(snake_case )) , ) lowercase = F''' {text}''' # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) lowercase = self.rust_tokenizer_class.from_pretrained( snake_case , use_fast=snake_case , add_prefix_space=snake_case , trim_offsets=snake_case ) lowercase = tokenizer_r(snake_case , return_offsets_mapping=snake_case , add_special_tokens=snake_case ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(snake_case ) + 1, 1 + len(snake_case ) + 1 + len(snake_case )) , ) lowercase = self.rust_tokenizer_class.from_pretrained( snake_case , use_fast=snake_case , add_prefix_space=snake_case , trim_offsets=snake_case ) lowercase = tokenizer_r(snake_case , return_offsets_mapping=snake_case , add_special_tokens=snake_case ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(snake_case ), 1 + len(snake_case ) + 1 + len(snake_case )) , ) lowercase = self.rust_tokenizer_class.from_pretrained( snake_case , use_fast=snake_case , add_prefix_space=snake_case , trim_offsets=snake_case ) lowercase = tokenizer_r(snake_case , return_offsets_mapping=snake_case , add_special_tokens=snake_case ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(snake_case )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(snake_case ), 1 + len(snake_case ) + 1 + len(snake_case )) , )
84
import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class UpperCAmelCase_ ( unittest.TestCase ): def __init__( self , _lowerCAmelCase , _lowerCAmelCase=7 , _lowerCAmelCase=3 , _lowerCAmelCase=18 , _lowerCAmelCase=30 , _lowerCAmelCase=400 , _lowerCAmelCase=True , _lowerCAmelCase=None , _lowerCAmelCase=True , ): UpperCAmelCase__ : List[str] = size if size is not None else {"""height""": 18, """width""": 18} UpperCAmelCase__ : Union[str, Any] = parent UpperCAmelCase__ : int = batch_size UpperCAmelCase__ : Tuple = num_channels UpperCAmelCase__ : Dict = image_size UpperCAmelCase__ : List[Any] = min_resolution UpperCAmelCase__ : str = max_resolution UpperCAmelCase__ : Union[str, Any] = do_resize UpperCAmelCase__ : Tuple = size UpperCAmelCase__ : int = do_normalize def __UpperCAmelCase ( self ): return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.8_8_6_6_4_4_3_6_3_4_0_3_3_2_0_3, 0.6_6_1_8_8_2_9_3_6_9_5_4_4_9_8_3, 0.3_8_9_1_7_4_6_4_0_1_7_8_6_8_0_4], [-0.6_0_4_2_5_5_9_1_4_6_8_8_1_1_0_4, -0.0_2_2_9_5_0_0_8_8_6_0_5_2_8_4_6_9, 0.5_4_2_3_7_9_7_3_6_9_0_0_3_2_9_6], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class UpperCAmelCase_ ( __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = ImageGPTImageProcessor if is_vision_available() else None def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = ImageGPTImageProcessingTester(self ) @property def __UpperCAmelCase ( self ): return self.image_processor_tester.prepare_image_processor_dict() def __UpperCAmelCase ( self ): UpperCAmelCase__ : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """clusters""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """do_normalize""" ) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 18} ) UpperCAmelCase__ : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) UpperCAmelCase__ : Optional[int] = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase , obj[key] ) ) else: self.assertEqual(obj[key] , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase__ : Union[str, Any] = os.path.join(_lowerCAmelCase , """image_processor.json""" ) image_processor_first.to_json_file(_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = self.image_processing_class.from_json_file(_lowerCAmelCase ).to_dict() UpperCAmelCase__ : Dict = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : str = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = self.image_processing_class.from_pretrained(_lowerCAmelCase ).to_dict() UpperCAmelCase__ : Tuple = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , _lowerCAmelCase ) @unittest.skip("""ImageGPT requires clusters at initialization""" ) def __UpperCAmelCase ( self ): pass def _lowerCamelCase ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ : Any = load_dataset("""hf-internal-testing/fixtures_image_utils""" , split="""test""" ) UpperCAmelCase__ : Dict = Image.open(dataset[4]["""file"""] ) UpperCAmelCase__ : Optional[Any] = Image.open(dataset[5]["""file"""] ) UpperCAmelCase__ : List[Any] = [imagea, imagea] return images @require_vision @require_torch class UpperCAmelCase_ ( unittest.TestCase ): @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = ImageGPTImageProcessor.from_pretrained("""openai/imagegpt-small""" ) UpperCAmelCase__ : int = prepare_images() # test non-batched UpperCAmelCase__ : List[str] = image_processing(images[0] , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 1024) ) UpperCAmelCase__ : List[Any] = [306, 191, 191] self.assertEqual(encoding.input_ids[0, :3].tolist() , _lowerCAmelCase ) # test batched UpperCAmelCase__ : List[str] = image_processing(_lowerCAmelCase , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 1024) ) UpperCAmelCase__ : Any = [303, 13, 13] self.assertEqual(encoding.input_ids[1, -3:].tolist() , _lowerCAmelCase )
79
0
def _a ( lowercase__ : str ): '''simple docstring''' assert column_title.isupper() SCREAMING_SNAKE_CASE__ : str = 0 SCREAMING_SNAKE_CASE__ : List[Any] = len(lowercase__ ) - 1 SCREAMING_SNAKE_CASE__ : str = 0 while index >= 0: SCREAMING_SNAKE_CASE__ : Union[str, Any] = (ord(column_title[index] ) - 64) * pow(26 , lowercase__ ) answer += value power += 1 index -= 1 return answer if __name__ == "__main__": from doctest import testmod testmod()
85
import os import unittest from transformers import MobileBertTokenizer, MobileBertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class UpperCAmelCase_ ( __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = MobileBertTokenizer __lowerCamelCase = MobileBertTokenizerFast __lowerCamelCase = True __lowerCamelCase = True __lowerCamelCase = filter_non_english __lowerCamelCase = 'google/mobilebert-uncased' def __UpperCAmelCase ( self ): super().setUp() UpperCAmelCase__ : Dict = [ """[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] UpperCAmelCase__ : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) UpperCAmelCase__ : List[str] = [ (tokenizer_def[0], self.pre_trained_model_path, tokenizer_def[2]) # else the 'google/' prefix is stripped for tokenizer_def in self.tokenizers_list ] def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : Tuple = """UNwant\u00E9d,running""" UpperCAmelCase__ : Union[str, Any] = """unwanted, running""" return input_text, output_text def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = self.tokenizer_class(self.vocab_file ) UpperCAmelCase__ : Tuple = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(_lowerCAmelCase , ["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [9, 6, 7, 12, 10, 11] ) def __UpperCAmelCase ( self ): if not self.test_rust_tokenizer: return UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : Dict = self.get_rust_tokenizer() UpperCAmelCase__ : List[str] = """UNwant\u00E9d,running""" UpperCAmelCase__ : Optional[int] = tokenizer.tokenize(_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = rust_tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = rust_tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.get_rust_tokenizer() UpperCAmelCase__ : Any = tokenizer.encode(_lowerCAmelCase ) UpperCAmelCase__ : str = rust_tokenizer.encode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) # With lower casing UpperCAmelCase__ : Tuple = self.get_tokenizer(do_lower_case=_lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.get_rust_tokenizer(do_lower_case=_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = """UNwant\u00E9d,running""" UpperCAmelCase__ : int = tokenizer.tokenize(_lowerCAmelCase ) UpperCAmelCase__ : Any = rust_tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = rust_tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = self.get_rust_tokenizer() UpperCAmelCase__ : List[str] = tokenizer.encode(_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = rust_tokenizer.encode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = BasicTokenizer() self.assertListEqual(tokenizer.tokenize("""ah\u535A\u63A8zz""" ) , ["""ah""", """\u535A""", """\u63A8""", """zz"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = BasicTokenizer(do_lower_case=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Union[str, Any] = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hällo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""h\u00E9llo"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[str] = BasicTokenizer(do_lower_case=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[str] = BasicTokenizer(do_lower_case=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : int = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HäLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HaLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = BasicTokenizer(do_lower_case=_lowerCAmelCase , never_split=["""[UNK]"""] ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? [UNK]""" ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?""", """[UNK]"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : int = ["""[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing"""] UpperCAmelCase__ : List[str] = {} for i, token in enumerate(_lowerCAmelCase ): UpperCAmelCase__ : Optional[Any] = i UpperCAmelCase__ : str = WordpieceTokenizer(vocab=_lowerCAmelCase , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""unwanted running""" ) , ["""un""", """##want""", """##ed""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.tokenize("""unwantedX running""" ) , ["""[UNK]""", """runn""", """##ing"""] ) def __UpperCAmelCase ( self ): self.assertTrue(_is_whitespace(""" """ ) ) self.assertTrue(_is_whitespace("""\t""" ) ) self.assertTrue(_is_whitespace("""\r""" ) ) self.assertTrue(_is_whitespace("""\n""" ) ) self.assertTrue(_is_whitespace("""\u00A0""" ) ) self.assertFalse(_is_whitespace("""A""" ) ) self.assertFalse(_is_whitespace("""-""" ) ) def __UpperCAmelCase ( self ): self.assertTrue(_is_control("""\u0005""" ) ) self.assertFalse(_is_control("""A""" ) ) self.assertFalse(_is_control(""" """ ) ) self.assertFalse(_is_control("""\t""" ) ) self.assertFalse(_is_control("""\r""" ) ) def __UpperCAmelCase ( self ): self.assertTrue(_is_punctuation("""-""" ) ) self.assertTrue(_is_punctuation("""$""" ) ) self.assertTrue(_is_punctuation("""`""" ) ) self.assertTrue(_is_punctuation(""".""" ) ) self.assertFalse(_is_punctuation("""A""" ) ) self.assertFalse(_is_punctuation(""" """ ) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = self.get_tokenizer() UpperCAmelCase__ : Union[str, Any] = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(_lowerCAmelCase ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) self.assertListEqual( [rust_tokenizer.tokenize(_lowerCAmelCase ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = self.tokenizer_class.from_pretrained("""google/mobilebert-uncased""" ) UpperCAmelCase__ : List[Any] = tokenizer.encode("""sequence builders""" , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer.encode("""multi-sequence build""" , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase ) UpperCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase , _lowerCAmelCase ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def __UpperCAmelCase ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): UpperCAmelCase__ : Any = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = f"A, naïve {tokenizer_r.mask_token} AllenNLP sentence." UpperCAmelCase__ : Optional[Any] = tokenizer_r.encode_plus( _lowerCAmelCase , return_attention_mask=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , ) UpperCAmelCase__ : Any = tokenizer_r.do_lower_case if hasattr(_lowerCAmelCase , """do_lower_case""" ) else False UpperCAmelCase__ : Optional[int] = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """A"""), ((1, 2), ""","""), ((3, 5), """na"""), ((5, 6), """##ï"""), ((6, 8), """##ve"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """Allen"""), ((21, 23), """##NL"""), ((23, 24), """##P"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """a"""), ((1, 2), ""","""), ((3, 8), """naive"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """allen"""), ((21, 23), """##nl"""), ((23, 24), """##p"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["""input_ids"""] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["""offset_mapping"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = ["""的""", """人""", """有"""] UpperCAmelCase__ : Tuple = """""".join(_lowerCAmelCase ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): UpperCAmelCase__ : Tuple = True UpperCAmelCase__ : Optional[Any] = self.tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = tokenizer_p.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer_r.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : Any = tokenizer_r.convert_ids_to_tokens(_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = tokenizer_p.convert_ids_to_tokens(_lowerCAmelCase ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : List[Any] = False UpperCAmelCase__ : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = tokenizer_r.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer_p.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = tokenizer_r.convert_ids_to_tokens(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = tokenizer_p.convert_ids_to_tokens(_lowerCAmelCase ) # it is expected that only the first Chinese character is not preceded by "##". UpperCAmelCase__ : List[str] = [ f"##{token}" if idx != 0 else token for idx, token in enumerate(_lowerCAmelCase ) ] self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase )
79
0
from __future__ import annotations __a :List[Any] = list[list[int]] # assigning initial values to the grid __a :Matrix = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution __a :Matrix = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def __snake_case ( __UpperCamelCase : Matrix ,__UpperCamelCase : int ,__UpperCamelCase : int ,__UpperCamelCase : int ): """simple docstring""" for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def __snake_case ( __UpperCamelCase : Matrix ): """simple docstring""" for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def __snake_case ( __UpperCamelCase : Matrix ): """simple docstring""" if location := find_empty_location(__UpperCamelCase ): A_ , A_ = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 ,10 ): if is_safe(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ): A_ = digit if sudoku(__UpperCamelCase ) is not None: return grid A_ = 0 return None def __snake_case ( __UpperCamelCase : Matrix ): """simple docstring""" for row in grid: for cell in row: print(__UpperCamelCase ,end=" " ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print('\nExample grid:\n' + '=' * 20) print_solution(example_grid) print('\nExample grid solution:') __a :int = sudoku(example_grid) if solution is not None: print_solution(solution) else: print('Cannot find a solution.')
86
import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg""" UpperCAmelCase__ : int = Image.open(requests.get(__lowerCamelCase , stream=__lowerCamelCase ).raw ).convert("""RGB""" ) UpperCAmelCase__ : Any = transforms.Compose( [ transforms.Resize((image_size, image_size) , interpolation=InterpolationMode.BICUBIC ), transforms.ToTensor(), transforms.Normalize((0.48_145_466, 0.4_578_275, 0.40_821_073) , (0.26_862_954, 0.26_130_258, 0.27_577_711) ), ] ) UpperCAmelCase__ : Optional[int] = transform(__lowerCamelCase ).unsqueeze(0 ).to(__lowerCamelCase ) return image def _lowerCamelCase ( __lowerCamelCase ) -> str: '''simple docstring''' if "visual_encoder" in key: UpperCAmelCase__ : Dict = re.sub("""visual_encoder*""" , """vision_model.encoder""" , __lowerCamelCase ) if "blocks" in key: UpperCAmelCase__ : Optional[Any] = re.sub(r"""blocks""" , """layers""" , __lowerCamelCase ) if "attn" in key: UpperCAmelCase__ : List[str] = re.sub(r"""attn""" , """self_attn""" , __lowerCamelCase ) if "norm1" in key: UpperCAmelCase__ : Union[str, Any] = re.sub(r"""norm1""" , """layer_norm1""" , __lowerCamelCase ) if "norm2" in key: UpperCAmelCase__ : Any = re.sub(r"""norm2""" , """layer_norm2""" , __lowerCamelCase ) if "encoder.norm" in key: UpperCAmelCase__ : Dict = re.sub(r"""encoder.norm""" , """post_layernorm""" , __lowerCamelCase ) if "encoder.patch_embed.proj" in key: UpperCAmelCase__ : List[str] = re.sub(r"""encoder.patch_embed.proj""" , """embeddings.patch_embedding""" , __lowerCamelCase ) if "encoder.pos_embed" in key: UpperCAmelCase__ : List[str] = re.sub(r"""encoder.pos_embed""" , """embeddings.position_embedding""" , __lowerCamelCase ) if "encoder.cls_token" in key: UpperCAmelCase__ : List[Any] = re.sub(r"""encoder.cls_token""" , """embeddings.class_embedding""" , __lowerCamelCase ) if "self_attn" in key: UpperCAmelCase__ : List[Any] = re.sub(r"""self_attn.proj""" , """self_attn.projection""" , __lowerCamelCase ) return key @torch.no_grad() def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase=None ) -> Tuple: '''simple docstring''' if config_path is not None: UpperCAmelCase__ : Any = BlipConfig.from_pretrained(__lowerCamelCase ) else: UpperCAmelCase__ : str = BlipConfig(projection_dim=512 , text_config={} , vision_config={} ) UpperCAmelCase__ : int = BlipForConditionalGeneration(__lowerCamelCase ).eval() UpperCAmelCase__ : Any = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth""" UpperCAmelCase__ : List[str] = blip_decoder(pretrained=__lowerCamelCase , image_size=384 , vit="""base""" ) UpperCAmelCase__ : Union[str, Any] = pt_model.eval() UpperCAmelCase__ : Optional[int] = pt_model.state_dict() for key in modified_state_dict.copy(): UpperCAmelCase__ : Dict = modified_state_dict.pop(__lowerCamelCase ) UpperCAmelCase__ : Union[str, Any] = rename_key(__lowerCamelCase ) UpperCAmelCase__ : List[str] = value hf_model.load_state_dict(__lowerCamelCase ) UpperCAmelCase__ : Tuple = 384 UpperCAmelCase__ : str = load_demo_image(image_size=__lowerCamelCase , device="""cpu""" ) UpperCAmelCase__ : str = BertTokenizer.from_pretrained("""bert-base-uncased""" ) UpperCAmelCase__ : Dict = tokenizer(["""a picture of"""] ).input_ids UpperCAmelCase__ : int = hf_model.generate(__lowerCamelCase , __lowerCamelCase ) assert out[0].tolist() == [3_0522, 1037, 3861, 1997, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] UpperCAmelCase__ : Any = hf_model.generate(__lowerCamelCase ) assert out[0].tolist() == [3_0522, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(__lowerCamelCase ) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' UpperCAmelCase__ : Union[str, Any] = ( """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth""" ) UpperCAmelCase__ : List[Any] = blip_vqa(pretrained=__lowerCamelCase , image_size=__lowerCamelCase , vit="""base""" ) vqa_model.eval() UpperCAmelCase__ : str = vqa_model.state_dict() for key in modified_state_dict.copy(): UpperCAmelCase__ : Dict = modified_state_dict.pop(__lowerCamelCase ) UpperCAmelCase__ : Dict = rename_key(__lowerCamelCase ) UpperCAmelCase__ : int = value UpperCAmelCase__ : List[str] = BlipForQuestionAnswering(__lowerCamelCase ) hf_vqa_model.load_state_dict(__lowerCamelCase ) UpperCAmelCase__ : Tuple = ["""How many dogs are in this image?"""] UpperCAmelCase__ : Union[str, Any] = tokenizer(__lowerCamelCase , return_tensors="""pt""" ).input_ids UpperCAmelCase__ : Optional[Any] = hf_vqa_model.generate(__lowerCamelCase , __lowerCamelCase ) print(tokenizer.decode(answer[0] ) ) assert tokenizer.decode(answer[0] ) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + """_vqa""" ) UpperCAmelCase__ : int = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth""" UpperCAmelCase__ : Any = blip_itm(pretrained=__lowerCamelCase , image_size=__lowerCamelCase , vit="""base""" ) itm_model.eval() UpperCAmelCase__ : List[Any] = itm_model.state_dict() for key in modified_state_dict.copy(): UpperCAmelCase__ : Dict = modified_state_dict.pop(__lowerCamelCase ) UpperCAmelCase__ : int = rename_key(__lowerCamelCase ) UpperCAmelCase__ : Any = value UpperCAmelCase__ : Optional[int] = BlipForImageTextRetrieval(__lowerCamelCase ) UpperCAmelCase__ : Union[str, Any] = ["""A picture of a woman with a dog sitting in a beach"""] UpperCAmelCase__ : List[Any] = tokenizer( __lowerCamelCase , return_tensors="""pt""" , padding="""max_length""" , truncation=__lowerCamelCase , max_length=35 , ).input_ids hf_itm_model.load_state_dict(__lowerCamelCase ) hf_itm_model.eval() UpperCAmelCase__ : List[str] = hf_itm_model(__lowerCamelCase , __lowerCamelCase , use_itm_head=__lowerCamelCase ) UpperCAmelCase__ : List[str] = hf_itm_model(__lowerCamelCase , __lowerCamelCase , use_itm_head=__lowerCamelCase ) assert out[0].item() == 0.2_110_687_494_277_954 assert torch.nn.functional.softmax(out_itm[0] , dim=1 )[:, 1].item() == 0.45_698_845_386_505_127 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + """_itm""" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : str = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") SCREAMING_SNAKE_CASE__ : List[Any] = parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
79
0
def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ ) -> int: """simple docstring""" return int((input_a, input_a).count(1 ) != 0 ) def SCREAMING_SNAKE_CASE ( ) -> None: """simple docstring""" assert or_gate(0 , 0 ) == 0 assert or_gate(0 , 1 ) == 1 assert or_gate(1 , 0 ) == 1 assert or_gate(1 , 1 ) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
87
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : Tuple = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : List[Any] = { """MIT/ast-finetuned-audioset-10-10-0.4593""": ( """https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json""" ), } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'audio-spectrogram-transformer' def __init__( self , _lowerCAmelCase=768 , _lowerCAmelCase=12 , _lowerCAmelCase=12 , _lowerCAmelCase=3072 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=1e-12 , _lowerCAmelCase=16 , _lowerCAmelCase=True , _lowerCAmelCase=10 , _lowerCAmelCase=10 , _lowerCAmelCase=1024 , _lowerCAmelCase=128 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : List[Any] = num_attention_heads UpperCAmelCase__ : Dict = intermediate_size UpperCAmelCase__ : Dict = hidden_act UpperCAmelCase__ : str = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : Tuple = initializer_range UpperCAmelCase__ : Dict = layer_norm_eps UpperCAmelCase__ : Optional[Any] = patch_size UpperCAmelCase__ : Tuple = qkv_bias UpperCAmelCase__ : Tuple = frequency_stride UpperCAmelCase__ : Union[str, Any] = time_stride UpperCAmelCase__ : Optional[Any] = max_length UpperCAmelCase__ : Optional[int] = num_mel_bins
79
0
"""simple docstring""" import torch from transformers import AutoModel class lowercase__ ( torch.nn.Module ): def __init__( self , SCREAMING_SNAKE_CASE="sayef/fsner-bert-base-uncased") -> str: super(SCREAMING_SNAKE_CASE , self).__init__() _lowerCamelCase : Union[str, Any] = AutoModel.from_pretrained(SCREAMING_SNAKE_CASE , return_dict=SCREAMING_SNAKE_CASE) _lowerCamelCase : List[str] = torch.nn.CosineSimilarity(3 , 1e-0_8) _lowerCamelCase : Optional[int] = torch.nn.Softmax(dim=1) def UpperCamelCase_ ( self , **SCREAMING_SNAKE_CASE) -> str: return self.bert(**SCREAMING_SNAKE_CASE).last_hidden_state def UpperCamelCase_ ( self , SCREAMING_SNAKE_CASE) -> Optional[Any]: return token_embeddings.sum(2 , keepdim=SCREAMING_SNAKE_CASE) def UpperCamelCase_ ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=1) -> Union[str, Any]: return self.softmax(T * self.cos(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE)) def UpperCamelCase_ ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) -> List[str]: _lowerCamelCase : str = W_supports["""sizes"""].tolist() _lowerCamelCase : int = W_supports["""start_token_id"""].item() _lowerCamelCase : str = W_supports["""end_token_id"""].item() del W_supports["sizes"] del W_supports["start_token_id"] del W_supports["end_token_id"] _lowerCamelCase : List[str] = self.BERT(**SCREAMING_SNAKE_CASE) _lowerCamelCase : Optional[Any] = self.BERT(**SCREAMING_SNAKE_CASE) _lowerCamelCase : Any = None _lowerCamelCase : List[Any] = None _lowerCamelCase : Any = W_supports["""input_ids"""] == start_token_id _lowerCamelCase : Any = W_supports["""input_ids"""] == end_token_id for i, size in enumerate(SCREAMING_SNAKE_CASE): if i == 0: _lowerCamelCase : List[str] = 0 else: _lowerCamelCase : Dict = support_sizes[i - 1] _lowerCamelCase : Union[str, Any] = S[s : s + size][start_token_masks[s : s + size]] _lowerCamelCase : Any = S[s : s + size][end_token_masks[s : s + size]] _lowerCamelCase : Any = torch.matmul(q[i] , s_start.T).sum(1).softmax(0) _lowerCamelCase : List[str] = torch.matmul(q[i] , s_end.T).sum(1).softmax(0) if p_starts is not None: _lowerCamelCase : str = torch.vstack((p_starts, p_start)) _lowerCamelCase : Optional[Any] = torch.vstack((p_ends, p_end)) else: _lowerCamelCase : Optional[Any] = p_start _lowerCamelCase : int = p_end return p_starts, p_ends
88
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) class UpperCAmelCase_ ( __lowerCamelCase ): def __init__( self , *_lowerCAmelCase , **_lowerCAmelCase ): warnings.warn( """The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use GLPNImageProcessor instead.""" , _lowerCAmelCase , ) super().__init__(*_lowerCAmelCase , **_lowerCAmelCase )
79
0
import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _lowerCamelCase( unittest.TestCase ): @property def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" torch.manual_seed(0) _lowercase : Dict = UNetaDModel( block_out_channels=(32, 64), layers_per_block=2, sample_size=32, in_channels=3, out_channels=3, down_block_types=('DownBlock2D', 'AttnDownBlock2D'), up_block_types=('AttnUpBlock2D', 'UpBlock2D'), ) return model def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : List[str] = self.dummy_uncond_unet _lowercase : Tuple = ScoreSdeVeScheduler() _lowercase : str = ScoreSdeVePipeline(unet=lowerCamelCase, scheduler=lowerCamelCase) sde_ve.to(lowerCamelCase) sde_ve.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Any = torch.manual_seed(0) _lowercase : Tuple = sde_ve(num_inference_steps=2, output_type='numpy', generator=lowerCamelCase).images _lowercase : Any = torch.manual_seed(0) _lowercase : Any = sde_ve(num_inference_steps=2, output_type='numpy', generator=lowerCamelCase, return_dict=lowerCamelCase)[ 0 ] _lowercase : Optional[Any] = image[0, -3:, -3:, -1] _lowercase : str = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _lowercase : Any = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1E-2 @slow @require_torch class _lowerCamelCase( unittest.TestCase ): def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : Union[str, Any] = 'google/ncsnpp-church-256' _lowercase : Dict = UNetaDModel.from_pretrained(lowerCamelCase) _lowercase : Dict = ScoreSdeVeScheduler.from_pretrained(lowerCamelCase) _lowercase : Union[str, Any] = ScoreSdeVePipeline(unet=lowerCamelCase, scheduler=lowerCamelCase) sde_ve.to(lowerCamelCase) sde_ve.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Dict = torch.manual_seed(0) _lowercase : str = sde_ve(num_inference_steps=10, output_type='numpy', generator=lowerCamelCase).images _lowercase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) _lowercase : Dict = np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2
89
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : int = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__ : List[str] = { """vocab_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt""", }, """tokenizer_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json""" ), """google/realm-orqa-nq-openqa""": ( """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-nq-reader""": ( """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-openqa""": ( """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-reader""": ( """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json""" ), }, } SCREAMING_SNAKE_CASE__ : Optional[Any] = { """google/realm-cc-news-pretrained-embedder""": 5_12, """google/realm-cc-news-pretrained-encoder""": 5_12, """google/realm-cc-news-pretrained-scorer""": 5_12, """google/realm-cc-news-pretrained-openqa""": 5_12, """google/realm-orqa-nq-openqa""": 5_12, """google/realm-orqa-nq-reader""": 5_12, """google/realm-orqa-wq-openqa""": 5_12, """google/realm-orqa-wq-reader""": 5_12, } SCREAMING_SNAKE_CASE__ : Optional[Any] = { """google/realm-cc-news-pretrained-embedder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-encoder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-scorer""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-reader""": {"""do_lower_case""": True}, """google/realm-orqa-wq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-wq-reader""": {"""do_lower_case""": True}, } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = PRETRAINED_INIT_CONFIGURATION __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = RealmTokenizer def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=True , _lowerCAmelCase="[UNK]" , _lowerCAmelCase="[SEP]" , _lowerCAmelCase="[PAD]" , _lowerCAmelCase="[CLS]" , _lowerCAmelCase="[MASK]" , _lowerCAmelCase=True , _lowerCAmelCase=None , **_lowerCAmelCase , ): super().__init__( _lowerCAmelCase , tokenizer_file=_lowerCAmelCase , do_lower_case=_lowerCAmelCase , unk_token=_lowerCAmelCase , sep_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , cls_token=_lowerCAmelCase , mask_token=_lowerCAmelCase , tokenize_chinese_chars=_lowerCAmelCase , strip_accents=_lowerCAmelCase , **_lowerCAmelCase , ) UpperCAmelCase__ : Tuple = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , _lowerCAmelCase ) != do_lower_case or normalizer_state.get("""strip_accents""" , _lowerCAmelCase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , _lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase__ : Any = getattr(_lowerCAmelCase , normalizer_state.pop("""type""" ) ) UpperCAmelCase__ : str = do_lower_case UpperCAmelCase__ : Tuple = strip_accents UpperCAmelCase__ : Tuple = tokenize_chinese_chars UpperCAmelCase__ : Union[str, Any] = normalizer_class(**_lowerCAmelCase ) UpperCAmelCase__ : Dict = do_lower_case def __UpperCAmelCase ( self , _lowerCAmelCase , **_lowerCAmelCase ): UpperCAmelCase__ : List[Any] = PaddingStrategy.MAX_LENGTH UpperCAmelCase__ : Optional[int] = text UpperCAmelCase__ : Optional[int] = kwargs.pop("""text_pair""" , _lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = kwargs.pop("""return_tensors""" , _lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = { """input_ids""": [], """attention_mask""": [], """token_type_ids""": [], } for idx, candidate_text in enumerate(_lowerCAmelCase ): if batch_text_pair is not None: UpperCAmelCase__ : str = batch_text_pair[idx] else: UpperCAmelCase__ : Any = None UpperCAmelCase__ : str = super().__call__(_lowerCAmelCase , _lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = encoded_candidates.get("""input_ids""" ) UpperCAmelCase__ : str = encoded_candidates.get("""attention_mask""" ) UpperCAmelCase__ : Union[str, Any] = encoded_candidates.get("""token_type_ids""" ) if encoded_input_ids is not None: output_data["input_ids"].append(_lowerCAmelCase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(_lowerCAmelCase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = {key: item for key, item in output_data.items() if len(_lowerCAmelCase ) != 0} return BatchEncoding(_lowerCAmelCase , tensor_type=_lowerCAmelCase ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase = None ): UpperCAmelCase__ : Any = [self.sep_token_id] UpperCAmelCase__ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase = None ): UpperCAmelCase__ : List[str] = self._tokenizer.model.save(_lowerCAmelCase , name=_lowerCAmelCase ) return tuple(_lowerCAmelCase )
79
0
'''simple docstring''' from __future__ import annotations import queue class a__ : '''simple docstring''' def __init__( self , lowerCamelCase_ ) -> int: lowerCAmelCase__ = data lowerCAmelCase__ = None lowerCAmelCase__ = None def _snake_case ( ) -> TreeNode: print('''\n********Press N to stop entering at any point of time********\n''' ) lowerCAmelCase__ = input('''Enter the value of the root node: ''' ).strip().lower() lowerCAmelCase__ = queue.Queue() lowerCAmelCase__ = TreeNode(int(A ) ) q.put(A ) while not q.empty(): lowerCAmelCase__ = q.get() lowerCAmelCase__ = F"""Enter the left node of {node_found.data}: """ lowerCAmelCase__ = input(A ).strip().lower() or '''n''' if check == "n": return tree_node lowerCAmelCase__ = TreeNode(int(A ) ) lowerCAmelCase__ = left_node q.put(A ) lowerCAmelCase__ = F"""Enter the right node of {node_found.data}: """ lowerCAmelCase__ = input(A ).strip().lower() or '''n''' if check == "n": return tree_node lowerCAmelCase__ = TreeNode(int(A ) ) lowerCAmelCase__ = right_node q.put(A ) raise def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return print(node.data , end=''',''' ) pre_order(node.left ) pre_order(node.right ) def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return in_order(node.left ) print(node.data , end=''',''' ) in_order(node.right ) def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return post_order(node.left ) post_order(node.right ) print(node.data , end=''',''' ) def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return lowerCAmelCase__ = queue.Queue() q.put(A ) while not q.empty(): lowerCAmelCase__ = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: q.put(node_dequeued.left ) if node_dequeued.right: q.put(node_dequeued.right ) def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return lowerCAmelCase__ = queue.Queue() q.put(A ) while not q.empty(): lowerCAmelCase__ = [] while not q.empty(): lowerCAmelCase__ = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: list_.append(node_dequeued.left ) if node_dequeued.right: list_.append(node_dequeued.right ) print() for node in list_: q.put(A ) def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return lowerCAmelCase__ = [] lowerCAmelCase__ = node while n or stack: while n: # start from root node, find its left child print(n.data , end=''',''' ) stack.append(A ) lowerCAmelCase__ = n.left # end of while means current node doesn't have left child lowerCAmelCase__ = stack.pop() # start to traverse its right child lowerCAmelCase__ = n.right def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return lowerCAmelCase__ = [] lowerCAmelCase__ = node while n or stack: while n: stack.append(A ) lowerCAmelCase__ = n.left lowerCAmelCase__ = stack.pop() print(n.data , end=''',''' ) lowerCAmelCase__ = n.right def _snake_case ( A ) -> None: if not isinstance(A , A ) or not node: return lowerCAmelCase__ , lowerCAmelCase__ = [], [] lowerCAmelCase__ = node stacka.append(A ) while stacka: # to find the reversed order of post order, store it in stack2 lowerCAmelCase__ = stacka.pop() if n.left: stacka.append(n.left ) if n.right: stacka.append(n.right ) stacka.append(A ) while stacka: # pop up from stack2 will be the post order print(stacka.pop().data , end=''',''' ) def _snake_case ( A = "" , A=50 , A="*" ) -> str: if not s: return "\n" + width * char lowerCAmelCase__ , lowerCAmelCase__ = divmod(width - len(A ) - 2 , 2 ) return F"""{left * char} {s} {(left + extra) * char}""" if __name__ == "__main__": import doctest doctest.testmod() print(prompt('''Binary Tree Traversals''')) __UpperCAmelCase = build_tree() print(prompt('''Pre Order Traversal''')) pre_order(node) print(prompt() + '''\n''') print(prompt('''In Order Traversal''')) in_order(node) print(prompt() + '''\n''') print(prompt('''Post Order Traversal''')) post_order(node) print(prompt() + '''\n''') print(prompt('''Level Order Traversal''')) level_order(node) print(prompt() + '''\n''') print(prompt('''Actual Level Order Traversal''')) level_order_actual(node) print('''*''' * 50 + '''\n''') print(prompt('''Pre Order Traversal - Iteration Version''')) pre_order_iter(node) print(prompt() + '''\n''') print(prompt('''In Order Traversal - Iteration Version''')) in_order_iter(node) print(prompt() + '''\n''') print(prompt('''Post Order Traversal - Iteration Version''')) post_order_iter(node) print(prompt())
90
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'facebook/bart-large-mnli' __lowerCamelCase = ( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) __lowerCamelCase = 'text_classifier' __lowerCamelCase = AutoTokenizer __lowerCamelCase = AutoModelForSequenceClassification __lowerCamelCase = ['text', ['text']] __lowerCamelCase = ['text'] def __UpperCAmelCase ( self ): super().setup() UpperCAmelCase__ : Optional[Any] = self.model.config UpperCAmelCase__ : Tuple = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("""entail""" ): UpperCAmelCase__ : Dict = int(_lowerCAmelCase ) if self.entailment_id == -1: raise ValueError("""Could not determine the entailment ID from the model config, please pass it at init.""" ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : List[Any] = labels return self.pre_processor( [text] * len(_lowerCAmelCase ) , [f"This example is {label}" for label in labels] , return_tensors="""pt""" , padding="""max_length""" , ) def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : str = outputs.logits UpperCAmelCase__ : List[Any] = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
79
0
"""simple docstring""" from __future__ import annotations from collections import deque from collections.abc import Iterator from dataclasses import dataclass @dataclass class lowerCAmelCase_ : '''simple docstring''' _lowerCamelCase: int _lowerCamelCase: int class lowerCAmelCase_ : '''simple docstring''' def __init__( self : Tuple ,A_ : int ) -> Union[str, Any]: A = [[] for _ in range(A_ )] A = size def __getitem__( self : str ,A_ : int ) -> Iterator[Edge]: return iter(self._graph[vertex] ) @property def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Any: return self._size def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,A_ : int ,A_ : int ,A_ : int ) -> Tuple: if weight not in (0, 1): raise ValueError('Edge weight must be either 0 or 1.' ) if to_vertex < 0 or to_vertex >= self.size: raise ValueError('Vertex indexes must be in [0; size).' ) self._graph[from_vertex].append(Edge(A_ ,A_ ) ) def _SCREAMING_SNAKE_CASE ( self : int ,A_ : int ,A_ : int ) -> int | None: A = deque([start_vertex] ) A = [None] * self.size A = 0 while queue: A = queue.popleft() A = distances[current_vertex] if current_distance is None: continue for edge in self[current_vertex]: A = current_distance + edge.weight A = distances[edge.destination_vertex] if ( isinstance(A_ ,A_ ) and new_distance >= dest_vertex_distance ): continue A = new_distance if edge.weight == 0: queue.appendleft(edge.destination_vertex ) else: queue.append(edge.destination_vertex ) if distances[finish_vertex] is None: raise ValueError('No path from start_vertex to finish_vertex.' ) return distances[finish_vertex] if __name__ == "__main__": import doctest doctest.testmod()
91
from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=13 , _lowerCAmelCase=30 , _lowerCAmelCase=2 , _lowerCAmelCase=3 , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=32 , _lowerCAmelCase=2 , _lowerCAmelCase=4 , _lowerCAmelCase=37 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=10 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=3 , _lowerCAmelCase=None , ): UpperCAmelCase__ : Tuple = parent UpperCAmelCase__ : Optional[int] = batch_size UpperCAmelCase__ : Union[str, Any] = image_size UpperCAmelCase__ : int = patch_size UpperCAmelCase__ : str = num_channels UpperCAmelCase__ : int = is_training UpperCAmelCase__ : List[str] = use_labels UpperCAmelCase__ : List[Any] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : Tuple = num_attention_heads UpperCAmelCase__ : Optional[int] = intermediate_size UpperCAmelCase__ : Optional[Any] = hidden_act UpperCAmelCase__ : int = hidden_dropout_prob UpperCAmelCase__ : int = attention_probs_dropout_prob UpperCAmelCase__ : List[str] = type_sequence_label_size UpperCAmelCase__ : Optional[int] = initializer_range UpperCAmelCase__ : Any = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) UpperCAmelCase__ : Any = (image_size // patch_size) ** 2 UpperCAmelCase__ : Tuple = num_patches + 1 def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ : List[str] = None if self.use_labels: UpperCAmelCase__ : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ : Union[str, Any] = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self ): return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : str = TFViTModel(config=_lowerCAmelCase ) UpperCAmelCase__ : str = model(_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase__ : Optional[Any] = self.image_size // 2 UpperCAmelCase__ : List[str] = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase__ : List[Any] = model(_lowerCAmelCase , interpolate_pos_encoding=_lowerCAmelCase , training=_lowerCAmelCase ) UpperCAmelCase__ : str = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : Tuple = self.type_sequence_label_size UpperCAmelCase__ : List[Any] = TFViTForImageClassification(_lowerCAmelCase ) UpperCAmelCase__ : List[str] = model(_lowerCAmelCase , labels=_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase__ : Tuple = self.image_size // 2 UpperCAmelCase__ : Union[str, Any] = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase__ : List[str] = model(_lowerCAmelCase , interpolate_pos_encoding=_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase__ : Union[str, Any] = 1 UpperCAmelCase__ : Optional[Any] = TFViTForImageClassification(_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase__ : List[str] = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : int = config_and_inputs UpperCAmelCase__ : int = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class UpperCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () __lowerCamelCase = ( {'feature-extraction': TFViTModel, 'image-classification': TFViTForImageClassification} if is_tf_available() else {} ) __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = TFViTModelTester(self ) UpperCAmelCase__ : int = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def __UpperCAmelCase ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def __UpperCAmelCase ( self ): pass @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def __UpperCAmelCase ( self ): pass def __UpperCAmelCase ( self ): UpperCAmelCase__ , UpperCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : str = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , tf.keras.layers.Layer ) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : Optional[int] = model_class(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ : Tuple = [*signature.parameters.keys()] UpperCAmelCase__ : str = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = TFViTModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(_lowerCAmelCase ) def _lowerCamelCase ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ : List[str] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class UpperCAmelCase_ ( unittest.TestCase ): @cached_property def __UpperCAmelCase ( self ): return ViTImageProcessor.from_pretrained("""google/vit-base-patch16-224""" ) if is_vision_available() else None @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = TFViTForImageClassification.from_pretrained("""google/vit-base-patch16-224""" ) UpperCAmelCase__ : List[Any] = self.default_image_processor UpperCAmelCase__ : Union[str, Any] = prepare_img() UpperCAmelCase__ : Optional[Any] = image_processor(images=_lowerCAmelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase__ : int = model(**_lowerCAmelCase ) # verify the logits UpperCAmelCase__ : Tuple = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) UpperCAmelCase__ : int = tf.constant([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ) tf.debugging.assert_near(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 )
79
0
'''simple docstring''' import argparse import os from pathlib import Path from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import PegasusConfig, PegasusForConditionalGeneration, PegasusTokenizer from transformers.models.pegasus.configuration_pegasus import DEFAULTS, task_specific_params UpperCamelCase_ = [ # replace left string with right string to get the relevant state_dict key (identical state dict to bart) ["""memory_attention""", """encoder_attn"""], ["""attention""", """attn"""], ["""/""", """."""], [""".LayerNorm.gamma""", """_layer_norm.weight"""], [""".LayerNorm.beta""", """_layer_norm.bias"""], ["""r.layer_""", """r.layers."""], ["""output_proj""", """out_proj"""], ["""ffn.dense_1.""", """fc2."""], ["""ffn.dense.""", """fc1."""], ["""ffn_layer_norm""", """final_layer_norm"""], ["""kernel""", """weight"""], ["""encoder_layer_norm.""", """encoder.layer_norm."""], ["""decoder_layer_norm.""", """decoder.layer_norm."""], ["""embeddings.weights""", """shared.weight"""], ] def _lowerCAmelCase ( __magic_name__ : str ) -> int: for pegasus_name, hf_name in PATTERNS: lowercase : Dict =k.replace(__magic_name__ , __magic_name__ ) return k def _lowerCAmelCase ( __magic_name__ : dict , __magic_name__ : dict ) -> PegasusForConditionalGeneration: lowercase : str =DEFAULTS.copy() cfg_kwargs.update(__magic_name__ ) lowercase : List[str] =PegasusConfig(**__magic_name__ ) lowercase : str =PegasusForConditionalGeneration(__magic_name__ ) lowercase : Dict =torch_model.model.state_dict() lowercase : Dict ={} for k, v in tf_weights.items(): lowercase : Union[str, Any] =rename_state_dict_key(__magic_name__ ) if new_k not in sd: raise ValueError(f'''could not find new key {new_k} in state dict. (converted from {k})''' ) if "dense" in k or "proj" in new_k: lowercase : Optional[Any] =v.T lowercase : Optional[int] =torch.tensor(__magic_name__ , dtype=sd[new_k].dtype ) assert v.shape == sd[new_k].shape, f'''{new_k}, {k}, {v.shape}, {sd[new_k].shape}''' # make sure embedding.padding_idx is respected lowercase : Any =torch.zeros_like(mapping['''shared.weight'''][cfg.pad_token_id + 1] ) lowercase : Tuple =mapping['''shared.weight'''] lowercase : Optional[int] =mapping['''shared.weight'''] lowercase : Optional[int] ={k: torch.zeros_like(__magic_name__ ) for k, v in sd.items() if k.endswith('''bias''' ) and k not in mapping} mapping.update(**__magic_name__ ) lowercase , lowercase : List[Any] =torch_model.model.load_state_dict(__magic_name__ , strict=__magic_name__ ) lowercase : List[str] =[ k for k in missing if k not in ['''encoder.embed_positions.weight''', '''decoder.embed_positions.weight'''] ] assert unexpected_missing == [], f'''no matches found for the following torch keys {unexpected_missing}''' assert extra == [], f'''no matches found for the following tf keys {extra}''' return torch_model def _lowerCAmelCase ( __magic_name__ : Dict="./ckpt/aeslc/model.ckpt-32000" ) -> Dict: lowercase : Dict =tf.train.list_variables(__magic_name__ ) lowercase : int ={} lowercase : Dict =['''Adafactor''', '''global_step'''] for name, shape in tqdm(__magic_name__ , desc='''converting tf checkpoint to dict''' ): lowercase : int =any(pat in name for pat in ignore_name ) if skip_key: continue lowercase : List[Any] =tf.train.load_variable(__magic_name__ , __magic_name__ ) lowercase : str =array return tf_weights def _lowerCAmelCase ( __magic_name__ : str , __magic_name__ : str ) -> List[Any]: # save tokenizer first lowercase : Tuple =Path(__magic_name__ ).parent.name lowercase : Optional[Any] =task_specific_params[f'''summarization_{dataset}''']['''max_position_embeddings'''] lowercase : List[Any] =PegasusTokenizer.from_pretrained('''sshleifer/pegasus''' , model_max_length=__magic_name__ ) assert tok.model_max_length == desired_max_model_length tok.save_pretrained(__magic_name__ ) # convert model lowercase : Tuple =get_tf_weights_as_numpy(__magic_name__ ) lowercase : str =task_specific_params[f'''summarization_{dataset}'''] if dataset == "large": lowercase : Union[str, Any] =task_specific_params lowercase : Any =convert_pegasus(__magic_name__ , __magic_name__ ) torch_model.save_pretrained(__magic_name__ ) lowercase : Dict =torch_model.state_dict() sd.pop('''model.decoder.embed_positions.weight''' ) sd.pop('''model.encoder.embed_positions.weight''' ) torch.save(__magic_name__ , Path(__magic_name__ ) / '''pytorch_model.bin''' ) if __name__ == "__main__": UpperCamelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument("""tf_ckpt_path""", type=str, help="""passed to tf.train.list_variables""") parser.add_argument("""save_dir""", default=None, type=str, help="""Path to the output PyTorch model.""") UpperCamelCase_ = parser.parse_args() if args.save_dir is None: UpperCamelCase_ = Path(args.tf_ckpt_path).parent.name UpperCamelCase_ = os.path.join("""pegasus""", dataset) convert_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir)
92
from functools import lru_cache @lru_cache def _lowerCamelCase ( __lowerCamelCase ) -> int: '''simple docstring''' if num < 0: raise ValueError("""Number should not be negative.""" ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
79
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" __magic_name__ :int = KandinskyVaaControlnetImgaImgPipeline __magic_name__ :Tuple = ["""image_embeds""", """negative_image_embeds""", """image""", """hint"""] __magic_name__ :List[Any] = ["""image_embeds""", """negative_image_embeds""", """image""", """hint"""] __magic_name__ :List[Any] = [ """generator""", """height""", """width""", """strength""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] __magic_name__ :Any = False @property def snake_case ( self ): '''simple docstring''' return 3_2 @property def snake_case ( self ): '''simple docstring''' return 3_2 @property def snake_case ( self ): '''simple docstring''' return self.time_input_dim @property def snake_case ( self ): '''simple docstring''' return self.time_input_dim * 4 @property def snake_case ( self ): '''simple docstring''' return 1_0_0 @property def snake_case ( self ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ :Union[str, Any] = { 'in_channels': 8, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'image_hint', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } lowerCAmelCase__ :int = UNetaDConditionModel(**__UpperCAmelCase ) return model @property def snake_case ( self ): '''simple docstring''' return { "block_out_channels": [3_2, 3_2, 6_4, 6_4], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 1_2, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def snake_case ( self ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ :Any = VQModel(**self.dummy_movq_kwargs ) return model def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = self.dummy_unet lowerCAmelCase__ :List[str] = self.dummy_movq lowerCAmelCase__ :List[Any] = { 'num_train_timesteps': 1_0_0_0, 'beta_schedule': 'linear', 'beta_start': 0.0_00_85, 'beta_end': 0.0_12, 'clip_sample': False, 'set_alpha_to_one': False, 'steps_offset': 0, 'prediction_type': 'epsilon', 'thresholding': False, } lowerCAmelCase__ :Union[str, Any] = DDIMScheduler(**__UpperCAmelCase ) lowerCAmelCase__ :List[Any] = { 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase=0 ): '''simple docstring''' lowerCAmelCase__ :Any = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(__UpperCAmelCase ) ).to(__UpperCAmelCase ) lowerCAmelCase__ :str = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( __UpperCAmelCase ) # create init_image lowerCAmelCase__ :int = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(__UpperCAmelCase ) ).to(__UpperCAmelCase ) lowerCAmelCase__ :Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase__ :Optional[int] = Image.fromarray(np.uinta(__UpperCAmelCase ) ).convert('RGB' ).resize((2_5_6, 2_5_6) ) # create hint lowerCAmelCase__ :str = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(__UpperCAmelCase ) ).to(__UpperCAmelCase ) if str(__UpperCAmelCase ).startswith('mps' ): lowerCAmelCase__ :Optional[int] = torch.manual_seed(__UpperCAmelCase ) else: lowerCAmelCase__ :Tuple = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) lowerCAmelCase__ :str = { 'image': init_image, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'hint': hint, 'generator': generator, 'height': 6_4, 'width': 6_4, 'num_inference_steps': 1_0, 'guidance_scale': 7.0, 'strength': 0.2, 'output_type': 'np', } return inputs def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = 'cpu' lowerCAmelCase__ :Optional[Any] = self.get_dummy_components() lowerCAmelCase__ :str = self.pipeline_class(**__UpperCAmelCase ) lowerCAmelCase__ :Optional[Any] = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) lowerCAmelCase__ :Dict = pipe(**self.get_dummy_inputs(__UpperCAmelCase ) ) lowerCAmelCase__ :List[str] = output.images lowerCAmelCase__ :Optional[Any] = pipe( **self.get_dummy_inputs(__UpperCAmelCase ) , return_dict=__UpperCAmelCase , )[0] lowerCAmelCase__ :Any = image[0, -3:, -3:, -1] lowerCAmelCase__ :Optional[int] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) lowerCAmelCase__ :List[Any] = np.array( [0.54_98_50_34, 0.55_50_93_65, 0.52_56_15_04, 0.5_57_04_94, 0.5_59_38_18, 0.5_26_39_79, 0.50_28_56_43, 0.5_06_98_46, 0.51_19_67_36] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy' ) lowerCAmelCase__ :Dict = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png' ) lowerCAmelCase__ :Dict = init_image.resize((5_1_2, 5_1_2) ) lowerCAmelCase__ :int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/hint_image_cat.png' ) lowerCAmelCase__ :Tuple = torch.from_numpy(np.array(__UpperCAmelCase ) ).float() / 2_55.0 lowerCAmelCase__ :Union[str, Any] = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) lowerCAmelCase__ :List[Any] = 'A robot, 4k photo' lowerCAmelCase__ :int = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-prior' , torch_dtype=torch.floataa ) pipe_prior.to(__UpperCAmelCase ) lowerCAmelCase__ :Tuple = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-controlnet-depth' , torch_dtype=torch.floataa ) lowerCAmelCase__ :int = pipeline.to(__UpperCAmelCase ) pipeline.set_progress_bar_config(disable=__UpperCAmelCase ) lowerCAmelCase__ :Tuple = torch.Generator(device='cpu' ).manual_seed(0 ) lowerCAmelCase__ , lowerCAmelCase__ :Any = pipe_prior( __UpperCAmelCase , image=__UpperCAmelCase , strength=0.85 , generator=__UpperCAmelCase , negative_prompt='' , ).to_tuple() lowerCAmelCase__ :int = pipeline( image=__UpperCAmelCase , image_embeds=__UpperCAmelCase , negative_image_embeds=__UpperCAmelCase , hint=__UpperCAmelCase , generator=__UpperCAmelCase , num_inference_steps=1_0_0 , height=5_1_2 , width=5_1_2 , strength=0.5 , output_type='np' , ) lowerCAmelCase__ :Any = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert_mean_pixel_difference(__UpperCAmelCase , __UpperCAmelCase )
93
import argparse import hashlib # hashlib is only used inside the Test class import struct class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase ): UpperCAmelCase__ : Any = data UpperCAmelCase__ : List[Any] = [0X6745_2301, 0Xefcd_ab89, 0X98ba_dcfe, 0X1032_5476, 0Xc3d2_e1f0] @staticmethod def __UpperCAmelCase ( _lowerCAmelCase , _lowerCAmelCase ): return ((n << b) | (n >> (32 - b))) & 0Xffff_ffff def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) UpperCAmelCase__ : Optional[int] = self.data + padding + struct.pack(""">Q""" , 8 * len(self.data ) ) return padded_data def __UpperCAmelCase ( self ): return [ self.padded_data[i : i + 64] for i in range(0 , len(self.padded_data ) , 64 ) ] def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : Dict = list(struct.unpack(""">16L""" , _lowerCAmelCase ) ) + [0] * 64 for i in range(16 , 80 ): UpperCAmelCase__ : Optional[int] = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) , 1 ) return w def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[str] = self.padding() UpperCAmelCase__ : List[str] = self.split_blocks() for block in self.blocks: UpperCAmelCase__ : Tuple = self.expand_block(_lowerCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.h for i in range(0 , 80 ): if 0 <= i < 20: UpperCAmelCase__ : Optional[int] = (b & c) | ((~b) & d) UpperCAmelCase__ : int = 0X5a82_7999 elif 20 <= i < 40: UpperCAmelCase__ : Tuple = b ^ c ^ d UpperCAmelCase__ : int = 0X6ed9_eba1 elif 40 <= i < 60: UpperCAmelCase__ : List[str] = (b & c) | (b & d) | (c & d) UpperCAmelCase__ : Tuple = 0X8f1b_bcdc elif 60 <= i < 80: UpperCAmelCase__ : int = b ^ c ^ d UpperCAmelCase__ : str = 0Xca62_c1d6 UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = ( self.rotate(_lowerCAmelCase , 5 ) + f + e + k + expanded_block[i] & 0Xffff_ffff, a, self.rotate(_lowerCAmelCase , 30 ), c, d, ) UpperCAmelCase__ : int = ( self.h[0] + a & 0Xffff_ffff, self.h[1] + b & 0Xffff_ffff, self.h[2] + c & 0Xffff_ffff, self.h[3] + d & 0Xffff_ffff, self.h[4] + e & 0Xffff_ffff, ) return ("{:08x}" * 5).format(*self.h ) def _lowerCamelCase ( ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase__ : Optional[Any] = B"""Test String""" assert SHAaHash(__lowerCamelCase ).final_hash() == hashlib.shaa(__lowerCamelCase ).hexdigest() # noqa: S324 def _lowerCamelCase ( ) -> str: '''simple docstring''' UpperCAmelCase__ : Optional[int] = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument("""--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) UpperCAmelCase__ : str = parser.parse_args() UpperCAmelCase__ : Union[str, Any] = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: UpperCAmelCase__ : List[Any] = f.read() else: UpperCAmelCase__ : int = bytes(__lowerCamelCase , """utf-8""" ) print(SHAaHash(__lowerCamelCase ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
79
0
'''simple docstring''' from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class UpperCAmelCase_ : """simple docstring""" UpperCamelCase_ = 42 UpperCamelCase_ = None UpperCamelCase_ = None def lowercase_ ( ) -> Node | None: """simple docstring""" lowercase : Tuple =Node(1 ) lowercase : List[Any] =Node(2 ) lowercase : str =Node(3 ) lowercase : Union[str, Any] =Node(4 ) lowercase : List[Any] =Node(5 ) return tree def lowercase_ ( __A : Node | None ) -> list[int]: """simple docstring""" return [root.data, *preorder(root.left ), *preorder(root.right )] if root else [] def lowercase_ ( __A : Node | None ) -> list[int]: """simple docstring""" return postorder(root.left ) + postorder(root.right ) + [root.data] if root else [] def lowercase_ ( __A : Node | None ) -> list[int]: """simple docstring""" return [*inorder(root.left ), root.data, *inorder(root.right )] if root else [] def lowercase_ ( __A : Node | None ) -> int: """simple docstring""" return (max(height(root.left ) , height(root.right ) ) + 1) if root else 0 def lowercase_ ( __A : Node | None ) -> Sequence[Node | None]: """simple docstring""" lowercase : list[Any] =[] if root is None: return output lowercase : Dict =deque([root] ) while process_queue: lowercase : int =process_queue.popleft() output.append(node.data ) if node.left: process_queue.append(node.left ) if node.right: process_queue.append(node.right ) return output def lowercase_ ( __A : Node | None , __A : int ) -> Sequence[Node | None]: """simple docstring""" lowercase : list[Any] =[] def populate_output(__A : Node | None , __A : int ) -> None: if not root: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.left , level - 1 ) populate_output(root.right , level - 1 ) populate_output(__A , __A ) return output def lowercase_ ( __A : Node | None , __A : int ) -> Sequence[Node | None]: """simple docstring""" lowercase : list[Any] =[] def populate_output(__A : Node | None , __A : int ) -> None: if root is None: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.right , level - 1 ) populate_output(root.left , level - 1 ) populate_output(__A , __A ) return output def lowercase_ ( __A : Node | None ) -> Sequence[Node | None] | list[Any]: """simple docstring""" if root is None: return [] lowercase : list[Sequence[Node | None]] =[] lowercase : str =0 lowercase : int =height(__A ) for h in range(1 , height_tree + 1 ): if not flag: output.append(get_nodes_from_left_to_right(__A , __A ) ) lowercase : Optional[Any] =1 else: output.append(get_nodes_from_right_to_left(__A , __A ) ) lowercase : Dict =0 return output def lowercase_ ( ) -> None: # Main function for testing. """simple docstring""" lowercase : int =make_tree() print(F'In-order Traversal: {inorder(__A )}' ) print(F'Pre-order Traversal: {preorder(__A )}' ) print(F'Post-order Traversal: {postorder(__A )}' , '''\n''' ) print(F'Height of Tree: {height(__A )}' , '''\n''' ) print('''Complete Level Order Traversal: ''' ) print(level_order(__A ) , '''\n''' ) print('''Level-wise order Traversal: ''' ) for level in range(1 , height(__A ) + 1 ): print(F'Level {level}:' , get_nodes_from_left_to_right(__A , level=__A ) ) print('''\nZigZag order Traversal: ''' ) print(zigzag(__A ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
94
from importlib import import_module from .logging import get_logger SCREAMING_SNAKE_CASE__ : Union[str, Any] = get_logger(__name__) class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : List[str] = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith("""__""" ): setattr(self , _lowerCAmelCase , getattr(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase__ : Tuple = module._original_module if isinstance(_lowerCAmelCase , _PatchedModuleObj ) else module class UpperCAmelCase_ : __lowerCamelCase = [] def __init__( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : str = obj UpperCAmelCase__ : List[str] = target UpperCAmelCase__ : List[str] = new UpperCAmelCase__ : Any = target.split(""".""" )[0] UpperCAmelCase__ : Union[str, Any] = {} UpperCAmelCase__ : str = attrs or [] def __enter__( self ): *UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.target.split(""".""" ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(_lowerCAmelCase ) ): try: UpperCAmelCase__ : Optional[int] = import_module(""".""".join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): UpperCAmelCase__ : Any = getattr(self.obj , _lowerCAmelCase ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(_lowerCAmelCase , _PatchedModuleObj ) and obj_attr._original_module is submodule) ): UpperCAmelCase__ : List[Any] = obj_attr # patch at top level setattr(self.obj , _lowerCAmelCase , _PatchedModuleObj(_lowerCAmelCase , attrs=self.attrs ) ) UpperCAmelCase__ : Optional[Any] = getattr(self.obj , _lowerCAmelCase ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(_lowerCAmelCase , _lowerCAmelCase , _PatchedModuleObj(getattr(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) , attrs=self.attrs ) ) UpperCAmelCase__ : Union[str, Any] = getattr(_lowerCAmelCase , _lowerCAmelCase ) # finally set the target attribute setattr(_lowerCAmelCase , _lowerCAmelCase , self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: UpperCAmelCase__ : Union[str, Any] = getattr(import_module(""".""".join(_lowerCAmelCase ) ) , _lowerCAmelCase ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , _lowerCAmelCase ) is attr_value: UpperCAmelCase__ : Optional[int] = getattr(self.obj , _lowerCAmelCase ) setattr(self.obj , _lowerCAmelCase , self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" UpperCAmelCase__ : Dict = globals()["""__builtins__"""][target_attr] setattr(self.obj , _lowerCAmelCase , self.new ) else: raise RuntimeError(f"Tried to patch attribute {target_attr} instead of a submodule." ) def __exit__( self , *_lowerCAmelCase ): for attr in list(self.original ): setattr(self.obj , _lowerCAmelCase , self.original.pop(_lowerCAmelCase ) ) def __UpperCAmelCase ( self ): self.__enter__() self._active_patches.append(self ) def __UpperCAmelCase ( self ): try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
79
0
"""simple docstring""" import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 lowerCamelCase_ = data_utils.TransfoXLTokenizer lowerCamelCase_ = data_utils.TransfoXLCorpus lowerCamelCase_ = data_utils lowerCamelCase_ = data_utils def snake_case ( A__ ,A__ ,A__ ,A__ ): if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(A__ ,"rb" ) as fp: UpperCAmelCase_ : List[str] = pickle.load(A__ ,encoding="latin1" ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) UpperCAmelCase_ : Dict = pytorch_dump_folder_path + "/" + VOCAB_FILES_NAMES["pretrained_vocab_file"] print(F"""Save vocabulary to {pytorch_vocab_dump_path}""" ) UpperCAmelCase_ : Tuple = corpus.vocab.__dict__ torch.save(A__ ,A__ ) UpperCAmelCase_ : Union[str, Any] = corpus.__dict__ corpus_dict_no_vocab.pop("vocab" ,A__ ) UpperCAmelCase_ : List[Any] = pytorch_dump_folder_path + "/" + CORPUS_NAME print(F"""Save dataset to {pytorch_dataset_dump_path}""" ) torch.save(A__ ,A__ ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model UpperCAmelCase_ : Optional[Any] = os.path.abspath(A__ ) UpperCAmelCase_ : List[Any] = os.path.abspath(A__ ) print(F"""Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.""" ) # Initialise PyTorch model if transfo_xl_config_file == "": UpperCAmelCase_ : Optional[Any] = TransfoXLConfig() else: UpperCAmelCase_ : Dict = TransfoXLConfig.from_json_file(A__ ) print(F"""Building PyTorch model from configuration: {config}""" ) UpperCAmelCase_ : Union[str, Any] = TransfoXLLMHeadModel(A__ ) UpperCAmelCase_ : Union[str, Any] = load_tf_weights_in_transfo_xl(A__ ,A__ ,A__ ) # Save pytorch-model UpperCAmelCase_ : Any = os.path.join(A__ ,A__ ) UpperCAmelCase_ : Union[str, Any] = os.path.join(A__ ,A__ ) print(F"""Save PyTorch model to {os.path.abspath(A__ )}""" ) torch.save(model.state_dict() ,A__ ) print(F"""Save configuration file to {os.path.abspath(A__ )}""" ) with open(A__ ,"w" ,encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the folder to store the PyTorch model or dataset/vocab.''', ) parser.add_argument( '''--tf_checkpoint_path''', default='''''', type=str, help='''An optional path to a TensorFlow checkpoint path to be converted.''', ) parser.add_argument( '''--transfo_xl_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained BERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--transfo_xl_dataset_file''', default='''''', type=str, help='''An optional dataset file to be converted in a vocabulary.''', ) lowerCamelCase_ = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
95
from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : str = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Any = { """huggingface/informer-tourism-monthly""": ( """https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json""" ), # See all Informer models at https://huggingface.co/models?filter=informer } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'informer' __lowerCamelCase = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', 'num_hidden_layers': 'encoder_layers', } def __init__( self , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = "student_t" , _lowerCAmelCase = "nll" , _lowerCAmelCase = 1 , _lowerCAmelCase = None , _lowerCAmelCase = "mean" , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = 64 , _lowerCAmelCase = 32 , _lowerCAmelCase = 32 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = True , _lowerCAmelCase = "gelu" , _lowerCAmelCase = 0.0_5 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 100 , _lowerCAmelCase = 0.0_2 , _lowerCAmelCase=True , _lowerCAmelCase = "prob" , _lowerCAmelCase = 5 , _lowerCAmelCase = True , **_lowerCAmelCase , ): # time series specific configuration UpperCAmelCase__ : List[str] = prediction_length UpperCAmelCase__ : Optional[Any] = context_length or prediction_length UpperCAmelCase__ : str = distribution_output UpperCAmelCase__ : int = loss UpperCAmelCase__ : Optional[Any] = input_size UpperCAmelCase__ : Any = num_time_features UpperCAmelCase__ : int = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] UpperCAmelCase__ : Union[str, Any] = scaling UpperCAmelCase__ : Optional[Any] = num_dynamic_real_features UpperCAmelCase__ : List[str] = num_static_real_features UpperCAmelCase__ : str = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(_lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The cardinality should be a list of the same length as `num_static_categorical_features`""" ) UpperCAmelCase__ : List[str] = cardinality else: UpperCAmelCase__ : Optional[Any] = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(_lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The embedding dimension should be a list of the same length as `num_static_categorical_features`""" ) UpperCAmelCase__ : str = embedding_dimension else: UpperCAmelCase__ : List[str] = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] UpperCAmelCase__ : Union[str, Any] = num_parallel_samples # Transformer architecture configuration UpperCAmelCase__ : Dict = input_size * len(self.lags_sequence ) + self._number_of_features UpperCAmelCase__ : Any = d_model UpperCAmelCase__ : int = encoder_attention_heads UpperCAmelCase__ : Optional[Any] = decoder_attention_heads UpperCAmelCase__ : int = encoder_ffn_dim UpperCAmelCase__ : Tuple = decoder_ffn_dim UpperCAmelCase__ : List[Any] = encoder_layers UpperCAmelCase__ : Optional[Any] = decoder_layers UpperCAmelCase__ : Tuple = dropout UpperCAmelCase__ : int = attention_dropout UpperCAmelCase__ : List[str] = activation_dropout UpperCAmelCase__ : Any = encoder_layerdrop UpperCAmelCase__ : Union[str, Any] = decoder_layerdrop UpperCAmelCase__ : Tuple = activation_function UpperCAmelCase__ : Dict = init_std UpperCAmelCase__ : str = use_cache # Informer UpperCAmelCase__ : Union[str, Any] = attention_type UpperCAmelCase__ : int = sampling_factor UpperCAmelCase__ : Any = distil super().__init__(is_encoder_decoder=_lowerCAmelCase , **_lowerCAmelCase ) @property def __UpperCAmelCase ( self ): return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
79
0
"""simple docstring""" def a ( __UpperCAmelCase : list[int] ) -> float: if not nums: # Makes sure that the list is not empty raise ValueError("""List is empty""" ) __magic_name__: Dict = sum(__UpperCAmelCase ) / len(__UpperCAmelCase ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(__UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
96
def _lowerCamelCase ( __lowerCamelCase ) -> bool: '''simple docstring''' if p < 2: raise ValueError("""p should not be less than 2!""" ) elif p == 2: return True UpperCAmelCase__ : Tuple = 4 UpperCAmelCase__ : Tuple = (1 << p) - 1 for _ in range(p - 2 ): UpperCAmelCase__ : List[str] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
79
0
import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def a ( snake_case__: List[Any] ): '''simple docstring''' if "cls_token" in name: lowercase_ = name.replace('''cls_token''' , '''vit.embeddings.cls_token''' ) if "mask_token" in name: lowercase_ = name.replace('''mask_token''' , '''decoder.mask_token''' ) if "decoder_pos_embed" in name: lowercase_ = name.replace('''decoder_pos_embed''' , '''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: lowercase_ = name.replace('''pos_embed''' , '''vit.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: lowercase_ = name.replace('''patch_embed.proj''' , '''vit.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: lowercase_ = name.replace('''patch_embed.norm''' , '''vit.embeddings.norm''' ) if "decoder_blocks" in name: lowercase_ = name.replace('''decoder_blocks''' , '''decoder.decoder_layers''' ) if "blocks" in name: lowercase_ = name.replace('''blocks''' , '''vit.encoder.layer''' ) if "attn.proj" in name: lowercase_ = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: lowercase_ = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: lowercase_ = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: lowercase_ = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: lowercase_ = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: lowercase_ = name.replace('''mlp.fc2''' , '''output.dense''' ) if "decoder_embed" in name: lowercase_ = name.replace('''decoder_embed''' , '''decoder.decoder_embed''' ) if "decoder_norm" in name: lowercase_ = name.replace('''decoder_norm''' , '''decoder.decoder_norm''' ) if "decoder_pred" in name: lowercase_ = name.replace('''decoder_pred''' , '''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name: lowercase_ = name.replace('''norm.weight''' , '''vit.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name: lowercase_ = name.replace('''norm.bias''' , '''vit.layernorm.bias''' ) return name def a ( snake_case__: Optional[int] , snake_case__: Any ): '''simple docstring''' for key in orig_state_dict.copy().keys(): lowercase_ = orig_state_dict.pop(snake_case__ ) if "qkv" in key: lowercase_ = key.split('''.''' ) lowercase_ = int(key_split[1] ) if "decoder_blocks" in key: lowercase_ = config.decoder_hidden_size lowercase_ = '''decoder.decoder_layers.''' if "weight" in key: lowercase_ = val[:dim, :] lowercase_ = val[dim : dim * 2, :] lowercase_ = val[-dim:, :] elif "bias" in key: lowercase_ = val[:dim] lowercase_ = val[dim : dim * 2] lowercase_ = val[-dim:] else: lowercase_ = config.hidden_size lowercase_ = '''vit.encoder.layer.''' if "weight" in key: lowercase_ = val[:dim, :] lowercase_ = val[dim : dim * 2, :] lowercase_ = val[-dim:, :] elif "bias" in key: lowercase_ = val[:dim] lowercase_ = val[dim : dim * 2] lowercase_ = val[-dim:] else: lowercase_ = val return orig_state_dict def a ( snake_case__: str , snake_case__: int ): '''simple docstring''' lowercase_ = ViTMAEConfig() if "large" in checkpoint_url: lowercase_ = 1_024 lowercase_ = 4_096 lowercase_ = 24 lowercase_ = 16 elif "huge" in checkpoint_url: lowercase_ = 14 lowercase_ = 1_280 lowercase_ = 5_120 lowercase_ = 32 lowercase_ = 16 lowercase_ = ViTMAEForPreTraining(snake_case__ ) lowercase_ = torch.hub.load_state_dict_from_url(snake_case__ , map_location='''cpu''' )['''model'''] lowercase_ = ViTMAEImageProcessor(size=config.image_size ) lowercase_ = convert_state_dict(snake_case__ , snake_case__ ) model.load_state_dict(snake_case__ ) model.eval() lowercase_ = '''https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg''' lowercase_ = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) lowercase_ = ViTMAEImageProcessor(size=config.image_size ) lowercase_ = image_processor(images=snake_case__ , return_tensors='''pt''' ) # forward pass torch.manual_seed(2 ) lowercase_ = model(**snake_case__ ) lowercase_ = outputs.logits if "large" in checkpoint_url: lowercase_ = torch.tensor( [[-0.7_3_0_9, -0.7_1_2_8, -1.0_1_6_9], [-1.0_1_6_1, -0.9_0_5_8, -1.1_8_7_8], [-1.0_4_7_8, -0.9_4_1_1, -1.1_9_1_1]] ) elif "huge" in checkpoint_url: lowercase_ = torch.tensor( [[-1.1_5_9_9, -0.9_1_9_9, -1.2_2_2_1], [-1.1_9_5_2, -0.9_2_6_9, -1.2_3_0_7], [-1.2_1_4_3, -0.9_3_3_7, -1.2_2_6_2]] ) else: lowercase_ = torch.tensor( [[-0.9_1_9_2, -0.8_4_8_1, -1.1_2_5_9], [-1.1_3_4_9, -1.0_0_3_4, -1.2_5_9_9], [-1.1_7_5_7, -1.0_4_2_9, -1.2_7_2_6]] ) # verify logits assert torch.allclose(logits[0, :3, :3] , snake_case__ , atol=1e-4 ) print(F'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(snake_case__ ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(snake_case__ ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth', type=str, help='URL of the checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) __a = parser.parse_args() convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
97
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE__ : Any = { """configuration_mobilevit""": ["""MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MobileViTConfig""", """MobileViTOnnxConfig"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : List[str] = ["""MobileViTFeatureExtractor"""] SCREAMING_SNAKE_CASE__ : Union[str, Any] = ["""MobileViTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Dict = [ """MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MobileViTForImageClassification""", """MobileViTForSemanticSegmentation""", """MobileViTModel""", """MobileViTPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Any = [ """TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFMobileViTForImageClassification""", """TFMobileViTForSemanticSegmentation""", """TFMobileViTModel""", """TFMobileViTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mobilevit import MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileViTConfig, MobileViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilevit import MobileViTFeatureExtractor from .image_processing_mobilevit import MobileViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilevit import ( MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel, MobileViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilevit import ( TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileViTForImageClassification, TFMobileViTForSemanticSegmentation, TFMobileViTModel, TFMobileViTPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
79
0
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device lowercase__ : Any = False class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" pass @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Optional[Any] ) -> Any: '''simple docstring''' _UpperCamelCase = VersatileDiffusionImageVariationPipeline.from_pretrained('''shi-labs/versatile-diffusion''' ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) _UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) _UpperCamelCase = torch.manual_seed(0 ) _UpperCamelCase = pipe( image=lowerCAmelCase__ , generator=lowerCAmelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' , ).images _UpperCamelCase = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _UpperCamelCase = np.array([0.0441, 0.0469, 0.0507, 0.0575, 0.0632, 0.0650, 0.0865, 0.0909, 0.0945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
98
from __future__ import annotations SCREAMING_SNAKE_CASE__ : List[str] = 8.988e9 # units = N * m^s * C^-2 def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> dict[str, float]: '''simple docstring''' UpperCAmelCase__ : int = abs(chargea * chargea ) if (force, chargea, chargea, distance).count(0 ) != 1: raise ValueError("""One and only one argument must be 0""" ) if distance < 0: raise ValueError("""Distance cannot be negative""" ) if force == 0: UpperCAmelCase__ : int = COULOMBS_CONSTANT * charge_product / (distance**2) return {"force": force} elif chargea == 0: UpperCAmelCase__ : str = abs(__lowerCamelCase ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge1": chargea} elif chargea == 0: UpperCAmelCase__ : Union[str, Any] = abs(__lowerCamelCase ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge2": chargea} elif distance == 0: UpperCAmelCase__ : Optional[Any] = (COULOMBS_CONSTANT * charge_product / abs(__lowerCamelCase )) ** 0.5 return {"distance": distance} raise ValueError("""Exactly one argument must be 0""" ) if __name__ == "__main__": import doctest doctest.testmod()
79
0
from typing import Dict, List, Optional, Tuple, Union import torch from ...models import AutoencoderKL, TransformeraDModel from ...schedulers import KarrasDiffusionSchedulers from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class __UpperCAmelCase ( __A ): """simple docstring""" def __init__( self , __A , __A , __A , __A = None , ): super().__init__() self.register_modules(transformer=__A , vae=__A , scheduler=__A ) # create a imagenet -> id dictionary for easier use __a = {} if idalabel is not None: for key, value in idalabel.items(): for label in value.split(""",""" ): __a = int(__A ) __a = dict(sorted(self.labels.items() ) ) def snake_case_ ( self , __A ): if not isinstance(__A , __A ): __a = list(__A ) for l in label: if l not in self.labels: raise ValueError( f'''{l} does not exist. Please make sure to select one of the following labels: \n {self.labels}.''' ) return [self.labels[l] for l in label] @torch.no_grad() def __call__( self , __A , __A = 4.0 , __A = None , __A = 50 , __A = "pil" , __A = True , ): __a = len(__A ) __a = self.transformer.config.sample_size __a = self.transformer.config.in_channels __a = randn_tensor( shape=(batch_size, latent_channels, latent_size, latent_size) , generator=__A , device=self.device , dtype=self.transformer.dtype , ) __a = torch.cat([latents] * 2 ) if guidance_scale > 1 else latents __a = torch.tensor(__A , device=self.device ).reshape(-1 ) __a = torch.tensor([1000] * batch_size , device=self.device ) __a = torch.cat([class_labels, class_null] , 0 ) if guidance_scale > 1 else class_labels # set step values self.scheduler.set_timesteps(__A ) for t in self.progress_bar(self.scheduler.timesteps ): if guidance_scale > 1: __a = latent_model_input[: len(__A ) // 2] __a = torch.cat([half, half] , dim=0 ) __a = self.scheduler.scale_model_input(__A , __A ) __a = t if not torch.is_tensor(__A ): # TODO: this requires sync between CPU and GPU. So try to pass timesteps as tensors if you can # This would be a good case for the `match` statement (Python 3.10+) __a = latent_model_input.device.type == """mps""" if isinstance(__A , __A ): __a = torch.floataa if is_mps else torch.floataa else: __a = torch.intaa if is_mps else torch.intaa __a = torch.tensor([timesteps] , dtype=__A , device=latent_model_input.device ) elif len(timesteps.shape ) == 0: __a = timesteps[None].to(latent_model_input.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML __a = timesteps.expand(latent_model_input.shape[0] ) # predict noise model_output __a = self.transformer( __A , timestep=__A , class_labels=__A ).sample # perform guidance if guidance_scale > 1: __a , __a = noise_pred[:, :latent_channels], noise_pred[:, latent_channels:] __a , __a = torch.split(__A , len(__A ) // 2 , dim=0 ) __a = uncond_eps + guidance_scale * (cond_eps - uncond_eps) __a = torch.cat([half_eps, half_eps] , dim=0 ) __a = torch.cat([eps, rest] , dim=1 ) # learned sigma if self.transformer.config.out_channels // 2 == latent_channels: __a , __a = torch.split(__A , __A , dim=1 ) else: __a = noise_pred # compute previous image: x_t -> x_t-1 __a = self.scheduler.step(__A , __A , __A ).prev_sample if guidance_scale > 1: __a , __a = latent_model_input.chunk(2 , dim=0 ) else: __a = latent_model_input __a = 1 / self.vae.config.scaling_factor * latents __a = self.vae.decode(__A ).sample __a = (samples / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 __a = samples.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __a = self.numpy_to_pil(__A ) if not return_dict: return (samples,) return ImagePipelineOutput(images=__A )
99
class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase ): # we need a list not a string, so do something to change the type UpperCAmelCase__ : Dict = arr.split(""",""" ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = [int(self.array[0] )] * len(self.array ) UpperCAmelCase__ : List[str] = [int(self.array[0] )] * len(self.array ) for i in range(1 , len(self.array ) ): UpperCAmelCase__ : Tuple = max( int(self.array[i] ) + sum_value[i - 1] , int(self.array[i] ) ) UpperCAmelCase__ : Union[str, Any] = max(sum_value[i] , rear[i - 1] ) return rear[len(self.array ) - 1] if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Tuple = input("""please input some numbers:""") SCREAMING_SNAKE_CASE__ : Dict = SubArray(whole_array) SCREAMING_SNAKE_CASE__ : Dict = array.solve_sub_array() print(("""the results is:""", re))
79
0
from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool from google.protobuf import symbol_database as _symbol_database from google.protobuf.internal import builder as _builder # @@protoc_insertion_point(imports) _A : Optional[int] = _symbol_database.Default() _A : Any = _descriptor_pool.Default().AddSerializedFile( b"""\n\x19sentencepiece_model.proto\x12\rsentencepiece\"\x80\x0c\n\x0bTrainerSpec\x12\r\n\x05input\x18\x01 \x03(\t\x12\x14\n\x0cinput_format\x18\x07 \x01(\t\x12\x14\n\x0cmodel_prefix\x18\x02 \x01(\t\x12\x41\n\nmodel_type\x18\x03 \x01(\x0e\x32$.sentencepiece.TrainerSpec.ModelType:\x07UNIGRAM\x12\x18\n\nvocab_size\x18\x04 \x01(\x05:\x04\x38\x30\x30\x30\x12\x17\n\x0f\x61\x63\x63\x65pt_language\x18\x05 \x03(\t\x12 \n\x15self_test_sample_size\x18\x06 \x01(\x05:\x01\x30\x12*\n\x1b\x65nable_differential_privacy\x18\x32 \x01(\x08:\x05\x66\x61lse\x12+\n differential_privacy_noise_level\x18\x33 \x01(\x02:\x01\x30\x12\x32\n\'differential_privacy_clipping_threshold\x18\x34 \x01(\x04:\x01\x30\x12\"\n\x12\x63haracter_coverage\x18\n \x01(\x02:\x06\x30.9995\x12\x1e\n\x13input_sentence_size\x18\x0b \x01(\x04:\x01\x30\x12$\n\x16shuffle_input_sentence\x18\x13 \x01(\x08:\x04true\x12 \n\x14mining_sentence_size\x18\x0c \x01(\x05\x42\x02\x18\x01\x12\"\n\x16training_sentence_size\x18\r \x01(\x05\x42\x02\x18\x01\x12(\n\x17seed_sentencepiece_size\x18\x0e \x01(\x05:\x07\x31\x30\x30\x30\x30\x30\x30\x12\x1e\n\x10shrinking_factor\x18\x0f \x01(\x02:\x04\x30.75\x12!\n\x13max_sentence_length\x18\x12 \x01(\x05:\x04\x34\x31\x39\x32\x12\x17\n\x0bnum_threads\x18\x10 \x01(\x05:\x02\x31\x36\x12\x1d\n\x12num_sub_iterations\x18\x11 \x01(\x05:\x01\x32\x12$\n\x18max_sentencepiece_length\x18\x14 \x01(\x05:\x02\x31\x36\x12%\n\x17split_by_unicode_script\x18\x15 \x01(\x08:\x04true\x12\x1d\n\x0fsplit_by_number\x18\x17 \x01(\x08:\x04true\x12!\n\x13split_by_whitespace\x18\x16 \x01(\x08:\x04true\x12)\n\x1atreat_whitespace_as_suffix\x18\x18 \x01(\x08:\x05\x66\x61lse\x12+\n\x1c\x61llow_whitespace_only_pieces\x18\x1a \x01(\x08:\x05\x66\x61lse\x12\x1b\n\x0csplit_digits\x18\x19 \x01(\x08:\x05\x66\x61lse\x12#\n\x19pretokenization_delimiter\x18\x35 \x01(\t:\x00\x12\x17\n\x0f\x63ontrol_symbols\x18\x1e \x03(\t\x12\x1c\n\x14user_defined_symbols\x18\x1f \x03(\t\x12\x16\n\x0erequired_chars\x18$ \x01(\t\x12\x1c\n\rbyte_fallback\x18# \x01(\x08:\x05\x66\x61lse\x12+\n\x1dvocabulary_output_piece_score\x18 \x01(\x08:\x04true\x12\x1e\n\x10hard_vocab_limit\x18! \x01(\x08:\x04true\x12\x1c\n\ruse_all_vocab\x18\" \x01(\x08:\x05\x66\x61lse\x12\x11\n\x06unk_id\x18( \x01(\x05:\x01\x30\x12\x11\n\x06\x62os_id\x18) \x01(\x05:\x01\x31\x12\x11\n\x06\x65os_id\x18* \x01(\x05:\x01\x32\x12\x12\n\x06pad_id\x18+ \x01(\x05:\x02-1\x12\x18\n\tunk_piece\x18- \x01(\t:\x05<unk>\x12\x16\n\tbos_piece\x18. \x01(\t:\x03<s>\x12\x17\n\teos_piece\x18/ \x01(\t:\x04</s>\x12\x18\n\tpad_piece\x18\x30 \x01(\t:\x05<pad>\x12\x1a\n\x0bunk_surface\x18, \x01(\t:\x05 \xe2\x81\x87 \x12+\n\x1ctrain_extremely_large_corpus\x18\x31 \x01(\x08:\x05\x66\x61lse\"5\n\tModelType\x12\x0b\n\x07UNIGRAM\x10\x01\x12\x07\n\x03\x42PE\x10\x02\x12\x08\n\x04WORD\x10\x03\x12\x08\n\x04\x43HAR\x10\x04*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\"\xd1\x01\n\x0eNormalizerSpec\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x1c\n\x14precompiled_charsmap\x18\x02 \x01(\x0c\x12\x1e\n\x10\x61\x64\x64_dummy_prefix\x18\x03 \x01(\x08:\x04true\x12&\n\x18remove_extra_whitespaces\x18\x04 \x01(\x08:\x04true\x12 \n\x12\x65scape_whitespaces\x18\x05 \x01(\x08:\x04true\x12\x1e\n\x16normalization_rule_tsv\x18\x06 \x01(\t*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\"y\n\x0cSelfTestData\x12\x33\n\x07samples\x18\x01 \x03(\x0b\x32\".sentencepiece.SelfTestData.Sample\x1a)\n\x06Sample\x12\r\n\x05input\x18\x01 \x01(\t\x12\x10\n\x08\x65xpected\x18\x02 \x01(\t*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\"\xfe\x03\n\nModelProto\x12\x37\n\x06pieces\x18\x01 \x03(\x0b\x32\'.sentencepiece.ModelProto.SentencePiece\x12\x30\n\x0ctrainer_spec\x18\x02 \x01(\x0b\x32\x1a.sentencepiece.TrainerSpec\x12\x36\n\x0fnormalizer_spec\x18\x03 \x01(\x0b\x32\x1d.sentencepiece.NormalizerSpec\x12\x33\n\x0eself_test_data\x18\x04 \x01(\x0b\x32\x1b.sentencepiece.SelfTestData\x12\x38\n\x11\x64\x65normalizer_spec\x18\x05 \x01(\x0b\x32\x1d.sentencepiece.NormalizerSpec\x1a\xd2\x01\n\rSentencePiece\x12\r\n\x05piece\x18\x01 \x01(\t\x12\r\n\x05score\x18\x02 \x01(\x02\x12\x42\n\x04type\x18\x03 \x01(\x0e\x32,.sentencepiece.ModelProto.SentencePiece.Type:\x06NORMAL\"T\n\x04Type\x12\n\n\x06NORMAL\x10\x01\x12\x0b\n\x07UNKNOWN\x10\x02\x12\x0b\n\x07\x43ONTROL\x10\x03\x12\x10\n\x0cUSER_DEFINED\x10\x04\x12\x08\n\x04\x42YTE\x10\x06\x12\n\n\x06UNUSED\x10\x05*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\x42\x02H\x03""" ) _A : Tuple = globals() _builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals) _builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, """sentencepiece_model_pb2""", _globals) if _descriptor._USE_C_DESCRIPTORS is False: _A : List[Any] = None _A : Any = b"""H\003""" # (generated by protobuf compiler, but `_TRAINERSPEC` is not defined) # _TRAINERSPEC.fields_by_name["mining_sentence_size"]._options = None # _TRAINERSPEC.fields_by_name["mining_sentence_size"]._serialized_options = b"\030\001" # _TRAINERSPEC.fields_by_name["training_sentence_size"]._options = None # _TRAINERSPEC.fields_by_name["training_sentence_size"]._serialized_options = b"\030\001" _A : Union[str, Any] = 45 _A : Optional[int] = 15_81 _A : int = 15_17 _A : List[Any] = 15_70 _A : Dict = 15_84 _A : List[str] = 17_93 _A : List[str] = 17_95 _A : List[Any] = 19_16 _A : List[str] = 18_64 _A : int = 19_05 _A : str = 19_19 _A : Any = 24_29 _A : Optional[Any] = 22_08 _A : Optional[Any] = 24_18 _A : str = 23_23 _A : Tuple = 24_07 # @@protoc_insertion_point(module_scope)
100
from ....configuration_utils import PretrainedConfig from ....utils import logging SCREAMING_SNAKE_CASE__ : List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Any = { """Visual-Attention-Network/van-base""": ( """https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json""" ), } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'van' def __init__( self , _lowerCAmelCase=224 , _lowerCAmelCase=3 , _lowerCAmelCase=[7, 3, 3, 3] , _lowerCAmelCase=[4, 2, 2, 2] , _lowerCAmelCase=[64, 128, 320, 512] , _lowerCAmelCase=[3, 3, 12, 3] , _lowerCAmelCase=[8, 8, 4, 4] , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=1e-6 , _lowerCAmelCase=1e-2 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) UpperCAmelCase__ : Tuple = image_size UpperCAmelCase__ : Optional[Any] = num_channels UpperCAmelCase__ : Optional[int] = patch_sizes UpperCAmelCase__ : int = strides UpperCAmelCase__ : Optional[int] = hidden_sizes UpperCAmelCase__ : str = depths UpperCAmelCase__ : Optional[Any] = mlp_ratios UpperCAmelCase__ : List[Any] = hidden_act UpperCAmelCase__ : Tuple = initializer_range UpperCAmelCase__ : Any = layer_norm_eps UpperCAmelCase__ : List[Any] = layer_scale_init_value UpperCAmelCase__ : int = drop_path_rate UpperCAmelCase__ : Dict = dropout_rate
79
0
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import AlignProcessor, EfficientNetImageProcessor @require_vision class __lowercase (unittest.TestCase ): """simple docstring""" def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ : List[Any] = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] SCREAMING_SNAKE_CASE_ : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) SCREAMING_SNAKE_CASE_ : Any = { 'do_resize': True, 'size': 2_0, 'do_center_crop': True, 'crop_size': 1_8, 'do_normalize': True, 'image_mean': [0.48_145_466, 0.4_578_275, 0.40_821_073], 'image_std': [0.26_862_954, 0.26_130_258, 0.27_577_711], } SCREAMING_SNAKE_CASE_ : List[str] = os.path.join(self.tmpdirname , lowerCAmelCase__ ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase__ ( self , **lowerCAmelCase__ ): """simple docstring""" return BertTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def UpperCamelCase__ ( self , **lowerCAmelCase__ ): """simple docstring""" return BertTokenizerFast.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def UpperCamelCase__ ( self , **lowerCAmelCase__ ): """simple docstring""" return EfficientNetImageProcessor.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def UpperCamelCase__ ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] SCREAMING_SNAKE_CASE_ : str = [Image.fromarray(np.moveaxis(lowerCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : Dict = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.get_image_processor() SCREAMING_SNAKE_CASE_ : List[Any] = AlignProcessor(tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ ) processor_slow.save_pretrained(self.tmpdirname ) SCREAMING_SNAKE_CASE_ : Optional[int] = AlignProcessor.from_pretrained(self.tmpdirname , use_fast=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : str = AlignProcessor(tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ ) processor_fast.save_pretrained(self.tmpdirname ) SCREAMING_SNAKE_CASE_ : List[Any] = AlignProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , lowerCAmelCase__ ) self.assertIsInstance(processor_fast.tokenizer , lowerCAmelCase__ ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , lowerCAmelCase__ ) self.assertIsInstance(processor_fast.image_processor , lowerCAmelCase__ ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = AlignProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) SCREAMING_SNAKE_CASE_ : Optional[int] = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) SCREAMING_SNAKE_CASE_ : Any = self.get_image_processor(do_normalize=lowerCAmelCase__ , padding_value=1.0 ) SCREAMING_SNAKE_CASE_ : List[str] = AlignProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=lowerCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowerCAmelCase__ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCAmelCase__ ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = self.get_image_processor() SCREAMING_SNAKE_CASE_ : Optional[int] = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : List[Any] = AlignProcessor(tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Optional[int] = self.prepare_image_inputs() SCREAMING_SNAKE_CASE_ : Optional[int] = image_processor(lowerCAmelCase__ , return_tensors='np' ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = processor(images=lowerCAmelCase__ , return_tensors='np' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = self.get_image_processor() SCREAMING_SNAKE_CASE_ : Optional[int] = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : Optional[Any] = AlignProcessor(tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : int = 'lower newer' SCREAMING_SNAKE_CASE_ : Tuple = processor(text=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Tuple = tokenizer(lowerCAmelCase__ , padding='max_length' , max_length=6_4 ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = self.get_image_processor() SCREAMING_SNAKE_CASE_ : Optional[Any] = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : Any = AlignProcessor(tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Any = 'lower newer' SCREAMING_SNAKE_CASE_ : List[str] = self.prepare_image_inputs() SCREAMING_SNAKE_CASE_ : Optional[Any] = processor(text=lowerCAmelCase__ , images=lowerCAmelCase__ ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'token_type_ids', 'attention_mask', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(lowerCAmelCase__ ): processor() def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = self.get_image_processor() SCREAMING_SNAKE_CASE_ : List[Any] = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : str = AlignProcessor(tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : List[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] SCREAMING_SNAKE_CASE_ : Union[str, Any] = processor.batch_decode(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : str = tokenizer.batch_decode(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = self.get_image_processor() SCREAMING_SNAKE_CASE_ : Tuple = self.get_tokenizer() SCREAMING_SNAKE_CASE_ : Optional[int] = AlignProcessor(tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Any = 'lower newer' SCREAMING_SNAKE_CASE_ : Optional[Any] = self.prepare_image_inputs() SCREAMING_SNAKE_CASE_ : Union[str, Any] = processor(text=lowerCAmelCase__ , images=lowerCAmelCase__ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
101
import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Any: '''simple docstring''' UpperCAmelCase__ : List[str] = s.rsplit(__lowerCamelCase , __lowerCamelCase ) return new.join(__lowerCamelCase ) def _lowerCamelCase ( __lowerCamelCase ) -> str: '''simple docstring''' # encoder.embeddings are double copied in original FLAVA return sum(param.float().sum() if """encoder.embeddings""" not in key else 0 for key, param in state_dict.items() ) def _lowerCamelCase ( __lowerCamelCase ) -> int: '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = {} UpperCAmelCase__ : Union[str, Any] = ["""group_1""", """group_2""", """group_3""", """group_4"""] for key, value in state_dict.items(): for group_key in group_keys: if group_key in key: UpperCAmelCase__ : Optional[Any] = key.replace(F"{group_key}." , F"{group_key}.group." ) if "res_path" in key: UpperCAmelCase__ : Optional[int] = key.replace("""res_path.""" , """res_path.path.""" ) if key.endswith(""".w""" ): UpperCAmelCase__ : List[Any] = rreplace(__lowerCamelCase , """.w""" , """.weight""" , 1 ) if key.endswith(""".b""" ): UpperCAmelCase__ : Optional[int] = rreplace(__lowerCamelCase , """.b""" , """.bias""" , 1 ) UpperCAmelCase__ : Union[str, Any] = value.float() return upgrade @torch.no_grad() def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=None , __lowerCamelCase=True ) -> str: '''simple docstring''' from dall_e import Encoder UpperCAmelCase__ : Dict = Encoder() if os.path.exists(__lowerCamelCase ): UpperCAmelCase__ : Optional[Any] = torch.load(__lowerCamelCase ) else: UpperCAmelCase__ : Tuple = torch.hub.load_state_dict_from_url(__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase__ : Any = ckpt.state_dict() encoder.load_state_dict(__lowerCamelCase ) if config_path is not None: UpperCAmelCase__ : Dict = FlavaImageCodebookConfig.from_pretrained(__lowerCamelCase ) else: UpperCAmelCase__ : Optional[Any] = FlavaImageCodebookConfig() UpperCAmelCase__ : Optional[Any] = FlavaImageCodebook(__lowerCamelCase ).eval() UpperCAmelCase__ : str = encoder.state_dict() UpperCAmelCase__ : Optional[int] = upgrade_state_dict(__lowerCamelCase ) hf_model.load_state_dict(__lowerCamelCase ) UpperCAmelCase__ : List[str] = hf_model.state_dict() UpperCAmelCase__ : Tuple = count_parameters(__lowerCamelCase ) UpperCAmelCase__ : int = count_parameters(__lowerCamelCase ) assert torch.allclose(__lowerCamelCase , __lowerCamelCase , atol=1E-3 ) if save_checkpoint: hf_model.save_pretrained(__lowerCamelCase ) else: return hf_state_dict if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Any = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to flava checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") SCREAMING_SNAKE_CASE__ : int = parser.parse_args() convert_dalle_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
79
0
"""simple docstring""" import argparse import collections import os import re import tempfile import pandas as pd from datasets import Dataset from huggingface_hub import hf_hub_download, upload_folder from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/update_metadata.py __magic_name__ : int = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. __magic_name__ : Any = direct_transformers_import(TRANSFORMERS_PATH) # Regexes that match TF/Flax/PT model names. __magic_name__ : Dict = re.compile(R"""TF(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") __magic_name__ : Tuple = re.compile(R"""Flax(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # Will match any TF or Flax model too so need to be in an else branch afterthe two previous regexes. __magic_name__ : Tuple = re.compile(R"""(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # Fill this with tuples (pipeline_tag, model_mapping, auto_model) __magic_name__ : Tuple = [ ("""pretraining""", """MODEL_FOR_PRETRAINING_MAPPING_NAMES""", """AutoModelForPreTraining"""), ("""feature-extraction""", """MODEL_MAPPING_NAMES""", """AutoModel"""), ("""audio-classification""", """MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForAudioClassification"""), ("""text-generation""", """MODEL_FOR_CAUSAL_LM_MAPPING_NAMES""", """AutoModelForCausalLM"""), ("""automatic-speech-recognition""", """MODEL_FOR_CTC_MAPPING_NAMES""", """AutoModelForCTC"""), ("""image-classification""", """MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForImageClassification"""), ("""image-segmentation""", """MODEL_FOR_IMAGE_SEGMENTATION_MAPPING_NAMES""", """AutoModelForImageSegmentation"""), ("""fill-mask""", """MODEL_FOR_MASKED_LM_MAPPING_NAMES""", """AutoModelForMaskedLM"""), ("""object-detection""", """MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES""", """AutoModelForObjectDetection"""), ( """zero-shot-object-detection""", """MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING_NAMES""", """AutoModelForZeroShotObjectDetection""", ), ("""question-answering""", """MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES""", """AutoModelForQuestionAnswering"""), ("""text2text-generation""", """MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES""", """AutoModelForSeq2SeqLM"""), ("""text-classification""", """MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForSequenceClassification"""), ("""automatic-speech-recognition""", """MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES""", """AutoModelForSpeechSeq2Seq"""), ( """table-question-answering""", """MODEL_FOR_TABLE_QUESTION_ANSWERING_MAPPING_NAMES""", """AutoModelForTableQuestionAnswering""", ), ("""token-classification""", """MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForTokenClassification"""), ("""multiple-choice""", """MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES""", """AutoModelForMultipleChoice"""), ( """next-sentence-prediction""", """MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES""", """AutoModelForNextSentencePrediction""", ), ( """audio-frame-classification""", """MODEL_FOR_AUDIO_FRAME_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForAudioFrameClassification""", ), ("""audio-xvector""", """MODEL_FOR_AUDIO_XVECTOR_MAPPING_NAMES""", """AutoModelForAudioXVector"""), ( """document-question-answering""", """MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES""", """AutoModelForDocumentQuestionAnswering""", ), ( """visual-question-answering""", """MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING_NAMES""", """AutoModelForVisualQuestionAnswering""", ), ("""image-to-text""", """MODEL_FOR_FOR_VISION_2_SEQ_MAPPING_NAMES""", """AutoModelForVision2Seq"""), ( """zero-shot-image-classification""", """MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForZeroShotImageClassification""", ), ("""depth-estimation""", """MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES""", """AutoModelForDepthEstimation"""), ("""video-classification""", """MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForVideoClassification"""), ("""mask-generation""", """MODEL_FOR_MASK_GENERATION_MAPPING_NAMES""", """AutoModelForMaskGeneration"""), ] def UpperCamelCase (SCREAMING_SNAKE_CASE ): UpperCamelCase : Any = re.finditer(""".+?(?:(?<=[a-z])(?=[A-Z])|(?<=[A-Z])(?=[A-Z][a-z])|$)""" , SCREAMING_SNAKE_CASE ) return [m.group(0 ) for m in matches] def UpperCamelCase (): UpperCamelCase : Union[str, Any] = transformers_module.models.auto.configuration_auto.CONFIG_MAPPING_NAMES UpperCamelCase : List[str] = { config.replace("""Config""" , """""" ): model_type for model_type, config in config_maping_names.items() } # Dictionaries flagging if each model prefix has a backend in PT/TF/Flax. UpperCamelCase : Dict = collections.defaultdict(SCREAMING_SNAKE_CASE ) UpperCamelCase : Union[str, Any] = collections.defaultdict(SCREAMING_SNAKE_CASE ) UpperCamelCase : Dict = collections.defaultdict(SCREAMING_SNAKE_CASE ) # Let's lookup through all transformers object (once) and find if models are supported by a given backend. for attr_name in dir(SCREAMING_SNAKE_CASE ): UpperCamelCase : Any = None if _re_tf_models.match(SCREAMING_SNAKE_CASE ) is not None: UpperCamelCase : List[str] = tf_models UpperCamelCase : List[str] = _re_tf_models.match(SCREAMING_SNAKE_CASE ).groups()[0] elif _re_flax_models.match(SCREAMING_SNAKE_CASE ) is not None: UpperCamelCase : int = flax_models UpperCamelCase : Tuple = _re_flax_models.match(SCREAMING_SNAKE_CASE ).groups()[0] elif _re_pt_models.match(SCREAMING_SNAKE_CASE ) is not None: UpperCamelCase : Optional[Any] = pt_models UpperCamelCase : str = _re_pt_models.match(SCREAMING_SNAKE_CASE ).groups()[0] if lookup_dict is not None: while len(SCREAMING_SNAKE_CASE ) > 0: if attr_name in model_prefix_to_model_type: UpperCamelCase : List[str] = True break # Try again after removing the last word in the name UpperCamelCase : int = """""".join(camel_case_split(SCREAMING_SNAKE_CASE )[:-1] ) UpperCamelCase : Any = set(list(pt_models.keys() ) + list(tf_models.keys() ) + list(flax_models.keys() ) ) UpperCamelCase : List[Any] = list(SCREAMING_SNAKE_CASE ) all_models.sort() UpperCamelCase : Optional[Any] = {"""model_type""": all_models} UpperCamelCase : List[Any] = [pt_models[t] for t in all_models] UpperCamelCase : Dict = [tf_models[t] for t in all_models] UpperCamelCase : int = [flax_models[t] for t in all_models] # Now let's use the auto-mapping names to make sure UpperCamelCase : Optional[Any] = {} for t in all_models: if t in transformers_module.models.auto.processing_auto.PROCESSOR_MAPPING_NAMES: UpperCamelCase : Optional[Any] = """AutoProcessor""" elif t in transformers_module.models.auto.tokenization_auto.TOKENIZER_MAPPING_NAMES: UpperCamelCase : int = """AutoTokenizer""" elif t in transformers_module.models.auto.feature_extraction_auto.FEATURE_EXTRACTOR_MAPPING_NAMES: UpperCamelCase : str = """AutoFeatureExtractor""" else: # Default to AutoTokenizer if a model has nothing, for backward compatibility. UpperCamelCase : Dict = """AutoTokenizer""" UpperCamelCase : int = [processors[t] for t in all_models] return pd.DataFrame(SCREAMING_SNAKE_CASE ) def UpperCamelCase (SCREAMING_SNAKE_CASE ): UpperCamelCase : List[Any] = [ transformers_module.models.auto.modeling_auto, transformers_module.models.auto.modeling_tf_auto, transformers_module.models.auto.modeling_flax_auto, ] for pipeline_tag, model_mapping, auto_class in PIPELINE_TAGS_AND_AUTO_MODELS: UpperCamelCase : Any = [model_mapping, f"""TF_{model_mapping}""", f"""FLAX_{model_mapping}"""] UpperCamelCase : str = [auto_class, f"""TF_{auto_class}""", f"""Flax_{auto_class}"""] # Loop through all three frameworks for module, cls, mapping in zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): # The type of pipeline may not exist in this framework if not hasattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): continue # First extract all model_names UpperCamelCase : List[str] = [] for name in getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).values(): if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): model_names.append(SCREAMING_SNAKE_CASE ) else: model_names.extend(list(SCREAMING_SNAKE_CASE ) ) # Add pipeline tag and auto model class for those models table.update({model_name: (pipeline_tag, cls) for model_name in model_names} ) return table def UpperCamelCase (SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): UpperCamelCase : Optional[Any] = get_frameworks_table() UpperCamelCase : str = Dataset.from_pandas(SCREAMING_SNAKE_CASE ) UpperCamelCase : Any = hf_hub_download( """huggingface/transformers-metadata""" , """pipeline_tags.json""" , repo_type="""dataset""" , token=SCREAMING_SNAKE_CASE ) UpperCamelCase : List[Any] = Dataset.from_json(SCREAMING_SNAKE_CASE ) UpperCamelCase : Tuple = { tags_dataset[i]["""model_class"""]: (tags_dataset[i]["""pipeline_tag"""], tags_dataset[i]["""auto_class"""]) for i in range(len(SCREAMING_SNAKE_CASE ) ) } UpperCamelCase : Any = update_pipeline_and_auto_class_table(SCREAMING_SNAKE_CASE ) # Sort the model classes to avoid some nondeterministic updates to create false update commits. UpperCamelCase : Tuple = sorted(table.keys() ) UpperCamelCase : str = pd.DataFrame( { """model_class""": model_classes, """pipeline_tag""": [table[m][0] for m in model_classes], """auto_class""": [table[m][1] for m in model_classes], } ) UpperCamelCase : Tuple = Dataset.from_pandas(SCREAMING_SNAKE_CASE ) with tempfile.TemporaryDirectory() as tmp_dir: frameworks_dataset.to_json(os.path.join(SCREAMING_SNAKE_CASE , """frameworks.json""" ) ) tags_dataset.to_json(os.path.join(SCREAMING_SNAKE_CASE , """pipeline_tags.json""" ) ) if commit_sha is not None: UpperCamelCase : Optional[Any] = ( f"""Update with commit {commit_sha}\n\nSee: """ f"""https://github.com/huggingface/transformers/commit/{commit_sha}""" ) else: UpperCamelCase : Optional[int] = """Update""" upload_folder( repo_id="""huggingface/transformers-metadata""" , folder_path=SCREAMING_SNAKE_CASE , repo_type="""dataset""" , token=SCREAMING_SNAKE_CASE , commit_message=SCREAMING_SNAKE_CASE , ) def UpperCamelCase (): UpperCamelCase : Dict = {tag: cls for tag, _, cls in PIPELINE_TAGS_AND_AUTO_MODELS} UpperCamelCase : str = transformers_module.pipelines.SUPPORTED_TASKS UpperCamelCase : Optional[Any] = [] for key in pipeline_tasks: if key not in in_table: UpperCamelCase : List[Any] = pipeline_tasks[key]["""pt"""] if isinstance(SCREAMING_SNAKE_CASE , (list, tuple) ): UpperCamelCase : Optional[Any] = model[0] UpperCamelCase : Optional[int] = model.__name__ if model not in in_table.values(): missing.append(SCREAMING_SNAKE_CASE ) if len(SCREAMING_SNAKE_CASE ) > 0: UpperCamelCase : Union[str, Any] = """, """.join(SCREAMING_SNAKE_CASE ) raise ValueError( """The following pipeline tags are not present in the `PIPELINE_TAGS_AND_AUTO_MODELS` constant inside """ f"""`utils/update_metadata.py`: {msg}. Please add them!""" ) if __name__ == "__main__": __magic_name__ : Dict = argparse.ArgumentParser() parser.add_argument("""--token""", type=str, help="""The token to use to push to the transformers-metadata dataset.""") parser.add_argument("""--commit_sha""", type=str, help="""The sha of the commit going with this update.""") parser.add_argument("""--check-only""", action="""store_true""", help="""Activate to just check all pipelines are present.""") __magic_name__ : List[Any] = parser.parse_args() if args.check_only: check_pipeline_tags() else: update_metadata(args.token, args.commit_sha)
102
def _lowerCamelCase ( __lowerCamelCase ) -> int: '''simple docstring''' return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def _lowerCamelCase ( __lowerCamelCase ) -> bool: '''simple docstring''' UpperCAmelCase__ : Any = 0 UpperCAmelCase__ : Union[str, Any] = number while duplicate > 0: UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = divmod(__lowerCamelCase , 10 ) fact_sum += factorial(__lowerCamelCase ) return fact_sum == number if __name__ == "__main__": print("""Program to check whether a number is a Krisnamurthy Number or not.""") SCREAMING_SNAKE_CASE__ : Optional[Any] = int(input("""Enter number: """).strip()) print( f'''{number} is {"" if krishnamurthy(number) else "not "}a Krishnamurthy Number.''' )
79
0
"""simple docstring""" import random from typing import Any def snake_case ( lowerCAmelCase_ ) -> list[Any]: for _ in range(len(lowerCAmelCase_ ) ): _snake_case = random.randint(0 , len(lowerCAmelCase_ ) - 1 ) _snake_case = random.randint(0 , len(lowerCAmelCase_ ) - 1 ) _snake_case , _snake_case = data[b], data[a] return data if __name__ == "__main__": snake_case = [0, 1, 2, 3, 4, 5, 6, 7] snake_case = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
103
def _lowerCamelCase ( __lowerCamelCase = 100_0000 ) -> int: '''simple docstring''' UpperCAmelCase__ : Tuple = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , __lowerCamelCase ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
79
0
"""simple docstring""" import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class UpperCamelCase__ ( _lowerCAmelCase ): """simple docstring""" @require_torch def snake_case__ ( self ) -> Dict: # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched A__ = "\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n " A__ = "\nmname = \"hf-internal-testing/tiny-random-bert\"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task=\"fill-mask\", model=mname)\nprint(\"success\")\n " A__ = "\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError(\"Offline mode is enabled, we shouldn't access internet\")\nsocket.socket = offline_socket\n " # Force fetching the files so that we can use the cache A__ = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(SCREAMING_SNAKE_CASE__ ) BertModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) BertTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ ) pipeline(task="fill-mask" , model=SCREAMING_SNAKE_CASE__ ) # baseline - just load from_pretrained with normal network A__ = [sys.executable, "-c", "\n".join([load, run, mock] )] # should succeed A__ = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files A__ = "1" A__ = subprocess.run(SCREAMING_SNAKE_CASE__ , env=SCREAMING_SNAKE_CASE__ , check=SCREAMING_SNAKE_CASE__ , capture_output=SCREAMING_SNAKE_CASE__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) @require_torch def snake_case__ ( self ) -> Optional[Any]: # python one-liner segments # this must be loaded before socket.socket is monkey-patched A__ = "\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n " A__ = "\nmname = \"hf-internal-testing/tiny-random-bert\"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task=\"fill-mask\", model=mname)\nprint(\"success\")\n " A__ = "\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error(\"Faking flaky internet\")\nsocket.socket = offline_socket\n " # Force fetching the files so that we can use the cache A__ = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(SCREAMING_SNAKE_CASE__ ) BertModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) BertTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ ) pipeline(task="fill-mask" , model=SCREAMING_SNAKE_CASE__ ) # baseline - just load from_pretrained with normal network A__ = [sys.executable, "-c", "\n".join([load, run, mock] )] # should succeed A__ = self.get_env() A__ = subprocess.run(SCREAMING_SNAKE_CASE__ , env=SCREAMING_SNAKE_CASE__ , check=SCREAMING_SNAKE_CASE__ , capture_output=SCREAMING_SNAKE_CASE__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) @require_torch def snake_case__ ( self ) -> Tuple: # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched A__ = "\nfrom transformers import BertConfig, BertModel, BertTokenizer\n " A__ = "\nmname = \"hf-internal-testing/tiny-random-bert-sharded\"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint(\"success\")\n " A__ = "\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError(\"Offline mode is enabled\")\nsocket.socket = offline_socket\n " # baseline - just load from_pretrained with normal network A__ = [sys.executable, "-c", "\n".join([load, run] )] # should succeed A__ = self.get_env() A__ = subprocess.run(SCREAMING_SNAKE_CASE__ , env=SCREAMING_SNAKE_CASE__ , check=SCREAMING_SNAKE_CASE__ , capture_output=SCREAMING_SNAKE_CASE__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) # next emulate no network A__ = [sys.executable, "-c", "\n".join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files A__ = "1" A__ = subprocess.run(SCREAMING_SNAKE_CASE__ , env=SCREAMING_SNAKE_CASE__ , check=SCREAMING_SNAKE_CASE__ , capture_output=SCREAMING_SNAKE_CASE__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) @require_torch def snake_case__ ( self ) -> Optional[int]: A__ = "\nfrom transformers import pipeline\n " A__ = "\nmname = \"hf-internal-testing/tiny-random-bert\"\npipe = pipeline(model=mname)\n " A__ = "\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error(\"Offline mode is enabled\")\nsocket.socket = offline_socket\n " A__ = self.get_env() A__ = "1" A__ = [sys.executable, "-c", "\n".join([load, mock, run] )] A__ = subprocess.run(SCREAMING_SNAKE_CASE__ , env=SCREAMING_SNAKE_CASE__ , check=SCREAMING_SNAKE_CASE__ , capture_output=SCREAMING_SNAKE_CASE__ ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( "You cannot infer task automatically within `pipeline` when using offline mode" , result.stderr.decode().replace("\n" , "" ) , ) @require_torch def snake_case__ ( self ) -> str: A__ = "\nfrom transformers import AutoModel\n " A__ = "\nmname = \"hf-internal-testing/test_dynamic_model\"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint(\"success\")\n " # baseline - just load from_pretrained with normal network A__ = [sys.executable, "-c", "\n".join([load, run] )] # should succeed A__ = self.get_env() A__ = subprocess.run(SCREAMING_SNAKE_CASE__ , env=SCREAMING_SNAKE_CASE__ , check=SCREAMING_SNAKE_CASE__ , capture_output=SCREAMING_SNAKE_CASE__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files A__ = "1" A__ = subprocess.run(SCREAMING_SNAKE_CASE__ , env=SCREAMING_SNAKE_CASE__ , check=SCREAMING_SNAKE_CASE__ , capture_output=SCREAMING_SNAKE_CASE__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() )
104
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Tuple = { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json""", # See all REALM models at https://huggingface.co/models?filter=realm } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'realm' def __init__( self , _lowerCAmelCase=30522 , _lowerCAmelCase=768 , _lowerCAmelCase=128 , _lowerCAmelCase=12 , _lowerCAmelCase=12 , _lowerCAmelCase=8 , _lowerCAmelCase=3072 , _lowerCAmelCase="gelu_new" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=512 , _lowerCAmelCase=2 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=1e-12 , _lowerCAmelCase=256 , _lowerCAmelCase=10 , _lowerCAmelCase=1e-3 , _lowerCAmelCase=5 , _lowerCAmelCase=320 , _lowerCAmelCase=13353718 , _lowerCAmelCase=5000 , _lowerCAmelCase=1 , _lowerCAmelCase=0 , _lowerCAmelCase=2 , **_lowerCAmelCase , ): super().__init__(pad_token_id=_lowerCAmelCase , bos_token_id=_lowerCAmelCase , eos_token_id=_lowerCAmelCase , **_lowerCAmelCase ) # Common config UpperCAmelCase__ : List[Any] = vocab_size UpperCAmelCase__ : Dict = max_position_embeddings UpperCAmelCase__ : Any = hidden_size UpperCAmelCase__ : str = retriever_proj_size UpperCAmelCase__ : Tuple = num_hidden_layers UpperCAmelCase__ : List[str] = num_attention_heads UpperCAmelCase__ : List[Any] = num_candidates UpperCAmelCase__ : str = intermediate_size UpperCAmelCase__ : str = hidden_act UpperCAmelCase__ : Optional[Any] = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : Union[str, Any] = initializer_range UpperCAmelCase__ : Any = type_vocab_size UpperCAmelCase__ : Optional[Any] = layer_norm_eps # Reader config UpperCAmelCase__ : str = span_hidden_size UpperCAmelCase__ : Union[str, Any] = max_span_width UpperCAmelCase__ : List[str] = reader_layer_norm_eps UpperCAmelCase__ : Dict = reader_beam_size UpperCAmelCase__ : Union[str, Any] = reader_seq_len # Retrieval config UpperCAmelCase__ : List[Any] = num_block_records UpperCAmelCase__ : List[Any] = searcher_beam_size
79
0
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. UpperCamelCase__ : str = {'''LayoutLMv2Config''', '''LayoutLMv3Config'''} @is_pipeline_test class lowerCAmelCase_ ( unittest.TestCase ): __a : Dict = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __a : List[Any] = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: __a : Union[str, Any] = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: __a : Any = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } @require_torch def snake_case ( self ): SCREAMING_SNAKE_CASE_ : List[Any] = pipeline( task='text-classification' ,model='hf-internal-testing/tiny-random-distilbert' ,framework='pt' ) SCREAMING_SNAKE_CASE_ : Optional[int] = text_classifier('This is great !' ) self.assertEqual(nested_simplify(snake_case__ ) ,[{'label': 'LABEL_0', 'score': 0.504}] ) SCREAMING_SNAKE_CASE_ : Any = text_classifier('This is great !' ,top_k=2 ) self.assertEqual( nested_simplify(snake_case__ ) ,[{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}] ) SCREAMING_SNAKE_CASE_ : Tuple = text_classifier(['This is great !', 'This is bad'] ,top_k=2 ) self.assertEqual( nested_simplify(snake_case__ ) ,[ [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}], [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}], ] ,) SCREAMING_SNAKE_CASE_ : Tuple = text_classifier('This is great !' ,top_k=1 ) self.assertEqual(nested_simplify(snake_case__ ) ,[{'label': 'LABEL_0', 'score': 0.504}] ) # Legacy behavior SCREAMING_SNAKE_CASE_ : Optional[Any] = text_classifier('This is great !' ,return_all_scores=snake_case__ ) self.assertEqual(nested_simplify(snake_case__ ) ,[{'label': 'LABEL_0', 'score': 0.504}] ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = text_classifier('This is great !' ,return_all_scores=snake_case__ ) self.assertEqual( nested_simplify(snake_case__ ) ,[[{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}]] ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = text_classifier(['This is great !', 'Something else'] ,return_all_scores=snake_case__ ) self.assertEqual( nested_simplify(snake_case__ ) ,[ [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}], [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}], ] ,) SCREAMING_SNAKE_CASE_ : Tuple = text_classifier(['This is great !', 'Something else'] ,return_all_scores=snake_case__ ) self.assertEqual( nested_simplify(snake_case__ ) ,[ {'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_0', 'score': 0.504}, ] ,) @require_torch def snake_case ( self ): import torch SCREAMING_SNAKE_CASE_ : Optional[Any] = pipeline( task='text-classification' ,model='hf-internal-testing/tiny-random-distilbert' ,framework='pt' ,device=torch.device('cpu' ) ,) SCREAMING_SNAKE_CASE_ : List[Any] = text_classifier('This is great !' ) self.assertEqual(nested_simplify(snake_case__ ) ,[{'label': 'LABEL_0', 'score': 0.504}] ) @require_tf def snake_case ( self ): SCREAMING_SNAKE_CASE_ : str = pipeline( task='text-classification' ,model='hf-internal-testing/tiny-random-distilbert' ,framework='tf' ) SCREAMING_SNAKE_CASE_ : int = text_classifier('This is great !' ) self.assertEqual(nested_simplify(snake_case__ ) ,[{'label': 'LABEL_0', 'score': 0.504}] ) @slow @require_torch def snake_case ( self ): SCREAMING_SNAKE_CASE_ : Tuple = pipeline('text-classification' ) SCREAMING_SNAKE_CASE_ : List[str] = text_classifier('This is great !' ) self.assertEqual(nested_simplify(snake_case__ ) ,[{'label': 'POSITIVE', 'score': 1.0}] ) SCREAMING_SNAKE_CASE_ : Optional[int] = text_classifier('This is bad !' ) self.assertEqual(nested_simplify(snake_case__ ) ,[{'label': 'NEGATIVE', 'score': 1.0}] ) SCREAMING_SNAKE_CASE_ : List[Any] = text_classifier('Birds are a type of animal' ) self.assertEqual(nested_simplify(snake_case__ ) ,[{'label': 'POSITIVE', 'score': 0.988}] ) @slow @require_tf def snake_case ( self ): SCREAMING_SNAKE_CASE_ : Optional[int] = pipeline('text-classification' ,framework='tf' ) SCREAMING_SNAKE_CASE_ : List[Any] = text_classifier('This is great !' ) self.assertEqual(nested_simplify(snake_case__ ) ,[{'label': 'POSITIVE', 'score': 1.0}] ) SCREAMING_SNAKE_CASE_ : Optional[int] = text_classifier('This is bad !' ) self.assertEqual(nested_simplify(snake_case__ ) ,[{'label': 'NEGATIVE', 'score': 1.0}] ) SCREAMING_SNAKE_CASE_ : Tuple = text_classifier('Birds are a type of animal' ) self.assertEqual(nested_simplify(snake_case__ ) ,[{'label': 'POSITIVE', 'score': 0.988}] ) def snake_case ( self ,snake_case__ ,snake_case__ ,snake_case__ ): SCREAMING_SNAKE_CASE_ : List[str] = TextClassificationPipeline(model=snake_case__ ,tokenizer=snake_case__ ) return text_classifier, ["HuggingFace is in", "This is another test"] def snake_case ( self ,snake_case__ ,snake_case__ ): SCREAMING_SNAKE_CASE_ : str = text_classifier.model # Small inputs because BartTokenizer tiny has maximum position embeddings = 22 SCREAMING_SNAKE_CASE_ : Dict = 'HuggingFace is in' SCREAMING_SNAKE_CASE_ : Optional[int] = text_classifier(snake_case__ ) self.assertEqual(nested_simplify(snake_case__ ) ,[{'label': ANY(snake_case__ ), 'score': ANY(snake_case__ )}] ) self.assertTrue(outputs[0]['label'] in model.config.idalabel.values() ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = ['HuggingFace is in ', 'Paris is in France'] SCREAMING_SNAKE_CASE_ : Optional[Any] = text_classifier(snake_case__ ) self.assertEqual( nested_simplify(snake_case__ ) ,[{'label': ANY(snake_case__ ), 'score': ANY(snake_case__ )}, {'label': ANY(snake_case__ ), 'score': ANY(snake_case__ )}] ,) self.assertTrue(outputs[0]['label'] in model.config.idalabel.values() ) self.assertTrue(outputs[1]['label'] in model.config.idalabel.values() ) # Forcing to get all results with `top_k=None` # This is NOT the legacy format SCREAMING_SNAKE_CASE_ : Tuple = text_classifier(snake_case__ ,top_k=snake_case__ ) SCREAMING_SNAKE_CASE_ : List[Any] = len(model.config.idalabel.values() ) self.assertEqual( nested_simplify(snake_case__ ) ,[[{'label': ANY(snake_case__ ), 'score': ANY(snake_case__ )}] * N, [{'label': ANY(snake_case__ ), 'score': ANY(snake_case__ )}] * N] ,) SCREAMING_SNAKE_CASE_ : List[str] = {'text': 'HuggingFace is in ', 'text_pair': 'Paris is in France'} SCREAMING_SNAKE_CASE_ : List[Any] = text_classifier(snake_case__ ) self.assertEqual( nested_simplify(snake_case__ ) ,{'label': ANY(snake_case__ ), 'score': ANY(snake_case__ )} ,) self.assertTrue(outputs['label'] in model.config.idalabel.values() ) # This might be used a text pair, but tokenizer + pipe interaction # makes it hard to understand that it's not using the pair properly # https://github.com/huggingface/transformers/issues/17305 # We disabled this usage instead as it was outputting wrong outputs. SCREAMING_SNAKE_CASE_ : List[str] = [['HuggingFace is in ', 'Paris is in France']] with self.assertRaises(snake_case__ ): text_classifier(snake_case__ ) # This used to be valid for doing text pairs # We're keeping it working because of backward compatibility SCREAMING_SNAKE_CASE_ : Tuple = text_classifier([[['HuggingFace is in ', 'Paris is in France']]] ) self.assertEqual( nested_simplify(snake_case__ ) ,[{'label': ANY(snake_case__ ), 'score': ANY(snake_case__ )}] ,) self.assertTrue(outputs[0]['label'] in model.config.idalabel.values() )
105
import gc import unittest from parameterized import parameterized from diffusers import FlaxUNetaDConditionModel from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp @slow @require_flax class UpperCAmelCase_ ( unittest.TestCase ): def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase ): return f"gaussian_noise_s={seed}_shape={'_'.join([str(_lowerCAmelCase ) for s in shape] )}.npy" def __UpperCAmelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def __UpperCAmelCase ( self , _lowerCAmelCase=0 , _lowerCAmelCase=(4, 4, 64, 64) , _lowerCAmelCase=False ): UpperCAmelCase__ : Optional[Any] = jnp.bfloataa if fpaa else jnp.floataa UpperCAmelCase__ : Union[str, Any] = jnp.array(load_hf_numpy(self.get_file_format(_lowerCAmelCase , _lowerCAmelCase ) ) , dtype=_lowerCAmelCase ) return image def __UpperCAmelCase ( self , _lowerCAmelCase=False , _lowerCAmelCase="CompVis/stable-diffusion-v1-4" ): UpperCAmelCase__ : int = jnp.bfloataa if fpaa else jnp.floataa UpperCAmelCase__ : Optional[Any] = """bf16""" if fpaa else None UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = FlaxUNetaDConditionModel.from_pretrained( _lowerCAmelCase , subfolder="""unet""" , dtype=_lowerCAmelCase , revision=_lowerCAmelCase ) return model, params def __UpperCAmelCase ( self , _lowerCAmelCase=0 , _lowerCAmelCase=(4, 77, 768) , _lowerCAmelCase=False ): UpperCAmelCase__ : Optional[int] = jnp.bfloataa if fpaa else jnp.floataa UpperCAmelCase__ : Optional[int] = jnp.array(load_hf_numpy(self.get_file_format(_lowerCAmelCase , _lowerCAmelCase ) ) , dtype=_lowerCAmelCase ) return hidden_states @parameterized.expand( [ # fmt: off [83, 4, [-0.2_3_2_3, -0.1_3_0_4, 0.0_8_1_3, -0.3_0_9_3, -0.0_9_1_9, -0.1_5_7_1, -0.1_1_2_5, -0.5_8_0_6]], [17, 0.5_5, [-0.0_8_3_1, -0.2_4_4_3, 0.0_9_0_1, -0.0_9_1_9, 0.3_3_9_6, 0.0_1_0_3, -0.3_7_4_3, 0.0_7_0_1]], [8, 0.8_9, [-0.4_8_6_3, 0.0_8_5_9, 0.0_8_7_5, -0.1_6_5_8, 0.9_1_9_9, -0.0_1_1_4, 0.4_8_3_9, 0.4_6_3_9]], [3, 1000, [-0.5_6_4_9, 0.2_4_0_2, -0.5_5_1_8, 0.1_2_4_8, 1.1_3_2_8, -0.2_4_4_3, -0.0_3_2_5, -1.0_0_7_8]], # fmt: on ] ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.get_unet_model(model_id="""CompVis/stable-diffusion-v1-4""" , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = self.get_latents(_lowerCAmelCase , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Dict = self.get_encoder_hidden_states(_lowerCAmelCase , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = model.apply( {"""params""": params} , _lowerCAmelCase , jnp.array(_lowerCAmelCase , dtype=jnp.intaa ) , encoder_hidden_states=_lowerCAmelCase , ).sample assert sample.shape == latents.shape UpperCAmelCase__ : Dict = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) UpperCAmelCase__ : List[Any] = jnp.array(_lowerCAmelCase , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, in the same hardware assert jnp.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-2 ) @parameterized.expand( [ # fmt: off [83, 4, [0.1_5_1_4, 0.0_8_0_7, 0.1_6_2_4, 0.1_0_1_6, -0.1_8_9_6, 0.0_2_6_3, 0.0_6_7_7, 0.2_3_1_0]], [17, 0.5_5, [0.1_1_6_4, -0.0_2_1_6, 0.0_1_7_0, 0.1_5_8_9, -0.3_1_2_0, 0.1_0_0_5, -0.0_5_8_1, -0.1_4_5_8]], [8, 0.8_9, [-0.1_7_5_8, -0.0_1_6_9, 0.1_0_0_4, -0.1_4_1_1, 0.1_3_1_2, 0.1_1_0_3, -0.1_9_9_6, 0.2_1_3_9]], [3, 1000, [0.1_2_1_4, 0.0_3_5_2, -0.0_7_3_1, -0.1_5_6_2, -0.0_9_9_4, -0.0_9_0_6, -0.2_3_4_0, -0.0_5_3_9]], # fmt: on ] ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.get_unet_model(model_id="""stabilityai/stable-diffusion-2""" , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = self.get_latents(_lowerCAmelCase , shape=(4, 4, 96, 96) , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Any = self.get_encoder_hidden_states(_lowerCAmelCase , shape=(4, 77, 1024) , fpaa=_lowerCAmelCase ) UpperCAmelCase__ : Dict = model.apply( {"""params""": params} , _lowerCAmelCase , jnp.array(_lowerCAmelCase , dtype=jnp.intaa ) , encoder_hidden_states=_lowerCAmelCase , ).sample assert sample.shape == latents.shape UpperCAmelCase__ : Any = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) UpperCAmelCase__ : Any = jnp.array(_lowerCAmelCase , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, on the same hardware assert jnp.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-2 )
79
0
import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __snake_case :Dict =logging.get_logger(__name__) class lowerCAmelCase__ ( _lowerCamelCase ): def __init__( self : Optional[Any] , *__UpperCamelCase : List[str] , **__UpperCamelCase : List[str] ) -> None: warnings.warn( 'The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use BeitImageProcessor instead.' , __UpperCamelCase , ) super().__init__(*__UpperCamelCase , **__UpperCamelCase )
106
import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class UpperCAmelCase_ ( unittest.TestCase ): def __init__( self , _lowerCAmelCase , _lowerCAmelCase=7 , _lowerCAmelCase=3 , _lowerCAmelCase=18 , _lowerCAmelCase=30 , _lowerCAmelCase=400 , _lowerCAmelCase=True , _lowerCAmelCase=None , _lowerCAmelCase=True , ): UpperCAmelCase__ : List[str] = size if size is not None else {"""height""": 18, """width""": 18} UpperCAmelCase__ : Union[str, Any] = parent UpperCAmelCase__ : int = batch_size UpperCAmelCase__ : Tuple = num_channels UpperCAmelCase__ : Dict = image_size UpperCAmelCase__ : List[Any] = min_resolution UpperCAmelCase__ : str = max_resolution UpperCAmelCase__ : Union[str, Any] = do_resize UpperCAmelCase__ : Tuple = size UpperCAmelCase__ : int = do_normalize def __UpperCAmelCase ( self ): return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.8_8_6_6_4_4_3_6_3_4_0_3_3_2_0_3, 0.6_6_1_8_8_2_9_3_6_9_5_4_4_9_8_3, 0.3_8_9_1_7_4_6_4_0_1_7_8_6_8_0_4], [-0.6_0_4_2_5_5_9_1_4_6_8_8_1_1_0_4, -0.0_2_2_9_5_0_0_8_8_6_0_5_2_8_4_6_9, 0.5_4_2_3_7_9_7_3_6_9_0_0_3_2_9_6], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class UpperCAmelCase_ ( __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = ImageGPTImageProcessor if is_vision_available() else None def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = ImageGPTImageProcessingTester(self ) @property def __UpperCAmelCase ( self ): return self.image_processor_tester.prepare_image_processor_dict() def __UpperCAmelCase ( self ): UpperCAmelCase__ : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCAmelCase , """clusters""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCAmelCase , """do_normalize""" ) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 18} ) UpperCAmelCase__ : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) UpperCAmelCase__ : Optional[int] = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase , obj[key] ) ) else: self.assertEqual(obj[key] , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase__ : Union[str, Any] = os.path.join(_lowerCAmelCase , """image_processor.json""" ) image_processor_first.to_json_file(_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = self.image_processing_class.from_json_file(_lowerCAmelCase ).to_dict() UpperCAmelCase__ : Dict = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : str = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = self.image_processing_class.from_pretrained(_lowerCAmelCase ).to_dict() UpperCAmelCase__ : Tuple = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , _lowerCAmelCase ) @unittest.skip("""ImageGPT requires clusters at initialization""" ) def __UpperCAmelCase ( self ): pass def _lowerCamelCase ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ : Any = load_dataset("""hf-internal-testing/fixtures_image_utils""" , split="""test""" ) UpperCAmelCase__ : Dict = Image.open(dataset[4]["""file"""] ) UpperCAmelCase__ : Optional[Any] = Image.open(dataset[5]["""file"""] ) UpperCAmelCase__ : List[Any] = [imagea, imagea] return images @require_vision @require_torch class UpperCAmelCase_ ( unittest.TestCase ): @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = ImageGPTImageProcessor.from_pretrained("""openai/imagegpt-small""" ) UpperCAmelCase__ : int = prepare_images() # test non-batched UpperCAmelCase__ : List[str] = image_processing(images[0] , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 1024) ) UpperCAmelCase__ : List[Any] = [306, 191, 191] self.assertEqual(encoding.input_ids[0, :3].tolist() , _lowerCAmelCase ) # test batched UpperCAmelCase__ : List[str] = image_processing(_lowerCAmelCase , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 1024) ) UpperCAmelCase__ : Any = [303, 13, 13] self.assertEqual(encoding.input_ids[1, -3:].tolist() , _lowerCAmelCase )
79
0
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TextaTextGenerationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, require_tf, require_torch from transformers.utils import is_torch_available from .test_pipelines_common import ANY if is_torch_available(): import torch @is_pipeline_test class lowercase_ ( unittest.TestCase ): """simple docstring""" __lowerCAmelCase = MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __lowerCAmelCase = TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING def __UpperCAmelCase ( self : Any, UpperCamelCase__ : str, UpperCamelCase__ : Optional[int], UpperCamelCase__ : Any ) -> List[str]: _A = TextaTextGenerationPipeline(model=UpperCamelCase__, tokenizer=UpperCamelCase__ ) return generator, ["Something to write", "Something else"] def __UpperCAmelCase ( self : Any, UpperCamelCase__ : Tuple, UpperCamelCase__ : Optional[Any] ) -> List[str]: _A = generator('Something there' ) self.assertEqual(UpperCamelCase__, [{'generated_text': ANY(UpperCamelCase__ )}] ) # These are encoder decoder, they don't just append to incoming string self.assertFalse(outputs[0]['generated_text'].startswith('Something there' ) ) _A = generator(['This is great !', 'Something else'], num_return_sequences=2, do_sample=UpperCamelCase__ ) self.assertEqual( UpperCamelCase__, [ [{'generated_text': ANY(UpperCamelCase__ )}, {'generated_text': ANY(UpperCamelCase__ )}], [{'generated_text': ANY(UpperCamelCase__ )}, {'generated_text': ANY(UpperCamelCase__ )}], ], ) _A = generator( ['This is great !', 'Something else'], num_return_sequences=2, batch_size=2, do_sample=UpperCamelCase__ ) self.assertEqual( UpperCamelCase__, [ [{'generated_text': ANY(UpperCamelCase__ )}, {'generated_text': ANY(UpperCamelCase__ )}], [{'generated_text': ANY(UpperCamelCase__ )}, {'generated_text': ANY(UpperCamelCase__ )}], ], ) with self.assertRaises(UpperCamelCase__ ): generator(4 ) @require_torch def __UpperCAmelCase ( self : Union[str, Any] ) -> Optional[Any]: _A = pipeline('text2text-generation', model='patrickvonplaten/t5-tiny-random', framework='pt' ) # do_sample=False necessary for reproducibility _A = generator('Something there', do_sample=UpperCamelCase__ ) self.assertEqual(UpperCamelCase__, [{'generated_text': ''}] ) _A = 3 _A = generator( 'Something there', num_return_sequences=UpperCamelCase__, num_beams=UpperCamelCase__, ) _A = [ {'generated_text': 'Beide Beide Beide Beide Beide Beide Beide Beide Beide'}, {'generated_text': 'Beide Beide Beide Beide Beide Beide Beide Beide'}, {'generated_text': ''}, ] self.assertEqual(UpperCamelCase__, UpperCamelCase__ ) _A = generator('This is a test', do_sample=UpperCamelCase__, num_return_sequences=2, return_tensors=UpperCamelCase__ ) self.assertEqual( UpperCamelCase__, [ {'generated_token_ids': ANY(torch.Tensor )}, {'generated_token_ids': ANY(torch.Tensor )}, ], ) _A = generator.model.config.eos_token_id _A = '<pad>' _A = generator( ['This is a test', 'This is a second test'], do_sample=UpperCamelCase__, num_return_sequences=2, batch_size=2, return_tensors=UpperCamelCase__, ) self.assertEqual( UpperCamelCase__, [ [ {'generated_token_ids': ANY(torch.Tensor )}, {'generated_token_ids': ANY(torch.Tensor )}, ], [ {'generated_token_ids': ANY(torch.Tensor )}, {'generated_token_ids': ANY(torch.Tensor )}, ], ], ) @require_tf def __UpperCAmelCase ( self : Optional[Any] ) -> Optional[int]: _A = pipeline('text2text-generation', model='patrickvonplaten/t5-tiny-random', framework='tf' ) # do_sample=False necessary for reproducibility _A = generator('Something there', do_sample=UpperCamelCase__ ) self.assertEqual(UpperCamelCase__, [{'generated_text': ''}] )
107
import os import unittest from transformers import MobileBertTokenizer, MobileBertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class UpperCAmelCase_ ( __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = MobileBertTokenizer __lowerCamelCase = MobileBertTokenizerFast __lowerCamelCase = True __lowerCamelCase = True __lowerCamelCase = filter_non_english __lowerCamelCase = 'google/mobilebert-uncased' def __UpperCAmelCase ( self ): super().setUp() UpperCAmelCase__ : Dict = [ """[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] UpperCAmelCase__ : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) UpperCAmelCase__ : List[str] = [ (tokenizer_def[0], self.pre_trained_model_path, tokenizer_def[2]) # else the 'google/' prefix is stripped for tokenizer_def in self.tokenizers_list ] def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : Tuple = """UNwant\u00E9d,running""" UpperCAmelCase__ : Union[str, Any] = """unwanted, running""" return input_text, output_text def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = self.tokenizer_class(self.vocab_file ) UpperCAmelCase__ : Tuple = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(_lowerCAmelCase , ["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [9, 6, 7, 12, 10, 11] ) def __UpperCAmelCase ( self ): if not self.test_rust_tokenizer: return UpperCAmelCase__ : Tuple = self.get_tokenizer() UpperCAmelCase__ : Dict = self.get_rust_tokenizer() UpperCAmelCase__ : List[str] = """UNwant\u00E9d,running""" UpperCAmelCase__ : Optional[int] = tokenizer.tokenize(_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = rust_tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = rust_tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.get_rust_tokenizer() UpperCAmelCase__ : Any = tokenizer.encode(_lowerCAmelCase ) UpperCAmelCase__ : str = rust_tokenizer.encode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) # With lower casing UpperCAmelCase__ : Tuple = self.get_tokenizer(do_lower_case=_lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.get_rust_tokenizer(do_lower_case=_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = """UNwant\u00E9d,running""" UpperCAmelCase__ : int = tokenizer.tokenize(_lowerCAmelCase ) UpperCAmelCase__ : Any = rust_tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = rust_tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = self.get_rust_tokenizer() UpperCAmelCase__ : List[str] = tokenizer.encode(_lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = rust_tokenizer.encode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = BasicTokenizer() self.assertListEqual(tokenizer.tokenize("""ah\u535A\u63A8zz""" ) , ["""ah""", """\u535A""", """\u63A8""", """zz"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = BasicTokenizer(do_lower_case=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Union[str, Any] = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hällo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""h\u00E9llo"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[str] = BasicTokenizer(do_lower_case=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[str] = BasicTokenizer(do_lower_case=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : int = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HäLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = BasicTokenizer(do_lower_case=_lowerCAmelCase , strip_accents=_lowerCAmelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HaLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = BasicTokenizer(do_lower_case=_lowerCAmelCase , never_split=["""[UNK]"""] ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? [UNK]""" ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?""", """[UNK]"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : int = ["""[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing"""] UpperCAmelCase__ : List[str] = {} for i, token in enumerate(_lowerCAmelCase ): UpperCAmelCase__ : Optional[Any] = i UpperCAmelCase__ : str = WordpieceTokenizer(vocab=_lowerCAmelCase , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""unwanted running""" ) , ["""un""", """##want""", """##ed""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.tokenize("""unwantedX running""" ) , ["""[UNK]""", """runn""", """##ing"""] ) def __UpperCAmelCase ( self ): self.assertTrue(_is_whitespace(""" """ ) ) self.assertTrue(_is_whitespace("""\t""" ) ) self.assertTrue(_is_whitespace("""\r""" ) ) self.assertTrue(_is_whitespace("""\n""" ) ) self.assertTrue(_is_whitespace("""\u00A0""" ) ) self.assertFalse(_is_whitespace("""A""" ) ) self.assertFalse(_is_whitespace("""-""" ) ) def __UpperCAmelCase ( self ): self.assertTrue(_is_control("""\u0005""" ) ) self.assertFalse(_is_control("""A""" ) ) self.assertFalse(_is_control(""" """ ) ) self.assertFalse(_is_control("""\t""" ) ) self.assertFalse(_is_control("""\r""" ) ) def __UpperCAmelCase ( self ): self.assertTrue(_is_punctuation("""-""" ) ) self.assertTrue(_is_punctuation("""$""" ) ) self.assertTrue(_is_punctuation("""`""" ) ) self.assertTrue(_is_punctuation(""".""" ) ) self.assertFalse(_is_punctuation("""A""" ) ) self.assertFalse(_is_punctuation(""" """ ) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = self.get_tokenizer() UpperCAmelCase__ : Union[str, Any] = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(_lowerCAmelCase ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) self.assertListEqual( [rust_tokenizer.tokenize(_lowerCAmelCase ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = self.tokenizer_class.from_pretrained("""google/mobilebert-uncased""" ) UpperCAmelCase__ : List[Any] = tokenizer.encode("""sequence builders""" , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer.encode("""multi-sequence build""" , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase ) UpperCAmelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase , _lowerCAmelCase ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def __UpperCAmelCase ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): UpperCAmelCase__ : Any = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = f"A, naïve {tokenizer_r.mask_token} AllenNLP sentence." UpperCAmelCase__ : Optional[Any] = tokenizer_r.encode_plus( _lowerCAmelCase , return_attention_mask=_lowerCAmelCase , return_token_type_ids=_lowerCAmelCase , return_offsets_mapping=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , ) UpperCAmelCase__ : Any = tokenizer_r.do_lower_case if hasattr(_lowerCAmelCase , """do_lower_case""" ) else False UpperCAmelCase__ : Optional[int] = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """A"""), ((1, 2), ""","""), ((3, 5), """na"""), ((5, 6), """##ï"""), ((6, 8), """##ve"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """Allen"""), ((21, 23), """##NL"""), ((23, 24), """##P"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """a"""), ((1, 2), ""","""), ((3, 8), """naive"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """allen"""), ((21, 23), """##nl"""), ((23, 24), """##p"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["""input_ids"""] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["""offset_mapping"""] ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Tuple = ["""的""", """人""", """有"""] UpperCAmelCase__ : Tuple = """""".join(_lowerCAmelCase ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): UpperCAmelCase__ : Tuple = True UpperCAmelCase__ : Optional[Any] = self.tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = tokenizer_p.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer_r.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : Any = tokenizer_r.convert_ids_to_tokens(_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = tokenizer_p.convert_ids_to_tokens(_lowerCAmelCase ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase__ : List[Any] = False UpperCAmelCase__ : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Tuple = self.tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = tokenizer_r.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = tokenizer_p.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = tokenizer_r.convert_ids_to_tokens(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = tokenizer_p.convert_ids_to_tokens(_lowerCAmelCase ) # it is expected that only the first Chinese character is not preceded by "##". UpperCAmelCase__ : List[str] = [ f"##{token}" if idx != 0 else token for idx, token in enumerate(_lowerCAmelCase ) ] self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase )
79
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __a: Union[str, Any] = logging.get_logger(__name__) __a: Tuple = { '''xlm-mlm-en-2048''': '''https://huggingface.co/xlm-mlm-en-2048/resolve/main/config.json''', '''xlm-mlm-ende-1024''': '''https://huggingface.co/xlm-mlm-ende-1024/resolve/main/config.json''', '''xlm-mlm-enfr-1024''': '''https://huggingface.co/xlm-mlm-enfr-1024/resolve/main/config.json''', '''xlm-mlm-enro-1024''': '''https://huggingface.co/xlm-mlm-enro-1024/resolve/main/config.json''', '''xlm-mlm-tlm-xnli15-1024''': '''https://huggingface.co/xlm-mlm-tlm-xnli15-1024/resolve/main/config.json''', '''xlm-mlm-xnli15-1024''': '''https://huggingface.co/xlm-mlm-xnli15-1024/resolve/main/config.json''', '''xlm-clm-enfr-1024''': '''https://huggingface.co/xlm-clm-enfr-1024/resolve/main/config.json''', '''xlm-clm-ende-1024''': '''https://huggingface.co/xlm-clm-ende-1024/resolve/main/config.json''', '''xlm-mlm-17-1280''': '''https://huggingface.co/xlm-mlm-17-1280/resolve/main/config.json''', '''xlm-mlm-100-1280''': '''https://huggingface.co/xlm-mlm-100-1280/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _lowerCamelCase = '''xlm''' _lowerCamelCase = { '''hidden_size''': '''emb_dim''', '''num_attention_heads''': '''n_heads''', '''num_hidden_layers''': '''n_layers''', '''n_words''': '''vocab_size''', # For backward compatibility } def __init__( self : str , lowerCamelCase : Optional[int]=3_0145 , lowerCamelCase : Dict=2048 , lowerCamelCase : Union[str, Any]=12 , lowerCamelCase : Dict=16 , lowerCamelCase : List[Any]=0.1 , lowerCamelCase : Any=0.1 , lowerCamelCase : List[Any]=True , lowerCamelCase : List[str]=False , lowerCamelCase : Any=False , lowerCamelCase : Any=False , lowerCamelCase : int=1 , lowerCamelCase : str=True , lowerCamelCase : int=512 , lowerCamelCase : Tuple=2048**-0.5 , lowerCamelCase : List[Any]=1E-12 , lowerCamelCase : str=0.02 , lowerCamelCase : Dict=0 , lowerCamelCase : Union[str, Any]=1 , lowerCamelCase : Any=2 , lowerCamelCase : Dict=3 , lowerCamelCase : Optional[int]=5 , lowerCamelCase : Optional[int]=True , lowerCamelCase : Optional[int]="first" , lowerCamelCase : Dict=True , lowerCamelCase : List[str]=None , lowerCamelCase : List[Any]=True , lowerCamelCase : List[str]=0.1 , lowerCamelCase : List[Any]=5 , lowerCamelCase : Optional[Any]=5 , lowerCamelCase : Tuple=0 , lowerCamelCase : str=0 , lowerCamelCase : int=2 , lowerCamelCase : int=0 , **lowerCamelCase : int , ) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = vocab_size _UpperCAmelCase = emb_dim _UpperCAmelCase = n_layers _UpperCAmelCase = n_heads _UpperCAmelCase = dropout _UpperCAmelCase = attention_dropout _UpperCAmelCase = gelu_activation _UpperCAmelCase = sinusoidal_embeddings _UpperCAmelCase = causal _UpperCAmelCase = asm _UpperCAmelCase = n_langs _UpperCAmelCase = use_lang_emb _UpperCAmelCase = layer_norm_eps _UpperCAmelCase = bos_index _UpperCAmelCase = eos_index _UpperCAmelCase = pad_index _UpperCAmelCase = unk_index _UpperCAmelCase = mask_index _UpperCAmelCase = is_encoder _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = embed_init_std _UpperCAmelCase = init_std _UpperCAmelCase = summary_type _UpperCAmelCase = summary_use_proj _UpperCAmelCase = summary_activation _UpperCAmelCase = summary_proj_to_labels _UpperCAmelCase = summary_first_dropout _UpperCAmelCase = start_n_top _UpperCAmelCase = end_n_top _UpperCAmelCase = mask_token_id _UpperCAmelCase = lang_id if "n_words" in kwargs: _UpperCAmelCase = kwargs["""n_words"""] super().__init__(pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , **lowerCamelCase ) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' @property def lowerCamelCase ( self : Any ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": _UpperCAmelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _UpperCAmelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
108
import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg""" UpperCAmelCase__ : int = Image.open(requests.get(__lowerCamelCase , stream=__lowerCamelCase ).raw ).convert("""RGB""" ) UpperCAmelCase__ : Any = transforms.Compose( [ transforms.Resize((image_size, image_size) , interpolation=InterpolationMode.BICUBIC ), transforms.ToTensor(), transforms.Normalize((0.48_145_466, 0.4_578_275, 0.40_821_073) , (0.26_862_954, 0.26_130_258, 0.27_577_711) ), ] ) UpperCAmelCase__ : Optional[int] = transform(__lowerCamelCase ).unsqueeze(0 ).to(__lowerCamelCase ) return image def _lowerCamelCase ( __lowerCamelCase ) -> str: '''simple docstring''' if "visual_encoder" in key: UpperCAmelCase__ : Dict = re.sub("""visual_encoder*""" , """vision_model.encoder""" , __lowerCamelCase ) if "blocks" in key: UpperCAmelCase__ : Optional[Any] = re.sub(r"""blocks""" , """layers""" , __lowerCamelCase ) if "attn" in key: UpperCAmelCase__ : List[str] = re.sub(r"""attn""" , """self_attn""" , __lowerCamelCase ) if "norm1" in key: UpperCAmelCase__ : Union[str, Any] = re.sub(r"""norm1""" , """layer_norm1""" , __lowerCamelCase ) if "norm2" in key: UpperCAmelCase__ : Any = re.sub(r"""norm2""" , """layer_norm2""" , __lowerCamelCase ) if "encoder.norm" in key: UpperCAmelCase__ : Dict = re.sub(r"""encoder.norm""" , """post_layernorm""" , __lowerCamelCase ) if "encoder.patch_embed.proj" in key: UpperCAmelCase__ : List[str] = re.sub(r"""encoder.patch_embed.proj""" , """embeddings.patch_embedding""" , __lowerCamelCase ) if "encoder.pos_embed" in key: UpperCAmelCase__ : List[str] = re.sub(r"""encoder.pos_embed""" , """embeddings.position_embedding""" , __lowerCamelCase ) if "encoder.cls_token" in key: UpperCAmelCase__ : List[Any] = re.sub(r"""encoder.cls_token""" , """embeddings.class_embedding""" , __lowerCamelCase ) if "self_attn" in key: UpperCAmelCase__ : List[Any] = re.sub(r"""self_attn.proj""" , """self_attn.projection""" , __lowerCamelCase ) return key @torch.no_grad() def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase=None ) -> Tuple: '''simple docstring''' if config_path is not None: UpperCAmelCase__ : Any = BlipConfig.from_pretrained(__lowerCamelCase ) else: UpperCAmelCase__ : str = BlipConfig(projection_dim=512 , text_config={} , vision_config={} ) UpperCAmelCase__ : int = BlipForConditionalGeneration(__lowerCamelCase ).eval() UpperCAmelCase__ : Any = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth""" UpperCAmelCase__ : List[str] = blip_decoder(pretrained=__lowerCamelCase , image_size=384 , vit="""base""" ) UpperCAmelCase__ : Union[str, Any] = pt_model.eval() UpperCAmelCase__ : Optional[int] = pt_model.state_dict() for key in modified_state_dict.copy(): UpperCAmelCase__ : Dict = modified_state_dict.pop(__lowerCamelCase ) UpperCAmelCase__ : Union[str, Any] = rename_key(__lowerCamelCase ) UpperCAmelCase__ : List[str] = value hf_model.load_state_dict(__lowerCamelCase ) UpperCAmelCase__ : Tuple = 384 UpperCAmelCase__ : str = load_demo_image(image_size=__lowerCamelCase , device="""cpu""" ) UpperCAmelCase__ : str = BertTokenizer.from_pretrained("""bert-base-uncased""" ) UpperCAmelCase__ : Dict = tokenizer(["""a picture of"""] ).input_ids UpperCAmelCase__ : int = hf_model.generate(__lowerCamelCase , __lowerCamelCase ) assert out[0].tolist() == [3_0522, 1037, 3861, 1997, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] UpperCAmelCase__ : Any = hf_model.generate(__lowerCamelCase ) assert out[0].tolist() == [3_0522, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(__lowerCamelCase ) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' UpperCAmelCase__ : Union[str, Any] = ( """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth""" ) UpperCAmelCase__ : List[Any] = blip_vqa(pretrained=__lowerCamelCase , image_size=__lowerCamelCase , vit="""base""" ) vqa_model.eval() UpperCAmelCase__ : str = vqa_model.state_dict() for key in modified_state_dict.copy(): UpperCAmelCase__ : Dict = modified_state_dict.pop(__lowerCamelCase ) UpperCAmelCase__ : Dict = rename_key(__lowerCamelCase ) UpperCAmelCase__ : int = value UpperCAmelCase__ : List[str] = BlipForQuestionAnswering(__lowerCamelCase ) hf_vqa_model.load_state_dict(__lowerCamelCase ) UpperCAmelCase__ : Tuple = ["""How many dogs are in this image?"""] UpperCAmelCase__ : Union[str, Any] = tokenizer(__lowerCamelCase , return_tensors="""pt""" ).input_ids UpperCAmelCase__ : Optional[Any] = hf_vqa_model.generate(__lowerCamelCase , __lowerCamelCase ) print(tokenizer.decode(answer[0] ) ) assert tokenizer.decode(answer[0] ) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + """_vqa""" ) UpperCAmelCase__ : int = """https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth""" UpperCAmelCase__ : Any = blip_itm(pretrained=__lowerCamelCase , image_size=__lowerCamelCase , vit="""base""" ) itm_model.eval() UpperCAmelCase__ : List[Any] = itm_model.state_dict() for key in modified_state_dict.copy(): UpperCAmelCase__ : Dict = modified_state_dict.pop(__lowerCamelCase ) UpperCAmelCase__ : int = rename_key(__lowerCamelCase ) UpperCAmelCase__ : Any = value UpperCAmelCase__ : Optional[int] = BlipForImageTextRetrieval(__lowerCamelCase ) UpperCAmelCase__ : Union[str, Any] = ["""A picture of a woman with a dog sitting in a beach"""] UpperCAmelCase__ : List[Any] = tokenizer( __lowerCamelCase , return_tensors="""pt""" , padding="""max_length""" , truncation=__lowerCamelCase , max_length=35 , ).input_ids hf_itm_model.load_state_dict(__lowerCamelCase ) hf_itm_model.eval() UpperCAmelCase__ : List[str] = hf_itm_model(__lowerCamelCase , __lowerCamelCase , use_itm_head=__lowerCamelCase ) UpperCAmelCase__ : List[str] = hf_itm_model(__lowerCamelCase , __lowerCamelCase , use_itm_head=__lowerCamelCase ) assert out[0].item() == 0.2_110_687_494_277_954 assert torch.nn.functional.softmax(out_itm[0] , dim=1 )[:, 1].item() == 0.45_698_845_386_505_127 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + """_itm""" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : str = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") SCREAMING_SNAKE_CASE__ : List[Any] = parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
79
0
'''simple docstring''' import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->UnCLIP class __a ( _snake_case ): __UpperCamelCase : torch.FloatTensor __UpperCamelCase : Optional[torch.FloatTensor] = None def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase=0.9_9_9 , __UpperCAmelCase="cosine" , ) -> Union[str, Any]: '''simple docstring''' if alpha_transform_type == "cosine": def alpha_bar_fn(__UpperCAmelCase ): return math.cos((t + 0.0_0_8) / 1.0_0_8 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(__UpperCAmelCase ): return math.exp(t * -1_2.0 ) else: raise ValueError(f"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) __SCREAMING_SNAKE_CASE = [] for i in range(__UpperCAmelCase ): __SCREAMING_SNAKE_CASE = i / num_diffusion_timesteps __SCREAMING_SNAKE_CASE = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__UpperCAmelCase ) / alpha_bar_fn(__UpperCAmelCase ) , __UpperCAmelCase ) ) return torch.tensor(__UpperCAmelCase , dtype=torch.floataa ) class __a ( _snake_case, _snake_case ): @register_to_config def __init__( self : Tuple ,lowerCamelCase : int = 1000 ,lowerCamelCase : str = "fixed_small_log" ,lowerCamelCase : bool = True ,lowerCamelCase : Optional[float] = 1.0 ,lowerCamelCase : str = "epsilon" ,lowerCamelCase : str = "squaredcos_cap_v2" ,): '''simple docstring''' if beta_schedule != "squaredcos_cap_v2": raise ValueError("""UnCLIPScheduler only supports `beta_schedule`: 'squaredcos_cap_v2'""" ) __SCREAMING_SNAKE_CASE = betas_for_alpha_bar(lowerCamelCase ) __SCREAMING_SNAKE_CASE = 1.0 - self.betas __SCREAMING_SNAKE_CASE = torch.cumprod(self.alphas ,dim=0 ) __SCREAMING_SNAKE_CASE = torch.tensor(1.0 ) # standard deviation of the initial noise distribution __SCREAMING_SNAKE_CASE = 1.0 # setable values __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = torch.from_numpy(np.arange(0 ,lowerCamelCase )[::-1].copy() ) __SCREAMING_SNAKE_CASE = variance_type def UpperCAmelCase__ ( self : Optional[Any] ,lowerCamelCase : torch.FloatTensor ,lowerCamelCase : Optional[int] = None ): '''simple docstring''' return sample def UpperCAmelCase__ ( self : str ,lowerCamelCase : int ,lowerCamelCase : Union[str, torch.device] = None ): '''simple docstring''' __SCREAMING_SNAKE_CASE = num_inference_steps __SCREAMING_SNAKE_CASE = (self.config.num_train_timesteps - 1) / (self.num_inference_steps - 1) __SCREAMING_SNAKE_CASE = (np.arange(0 ,lowerCamelCase ) * step_ratio).round()[::-1].copy().astype(np.intaa ) __SCREAMING_SNAKE_CASE = torch.from_numpy(lowerCamelCase ).to(lowerCamelCase ) def UpperCAmelCase__ ( self : List[str] ,lowerCamelCase : Dict ,lowerCamelCase : Optional[Any]=None ,lowerCamelCase : List[str]=None ,lowerCamelCase : Any=None ): '''simple docstring''' if prev_timestep is None: __SCREAMING_SNAKE_CASE = t - 1 __SCREAMING_SNAKE_CASE = self.alphas_cumprod[t] __SCREAMING_SNAKE_CASE = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one __SCREAMING_SNAKE_CASE = 1 - alpha_prod_t __SCREAMING_SNAKE_CASE = 1 - alpha_prod_t_prev if prev_timestep == t - 1: __SCREAMING_SNAKE_CASE = self.betas[t] else: __SCREAMING_SNAKE_CASE = 1 - alpha_prod_t / alpha_prod_t_prev # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample __SCREAMING_SNAKE_CASE = beta_prod_t_prev / beta_prod_t * beta if variance_type is None: __SCREAMING_SNAKE_CASE = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small_log": __SCREAMING_SNAKE_CASE = torch.log(torch.clamp(lowerCamelCase ,min=1E-2_0 ) ) __SCREAMING_SNAKE_CASE = torch.exp(0.5 * variance ) elif variance_type == "learned_range": # NOTE difference with DDPM scheduler __SCREAMING_SNAKE_CASE = variance.log() __SCREAMING_SNAKE_CASE = beta.log() __SCREAMING_SNAKE_CASE = (predicted_variance + 1) / 2 __SCREAMING_SNAKE_CASE = frac * max_log + (1 - frac) * min_log return variance def UpperCAmelCase__ ( self : str ,lowerCamelCase : torch.FloatTensor ,lowerCamelCase : int ,lowerCamelCase : torch.FloatTensor ,lowerCamelCase : Optional[int] = None ,lowerCamelCase : Tuple=None ,lowerCamelCase : bool = True ,): '''simple docstring''' __SCREAMING_SNAKE_CASE = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type == "learned_range": __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = torch.split(lowerCamelCase ,sample.shape[1] ,dim=1 ) else: __SCREAMING_SNAKE_CASE = None # 1. compute alphas, betas if prev_timestep is None: __SCREAMING_SNAKE_CASE = t - 1 __SCREAMING_SNAKE_CASE = self.alphas_cumprod[t] __SCREAMING_SNAKE_CASE = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one __SCREAMING_SNAKE_CASE = 1 - alpha_prod_t __SCREAMING_SNAKE_CASE = 1 - alpha_prod_t_prev if prev_timestep == t - 1: __SCREAMING_SNAKE_CASE = self.betas[t] __SCREAMING_SNAKE_CASE = self.alphas[t] else: __SCREAMING_SNAKE_CASE = 1 - alpha_prod_t / alpha_prod_t_prev __SCREAMING_SNAKE_CASE = 1 - beta # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": __SCREAMING_SNAKE_CASE = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": __SCREAMING_SNAKE_CASE = model_output else: raise ValueError( f"""prediction_type given as {self.config.prediction_type} must be one of `epsilon` or `sample`""" """ for the UnCLIPScheduler.""" ) # 3. Clip "predicted x_0" if self.config.clip_sample: __SCREAMING_SNAKE_CASE = torch.clamp( lowerCamelCase ,-self.config.clip_sample_range ,self.config.clip_sample_range ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf __SCREAMING_SNAKE_CASE = (alpha_prod_t_prev ** 0.5 * beta) / beta_prod_t __SCREAMING_SNAKE_CASE = alpha ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf __SCREAMING_SNAKE_CASE = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise __SCREAMING_SNAKE_CASE = 0 if t > 0: __SCREAMING_SNAKE_CASE = randn_tensor( model_output.shape ,dtype=model_output.dtype ,generator=lowerCamelCase ,device=model_output.device ) __SCREAMING_SNAKE_CASE = self._get_variance( lowerCamelCase ,predicted_variance=lowerCamelCase ,prev_timestep=lowerCamelCase ,) if self.variance_type == "fixed_small_log": __SCREAMING_SNAKE_CASE = variance elif self.variance_type == "learned_range": __SCREAMING_SNAKE_CASE = (0.5 * variance).exp() else: raise ValueError( f"""variance_type given as {self.variance_type} must be one of `fixed_small_log` or `learned_range`""" """ for the UnCLIPScheduler.""" ) __SCREAMING_SNAKE_CASE = variance * variance_noise __SCREAMING_SNAKE_CASE = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return UnCLIPSchedulerOutput(prev_sample=lowerCamelCase ,pred_original_sample=lowerCamelCase ) def UpperCAmelCase__ ( self : Any ,lowerCamelCase : torch.FloatTensor ,lowerCamelCase : torch.FloatTensor ,lowerCamelCase : torch.IntTensor ,): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.alphas_cumprod.to(device=original_samples.device ,dtype=original_samples.dtype ) __SCREAMING_SNAKE_CASE = timesteps.to(original_samples.device ) __SCREAMING_SNAKE_CASE = alphas_cumprod[timesteps] ** 0.5 __SCREAMING_SNAKE_CASE = sqrt_alpha_prod.flatten() while len(sqrt_alpha_prod.shape ) < len(original_samples.shape ): __SCREAMING_SNAKE_CASE = sqrt_alpha_prod.unsqueeze(-1 ) __SCREAMING_SNAKE_CASE = (1 - alphas_cumprod[timesteps]) ** 0.5 __SCREAMING_SNAKE_CASE = sqrt_one_minus_alpha_prod.flatten() while len(sqrt_one_minus_alpha_prod.shape ) < len(original_samples.shape ): __SCREAMING_SNAKE_CASE = sqrt_one_minus_alpha_prod.unsqueeze(-1 ) __SCREAMING_SNAKE_CASE = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples
109
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : Tuple = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : List[Any] = { """MIT/ast-finetuned-audioset-10-10-0.4593""": ( """https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json""" ), } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'audio-spectrogram-transformer' def __init__( self , _lowerCAmelCase=768 , _lowerCAmelCase=12 , _lowerCAmelCase=12 , _lowerCAmelCase=3072 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=1e-12 , _lowerCAmelCase=16 , _lowerCAmelCase=True , _lowerCAmelCase=10 , _lowerCAmelCase=10 , _lowerCAmelCase=1024 , _lowerCAmelCase=128 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : List[Any] = num_attention_heads UpperCAmelCase__ : Dict = intermediate_size UpperCAmelCase__ : Dict = hidden_act UpperCAmelCase__ : str = hidden_dropout_prob UpperCAmelCase__ : str = attention_probs_dropout_prob UpperCAmelCase__ : Tuple = initializer_range UpperCAmelCase__ : Dict = layer_norm_eps UpperCAmelCase__ : Optional[Any] = patch_size UpperCAmelCase__ : Tuple = qkv_bias UpperCAmelCase__ : Tuple = frequency_stride UpperCAmelCase__ : Union[str, Any] = time_stride UpperCAmelCase__ : Optional[Any] = max_length UpperCAmelCase__ : Optional[int] = num_mel_bins
79
0
"""simple docstring""" import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase__ = get_tests_dir('fixtures/test_sentencepiece.model') UpperCamelCase__ = get_tests_dir('fixtures/test_sentencepiece_bpe.model') UpperCamelCase__ = 'pt' if is_torch_available() else 'tf' @require_sentencepiece @require_tokenizers class a ( lowercase , unittest.TestCase ): UpperCamelCase : Optional[int] = CamembertTokenizer UpperCamelCase : Optional[int] = CamembertTokenizerFast UpperCamelCase : str = True UpperCamelCase : Tuple = True def __snake_case ( self ): super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase__ : int = CamembertTokenizer(UpperCamelCase_ ) tokenizer.save_pretrained(self.tmpdirname ) def __snake_case ( self ): UpperCAmelCase__ : Optional[int] = '<pad>' UpperCAmelCase__ : Tuple = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCamelCase_ ) , UpperCamelCase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCamelCase_ ) , UpperCamelCase_ ) def __snake_case ( self ): UpperCAmelCase__ : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>NOTUSED' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-1] , '<mask>' ) self.assertEqual(len(UpperCamelCase_ ) , 1_004 ) def __snake_case ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_005 ) def __snake_case ( self ): UpperCAmelCase__ : Dict = CamembertTokenizer(UpperCamelCase_ ) tokenizer.save_pretrained(self.tmpdirname ) UpperCAmelCase__ : str = CamembertTokenizerFast.from_pretrained(self.tmpdirname ) UpperCAmelCase__ : Union[str, Any] = 'I was born in 92000, and this is falsé.' UpperCAmelCase__ : Optional[Any] = tokenizer.encode(UpperCamelCase_ ) UpperCAmelCase__ : Dict = rust_tokenizer.encode(UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) UpperCAmelCase__ : str = tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) UpperCAmelCase__ : Union[str, Any] = rust_tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) UpperCAmelCase__ : List[Any] = tokenizer.convert_ids_to_tokens(UpperCamelCase_ ) UpperCAmelCase__ : str = rust_tokenizer.tokenize(UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) def __snake_case ( self ): if not self.test_rust_tokenizer: return UpperCAmelCase__ : Any = self.get_tokenizer() UpperCAmelCase__ : List[str] = self.get_rust_tokenizer() UpperCAmelCase__ : List[str] = 'I was born in 92000, and this is falsé.' UpperCAmelCase__ : List[str] = tokenizer.tokenize(UpperCamelCase_ ) UpperCAmelCase__ : str = rust_tokenizer.tokenize(UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) UpperCAmelCase__ : Optional[Any] = tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) UpperCAmelCase__ : Union[str, Any] = rust_tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) UpperCAmelCase__ : str = self.get_rust_tokenizer() UpperCAmelCase__ : Union[str, Any] = tokenizer.encode(UpperCamelCase_ ) UpperCAmelCase__ : List[str] = rust_tokenizer.encode(UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) @slow def __snake_case ( self ): # fmt: off UpperCAmelCase__ : Any = {'input_ids': [[5, 54, 7_196, 297, 30, 23, 776, 18, 11, 3_215, 3_705, 8_252, 22, 3_164, 1_181, 2_116, 29, 16, 813, 25, 791, 3_314, 20, 3_446, 38, 27_575, 120, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 468, 17, 11, 9_088, 20, 1_517, 8, 22_804, 18_818, 10, 38, 629, 607, 607, 142, 19, 7_196, 867, 56, 10_326, 24, 2_267, 20, 416, 5_072, 15_612, 233, 734, 7, 2_399, 27, 16, 3_015, 1_649, 7, 24, 20, 4_338, 2_399, 27, 13, 3_400, 14, 13, 6_189, 8, 930, 9, 6]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. UpperCAmelCase__ : Union[str, Any] = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=UpperCamelCase_ , model_name='camembert-base' , revision='3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf' , sequences=UpperCamelCase_ , )
110
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) class UpperCAmelCase_ ( __lowerCamelCase ): def __init__( self , *_lowerCAmelCase , **_lowerCAmelCase ): warnings.warn( """The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use GLPNImageProcessor instead.""" , _lowerCAmelCase , ) super().__init__(*_lowerCAmelCase , **_lowerCAmelCase )
79
0
"""simple docstring""" from collections import deque from math import floor from random import random from time import time class SCREAMING_SNAKE_CASE__ : def __init__( self : Tuple ): lowerCAmelCase = {} def __lowercase ( self : Optional[Any] , lowerCAmelCase : List[Any] , lowerCAmelCase : List[str] , lowerCAmelCase : Union[str, Any]=1 ): if self.graph.get(_lowerCAmelCase ): if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: lowerCAmelCase = [[w, v]] if not self.graph.get(_lowerCAmelCase ): lowerCAmelCase = [] def __lowercase ( self : Optional[Any] ): return list(self.graph ) def __lowercase ( self : Dict , lowerCAmelCase : Optional[Any] , lowerCAmelCase : Optional[int] ): if self.graph.get(_lowerCAmelCase ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(_lowerCAmelCase ) def __lowercase ( self : Optional[int] , lowerCAmelCase : List[str]=-2 , lowerCAmelCase : List[str]=-1 ): if s == d: return [] lowerCAmelCase = [] lowerCAmelCase = [] if s == -2: lowerCAmelCase = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) lowerCAmelCase = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowerCAmelCase = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(_lowerCAmelCase ) return visited else: stack.append(node[1] ) visited.append(node[1] ) lowerCAmelCase = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(_lowerCAmelCase ) != 0: lowerCAmelCase = stack[len(_lowerCAmelCase ) - 1] else: lowerCAmelCase = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return visited def __lowercase ( self : int , lowerCAmelCase : Optional[int]=-1 ): if c == -1: lowerCAmelCase = floor(random() * 1_0000 ) + 10 for i in range(_lowerCAmelCase ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): lowerCAmelCase = floor(random() * c ) + 1 if n != i: self.add_pair(_lowerCAmelCase , _lowerCAmelCase , 1 ) def __lowercase ( self : Dict , lowerCAmelCase : str=-2 ): lowerCAmelCase = deque() lowerCAmelCase = [] if s == -2: lowerCAmelCase = list(self.graph )[0] d.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) while d: lowerCAmelCase = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def __lowercase ( self : str , lowerCAmelCase : Optional[Any] ): lowerCAmelCase = 0 for x in self.graph: for y in self.graph[x]: if y[1] == u: count += 1 return count def __lowercase ( self : Optional[int] , lowerCAmelCase : List[str] ): return len(self.graph[u] ) def __lowercase ( self : Union[str, Any] , lowerCAmelCase : str=-2 ): lowerCAmelCase = [] lowerCAmelCase = [] if s == -2: lowerCAmelCase = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) lowerCAmelCase = s lowerCAmelCase = [] while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowerCAmelCase = s for node in self.graph[s]: if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) lowerCAmelCase = node[1] break # check if all the children are visited if s == ss: sorted_nodes.append(stack.pop() ) if len(_lowerCAmelCase ) != 0: lowerCAmelCase = stack[len(_lowerCAmelCase ) - 1] else: lowerCAmelCase = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return sorted_nodes def __lowercase ( self : Union[str, Any] ): lowerCAmelCase = [] lowerCAmelCase = [] lowerCAmelCase = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) lowerCAmelCase = -2 lowerCAmelCase = [] lowerCAmelCase = s lowerCAmelCase = False lowerCAmelCase = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowerCAmelCase = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): lowerCAmelCase = len(_lowerCAmelCase ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) lowerCAmelCase = node[1] break # check if all the children are visited if s == ss: stack.pop() lowerCAmelCase = True if len(_lowerCAmelCase ) != 0: lowerCAmelCase = stack[len(_lowerCAmelCase ) - 1] else: lowerCAmelCase = False indirect_parents.append(_lowerCAmelCase ) lowerCAmelCase = s lowerCAmelCase = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return list(_lowerCAmelCase ) def __lowercase ( self : Tuple ): lowerCAmelCase = [] lowerCAmelCase = [] lowerCAmelCase = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) lowerCAmelCase = -2 lowerCAmelCase = [] lowerCAmelCase = s lowerCAmelCase = False lowerCAmelCase = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowerCAmelCase = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): lowerCAmelCase = len(_lowerCAmelCase ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) lowerCAmelCase = node[1] break # check if all the children are visited if s == ss: stack.pop() lowerCAmelCase = True if len(_lowerCAmelCase ) != 0: lowerCAmelCase = stack[len(_lowerCAmelCase ) - 1] else: lowerCAmelCase = False indirect_parents.append(_lowerCAmelCase ) lowerCAmelCase = s lowerCAmelCase = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return False def __lowercase ( self : Optional[int] , lowerCAmelCase : Tuple=-2 , lowerCAmelCase : Optional[Any]=-1 ): lowerCAmelCase = time() self.dfs(_lowerCAmelCase , _lowerCAmelCase ) lowerCAmelCase = time() return end - begin def __lowercase ( self : Union[str, Any] , lowerCAmelCase : Any=-2 ): lowerCAmelCase = time() self.bfs(_lowerCAmelCase ) lowerCAmelCase = time() return end - begin class SCREAMING_SNAKE_CASE__ : def __init__( self : Optional[int] ): lowerCAmelCase = {} def __lowercase ( self : List[Any] , lowerCAmelCase : Dict , lowerCAmelCase : Dict , lowerCAmelCase : str=1 ): # check if the u exists if self.graph.get(_lowerCAmelCase ): # if there already is a edge if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: # if u does not exist lowerCAmelCase = [[w, v]] # add the other way if self.graph.get(_lowerCAmelCase ): # if there already is a edge if self.graph[v].count([w, u] ) == 0: self.graph[v].append([w, u] ) else: # if u does not exist lowerCAmelCase = [[w, u]] def __lowercase ( self : Union[str, Any] , lowerCAmelCase : str , lowerCAmelCase : List[Any] ): if self.graph.get(_lowerCAmelCase ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(_lowerCAmelCase ) # the other way round if self.graph.get(_lowerCAmelCase ): for _ in self.graph[v]: if _[1] == u: self.graph[v].remove(_lowerCAmelCase ) def __lowercase ( self : int , lowerCAmelCase : Tuple=-2 , lowerCAmelCase : int=-1 ): if s == d: return [] lowerCAmelCase = [] lowerCAmelCase = [] if s == -2: lowerCAmelCase = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) lowerCAmelCase = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowerCAmelCase = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(_lowerCAmelCase ) return visited else: stack.append(node[1] ) visited.append(node[1] ) lowerCAmelCase = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(_lowerCAmelCase ) != 0: lowerCAmelCase = stack[len(_lowerCAmelCase ) - 1] else: lowerCAmelCase = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return visited def __lowercase ( self : int , lowerCAmelCase : List[Any]=-1 ): if c == -1: lowerCAmelCase = floor(random() * 1_0000 ) + 10 for i in range(_lowerCAmelCase ): # every vertex has max 100 edges for _ in range(floor(random() * 102 ) + 1 ): lowerCAmelCase = floor(random() * c ) + 1 if n != i: self.add_pair(_lowerCAmelCase , _lowerCAmelCase , 1 ) def __lowercase ( self : Tuple , lowerCAmelCase : List[str]=-2 ): lowerCAmelCase = deque() lowerCAmelCase = [] if s == -2: lowerCAmelCase = list(self.graph )[0] d.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) while d: lowerCAmelCase = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def __lowercase ( self : Optional[int] , lowerCAmelCase : List[str] ): return len(self.graph[u] ) def __lowercase ( self : str ): lowerCAmelCase = [] lowerCAmelCase = [] lowerCAmelCase = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) lowerCAmelCase = -2 lowerCAmelCase = [] lowerCAmelCase = s lowerCAmelCase = False lowerCAmelCase = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowerCAmelCase = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): lowerCAmelCase = len(_lowerCAmelCase ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) lowerCAmelCase = node[1] break # check if all the children are visited if s == ss: stack.pop() lowerCAmelCase = True if len(_lowerCAmelCase ) != 0: lowerCAmelCase = stack[len(_lowerCAmelCase ) - 1] else: lowerCAmelCase = False indirect_parents.append(_lowerCAmelCase ) lowerCAmelCase = s lowerCAmelCase = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return list(_lowerCAmelCase ) def __lowercase ( self : Optional[Any] ): lowerCAmelCase = [] lowerCAmelCase = [] lowerCAmelCase = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) lowerCAmelCase = -2 lowerCAmelCase = [] lowerCAmelCase = s lowerCAmelCase = False lowerCAmelCase = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: lowerCAmelCase = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): lowerCAmelCase = len(_lowerCAmelCase ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) lowerCAmelCase = node[1] break # check if all the children are visited if s == ss: stack.pop() lowerCAmelCase = True if len(_lowerCAmelCase ) != 0: lowerCAmelCase = stack[len(_lowerCAmelCase ) - 1] else: lowerCAmelCase = False indirect_parents.append(_lowerCAmelCase ) lowerCAmelCase = s lowerCAmelCase = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return False def __lowercase ( self : Union[str, Any] ): return list(self.graph ) def __lowercase ( self : Union[str, Any] , lowerCAmelCase : List[str]=-2 , lowerCAmelCase : Tuple=-1 ): lowerCAmelCase = time() self.dfs(_lowerCAmelCase , _lowerCAmelCase ) lowerCAmelCase = time() return end - begin def __lowercase ( self : Union[str, Any] , lowerCAmelCase : Optional[int]=-2 ): lowerCAmelCase = time() self.bfs(_lowerCAmelCase ) lowerCAmelCase = time() return end - begin
169
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : int = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__ : List[str] = { """vocab_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt""", }, """tokenizer_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json""" ), """google/realm-orqa-nq-openqa""": ( """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-nq-reader""": ( """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-openqa""": ( """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-reader""": ( """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json""" ), }, } SCREAMING_SNAKE_CASE__ : Optional[Any] = { """google/realm-cc-news-pretrained-embedder""": 5_12, """google/realm-cc-news-pretrained-encoder""": 5_12, """google/realm-cc-news-pretrained-scorer""": 5_12, """google/realm-cc-news-pretrained-openqa""": 5_12, """google/realm-orqa-nq-openqa""": 5_12, """google/realm-orqa-nq-reader""": 5_12, """google/realm-orqa-wq-openqa""": 5_12, """google/realm-orqa-wq-reader""": 5_12, } SCREAMING_SNAKE_CASE__ : Optional[Any] = { """google/realm-cc-news-pretrained-embedder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-encoder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-scorer""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-reader""": {"""do_lower_case""": True}, """google/realm-orqa-wq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-wq-reader""": {"""do_lower_case""": True}, } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = PRETRAINED_INIT_CONFIGURATION __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = RealmTokenizer def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=True , _lowerCAmelCase="[UNK]" , _lowerCAmelCase="[SEP]" , _lowerCAmelCase="[PAD]" , _lowerCAmelCase="[CLS]" , _lowerCAmelCase="[MASK]" , _lowerCAmelCase=True , _lowerCAmelCase=None , **_lowerCAmelCase , ): super().__init__( _lowerCAmelCase , tokenizer_file=_lowerCAmelCase , do_lower_case=_lowerCAmelCase , unk_token=_lowerCAmelCase , sep_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , cls_token=_lowerCAmelCase , mask_token=_lowerCAmelCase , tokenize_chinese_chars=_lowerCAmelCase , strip_accents=_lowerCAmelCase , **_lowerCAmelCase , ) UpperCAmelCase__ : Tuple = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , _lowerCAmelCase ) != do_lower_case or normalizer_state.get("""strip_accents""" , _lowerCAmelCase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , _lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase__ : Any = getattr(_lowerCAmelCase , normalizer_state.pop("""type""" ) ) UpperCAmelCase__ : str = do_lower_case UpperCAmelCase__ : Tuple = strip_accents UpperCAmelCase__ : Tuple = tokenize_chinese_chars UpperCAmelCase__ : Union[str, Any] = normalizer_class(**_lowerCAmelCase ) UpperCAmelCase__ : Dict = do_lower_case def __UpperCAmelCase ( self , _lowerCAmelCase , **_lowerCAmelCase ): UpperCAmelCase__ : List[Any] = PaddingStrategy.MAX_LENGTH UpperCAmelCase__ : Optional[int] = text UpperCAmelCase__ : Optional[int] = kwargs.pop("""text_pair""" , _lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = kwargs.pop("""return_tensors""" , _lowerCAmelCase ) UpperCAmelCase__ : Optional[Any] = { """input_ids""": [], """attention_mask""": [], """token_type_ids""": [], } for idx, candidate_text in enumerate(_lowerCAmelCase ): if batch_text_pair is not None: UpperCAmelCase__ : str = batch_text_pair[idx] else: UpperCAmelCase__ : Any = None UpperCAmelCase__ : str = super().__call__(_lowerCAmelCase , _lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = encoded_candidates.get("""input_ids""" ) UpperCAmelCase__ : str = encoded_candidates.get("""attention_mask""" ) UpperCAmelCase__ : Union[str, Any] = encoded_candidates.get("""token_type_ids""" ) if encoded_input_ids is not None: output_data["input_ids"].append(_lowerCAmelCase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(_lowerCAmelCase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = {key: item for key, item in output_data.items() if len(_lowerCAmelCase ) != 0} return BatchEncoding(_lowerCAmelCase , tensor_type=_lowerCAmelCase ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase = None ): UpperCAmelCase__ : Any = [self.sep_token_id] UpperCAmelCase__ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase = None ): UpperCAmelCase__ : List[str] = self._tokenizer.model.save(_lowerCAmelCase , name=_lowerCAmelCase ) return tuple(_lowerCAmelCase )
79
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class lowercase : _SCREAMING_SNAKE_CASE = BlenderbotConfig _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = 'gelu' def __init__( self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=False , lowercase=99 , lowercase=32 , lowercase=2 , lowercase=4 , lowercase=37 , lowercase=0.1 , lowercase=0.1 , lowercase=20 , lowercase=2 , lowercase=1 , lowercase=0 , ) -> Union[str, Any]: lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = seq_length lowerCAmelCase = is_training lowerCAmelCase = use_labels lowerCAmelCase = vocab_size lowerCAmelCase = hidden_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = max_position_embeddings lowerCAmelCase = eos_token_id lowerCAmelCase = pad_token_id lowerCAmelCase = bos_token_id def _snake_case ( self ) -> Optional[int]: lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) lowerCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) lowerCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) lowerCAmelCase = prepare_blenderbot_inputs_dict(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return config, inputs_dict def _snake_case ( self , lowercase , lowercase ) -> int: lowerCAmelCase = TFBlenderbotModel(config=_lowerCAmelCase ).get_decoder() lowerCAmelCase = inputs_dict["""input_ids"""] lowerCAmelCase = input_ids[:1, :] lowerCAmelCase = inputs_dict["""attention_mask"""][:1, :] lowerCAmelCase = inputs_dict["""head_mask"""] lowerCAmelCase = 1 # first forward pass lowerCAmelCase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , head_mask=_lowerCAmelCase , use_cache=_lowerCAmelCase ) lowerCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowerCAmelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and lowerCAmelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) lowerCAmelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) lowerCAmelCase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )[0] lowerCAmelCase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , past_key_values=_lowerCAmelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice lowerCAmelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) lowerCAmelCase = output_from_no_past[:, -3:, random_slice_idx] lowerCAmelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_lowerCAmelCase , _lowerCAmelCase , rtol=1e-3 ) def UpperCAmelCase__ ( SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : Optional[Any] , SCREAMING_SNAKE_CASE : Optional[Any]=None , SCREAMING_SNAKE_CASE : List[str]=None , SCREAMING_SNAKE_CASE : str=None , SCREAMING_SNAKE_CASE : List[str]=None , SCREAMING_SNAKE_CASE : int=None , ): '''simple docstring''' if attention_mask is None: lowerCAmelCase = tf.cast(tf.math.not_equal(__lowerCamelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: lowerCAmelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: lowerCAmelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: lowerCAmelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: lowerCAmelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class lowercase ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): _SCREAMING_SNAKE_CASE = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () _SCREAMING_SNAKE_CASE = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () _SCREAMING_SNAKE_CASE = ( { 'conversational': TFBlenderbotForConditionalGeneration, 'feature-extraction': TFBlenderbotModel, 'summarization': TFBlenderbotForConditionalGeneration, 'text2text-generation': TFBlenderbotForConditionalGeneration, 'translation': TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) _SCREAMING_SNAKE_CASE = True _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = False def _snake_case ( self ) -> List[Any]: lowerCAmelCase = TFBlenderbotModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=_lowerCAmelCase ) def _snake_case ( self ) -> Optional[Any]: self.config_tester.run_common_tests() def _snake_case ( self ) -> List[Any]: lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_lowerCAmelCase ) @require_tokenizers @require_tf class lowercase ( unittest.TestCase ): _SCREAMING_SNAKE_CASE = ['My friends are cool but they eat too many carbs.'] _SCREAMING_SNAKE_CASE = 'facebook/blenderbot-400M-distill' @cached_property def _snake_case ( self ) -> int: return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def _snake_case ( self ) -> Tuple: lowerCAmelCase = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def _snake_case ( self ) -> Optional[int]: lowerCAmelCase = self.tokenizer(self.src_text , return_tensors="""tf""" ) lowerCAmelCase = self.model.generate( model_inputs.input_ids , ) lowerCAmelCase = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=_lowerCAmelCase )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
532
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'facebook/bart-large-mnli' __lowerCamelCase = ( 'This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which ' 'should be the text to classify, and `labels`, which should be the list of labels to use for classification. ' 'It returns the most likely label in the list of provided `labels` for the input text.' ) __lowerCamelCase = 'text_classifier' __lowerCamelCase = AutoTokenizer __lowerCamelCase = AutoModelForSequenceClassification __lowerCamelCase = ['text', ['text']] __lowerCamelCase = ['text'] def __UpperCAmelCase ( self ): super().setup() UpperCAmelCase__ : Optional[Any] = self.model.config UpperCAmelCase__ : Tuple = -1 for idx, label in config.idalabel.items(): if label.lower().startswith("""entail""" ): UpperCAmelCase__ : Dict = int(_lowerCAmelCase ) if self.entailment_id == -1: raise ValueError("""Could not determine the entailment ID from the model config, please pass it at init.""" ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : List[Any] = labels return self.pre_processor( [text] * len(_lowerCAmelCase ) , [f"This example is {label}" for label in labels] , return_tensors="""pt""" , padding="""max_length""" , ) def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : str = outputs.logits UpperCAmelCase__ : List[Any] = torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
79
0
"""simple docstring""" import math _UpperCAmelCase = 1_0 _UpperCAmelCase = 7 _UpperCAmelCase = BALLS_PER_COLOUR * NUM_COLOURS def __magic_name__ ( lowercase = 20 ): SCREAMING_SNAKE_CASE_: Dict =math.comb(__lowerCamelCase , __lowerCamelCase ) SCREAMING_SNAKE_CASE_: List[Any] =math.comb(NUM_BALLS - BALLS_PER_COLOUR , __lowerCamelCase ) SCREAMING_SNAKE_CASE_: Tuple =NUM_COLOURS * (1 - missing_colour / total) return f'''{result:.9f}''' if __name__ == "__main__": print(solution(2_0))
409
from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=13 , _lowerCAmelCase=30 , _lowerCAmelCase=2 , _lowerCAmelCase=3 , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=32 , _lowerCAmelCase=2 , _lowerCAmelCase=4 , _lowerCAmelCase=37 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=10 , _lowerCAmelCase=0.0_2 , _lowerCAmelCase=3 , _lowerCAmelCase=None , ): UpperCAmelCase__ : Tuple = parent UpperCAmelCase__ : Optional[int] = batch_size UpperCAmelCase__ : Union[str, Any] = image_size UpperCAmelCase__ : int = patch_size UpperCAmelCase__ : str = num_channels UpperCAmelCase__ : int = is_training UpperCAmelCase__ : List[str] = use_labels UpperCAmelCase__ : List[Any] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : Tuple = num_attention_heads UpperCAmelCase__ : Optional[int] = intermediate_size UpperCAmelCase__ : Optional[Any] = hidden_act UpperCAmelCase__ : int = hidden_dropout_prob UpperCAmelCase__ : int = attention_probs_dropout_prob UpperCAmelCase__ : List[str] = type_sequence_label_size UpperCAmelCase__ : Optional[int] = initializer_range UpperCAmelCase__ : Any = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) UpperCAmelCase__ : Any = (image_size // patch_size) ** 2 UpperCAmelCase__ : Tuple = num_patches + 1 def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ : List[str] = None if self.use_labels: UpperCAmelCase__ : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ : Union[str, Any] = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self ): return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : str = TFViTModel(config=_lowerCAmelCase ) UpperCAmelCase__ : str = model(_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase__ : Optional[Any] = self.image_size // 2 UpperCAmelCase__ : List[str] = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase__ : List[Any] = model(_lowerCAmelCase , interpolate_pos_encoding=_lowerCAmelCase , training=_lowerCAmelCase ) UpperCAmelCase__ : str = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def __UpperCAmelCase ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase__ : Tuple = self.type_sequence_label_size UpperCAmelCase__ : List[Any] = TFViTForImageClassification(_lowerCAmelCase ) UpperCAmelCase__ : List[str] = model(_lowerCAmelCase , labels=_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. UpperCAmelCase__ : Tuple = self.image_size // 2 UpperCAmelCase__ : Union[str, Any] = pixel_values[:, :, :image_size, :image_size] UpperCAmelCase__ : List[str] = model(_lowerCAmelCase , interpolate_pos_encoding=_lowerCAmelCase , training=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase__ : Union[str, Any] = 1 UpperCAmelCase__ : Optional[Any] = TFViTForImageClassification(_lowerCAmelCase ) UpperCAmelCase__ : List[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase__ : List[str] = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : int = config_and_inputs UpperCAmelCase__ : int = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class UpperCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): __lowerCamelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () __lowerCamelCase = ( {'feature-extraction': TFViTModel, 'image-classification': TFViTForImageClassification} if is_tf_available() else {} ) __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = TFViTModelTester(self ) UpperCAmelCase__ : int = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def __UpperCAmelCase ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def __UpperCAmelCase ( self ): pass @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def __UpperCAmelCase ( self ): pass def __UpperCAmelCase ( self ): UpperCAmelCase__ , UpperCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : str = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , tf.keras.layers.Layer ) ) def __UpperCAmelCase ( self ): UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : Optional[int] = model_class(_lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ : Tuple = [*signature.parameters.keys()] UpperCAmelCase__ : str = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = TFViTModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(_lowerCAmelCase ) def _lowerCamelCase ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ : List[str] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class UpperCAmelCase_ ( unittest.TestCase ): @cached_property def __UpperCAmelCase ( self ): return ViTImageProcessor.from_pretrained("""google/vit-base-patch16-224""" ) if is_vision_available() else None @slow def __UpperCAmelCase ( self ): UpperCAmelCase__ : Optional[int] = TFViTForImageClassification.from_pretrained("""google/vit-base-patch16-224""" ) UpperCAmelCase__ : List[Any] = self.default_image_processor UpperCAmelCase__ : Union[str, Any] = prepare_img() UpperCAmelCase__ : Optional[Any] = image_processor(images=_lowerCAmelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase__ : int = model(**_lowerCAmelCase ) # verify the logits UpperCAmelCase__ : Tuple = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) UpperCAmelCase__ : int = tf.constant([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ) tf.debugging.assert_near(outputs.logits[0, :3] , _lowerCAmelCase , atol=1e-4 )
79
0
import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse("0.8.3"): raise Exception("requires gluonnlp == 0.8.3") if version.parse(mx.__version__) != version.parse("1.5.0"): raise Exception("requires mxnet == 1.5.0") logging.set_verbosity_info() lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = """The Nymphenburg Palace is a beautiful palace in Munich!""" def A_ ( lowercase_ , lowercase_ ) -> Optional[Any]: _snake_case : Tuple = { """attention_cell""": """multi_head""", """num_layers""": 4, """units""": 1024, """hidden_size""": 768, """max_length""": 512, """num_heads""": 8, """scaled""": True, """dropout""": 0.1, """use_residual""": True, """embed_size""": 1024, """embed_dropout""": 0.1, """word_embed""": None, """layer_norm_eps""": 1E-5, """token_type_vocab_size""": 2, } _snake_case : Tuple = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py _snake_case : str = BERTEncoder( attention_cell=predefined_args['''attention_cell'''] , num_layers=predefined_args['''num_layers'''] , units=predefined_args['''units'''] , hidden_size=predefined_args['''hidden_size'''] , max_length=predefined_args['''max_length'''] , num_heads=predefined_args['''num_heads'''] , scaled=predefined_args['''scaled'''] , dropout=predefined_args['''dropout'''] , output_attention=__lowerCamelCase , output_all_encodings=__lowerCamelCase , use_residual=predefined_args['''use_residual'''] , activation=predefined_args.get('''activation''' , '''gelu''' ) , layer_norm_eps=predefined_args.get('''layer_norm_eps''' , __lowerCamelCase ) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later _snake_case : List[Any] = """openwebtext_ccnews_stories_books_cased""" # Specify download folder to Gluonnlp's vocab _snake_case : List[str] = os.path.join(get_home_dir() , '''models''' ) _snake_case : Tuple = _load_vocab(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , cls=__lowerCamelCase ) _snake_case : Union[str, Any] = nlp.model.BERTModel( __lowerCamelCase , len(__lowerCamelCase ) , units=predefined_args['''units'''] , embed_size=predefined_args['''embed_size'''] , embed_dropout=predefined_args['''embed_dropout'''] , word_embed=predefined_args['''word_embed'''] , use_pooler=__lowerCamelCase , use_token_type_embed=__lowerCamelCase , token_type_vocab_size=predefined_args['''token_type_vocab_size'''] , use_classifier=__lowerCamelCase , use_decoder=__lowerCamelCase , ) original_bort.load_parameters(__lowerCamelCase , cast_dtype=__lowerCamelCase , ignore_extra=__lowerCamelCase ) _snake_case : List[Any] = original_bort._collect_params_with_prefix() # Build our config 🤗 _snake_case : Optional[Any] = { """architectures""": ["""BertForMaskedLM"""], """attention_probs_dropout_prob""": predefined_args["""dropout"""], """hidden_act""": """gelu""", """hidden_dropout_prob""": predefined_args["""dropout"""], """hidden_size""": predefined_args["""embed_size"""], """initializer_range""": 0.02, """intermediate_size""": predefined_args["""hidden_size"""], """layer_norm_eps""": predefined_args["""layer_norm_eps"""], """max_position_embeddings""": predefined_args["""max_length"""], """model_type""": """bort""", """num_attention_heads""": predefined_args["""num_heads"""], """num_hidden_layers""": predefined_args["""num_layers"""], """pad_token_id""": 1, # 2 = BERT, 1 = RoBERTa """type_vocab_size""": 1, # 2 = BERT, 1 = RoBERTa """vocab_size""": len(__lowerCamelCase ), } _snake_case : str = BertConfig.from_dict(__lowerCamelCase ) _snake_case : Union[str, Any] = BertForMaskedLM(__lowerCamelCase ) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(lowercase_ ) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy() ) ) # Check param shapes and map new HF param back def check_and_map_params(lowercase_ , lowercase_ ): _snake_case : List[str] = hf_param.shape _snake_case : Any = to_torch(params[gluon_param] ) _snake_case : Union[str, Any] = gluon_param.shape assert ( shape_hf == shape_gluon ), f'''The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers''' return gluon_param _snake_case : Dict = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , '''word_embed.0.weight''' ) _snake_case : Union[str, Any] = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , '''encoder.position_weight''' ) _snake_case : Dict = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , '''encoder.layer_norm.beta''' ) _snake_case : List[Any] = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , '''encoder.layer_norm.gamma''' ) # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) _snake_case : Tuple = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data ) for i in range(hf_bort_config.num_hidden_layers ): _snake_case : BertLayer = hf_bort_model.bert.encoder.layer[i] # self attention _snake_case : BertSelfAttention = layer.attention.self _snake_case : Union[str, Any] = check_and_map_params( self_attn.key.bias.data , f'''encoder.transformer_cells.{i}.attention_cell.proj_key.bias''' ) _snake_case : int = check_and_map_params( self_attn.key.weight.data , f'''encoder.transformer_cells.{i}.attention_cell.proj_key.weight''' ) _snake_case : Optional[int] = check_and_map_params( self_attn.query.bias.data , f'''encoder.transformer_cells.{i}.attention_cell.proj_query.bias''' ) _snake_case : Dict = check_and_map_params( self_attn.query.weight.data , f'''encoder.transformer_cells.{i}.attention_cell.proj_query.weight''' ) _snake_case : Tuple = check_and_map_params( self_attn.value.bias.data , f'''encoder.transformer_cells.{i}.attention_cell.proj_value.bias''' ) _snake_case : str = check_and_map_params( self_attn.value.weight.data , f'''encoder.transformer_cells.{i}.attention_cell.proj_value.weight''' ) # self attention output _snake_case : BertSelfOutput = layer.attention.output _snake_case : Optional[Any] = check_and_map_params( self_output.dense.bias , f'''encoder.transformer_cells.{i}.proj.bias''' ) _snake_case : int = check_and_map_params( self_output.dense.weight , f'''encoder.transformer_cells.{i}.proj.weight''' ) _snake_case : int = check_and_map_params( self_output.LayerNorm.bias , f'''encoder.transformer_cells.{i}.layer_norm.beta''' ) _snake_case : List[Any] = check_and_map_params( self_output.LayerNorm.weight , f'''encoder.transformer_cells.{i}.layer_norm.gamma''' ) # intermediate _snake_case : BertIntermediate = layer.intermediate _snake_case : Optional[int] = check_and_map_params( intermediate.dense.bias , f'''encoder.transformer_cells.{i}.ffn.ffn_1.bias''' ) _snake_case : List[Any] = check_and_map_params( intermediate.dense.weight , f'''encoder.transformer_cells.{i}.ffn.ffn_1.weight''' ) # output _snake_case : BertOutput = layer.output _snake_case : str = check_and_map_params( bert_output.dense.bias , f'''encoder.transformer_cells.{i}.ffn.ffn_2.bias''' ) _snake_case : List[str] = check_and_map_params( bert_output.dense.weight , f'''encoder.transformer_cells.{i}.ffn.ffn_2.weight''' ) _snake_case : Optional[Any] = check_and_map_params( bert_output.LayerNorm.bias , f'''encoder.transformer_cells.{i}.ffn.layer_norm.beta''' ) _snake_case : Dict = check_and_map_params( bert_output.LayerNorm.weight , f'''encoder.transformer_cells.{i}.ffn.layer_norm.gamma''' ) # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models _snake_case : Optional[int] = RobertaTokenizer.from_pretrained('''roberta-base''' ) _snake_case : Any = tokenizer.encode_plus(__lowerCamelCase )["""input_ids"""] # Get gluon output _snake_case : Dict = mx.nd.array([input_ids] ) _snake_case : Union[str, Any] = original_bort(inputs=__lowerCamelCase , token_types=[] ) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(__lowerCamelCase ) _snake_case : int = BertModel.from_pretrained(__lowerCamelCase ) hf_bort_model.eval() _snake_case : int = tokenizer.encode_plus(__lowerCamelCase , return_tensors='''pt''' ) _snake_case : Dict = hf_bort_model(**__lowerCamelCase )[0] _snake_case : Union[str, Any] = output_gluon[0].asnumpy() _snake_case : Tuple = output_hf[0].detach().numpy() _snake_case : str = np.max(np.abs(hf_layer - gluon_layer ) ).item() _snake_case : Dict = np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1E-3 ) if success: print('''✔️ Both model do output the same tensors''' ) else: print('''❌ Both model do **NOT** output the same tensors''' ) print('''Absolute difference is:''' , __lowerCamelCase ) if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--bort_checkpoint_path", default=None, type=str, required=True, help="Path the official Bort params file." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) lowerCAmelCase_ = parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
326
from functools import lru_cache @lru_cache def _lowerCamelCase ( __lowerCamelCase ) -> int: '''simple docstring''' if num < 0: raise ValueError("""Number should not be negative.""" ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
79
0
import argparse import json from collections import OrderedDict from functools import partial from pathlib import Path import timm import torch from huggingface_hub import hf_hub_download from transformers import LevitConfig, LevitForImageClassificationWithTeacher, LevitImageProcessor from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE :List[str] = logging.get_logger() def _lowerCAmelCase ( lowerCAmelCase_ :int , lowerCAmelCase_ :str , lowerCAmelCase_ :Union[str, Any] , lowerCAmelCase_ :Optional[Any] , lowerCAmelCase_ :Dict = True )->Any: '''simple docstring''' print(F'''Converting {name}...''' ) with torch.no_grad(): if hidden_sizes == 128: if name[-1] == "S": snake_case_ = timm.create_model("levit_128s" , pretrained=__lowerCamelCase ) else: snake_case_ = timm.create_model("levit_128" , pretrained=__lowerCamelCase ) if hidden_sizes == 192: snake_case_ = timm.create_model("levit_192" , pretrained=__lowerCamelCase ) if hidden_sizes == 256: snake_case_ = timm.create_model("levit_256" , pretrained=__lowerCamelCase ) if hidden_sizes == 384: snake_case_ = timm.create_model("levit_384" , pretrained=__lowerCamelCase ) from_model.eval() snake_case_ = LevitForImageClassificationWithTeacher(__lowerCamelCase ).eval() snake_case_ = OrderedDict() snake_case_ = from_model.state_dict() snake_case_ = list(from_model.state_dict().keys() ) snake_case_ = list(our_model.state_dict().keys() ) print(len(__lowerCamelCase ) , len(__lowerCamelCase ) ) for i in range(len(__lowerCamelCase ) ): snake_case_ = weights[og_keys[i]] our_model.load_state_dict(__lowerCamelCase ) snake_case_ = torch.randn((2, 3, 224, 224) ) snake_case_ = from_model(__lowerCamelCase ) snake_case_ = our_model(__lowerCamelCase ).logits assert torch.allclose(__lowerCamelCase , __lowerCamelCase ), "The model logits don't match the original one." snake_case_ = name print(__lowerCamelCase ) if push_to_hub: our_model.save_pretrained(save_directory / checkpoint_name ) snake_case_ = LevitImageProcessor() image_processor.save_pretrained(save_directory / checkpoint_name ) print(F'''Pushed {checkpoint_name}''' ) def _lowerCAmelCase ( lowerCAmelCase_ :List[Any] , lowerCAmelCase_ :Dict = None , lowerCAmelCase_ :List[Any] = True )->List[str]: '''simple docstring''' snake_case_ = """imagenet-1k-id2label.json""" snake_case_ = 1_000 snake_case_ = (1, num_labels) snake_case_ = """huggingface/label-files""" snake_case_ = num_labels snake_case_ = json.load(open(hf_hub_download(__lowerCamelCase , __lowerCamelCase , repo_type="dataset" ) , "r" ) ) snake_case_ = {int(__lowerCamelCase ): v for k, v in idalabel.items()} snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} snake_case_ = partial(__lowerCamelCase , num_labels=__lowerCamelCase , idalabel=__lowerCamelCase , labelaid=__lowerCamelCase ) snake_case_ = { """levit-128S""": 128, """levit-128""": 128, """levit-192""": 192, """levit-256""": 256, """levit-384""": 384, } snake_case_ = { """levit-128S""": ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 6, 8] , depths=[2, 3, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), """levit-128""": ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 8, 12] , depths=[4, 4, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), """levit-192""": ImageNetPreTrainedConfig( hidden_sizes=[192, 288, 384] , num_attention_heads=[3, 5, 6] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), """levit-256""": ImageNetPreTrainedConfig( hidden_sizes=[256, 384, 512] , num_attention_heads=[4, 6, 8] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), """levit-384""": ImageNetPreTrainedConfig( hidden_sizes=[384, 512, 768] , num_attention_heads=[6, 9, 12] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0.1 , ), } if model_name: convert_weight_and_push( names_to_hidden_sizes[model_name] , __lowerCamelCase , names_to_config[model_name] , __lowerCamelCase , __lowerCamelCase ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(names_to_hidden_sizes[model_name] , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) return config, expected_shape if __name__ == "__main__": SCREAMING_SNAKE_CASE :Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default=None, type=str, help='''The name of the model you wish to convert, it must be one of the supported Levit* architecture,''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''levit-dump-folder/''', type=Path, required=False, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--push_to_hub''', action='''store_true''', help='''Push model and image processor to the hub''') parser.add_argument( '''--no-push_to_hub''', dest='''push_to_hub''', action='''store_false''', help='''Do not push model and image processor to the hub''', ) SCREAMING_SNAKE_CASE :str = parser.parse_args() SCREAMING_SNAKE_CASE :Path = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
283
import argparse import hashlib # hashlib is only used inside the Test class import struct class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase ): UpperCAmelCase__ : Any = data UpperCAmelCase__ : List[Any] = [0X6745_2301, 0Xefcd_ab89, 0X98ba_dcfe, 0X1032_5476, 0Xc3d2_e1f0] @staticmethod def __UpperCAmelCase ( _lowerCAmelCase , _lowerCAmelCase ): return ((n << b) | (n >> (32 - b))) & 0Xffff_ffff def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[Any] = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) UpperCAmelCase__ : Optional[int] = self.data + padding + struct.pack(""">Q""" , 8 * len(self.data ) ) return padded_data def __UpperCAmelCase ( self ): return [ self.padded_data[i : i + 64] for i in range(0 , len(self.padded_data ) , 64 ) ] def __UpperCAmelCase ( self , _lowerCAmelCase ): UpperCAmelCase__ : Dict = list(struct.unpack(""">16L""" , _lowerCAmelCase ) ) + [0] * 64 for i in range(16 , 80 ): UpperCAmelCase__ : Optional[int] = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) , 1 ) return w def __UpperCAmelCase ( self ): UpperCAmelCase__ : List[str] = self.padding() UpperCAmelCase__ : List[str] = self.split_blocks() for block in self.blocks: UpperCAmelCase__ : Tuple = self.expand_block(_lowerCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.h for i in range(0 , 80 ): if 0 <= i < 20: UpperCAmelCase__ : Optional[int] = (b & c) | ((~b) & d) UpperCAmelCase__ : int = 0X5a82_7999 elif 20 <= i < 40: UpperCAmelCase__ : Tuple = b ^ c ^ d UpperCAmelCase__ : int = 0X6ed9_eba1 elif 40 <= i < 60: UpperCAmelCase__ : List[str] = (b & c) | (b & d) | (c & d) UpperCAmelCase__ : Tuple = 0X8f1b_bcdc elif 60 <= i < 80: UpperCAmelCase__ : int = b ^ c ^ d UpperCAmelCase__ : str = 0Xca62_c1d6 UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = ( self.rotate(_lowerCAmelCase , 5 ) + f + e + k + expanded_block[i] & 0Xffff_ffff, a, self.rotate(_lowerCAmelCase , 30 ), c, d, ) UpperCAmelCase__ : int = ( self.h[0] + a & 0Xffff_ffff, self.h[1] + b & 0Xffff_ffff, self.h[2] + c & 0Xffff_ffff, self.h[3] + d & 0Xffff_ffff, self.h[4] + e & 0Xffff_ffff, ) return ("{:08x}" * 5).format(*self.h ) def _lowerCamelCase ( ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase__ : Optional[Any] = B"""Test String""" assert SHAaHash(__lowerCamelCase ).final_hash() == hashlib.shaa(__lowerCamelCase ).hexdigest() # noqa: S324 def _lowerCamelCase ( ) -> str: '''simple docstring''' UpperCAmelCase__ : Optional[int] = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument("""--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) UpperCAmelCase__ : str = parser.parse_args() UpperCAmelCase__ : Union[str, Any] = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: UpperCAmelCase__ : List[Any] = f.read() else: UpperCAmelCase__ : int = bytes(__lowerCamelCase , """utf-8""" ) print(SHAaHash(__lowerCamelCase ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
79
0
# Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING import numpy as np import pyarrow as pa from .. import config from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import torch class _SCREAMING_SNAKE_CASE ( TensorFormatter[Mapping, '''torch.Tensor''', Mapping] ): def __init__(self , UpperCAmelCase=None , **UpperCAmelCase): '''simple docstring''' super().__init__(features=_lowerCAmelCase) __UpperCAmelCase =torch_tensor_kwargs import torch # noqa import torch at initialization def A__ (self , UpperCAmelCase): '''simple docstring''' import torch if isinstance(_lowerCAmelCase , _lowerCAmelCase) and column: if all( isinstance(_lowerCAmelCase , torch.Tensor) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column): return torch.stack(_lowerCAmelCase) return column def A__ (self , UpperCAmelCase): '''simple docstring''' import torch if isinstance(_lowerCAmelCase , (str, bytes, type(_lowerCAmelCase))): return value elif isinstance(_lowerCAmelCase , (np.character, np.ndarray)) and np.issubdtype(value.dtype , np.character): return value.tolist() __UpperCAmelCase ={} if isinstance(_lowerCAmelCase , (np.number, np.ndarray)) and np.issubdtype(value.dtype , np.integer): __UpperCAmelCase ={"""dtype""": torch.intaa} elif isinstance(_lowerCAmelCase , (np.number, np.ndarray)) and np.issubdtype(value.dtype , np.floating): __UpperCAmelCase ={"""dtype""": torch.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(_lowerCAmelCase , PIL.Image.Image): __UpperCAmelCase =np.asarray(_lowerCAmelCase) return torch.tensor(_lowerCAmelCase , **{**default_dtype, **self.torch_tensor_kwargs}) def A__ (self , UpperCAmelCase): '''simple docstring''' import torch # support for torch, tf, jax etc. if hasattr(_lowerCAmelCase , '''__array__''') and not isinstance(_lowerCAmelCase , torch.Tensor): __UpperCAmelCase =data_struct.__array__() # support for nested types like struct of list of struct if isinstance(_lowerCAmelCase , np.ndarray): if data_struct.dtype == object: # torch tensors cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(_lowerCAmelCase) for substruct in data_struct]) elif isinstance(_lowerCAmelCase , (list, tuple)): return self._consolidate([self.recursive_tensorize(_lowerCAmelCase) for substruct in data_struct]) return self._tensorize(_lowerCAmelCase) def A__ (self , UpperCAmelCase): '''simple docstring''' return map_nested(self._recursive_tensorize , _lowerCAmelCase , map_list=_lowerCAmelCase) def A__ (self , UpperCAmelCase): '''simple docstring''' __UpperCAmelCase =self.numpy_arrow_extractor().extract_row(_lowerCAmelCase) __UpperCAmelCase =self.python_features_decoder.decode_row(_lowerCAmelCase) return self.recursive_tensorize(_lowerCAmelCase) def A__ (self , UpperCAmelCase): '''simple docstring''' __UpperCAmelCase =self.numpy_arrow_extractor().extract_column(_lowerCAmelCase) __UpperCAmelCase =self.python_features_decoder.decode_column(_lowerCAmelCase , pa_table.column_names[0]) __UpperCAmelCase =self.recursive_tensorize(_lowerCAmelCase) __UpperCAmelCase =self._consolidate(_lowerCAmelCase) return column def A__ (self , UpperCAmelCase): '''simple docstring''' __UpperCAmelCase =self.numpy_arrow_extractor().extract_batch(_lowerCAmelCase) __UpperCAmelCase =self.python_features_decoder.decode_batch(_lowerCAmelCase) __UpperCAmelCase =self.recursive_tensorize(_lowerCAmelCase) for column_name in batch: __UpperCAmelCase =self._consolidate(batch[column_name]) return batch
132
from importlib import import_module from .logging import get_logger SCREAMING_SNAKE_CASE__ : Union[str, Any] = get_logger(__name__) class UpperCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : List[str] = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith("""__""" ): setattr(self , _lowerCAmelCase , getattr(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase__ : Tuple = module._original_module if isinstance(_lowerCAmelCase , _PatchedModuleObj ) else module class UpperCAmelCase_ : __lowerCamelCase = [] def __init__( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=None ): UpperCAmelCase__ : str = obj UpperCAmelCase__ : List[str] = target UpperCAmelCase__ : List[str] = new UpperCAmelCase__ : Any = target.split(""".""" )[0] UpperCAmelCase__ : Union[str, Any] = {} UpperCAmelCase__ : str = attrs or [] def __enter__( self ): *UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.target.split(""".""" ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(_lowerCAmelCase ) ): try: UpperCAmelCase__ : Optional[int] = import_module(""".""".join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): UpperCAmelCase__ : Any = getattr(self.obj , _lowerCAmelCase ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(_lowerCAmelCase , _PatchedModuleObj ) and obj_attr._original_module is submodule) ): UpperCAmelCase__ : List[Any] = obj_attr # patch at top level setattr(self.obj , _lowerCAmelCase , _PatchedModuleObj(_lowerCAmelCase , attrs=self.attrs ) ) UpperCAmelCase__ : Optional[Any] = getattr(self.obj , _lowerCAmelCase ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(_lowerCAmelCase , _lowerCAmelCase , _PatchedModuleObj(getattr(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) , attrs=self.attrs ) ) UpperCAmelCase__ : Union[str, Any] = getattr(_lowerCAmelCase , _lowerCAmelCase ) # finally set the target attribute setattr(_lowerCAmelCase , _lowerCAmelCase , self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: UpperCAmelCase__ : Union[str, Any] = getattr(import_module(""".""".join(_lowerCAmelCase ) ) , _lowerCAmelCase ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , _lowerCAmelCase ) is attr_value: UpperCAmelCase__ : Optional[int] = getattr(self.obj , _lowerCAmelCase ) setattr(self.obj , _lowerCAmelCase , self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" UpperCAmelCase__ : Dict = globals()["""__builtins__"""][target_attr] setattr(self.obj , _lowerCAmelCase , self.new ) else: raise RuntimeError(f"Tried to patch attribute {target_attr} instead of a submodule." ) def __exit__( self , *_lowerCAmelCase ): for attr in list(self.original ): setattr(self.obj , _lowerCAmelCase , self.original.pop(_lowerCAmelCase ) ) def __UpperCAmelCase ( self ): self.__enter__() self._active_patches.append(self ) def __UpperCAmelCase ( self ): try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
79
0
from __future__ import annotations from typing import Any class A__ : """simple docstring""" def __init__( self : int , lowerCamelCase__ : Optional[int] ): a__ : int = num_of_nodes a__ : list[list[int]] = [] a__ : dict[int, int] = {} def _UpperCamelCase( self : Optional[Any] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : List[Any] , lowerCamelCase__ : Optional[Any] ): self.m_edges.append([u_node, v_node, weight] ) def _UpperCamelCase( self : List[Any] , lowerCamelCase__ : Optional[int] ): if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node] ) def _UpperCamelCase( self : str , lowerCamelCase__ : Optional[Any] ): if self.m_component[u_node] != u_node: for k in self.m_component: a__ : List[str] = self.find_component(_lowerCAmelCase ) def _UpperCamelCase( self : Optional[int] , lowerCamelCase__ : int , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : int ): if component_size[u_node] <= component_size[v_node]: a__ : Tuple = v_node component_size[v_node] += component_size[u_node] self.set_component(_lowerCAmelCase ) elif component_size[u_node] >= component_size[v_node]: a__ : Dict = self.find_component(_lowerCAmelCase ) component_size[u_node] += component_size[v_node] self.set_component(_lowerCAmelCase ) def _UpperCamelCase( self : str ): a__ : List[str] = [] a__ : Tuple = 0 a__ : list[Any] = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes ): self.m_component.update({node: node} ) component_size.append(1 ) a__ : Optional[Any] = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: a__ : Optional[Any] = edge a__ : List[str] = self.m_component[u] a__ : Optional[Any] = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): a__ : Union[str, Any] = [u, v, w] for edge in minimum_weight_edge: if isinstance(_lowerCAmelCase , _lowerCAmelCase ): a__ : Tuple = edge a__ : str = self.m_component[u] a__ : Union[str, Any] = self.m_component[v] if u_component != v_component: mst_weight += w self.union(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) print(f'''Added edge [{u} - {v}]\nAdded weight: {w}\n''' ) num_of_components -= 1 a__ : int = [-1] * self.m_num_of_nodes print(f'''The total weight of the minimal spanning tree is: {mst_weight}''' ) def UpperCamelCase_ ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
37
from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : str = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Any = { """huggingface/informer-tourism-monthly""": ( """https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json""" ), # See all Informer models at https://huggingface.co/models?filter=informer } class UpperCAmelCase_ ( __lowerCamelCase ): __lowerCamelCase = 'informer' __lowerCamelCase = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', 'num_hidden_layers': 'encoder_layers', } def __init__( self , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = "student_t" , _lowerCAmelCase = "nll" , _lowerCAmelCase = 1 , _lowerCAmelCase = None , _lowerCAmelCase = "mean" , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = 0 , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = 64 , _lowerCAmelCase = 32 , _lowerCAmelCase = 32 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = 2 , _lowerCAmelCase = True , _lowerCAmelCase = "gelu" , _lowerCAmelCase = 0.0_5 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 100 , _lowerCAmelCase = 0.0_2 , _lowerCAmelCase=True , _lowerCAmelCase = "prob" , _lowerCAmelCase = 5 , _lowerCAmelCase = True , **_lowerCAmelCase , ): # time series specific configuration UpperCAmelCase__ : List[str] = prediction_length UpperCAmelCase__ : Optional[Any] = context_length or prediction_length UpperCAmelCase__ : str = distribution_output UpperCAmelCase__ : int = loss UpperCAmelCase__ : Optional[Any] = input_size UpperCAmelCase__ : Any = num_time_features UpperCAmelCase__ : int = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] UpperCAmelCase__ : Union[str, Any] = scaling UpperCAmelCase__ : Optional[Any] = num_dynamic_real_features UpperCAmelCase__ : List[str] = num_static_real_features UpperCAmelCase__ : str = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(_lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The cardinality should be a list of the same length as `num_static_categorical_features`""" ) UpperCAmelCase__ : List[str] = cardinality else: UpperCAmelCase__ : Optional[Any] = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(_lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The embedding dimension should be a list of the same length as `num_static_categorical_features`""" ) UpperCAmelCase__ : str = embedding_dimension else: UpperCAmelCase__ : List[str] = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] UpperCAmelCase__ : Union[str, Any] = num_parallel_samples # Transformer architecture configuration UpperCAmelCase__ : Dict = input_size * len(self.lags_sequence ) + self._number_of_features UpperCAmelCase__ : Any = d_model UpperCAmelCase__ : int = encoder_attention_heads UpperCAmelCase__ : Optional[Any] = decoder_attention_heads UpperCAmelCase__ : int = encoder_ffn_dim UpperCAmelCase__ : Tuple = decoder_ffn_dim UpperCAmelCase__ : List[Any] = encoder_layers UpperCAmelCase__ : Optional[Any] = decoder_layers UpperCAmelCase__ : Tuple = dropout UpperCAmelCase__ : int = attention_dropout UpperCAmelCase__ : List[str] = activation_dropout UpperCAmelCase__ : Any = encoder_layerdrop UpperCAmelCase__ : Union[str, Any] = decoder_layerdrop UpperCAmelCase__ : Tuple = activation_function UpperCAmelCase__ : Dict = init_std UpperCAmelCase__ : str = use_cache # Informer UpperCAmelCase__ : Union[str, Any] = attention_type UpperCAmelCase__ : int = sampling_factor UpperCAmelCase__ : Any = distil super().__init__(is_encoder_decoder=_lowerCAmelCase , **_lowerCAmelCase ) @property def __UpperCAmelCase ( self ): return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
79
0
'''simple docstring''' from __future__ import annotations def UpperCamelCase ( lowercase_ : Optional[int] , lowercase_ : Any , lowercase_ : str , ) -> tuple: '''simple docstring''' if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif electron_conc < 0: raise ValueError('''Electron concentration cannot be negative in a semiconductor''' ) elif hole_conc < 0: raise ValueError('''Hole concentration cannot be negative in a semiconductor''' ) elif intrinsic_conc < 0: raise ValueError( '''Intrinsic concentration cannot be negative in a semiconductor''' ) elif electron_conc == 0: return ( "electron_conc", intrinsic_conc**2 / hole_conc, ) elif hole_conc == 0: return ( "hole_conc", intrinsic_conc**2 / electron_conc, ) elif intrinsic_conc == 0: return ( "intrinsic_conc", (electron_conc * hole_conc) ** 0.5, ) else: return (-1, -1) if __name__ == "__main__": import doctest doctest.testmod()
72
def _lowerCamelCase ( __lowerCamelCase ) -> bool: '''simple docstring''' if p < 2: raise ValueError("""p should not be less than 2!""" ) elif p == 2: return True UpperCAmelCase__ : Tuple = 4 UpperCAmelCase__ : Tuple = (1 << p) - 1 for _ in range(p - 2 ): UpperCAmelCase__ : List[str] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
79
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase : Tuple = logging.get_logger(__name__) UpperCAmelCase : List[Any] = { """MIT/ast-finetuned-audioset-10-10-0.4593""": ( """https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json""" ), } class lowerCAmelCase__ ( __lowerCamelCase ): """simple docstring""" lowerCAmelCase__ = "audio-spectrogram-transformer" def __init__( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Dict=768 , __SCREAMING_SNAKE_CASE : Tuple=12 , __SCREAMING_SNAKE_CASE : Any=12 , __SCREAMING_SNAKE_CASE : Dict=3_072 , __SCREAMING_SNAKE_CASE : Dict="gelu" , __SCREAMING_SNAKE_CASE : List[str]=0.0 , __SCREAMING_SNAKE_CASE : int=0.0 , __SCREAMING_SNAKE_CASE : Tuple=0.02 , __SCREAMING_SNAKE_CASE : Any=1E-12 , __SCREAMING_SNAKE_CASE : int=16 , __SCREAMING_SNAKE_CASE : Union[str, Any]=True , __SCREAMING_SNAKE_CASE : List[Any]=10 , __SCREAMING_SNAKE_CASE : Optional[int]=10 , __SCREAMING_SNAKE_CASE : List[str]=1_024 , __SCREAMING_SNAKE_CASE : str=128 , **__SCREAMING_SNAKE_CASE : Union[str, Any] , ) -> Any: """simple docstring""" super().__init__(**_lowerCAmelCase ) __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = layer_norm_eps __SCREAMING_SNAKE_CASE = patch_size __SCREAMING_SNAKE_CASE = qkv_bias __SCREAMING_SNAKE_CASE = frequency_stride __SCREAMING_SNAKE_CASE = time_stride __SCREAMING_SNAKE_CASE = max_length __SCREAMING_SNAKE_CASE = num_mel_bins
627
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE__ : Any = { """configuration_mobilevit""": ["""MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MobileViTConfig""", """MobileViTOnnxConfig"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : List[str] = ["""MobileViTFeatureExtractor"""] SCREAMING_SNAKE_CASE__ : Union[str, Any] = ["""MobileViTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Dict = [ """MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MobileViTForImageClassification""", """MobileViTForSemanticSegmentation""", """MobileViTModel""", """MobileViTPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Any = [ """TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFMobileViTForImageClassification""", """TFMobileViTForSemanticSegmentation""", """TFMobileViTModel""", """TFMobileViTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mobilevit import MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileViTConfig, MobileViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilevit import MobileViTFeatureExtractor from .image_processing_mobilevit import MobileViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilevit import ( MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel, MobileViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilevit import ( TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileViTForImageClassification, TFMobileViTForSemanticSegmentation, TFMobileViTModel, TFMobileViTPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
79
0
'''simple docstring''' import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class UpperCAmelCase__ ( unittest.TestCase ): def __init__( self : Optional[Any],__A : Dict,__A : int=7,__A : Union[str, Any]=3,__A : Union[str, Any]=1_8,__A : Dict=3_0,__A : List[Any]=4_0_0,__A : str=True,__A : Dict=None,__A : Optional[Any]=True,): _lowerCamelCase : List[str] = size if size is not None else {"""height""": 1_8, """width""": 1_8} _lowerCamelCase : Union[str, Any] = parent _lowerCamelCase : int = batch_size _lowerCamelCase : Tuple = num_channels _lowerCamelCase : Dict = image_size _lowerCamelCase : List[Any] = min_resolution _lowerCamelCase : str = max_resolution _lowerCamelCase : Union[str, Any] = do_resize _lowerCamelCase : Tuple = size _lowerCamelCase : int = do_normalize def lowerCamelCase_ ( self : int ): return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.8866443634033203, 0.6618829369544983, 0.3891746401786804], [-0.6042559146881104, -0.02295008860528469, 0.5423797369003296], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class UpperCAmelCase__ ( __lowerCamelCase , unittest.TestCase ): lowerCAmelCase_ = ImageGPTImageProcessor if is_vision_available() else None def lowerCamelCase_ ( self : Dict ): _lowerCamelCase : Any = ImageGPTImageProcessingTester(self ) @property def lowerCamelCase_ ( self : List[str] ): return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase_ ( self : Union[str, Any] ): _lowerCamelCase : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCAmelCase,"clusters" ) ) self.assertTrue(hasattr(_lowerCAmelCase,"do_resize" ) ) self.assertTrue(hasattr(_lowerCAmelCase,"size" ) ) self.assertTrue(hasattr(_lowerCAmelCase,"do_normalize" ) ) def lowerCamelCase_ ( self : Union[str, Any] ): _lowerCamelCase : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size,{"height": 1_8, "width": 1_8} ) _lowerCamelCase : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict,size=4_2 ) self.assertEqual(image_processor.size,{"height": 4_2, "width": 4_2} ) def lowerCamelCase_ ( self : Optional[int] ): _lowerCamelCase : Optional[int] = self.image_processing_class(**self.image_processor_dict ) _lowerCamelCase : Optional[int] = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase,obj[key] ) ) else: self.assertEqual(obj[key],_lowerCAmelCase ) def lowerCamelCase_ ( self : Any ): _lowerCamelCase : Optional[int] = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : Union[str, Any] = os.path.join(_lowerCAmelCase,"image_processor.json" ) image_processor_first.to_json_file(_lowerCAmelCase ) _lowerCamelCase : Optional[Any] = self.image_processing_class.from_json_file(_lowerCAmelCase ).to_dict() _lowerCamelCase : Dict = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase,image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key],_lowerCAmelCase ) def lowerCamelCase_ ( self : str ): _lowerCamelCase : str = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(_lowerCAmelCase ) _lowerCamelCase : List[Any] = self.image_processing_class.from_pretrained(_lowerCAmelCase ).to_dict() _lowerCamelCase : Tuple = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(_lowerCAmelCase,image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key],_lowerCAmelCase ) @unittest.skip("ImageGPT requires clusters at initialization" ) def lowerCamelCase_ ( self : Optional[Any] ): pass def A_ ( ): """simple docstring""" _lowerCamelCase : Any = load_dataset("hf-internal-testing/fixtures_image_utils" , split="test" ) _lowerCamelCase : Dict = Image.open(dataset[4]["file"] ) _lowerCamelCase : Optional[Any] = Image.open(dataset[5]["file"] ) _lowerCamelCase : List[Any] = [imagea, imagea] return images @require_vision @require_torch class UpperCAmelCase__ ( unittest.TestCase ): @slow def lowerCamelCase_ ( self : Union[str, Any] ): _lowerCamelCase : Tuple = ImageGPTImageProcessor.from_pretrained("openai/imagegpt-small" ) _lowerCamelCase : int = prepare_images() # test non-batched _lowerCamelCase : List[str] = image_processing(images[0],return_tensors="pt" ) self.assertIsInstance(encoding.input_ids,torch.LongTensor ) self.assertEqual(encoding.input_ids.shape,(1, 1_0_2_4) ) _lowerCamelCase : List[Any] = [3_0_6, 1_9_1, 1_9_1] self.assertEqual(encoding.input_ids[0, :3].tolist(),_lowerCAmelCase ) # test batched _lowerCamelCase : List[str] = image_processing(_lowerCAmelCase,return_tensors="pt" ) self.assertIsInstance(encoding.input_ids,torch.LongTensor ) self.assertEqual(encoding.input_ids.shape,(2, 1_0_2_4) ) _lowerCamelCase : Any = [3_0_3, 1_3, 1_3] self.assertEqual(encoding.input_ids[1, -3:].tolist(),_lowerCAmelCase )
44
from __future__ import annotations SCREAMING_SNAKE_CASE__ : List[str] = 8.988e9 # units = N * m^s * C^-2 def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> dict[str, float]: '''simple docstring''' UpperCAmelCase__ : int = abs(chargea * chargea ) if (force, chargea, chargea, distance).count(0 ) != 1: raise ValueError("""One and only one argument must be 0""" ) if distance < 0: raise ValueError("""Distance cannot be negative""" ) if force == 0: UpperCAmelCase__ : int = COULOMBS_CONSTANT * charge_product / (distance**2) return {"force": force} elif chargea == 0: UpperCAmelCase__ : str = abs(__lowerCamelCase ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge1": chargea} elif chargea == 0: UpperCAmelCase__ : Union[str, Any] = abs(__lowerCamelCase ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge2": chargea} elif distance == 0: UpperCAmelCase__ : Optional[Any] = (COULOMBS_CONSTANT * charge_product / abs(__lowerCamelCase )) ** 0.5 return {"distance": distance} raise ValueError("""Exactly one argument must be 0""" ) if __name__ == "__main__": import doctest doctest.testmod()
79
0